https://launchpad.net/ubuntu/+archive/test-rebuild-20201216-hirsute/+build/20426680 RUN: /usr/share/launchpad-buildd/bin/builder-prep Kernel version: Linux bos02-s390x-015 4.15.0-134-generic #138-Ubuntu SMP Fri Jan 15 10:53:04 UTC 2021 s390x Buildd toolchain package versions: launchpad-buildd_194~476~ubuntu18.04.1 python3-lpbuildd_194~476~ubuntu18.04.1 sbuild_0.75.0-1ubuntu1 bzr-builder_0.7.3+bzr174~ppa13~ubuntu16.04.1 bzr_2.7.0+bzr6622-10 git-build-recipe_0.3.6~git201906051340.ff11471~ubuntu18.04.1 git_1:2.17.1-1ubuntu0.7 dpkg-dev_1.19.0.5ubuntu2.3 python-debian_0.1.32 python3-debian_0.1.32. Syncing the system clock with the buildd NTP service... 29 Jan 19:28:00 ntpdate[1719]: adjust time server 10.211.37.1 offset 0.021088 sec RUN: /usr/share/launchpad-buildd/bin/in-target unpack-chroot --backend=chroot --series=hirsute --arch=s390x PACKAGEBUILD-20426680 --image-type chroot /home/buildd/filecache-default/9aaab3d1c3367126bc5b66d2474faf9141147628 Creating target for build PACKAGEBUILD-20426680 RUN: /usr/share/launchpad-buildd/bin/in-target mount-chroot --backend=chroot --series=hirsute --arch=s390x PACKAGEBUILD-20426680 Starting target for build PACKAGEBUILD-20426680 RUN: /usr/share/launchpad-buildd/bin/in-target override-sources-list --backend=chroot --series=hirsute --arch=s390x PACKAGEBUILD-20426680 'deb http://ppa.launchpad.net/ubuntu-toolchain-r/ppa/ubuntu hirsute main' 'deb http://ftpmaster.internal/ubuntu hirsute main universe' Overriding sources.list in build-PACKAGEBUILD-20426680 RUN: /usr/share/launchpad-buildd/bin/in-target add-trusted-keys --backend=chroot --series=hirsute --arch=s390x PACKAGEBUILD-20426680 Adding trusted keys to build-PACKAGEBUILD-20426680 Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). OK Warning: apt-key is deprecated. Manage keyring files in trusted.gpg.d instead (see apt-key(8)). /etc/apt/trusted.gpg -------------------- pub rsa1024 2009-10-22 [SC] 60C3 1780 3A41 BA51 845E 371A 1E93 77A2 BA9E F27F uid [ unknown] Launchpad Toolchain builds /etc/apt/trusted.gpg.d/ubuntu-keyring-2012-cdimage.gpg ------------------------------------------------------ pub rsa4096 2012-05-11 [SC] 8439 38DF 228D 22F7 B374 2BC0 D94A A3F0 EFE2 1092 uid [ unknown] Ubuntu CD Image Automatic Signing Key (2012) /etc/apt/trusted.gpg.d/ubuntu-keyring-2018-archive.gpg ------------------------------------------------------ pub rsa4096 2018-09-17 [SC] F6EC B376 2474 EDA9 D21B 7022 8719 20D1 991B C93C uid [ unknown] Ubuntu Archive Automatic Signing Key (2018) RUN: /usr/share/launchpad-buildd/bin/in-target update-debian-chroot --backend=chroot --series=hirsute --arch=s390x PACKAGEBUILD-20426680 Updating target for build PACKAGEBUILD-20426680 Get:1 http://ppa.launchpad.net/ubuntu-toolchain-r/ppa/ubuntu hirsute InRelease [23.8 kB] Get:2 http://ftpmaster.internal/ubuntu hirsute InRelease [269 kB] Get:3 http://ppa.launchpad.net/ubuntu-toolchain-r/ppa/ubuntu hirsute/main s390x Packages [16.8 kB] Get:4 http://ftpmaster.internal/ubuntu hirsute/main s390x Packages [1333 kB] Get:5 http://ppa.launchpad.net/ubuntu-toolchain-r/ppa/ubuntu hirsute/main Translation-en [10.0 kB] Get:6 http://ftpmaster.internal/ubuntu hirsute/main Translation-en [513 kB] Get:7 http://ftpmaster.internal/ubuntu hirsute/universe s390x Packages [12.3 MB] Get:8 http://ftpmaster.internal/ubuntu hirsute/universe Translation-en [5422 kB] Fetched 19.9 MB in 4s (5454 kB/s) Reading package lists... Reading package lists... Building dependency tree... Reading state information... Calculating upgrade... The following packages were automatically installed and are no longer required: libisl22 libperl5.30 perl-modules-5.30 Use 'sudo apt autoremove' to remove them. The following NEW packages will be installed: libisl23 libperl5.32 libxxhash0 lto-disabled-list perl-modules-5.32 The following packages will be upgraded: adduser apt base-files base-passwd bash binutils binutils-common binutils-s390x-linux-gnu bsdutils coreutils cpp-10 dash dpkg dpkg-dev fakeroot g++-10 gcc-10 gcc-10-base gpg gpg-agent gpgconf gpgv grep gzip init init-system-helpers libacl1 libapparmor1 libapt-pkg6.0 libasan6 libatomic1 libattr1 libaudit-common libaudit1 libbinutils libblkid1 libc-bin libc-dev-bin libc6 libc6-dev libcap-ng0 libcap2 libcc1-0 libcrypt-dev libcrypt1 libcryptsetup12 libctf-nobfd0 libctf0 libdebconfclient0 libdevmapper1.02.1 libdpkg-perl libfakeroot libgcc-10-dev libgcc-s1 libgcrypt20 libgdbm-compat4 libgdbm6 libgnutls30 libgomp1 libgssapi-krb5-2 libidn2-0 libip4tc2 libitm1 libk5crypto3 libkrb5-3 libkrb5support0 liblockfile-bin liblockfile1 liblz4-1 libmount1 libmpc3 libncurses6 libncursesw6 libnpth0 libp11-kit0 libpcre2-8-0 libreadline8 libseccomp2 libselinux1 libsemanage-common libsemanage1 libsmartcols1 libsqlite3-0 libssl1.1 libstdc++-10-dev libstdc++6 libsystemd0 libtinfo6 libtirpc-common libtirpc-dev libtirpc3 libubsan1 libudev1 libuuid1 libzstd1 linux-libc-dev login mount ncurses-base ncurses-bin openssl passwd patch perl perl-base readline-common sensible-utils systemd systemd-sysv systemd-timesyncd sysvinit-utils tar tzdata util-linux 114 upgraded, 5 newly installed, 0 to remove and 0 not upgraded. Need to get 79.0 MB of archives. After this operation, 44.4 MB of additional disk space will be used. Get:1 http://ftpmaster.internal/ubuntu hirsute/main s390x libcrypt-dev s390x 1:4.4.17-1ubuntu1 [111 kB] Get:2 http://ftpmaster.internal/ubuntu hirsute/main s390x libc6-dev s390x 2.32-0ubuntu6 [1745 kB] Get:3 http://ftpmaster.internal/ubuntu hirsute/main s390x libc-dev-bin s390x 2.32-0ubuntu6 [29.8 kB] Get:4 http://ftpmaster.internal/ubuntu hirsute/main s390x libcrypt1 s390x 1:4.4.17-1ubuntu1 [84.3 kB] Get:5 http://ftpmaster.internal/ubuntu hirsute/main s390x linux-libc-dev s390x 5.8.0-36.40+21.04.1 [1138 kB] Get:6 http://ftpmaster.internal/ubuntu hirsute/main s390x libtirpc-common all 1.3.1-1 [7452 B] Get:7 http://ftpmaster.internal/ubuntu hirsute/main s390x libk5crypto3 s390x 1.18.3-4 [77.0 kB] Get:8 http://ftpmaster.internal/ubuntu hirsute/main s390x libgssapi-krb5-2 s390x 1.18.3-4 [122 kB] Get:9 http://ftpmaster.internal/ubuntu hirsute/main s390x libkrb5-3 s390x 1.18.3-4 [368 kB] Get:10 http://ftpmaster.internal/ubuntu hirsute/main s390x libkrb5support0 s390x 1.18.3-4 [31.4 kB] Get:11 http://ftpmaster.internal/ubuntu hirsute/main s390x libssl1.1 s390x 1.1.1f-1ubuntu5 [1044 kB] Get:12 http://ftpmaster.internal/ubuntu hirsute/main s390x libtirpc-dev s390x 1.3.1-1 [190 kB] Get:13 http://ftpmaster.internal/ubuntu hirsute/main s390x libtirpc3 s390x 1.3.1-1 [75.1 kB] Get:14 http://ftpmaster.internal/ubuntu hirsute/main s390x libisl23 s390x 0.23-1 [689 kB] Get:15 http://ftpmaster.internal/ubuntu hirsute/main s390x libcc1-0 s390x 10.2.1-6ubuntu1 [47.7 kB] Get:16 http://ftpmaster.internal/ubuntu hirsute/main s390x libctf0 s390x 2.35.50.20210106-1ubuntu2 [108 kB] Get:17 http://ftpmaster.internal/ubuntu hirsute/main s390x libctf-nobfd0 s390x 2.35.50.20210106-1ubuntu2 [110 kB] Get:18 http://ftpmaster.internal/ubuntu hirsute/main s390x binutils-s390x-linux-gnu s390x 2.35.50.20210106-1ubuntu2 [1776 kB] Get:19 http://ftpmaster.internal/ubuntu hirsute/main s390x libbinutils s390x 2.35.50.20210106-1ubuntu2 [443 kB] Get:20 http://ftpmaster.internal/ubuntu hirsute/main s390x binutils-common s390x 2.35.50.20210106-1ubuntu2 [216 kB] Get:21 http://ftpmaster.internal/ubuntu hirsute/main s390x binutils s390x 2.35.50.20210106-1ubuntu2 [3360 B] Get:22 http://ftpmaster.internal/ubuntu hirsute/main s390x gcc-10-base s390x 10.2.1-6ubuntu1 [19.7 kB] Get:23 http://ftpmaster.internal/ubuntu hirsute/main s390x libgcc-s1 s390x 10.2.1-6ubuntu1 [30.3 kB] Get:24 http://ftpmaster.internal/ubuntu hirsute/main s390x libgomp1 s390x 10.2.1-6ubuntu1 [115 kB] Get:25 http://ftpmaster.internal/ubuntu hirsute/main s390x libitm1 s390x 10.2.1-6ubuntu1 [28.6 kB] Get:26 http://ftpmaster.internal/ubuntu hirsute/main s390x libatomic1 s390x 10.2.1-6ubuntu1 [8384 B] Get:27 http://ftpmaster.internal/ubuntu hirsute/main s390x libasan6 s390x 10.2.1-6ubuntu1 [2280 kB] Get:28 http://ftpmaster.internal/ubuntu hirsute/main s390x libubsan1 s390x 10.2.1-6ubuntu1 [907 kB] Get:29 http://ftpmaster.internal/ubuntu hirsute/main s390x g++-10 s390x 10.2.1-6ubuntu1 [8825 kB] Get:30 http://ftpmaster.internal/ubuntu hirsute/main s390x libstdc++-10-dev s390x 10.2.1-6ubuntu1 [1738 kB] Get:31 http://ftpmaster.internal/ubuntu hirsute/main s390x libgcc-10-dev s390x 10.2.1-6ubuntu1 [776 kB] Get:32 http://ftpmaster.internal/ubuntu hirsute/main s390x gcc-10 s390x 10.2.1-6ubuntu1 [15.1 MB] Get:33 http://ftpmaster.internal/ubuntu hirsute/main s390x cpp-10 s390x 10.2.1-6ubuntu1 [7447 kB] Get:34 http://ftpmaster.internal/ubuntu hirsute/main s390x libstdc++6 s390x 10.2.1-6ubuntu1 [520 kB] Get:35 http://ftpmaster.internal/ubuntu hirsute/main s390x libmpc3 s390x 1.2.0-1 [43.6 kB] Get:36 http://ftpmaster.internal/ubuntu hirsute/main s390x libzstd1 s390x 1.4.8+dfsg-1 [312 kB] Get:37 http://ftpmaster.internal/ubuntu hirsute/main s390x libc6 s390x 2.32-0ubuntu6 [2332 kB] Get:38 http://ftpmaster.internal/ubuntu hirsute/main s390x base-files s390x 11ubuntu16 [60.4 kB] Get:39 http://ftpmaster.internal/ubuntu hirsute/main s390x bash s390x 5.1-1ubuntu1 [751 kB] Get:40 http://ftpmaster.internal/ubuntu hirsute/main s390x bsdutils s390x 1:2.36.1-1ubuntu2 [85.9 kB] Get:41 http://ftpmaster.internal/ubuntu hirsute/main s390x coreutils s390x 8.32-4ubuntu2 [1459 kB] Get:42 http://ftpmaster.internal/ubuntu hirsute/main s390x tar s390x 1.32+dfsg-1 [307 kB] Get:43 http://ftpmaster.internal/ubuntu hirsute/main s390x dpkg s390x 1.20.7.1ubuntu2 [1263 kB] Get:44 http://ftpmaster.internal/ubuntu hirsute/main s390x dash s390x 0.5.11+git20200708+dd9ef66+really0.5.11+git20200708+dd9ef66-5ubuntu1 [112 kB] Get:45 http://ftpmaster.internal/ubuntu hirsute/main s390x grep s390x 3.6-1 [161 kB] Get:46 http://ftpmaster.internal/ubuntu hirsute/main s390x gzip s390x 1.10-2ubuntu3 [101 kB] Get:47 http://ftpmaster.internal/ubuntu hirsute/main s390x login s390x 1:4.8.1-1ubuntu8 [219 kB] Get:48 http://ftpmaster.internal/ubuntu hirsute/main s390x libncurses6 s390x 6.2+20201114-2 [106 kB] Get:49 http://ftpmaster.internal/ubuntu hirsute/main s390x libncursesw6 s390x 6.2+20201114-2 [136 kB] Get:50 http://ftpmaster.internal/ubuntu hirsute/main s390x libtinfo6 s390x 6.2+20201114-2 [92.4 kB] Get:51 http://ftpmaster.internal/ubuntu hirsute/main s390x ncurses-bin s390x 6.2+20201114-2 [176 kB] Get:52 http://ftpmaster.internal/ubuntu hirsute/main s390x perl-modules-5.32 all 5.32.0-6 [2754 kB] Get:53 http://ftpmaster.internal/ubuntu hirsute/main s390x libperl5.32 s390x 5.32.0-6 [3947 kB] Get:54 http://ftpmaster.internal/ubuntu hirsute/main s390x perl s390x 5.32.0-6 [225 kB] Get:55 http://ftpmaster.internal/ubuntu hirsute/main s390x perl-base s390x 5.32.0-6 [1574 kB] Get:56 http://ftpmaster.internal/ubuntu hirsute/main s390x libgdbm6 s390x 1.19-2 [28.1 kB] Get:57 http://ftpmaster.internal/ubuntu hirsute/main s390x libgdbm-compat4 s390x 1.19-2 [6320 B] Get:58 http://ftpmaster.internal/ubuntu hirsute/main s390x util-linux s390x 2.36.1-1ubuntu2 [1072 kB] Get:59 http://ftpmaster.internal/ubuntu hirsute/main s390x libdebconfclient0 s390x 0.256ubuntu1 [6076 B] Get:60 http://ftpmaster.internal/ubuntu hirsute/main s390x base-passwd s390x 3.5.48 [48.5 kB] Get:61 http://ftpmaster.internal/ubuntu hirsute/main s390x init-system-helpers all 1.60 [37.8 kB] Get:62 http://ftpmaster.internal/ubuntu hirsute/main s390x libc-bin s390x 2.32-0ubuntu6 [599 kB] Get:63 http://ftpmaster.internal/ubuntu hirsute/main s390x ncurses-base all 6.2+20201114-2 [18.4 kB] Get:64 http://ftpmaster.internal/ubuntu hirsute/main s390x sysvinit-utils s390x 2.96-5ubuntu1 [20.6 kB] Get:65 http://ftpmaster.internal/ubuntu hirsute/main s390x libxxhash0 s390x 0.8.0-2 [18.3 kB] Get:66 http://ftpmaster.internal/ubuntu hirsute/main s390x libgcrypt20 s390x 1.8.7-2ubuntu1 [393 kB] Get:67 http://ftpmaster.internal/ubuntu hirsute/main s390x liblz4-1 s390x 1.9.3-0ubuntu1 [57.8 kB] Get:68 http://ftpmaster.internal/ubuntu hirsute/main s390x systemd-sysv s390x 247.1-4ubuntu1 [10.3 kB] Get:69 http://ftpmaster.internal/ubuntu hirsute/main s390x libaudit-common all 1:2.8.5-3ubuntu3 [4048 B] Get:70 http://ftpmaster.internal/ubuntu hirsute/main s390x libcap-ng0 s390x 0.7.9-2.2build1 [11.0 kB] Get:71 http://ftpmaster.internal/ubuntu hirsute/main s390x libaudit1 s390x 1:2.8.5-3ubuntu3 [38.1 kB] Get:72 http://ftpmaster.internal/ubuntu hirsute/main s390x libpcre2-8-0 s390x 10.35-2ubuntu1 [110 kB] Get:73 http://ftpmaster.internal/ubuntu hirsute/main s390x libselinux1 s390x 3.1-2build2 [74.9 kB] Get:74 http://ftpmaster.internal/ubuntu hirsute/main s390x libsemanage-common all 3.1-1build2 [10.0 kB] Get:75 http://ftpmaster.internal/ubuntu hirsute/main s390x libsemanage1 s390x 3.1-1build2 [88.5 kB] Get:76 http://ftpmaster.internal/ubuntu hirsute/main s390x passwd s390x 1:4.8.1-1ubuntu8 [794 kB] Get:77 http://ftpmaster.internal/ubuntu hirsute/main s390x adduser all 3.118ubuntu5 [156 kB] Get:78 http://ftpmaster.internal/ubuntu hirsute/main s390x systemd-timesyncd s390x 247.1-4ubuntu1 [27.1 kB] Get:79 http://ftpmaster.internal/ubuntu hirsute/main s390x libapparmor1 s390x 3.0.0-0ubuntu5 [35.9 kB] Get:80 http://ftpmaster.internal/ubuntu hirsute/main s390x libblkid1 s390x 2.36.1-1ubuntu2 [128 kB] Get:81 http://ftpmaster.internal/ubuntu hirsute/main s390x libudev1 s390x 247.1-4ubuntu1 [75.0 kB] Get:82 http://ftpmaster.internal/ubuntu hirsute/main s390x libdevmapper1.02.1 s390x 2:1.02.167-1ubuntu4 [122 kB] Get:83 http://ftpmaster.internal/ubuntu hirsute/main s390x libuuid1 s390x 2.36.1-1ubuntu2 [21.6 kB] Get:84 http://ftpmaster.internal/ubuntu hirsute/main s390x libcryptsetup12 s390x 2:2.3.4-1ubuntu1 [184 kB] Get:85 http://ftpmaster.internal/ubuntu hirsute/main s390x libidn2-0 s390x 2.3.0-5 [51.2 kB] Get:86 http://ftpmaster.internal/ubuntu hirsute/main s390x libp11-kit0 s390x 0.23.22-1 [210 kB] Get:87 http://ftpmaster.internal/ubuntu hirsute/main s390x libgnutls30 s390x 3.7.0-5ubuntu1 [826 kB] Get:88 http://ftpmaster.internal/ubuntu hirsute/main s390x libip4tc2 s390x 1.8.5-3ubuntu4 [20.0 kB] Get:89 http://ftpmaster.internal/ubuntu hirsute/main s390x libseccomp2 s390x 2.4.3-1ubuntu6 [55.9 kB] Get:90 http://ftpmaster.internal/ubuntu hirsute/main s390x mount s390x 2.36.1-1ubuntu2 [119 kB] Get:91 http://ftpmaster.internal/ubuntu hirsute/main s390x systemd s390x 247.1-4ubuntu1 [4015 kB] Get:92 http://ftpmaster.internal/ubuntu hirsute/main s390x libsystemd0 s390x 247.1-4ubuntu1 [296 kB] Get:93 http://ftpmaster.internal/ubuntu hirsute/main s390x libapt-pkg6.0 s390x 2.1.18 [897 kB] Get:94 http://ftpmaster.internal/ubuntu hirsute/main s390x gpgv s390x 2.2.20-1ubuntu2 [218 kB] Get:95 http://ftpmaster.internal/ubuntu hirsute/main s390x apt s390x 2.1.18 [1306 kB] Get:96 http://ftpmaster.internal/ubuntu hirsute/main s390x init s390x 1.60 [6040 B] Get:97 http://ftpmaster.internal/ubuntu hirsute/main s390x libacl1 s390x 2.2.53-9 [17.6 kB] Get:98 http://ftpmaster.internal/ubuntu hirsute/main s390x libattr1 s390x 1:2.4.48-6 [12.8 kB] Get:99 http://ftpmaster.internal/ubuntu hirsute/main s390x libmount1 s390x 2.36.1-1ubuntu2 [142 kB] Get:100 http://ftpmaster.internal/ubuntu hirsute/main s390x libsmartcols1 s390x 2.36.1-1ubuntu2 [94.2 kB] Get:101 http://ftpmaster.internal/ubuntu hirsute/main s390x sensible-utils all 0.0.14 [13.5 kB] Get:102 http://ftpmaster.internal/ubuntu hirsute/main s390x libcap2 s390x 1:2.44-1 [18.3 kB] Get:103 http://ftpmaster.internal/ubuntu hirsute/main s390x readline-common all 8.1-1 [54.1 kB] Get:104 http://ftpmaster.internal/ubuntu hirsute/main s390x libreadline8 s390x 8.1-1 [143 kB] Get:105 http://ftpmaster.internal/ubuntu hirsute/main s390x libsqlite3-0 s390x 3.34.0-1 [650 kB] Get:106 http://ftpmaster.internal/ubuntu hirsute/main s390x openssl s390x 1.1.1f-1ubuntu5 [615 kB] Get:107 http://ftpmaster.internal/ubuntu hirsute/main s390x tzdata all 2021a-1ubuntu1 [294 kB] Get:108 http://ftpmaster.internal/ubuntu hirsute/main s390x dpkg-dev all 1.20.7.1ubuntu2 [935 kB] Get:109 http://ftpmaster.internal/ubuntu hirsute/main s390x libdpkg-perl all 1.20.7.1ubuntu2 [233 kB] Get:110 http://ftpmaster.internal/ubuntu hirsute/main s390x patch s390x 2.7.6-7 [114 kB] Get:111 http://ftpmaster.internal/ubuntu hirsute/main s390x lto-disabled-list all 1 [1996 B] Get:112 http://ftpmaster.internal/ubuntu hirsute/main s390x libfakeroot s390x 1.25.3-1.1 [27.5 kB] Get:113 http://ftpmaster.internal/ubuntu hirsute/main s390x fakeroot s390x 1.25.3-1.1 [62.7 kB] Get:114 http://ftpmaster.internal/ubuntu hirsute/main s390x libnpth0 s390x 1.6-3 [8048 B] Get:115 http://ftpmaster.internal/ubuntu hirsute/main s390x gpg s390x 2.2.20-1ubuntu2 [511 kB] Get:116 http://ftpmaster.internal/ubuntu hirsute/main s390x gpgconf s390x 2.2.20-1ubuntu2 [138 kB] Get:117 http://ftpmaster.internal/ubuntu hirsute/main s390x gpg-agent s390x 2.2.20-1ubuntu2 [256 kB] Get:118 http://ftpmaster.internal/ubuntu hirsute/main s390x liblockfile-bin s390x 1.17-1 [11.7 kB] Get:119 http://ftpmaster.internal/ubuntu hirsute/main s390x liblockfile1 s390x 1.17-1 [6612 B] debconf: delaying package configuration, since apt-utils is not installed Fetched 79.0 MB in 2s (34.6 MB/s) (Reading database ... 12766 files and directories currently installed.) Preparing to unpack .../libcrypt-dev_1%3a4.4.17-1ubuntu1_s390x.deb ... Unpacking libcrypt-dev:s390x (1:4.4.17-1ubuntu1) over (1:4.4.16-1ubuntu1) ... Preparing to unpack .../libc6-dev_2.32-0ubuntu6_s390x.deb ... Unpacking libc6-dev:s390x (2.32-0ubuntu6) over (2.32-0ubuntu3) ... Preparing to unpack .../libc-dev-bin_2.32-0ubuntu6_s390x.deb ... Unpacking libc-dev-bin (2.32-0ubuntu6) over (2.32-0ubuntu3) ... Preparing to unpack .../libcrypt1_1%3a4.4.17-1ubuntu1_s390x.deb ... Unpacking libcrypt1:s390x (1:4.4.17-1ubuntu1) over (1:4.4.16-1ubuntu1) ... Setting up libcrypt1:s390x (1:4.4.17-1ubuntu1) ... (Reading database ... 12765 files and directories currently installed.) Preparing to unpack .../00-linux-libc-dev_5.8.0-36.40+21.04.1_s390x.deb ... Unpacking linux-libc-dev:s390x (5.8.0-36.40+21.04.1) over (5.8.0-25.26) ... Preparing to unpack .../01-libtirpc-common_1.3.1-1_all.deb ... Unpacking libtirpc-common (1.3.1-1) over (1.2.6-1build1) ... Preparing to unpack .../02-libk5crypto3_1.18.3-4_s390x.deb ... Unpacking libk5crypto3:s390x (1.18.3-4) over (1.17-10) ... Preparing to unpack .../03-libgssapi-krb5-2_1.18.3-4_s390x.deb ... Unpacking libgssapi-krb5-2:s390x (1.18.3-4) over (1.17-10) ... Preparing to unpack .../04-libkrb5-3_1.18.3-4_s390x.deb ... Unpacking libkrb5-3:s390x (1.18.3-4) over (1.17-10) ... Preparing to unpack .../05-libkrb5support0_1.18.3-4_s390x.deb ... Unpacking libkrb5support0:s390x (1.18.3-4) over (1.17-10) ... Preparing to unpack .../06-libssl1.1_1.1.1f-1ubuntu5_s390x.deb ... Unpacking libssl1.1:s390x (1.1.1f-1ubuntu5) over (1.1.1f-1ubuntu4) ... Preparing to unpack .../07-libtirpc-dev_1.3.1-1_s390x.deb ... Unpacking libtirpc-dev:s390x (1.3.1-1) over (1.2.6-1build1) ... Preparing to unpack .../08-libtirpc3_1.3.1-1_s390x.deb ... Unpacking libtirpc3:s390x (1.3.1-1) over (1.2.6-1build1) ... Selecting previously unselected package libisl23:s390x. Preparing to unpack .../09-libisl23_0.23-1_s390x.deb ... Unpacking libisl23:s390x (0.23-1) ... Preparing to unpack .../10-libcc1-0_10.2.1-6ubuntu1_s390x.deb ... Unpacking libcc1-0:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../11-libctf0_2.35.50.20210106-1ubuntu2_s390x.deb ... Unpacking libctf0:s390x (2.35.50.20210106-1ubuntu2) over (2.35.1-1ubuntu1) ... Preparing to unpack .../12-libctf-nobfd0_2.35.50.20210106-1ubuntu2_s390x.deb ... Unpacking libctf-nobfd0:s390x (2.35.50.20210106-1ubuntu2) over (2.35.1-1ubuntu1) ... Preparing to unpack .../13-binutils-s390x-linux-gnu_2.35.50.20210106-1ubuntu2_s390x.deb ... Unpacking binutils-s390x-linux-gnu (2.35.50.20210106-1ubuntu2) over (2.35.1-1ubuntu1) ... Preparing to unpack .../14-libbinutils_2.35.50.20210106-1ubuntu2_s390x.deb ... Unpacking libbinutils:s390x (2.35.50.20210106-1ubuntu2) over (2.35.1-1ubuntu1) ... Preparing to unpack .../15-binutils-common_2.35.50.20210106-1ubuntu2_s390x.deb ... Unpacking binutils-common:s390x (2.35.50.20210106-1ubuntu2) over (2.35.1-1ubuntu1) ... Preparing to unpack .../16-binutils_2.35.50.20210106-1ubuntu2_s390x.deb ... Unpacking binutils (2.35.50.20210106-1ubuntu2) over (2.35.1-1ubuntu1) ... Preparing to unpack .../17-gcc-10-base_10.2.1-6ubuntu1_s390x.deb ... Unpacking gcc-10-base:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Setting up gcc-10-base:s390x (10.2.1-6ubuntu1) ... (Reading database ... 12773 files and directories currently installed.) Preparing to unpack .../libgcc-s1_10.2.1-6ubuntu1_s390x.deb ... Unpacking libgcc-s1:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Setting up libgcc-s1:s390x (10.2.1-6ubuntu1) ... (Reading database ... 12773 files and directories currently installed.) Preparing to unpack .../00-libgomp1_10.2.1-6ubuntu1_s390x.deb ... Unpacking libgomp1:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../01-libitm1_10.2.1-6ubuntu1_s390x.deb ... Unpacking libitm1:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../02-libatomic1_10.2.1-6ubuntu1_s390x.deb ... Unpacking libatomic1:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../03-libasan6_10.2.1-6ubuntu1_s390x.deb ... Unpacking libasan6:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../04-libubsan1_10.2.1-6ubuntu1_s390x.deb ... Unpacking libubsan1:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../05-g++-10_10.2.1-6ubuntu1_s390x.deb ... Unpacking g++-10 (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../06-libstdc++-10-dev_10.2.1-6ubuntu1_s390x.deb ... Unpacking libstdc++-10-dev:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../07-libgcc-10-dev_10.2.1-6ubuntu1_s390x.deb ... Unpacking libgcc-10-dev:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../08-gcc-10_10.2.1-6ubuntu1_s390x.deb ... Unpacking gcc-10 (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../09-cpp-10_10.2.1-6ubuntu1_s390x.deb ... Unpacking cpp-10 (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Preparing to unpack .../10-libstdc++6_10.2.1-6ubuntu1_s390x.deb ... Unpacking libstdc++6:s390x (10.2.1-6ubuntu1) over (10.2.0-13ubuntu1) ... Setting up libstdc++6:s390x (10.2.1-6ubuntu1) ... (Reading database ... 12773 files and directories currently installed.) Preparing to unpack .../libmpc3_1.2.0-1_s390x.deb ... Unpacking libmpc3:s390x (1.2.0-1) over (1.2.0~rc1-1) ... Preparing to unpack .../libzstd1_1.4.8+dfsg-1_s390x.deb ... Unpacking libzstd1:s390x (1.4.8+dfsg-1) over (1.4.5+dfsg-4) ... Setting up libzstd1:s390x (1.4.8+dfsg-1) ... (Reading database ... 12773 files and directories currently installed.) Preparing to unpack .../libc6_2.32-0ubuntu6_s390x.deb ... Unpacking libc6:s390x (2.32-0ubuntu6) over (2.32-0ubuntu3) ... Setting up libc6:s390x (2.32-0ubuntu6) ... (Reading database ... 12773 files and directories currently installed.) Preparing to unpack .../base-files_11ubuntu16_s390x.deb ... Unpacking base-files (11ubuntu16) over (11ubuntu14) ... Setting up base-files (11ubuntu16) ... Installing new version of config file /etc/issue ... Installing new version of config file /etc/issue.net ... Installing new version of config file /etc/lsb-release ... (Reading database ... 12773 files and directories currently installed.) Preparing to unpack .../bash_5.1-1ubuntu1_s390x.deb ... Unpacking bash (5.1-1ubuntu1) over (5.0-6ubuntu2) ... Setting up bash (5.1-1ubuntu1) ... update-alternatives: using /usr/share/man/man7/bash-builtins.7.gz to provide /usr/share/man/man7/builtins.7.gz (builtins.7.gz) in auto mode (Reading database ... 12773 files and directories currently installed.) Preparing to unpack .../bsdutils_1%3a2.36.1-1ubuntu2_s390x.deb ... Unpacking bsdutils (1:2.36.1-1ubuntu2) over (1:2.36-3ubuntu1) ... Setting up bsdutils (1:2.36.1-1ubuntu2) ... (Reading database ... 12773 files and directories currently installed.) Preparing to unpack .../coreutils_8.32-4ubuntu2_s390x.deb ... Unpacking coreutils (8.32-4ubuntu2) over (8.32-3ubuntu1) ... Setting up coreutils (8.32-4ubuntu2) ... (Reading database ... 12773 files and directories currently installed.) Preparing to unpack .../tar_1.32+dfsg-1_s390x.deb ... Unpacking tar (1.32+dfsg-1) over (1.30+dfsg-7) ... Setting up tar (1.32+dfsg-1) ... (Reading database ... 12773 files and directories currently installed.) Preparing to unpack .../dpkg_1.20.7.1ubuntu2_s390x.deb ... Unpacking dpkg (1.20.7.1ubuntu2) over (1.20.5ubuntu2) ... Setting up dpkg (1.20.7.1ubuntu2) ... Installing new version of config file /etc/cron.daily/dpkg ... (Reading database ... 12787 files and directories currently installed.) Preparing to unpack .../dash_0.5.11+git20200708+dd9ef66+really0.5.11+git20200708+dd9ef66-5ubuntu1_s390x.deb ... Unpacking dash (0.5.11+git20200708+dd9ef66+really0.5.11+git20200708+dd9ef66-5ubuntu1) over (0.5.10.2-7) ... Setting up dash (0.5.11+git20200708+dd9ef66+really0.5.11+git20200708+dd9ef66-5ubuntu1) ... (Reading database ... 12787 files and directories currently installed.) Preparing to unpack .../archives/grep_3.6-1_s390x.deb ... Unpacking grep (3.6-1) over (3.4-1) ... Setting up grep (3.6-1) ... (Reading database ... 12788 files and directories currently installed.) Preparing to unpack .../gzip_1.10-2ubuntu3_s390x.deb ... Unpacking gzip (1.10-2ubuntu3) over (1.10-2ubuntu1) ... Setting up gzip (1.10-2ubuntu3) ... (Reading database ... 12788 files and directories currently installed.) Preparing to unpack .../login_1%3a4.8.1-1ubuntu8_s390x.deb ... Unpacking login (1:4.8.1-1ubuntu8) over (1:4.8.1-1ubuntu6) ... Setting up login (1:4.8.1-1ubuntu8) ... Installing new version of config file /etc/login.defs ... (Reading database ... 12788 files and directories currently installed.) Preparing to unpack .../libncurses6_6.2+20201114-2_s390x.deb ... Unpacking libncurses6:s390x (6.2+20201114-2) over (6.2-1) ... Preparing to unpack .../libncursesw6_6.2+20201114-2_s390x.deb ... Unpacking libncursesw6:s390x (6.2+20201114-2) over (6.2-1) ... Preparing to unpack .../libtinfo6_6.2+20201114-2_s390x.deb ... Unpacking libtinfo6:s390x (6.2+20201114-2) over (6.2-1) ... Setting up libtinfo6:s390x (6.2+20201114-2) ... (Reading database ... 12788 files and directories currently installed.) Preparing to unpack .../ncurses-bin_6.2+20201114-2_s390x.deb ... Unpacking ncurses-bin (6.2+20201114-2) over (6.2-1) ... Setting up ncurses-bin (6.2+20201114-2) ... (Reading database ... 12788 files and directories currently installed.) Preparing to unpack .../perl_5.32.0-6_s390x.deb ... Unpacking perl (5.32.0-6) over (5.30.3-4) ... Selecting previously unselected package perl-modules-5.32. Preparing to unpack .../perl-modules-5.32_5.32.0-6_all.deb ... Unpacking perl-modules-5.32 (5.32.0-6) ... Selecting previously unselected package libperl5.32:s390x. Preparing to unpack .../libperl5.32_5.32.0-6_s390x.deb ... Unpacking libperl5.32:s390x (5.32.0-6) ... Preparing to unpack .../perl-base_5.32.0-6_s390x.deb ... Unpacking perl-base (5.32.0-6) over (5.30.3-4) ... Setting up perl-base (5.32.0-6) ... (Reading database ... 14708 files and directories currently installed.) Preparing to unpack .../libgdbm6_1.19-2_s390x.deb ... Unpacking libgdbm6:s390x (1.19-2) over (1.18.1-5.1) ... Preparing to unpack .../libgdbm-compat4_1.19-2_s390x.deb ... Unpacking libgdbm-compat4:s390x (1.19-2) over (1.18.1-5.1) ... Preparing to unpack .../util-linux_2.36.1-1ubuntu2_s390x.deb ... Unpacking util-linux (2.36.1-1ubuntu2) over (2.36-3ubuntu1) ... Setting up util-linux (2.36.1-1ubuntu2) ... (Reading database ... 14709 files and directories currently installed.) Preparing to unpack .../libdebconfclient0_0.256ubuntu1_s390x.deb ... Unpacking libdebconfclient0:s390x (0.256ubuntu1) over (0.252ubuntu1) ... Setting up libdebconfclient0:s390x (0.256ubuntu1) ... (Reading database ... 14709 files and directories currently installed.) Preparing to unpack .../base-passwd_3.5.48_s390x.deb ... Unpacking base-passwd (3.5.48) over (3.5.47) ... Setting up base-passwd (3.5.48) ... Changing home-directory of irc from /var/run/ircd to /run/ircd 1 changes have been made, rewriting files Writing passwd-file to /etc/passwd Writing shadow-file to /etc/shadow Writing group-file to /etc/group (Reading database ... 14709 files and directories currently installed.) Preparing to unpack .../init-system-helpers_1.60_all.deb ... Unpacking init-system-helpers (1.60) over (1.58) ... Setting up init-system-helpers (1.60) ... (Reading database ... 14708 files and directories currently installed.) Preparing to unpack .../libc-bin_2.32-0ubuntu6_s390x.deb ... Unpacking libc-bin (2.32-0ubuntu6) over (2.32-0ubuntu3) ... Setting up libc-bin (2.32-0ubuntu6) ... (Reading database ... 14708 files and directories currently installed.) Preparing to unpack .../ncurses-base_6.2+20201114-2_all.deb ... Unpacking ncurses-base (6.2+20201114-2) over (6.2-1) ... Setting up ncurses-base (6.2+20201114-2) ... (Reading database ... 14708 files and directories currently installed.) Preparing to unpack .../sysvinit-utils_2.96-5ubuntu1_s390x.deb ... Unpacking sysvinit-utils (2.96-5ubuntu1) over (2.96-3ubuntu1) ... Setting up sysvinit-utils (2.96-5ubuntu1) ... Selecting previously unselected package libxxhash0:s390x. (Reading database ... 14708 files and directories currently installed.) Preparing to unpack .../libxxhash0_0.8.0-2_s390x.deb ... Unpacking libxxhash0:s390x (0.8.0-2) ... Setting up libxxhash0:s390x (0.8.0-2) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libgcrypt20_1.8.7-2ubuntu1_s390x.deb ... Unpacking libgcrypt20:s390x (1.8.7-2ubuntu1) over (1.8.5-5ubuntu2) ... Setting up libgcrypt20:s390x (1.8.7-2ubuntu1) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../liblz4-1_1.9.3-0ubuntu1_s390x.deb ... Unpacking liblz4-1:s390x (1.9.3-0ubuntu1) over (1.9.2-2) ... Setting up liblz4-1:s390x (1.9.3-0ubuntu1) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../systemd-sysv_247.1-4ubuntu1_s390x.deb ... Unpacking systemd-sysv (247.1-4ubuntu1) over (246.6-1ubuntu1) ... Preparing to unpack .../libaudit-common_1%3a2.8.5-3ubuntu3_all.deb ... Unpacking libaudit-common (1:2.8.5-3ubuntu3) over (1:2.8.5-3ubuntu1) ... Setting up libaudit-common (1:2.8.5-3ubuntu3) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libcap-ng0_0.7.9-2.2build1_s390x.deb ... Unpacking libcap-ng0:s390x (0.7.9-2.2build1) over (0.7.9-2.2) ... Setting up libcap-ng0:s390x (0.7.9-2.2build1) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libaudit1_1%3a2.8.5-3ubuntu3_s390x.deb ... Unpacking libaudit1:s390x (1:2.8.5-3ubuntu3) over (1:2.8.5-3ubuntu1) ... Setting up libaudit1:s390x (1:2.8.5-3ubuntu3) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libpcre2-8-0_10.35-2ubuntu1_s390x.deb ... Unpacking libpcre2-8-0:s390x (10.35-2ubuntu1) over (10.34-7) ... Setting up libpcre2-8-0:s390x (10.35-2ubuntu1) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libselinux1_3.1-2build2_s390x.deb ... Unpacking libselinux1:s390x (3.1-2build2) over (3.1-2) ... Setting up libselinux1:s390x (3.1-2build2) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libsemanage-common_3.1-1build2_all.deb ... Unpacking libsemanage-common (3.1-1build2) over (3.1-1) ... Setting up libsemanage-common (3.1-1build2) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libsemanage1_3.1-1build2_s390x.deb ... Unpacking libsemanage1:s390x (3.1-1build2) over (3.1-1) ... Setting up libsemanage1:s390x (3.1-1build2) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../passwd_1%3a4.8.1-1ubuntu8_s390x.deb ... Unpacking passwd (1:4.8.1-1ubuntu8) over (1:4.8.1-1ubuntu6) ... Setting up passwd (1:4.8.1-1ubuntu8) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../adduser_3.118ubuntu5_all.deb ... Unpacking adduser (3.118ubuntu5) over (3.118ubuntu2) ... Setting up adduser (3.118ubuntu5) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../systemd-timesyncd_247.1-4ubuntu1_s390x.deb ... Unpacking systemd-timesyncd (247.1-4ubuntu1) over (246.6-1ubuntu1) ... Preparing to unpack .../libapparmor1_3.0.0-0ubuntu5_s390x.deb ... Unpacking libapparmor1:s390x (3.0.0-0ubuntu5) over (3.0.0-0ubuntu1) ... Preparing to unpack .../libblkid1_2.36.1-1ubuntu2_s390x.deb ... Unpacking libblkid1:s390x (2.36.1-1ubuntu2) over (2.36-3ubuntu1) ... Setting up libblkid1:s390x (2.36.1-1ubuntu2) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libudev1_247.1-4ubuntu1_s390x.deb ... Unpacking libudev1:s390x (247.1-4ubuntu1) over (246.6-1ubuntu1) ... Setting up libudev1:s390x (247.1-4ubuntu1) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libdevmapper1.02.1_2%3a1.02.167-1ubuntu4_s390x.deb ... Unpacking libdevmapper1.02.1:s390x (2:1.02.167-1ubuntu4) over (2:1.02.167-1ubuntu3) ... Preparing to unpack .../libuuid1_2.36.1-1ubuntu2_s390x.deb ... Unpacking libuuid1:s390x (2.36.1-1ubuntu2) over (2.36-3ubuntu1) ... Setting up libuuid1:s390x (2.36.1-1ubuntu2) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libcryptsetup12_2%3a2.3.4-1ubuntu1_s390x.deb ... Unpacking libcryptsetup12:s390x (2:2.3.4-1ubuntu1) over (2:2.3.3-1ubuntu6) ... Preparing to unpack .../libidn2-0_2.3.0-5_s390x.deb ... Unpacking libidn2-0:s390x (2.3.0-5) over (2.3.0-1) ... Setting up libidn2-0:s390x (2.3.0-5) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libp11-kit0_0.23.22-1_s390x.deb ... Unpacking libp11-kit0:s390x (0.23.22-1) over (0.23.21-2build1) ... Setting up libp11-kit0:s390x (0.23.22-1) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libgnutls30_3.7.0-5ubuntu1_s390x.deb ... Unpacking libgnutls30:s390x (3.7.0-5ubuntu1) over (3.6.15-4ubuntu2) ... Setting up libgnutls30:s390x (3.7.0-5ubuntu1) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../libip4tc2_1.8.5-3ubuntu4_s390x.deb ... Unpacking libip4tc2:s390x (1.8.5-3ubuntu4) over (1.8.5-3ubuntu1) ... Preparing to unpack .../libseccomp2_2.4.3-1ubuntu6_s390x.deb ... Unpacking libseccomp2:s390x (2.4.3-1ubuntu6) over (2.4.3-1ubuntu4) ... Setting up libseccomp2:s390x (2.4.3-1ubuntu6) ... (Reading database ... 14713 files and directories currently installed.) Preparing to unpack .../mount_2.36.1-1ubuntu2_s390x.deb ... Unpacking mount (2.36.1-1ubuntu2) over (2.36-3ubuntu1) ... Preparing to unpack .../systemd_247.1-4ubuntu1_s390x.deb ... Unpacking systemd (247.1-4ubuntu1) over (246.6-1ubuntu1) ... Preparing to unpack .../libsystemd0_247.1-4ubuntu1_s390x.deb ... Unpacking libsystemd0:s390x (247.1-4ubuntu1) over (246.6-1ubuntu1) ... Setting up libsystemd0:s390x (247.1-4ubuntu1) ... (Reading database ... 14721 files and directories currently installed.) Preparing to unpack .../libapt-pkg6.0_2.1.18_s390x.deb ... Unpacking libapt-pkg6.0:s390x (2.1.18) over (2.1.10) ... Setting up libapt-pkg6.0:s390x (2.1.18) ... (Reading database ... 14721 files and directories currently installed.) Preparing to unpack .../gpgv_2.2.20-1ubuntu2_s390x.deb ... Unpacking gpgv (2.2.20-1ubuntu2) over (2.2.20-1ubuntu1) ... Setting up gpgv (2.2.20-1ubuntu2) ... (Reading database ... 14721 files and directories currently installed.) Preparing to unpack .../archives/apt_2.1.18_s390x.deb ... Unpacking apt (2.1.18) over (2.1.10) ... Setting up apt (2.1.18) ... Installing new version of config file /etc/kernel/postinst.d/apt-auto-removal ... Setting up libapparmor1:s390x (3.0.0-0ubuntu5) ... Setting up libdevmapper1.02.1:s390x (2:1.02.167-1ubuntu4) ... Setting up libssl1.1:s390x (1.1.1f-1ubuntu5) ... Setting up libcryptsetup12:s390x (2:2.3.4-1ubuntu1) ... Setting up libip4tc2:s390x (1.8.5-3ubuntu4) ... Setting up mount (2.36.1-1ubuntu2) ... Setting up systemd-timesyncd (247.1-4ubuntu1) ... Setting up systemd (247.1-4ubuntu1) ... Installing new version of config file /etc/systemd/logind.conf ... Installing new version of config file /etc/systemd/resolved.conf ... Installing new version of config file /etc/systemd/system.conf ... Initializing machine ID from random generator. Removing obsolete conffile /etc/pam.d/systemd-user ... Setting up systemd-sysv (247.1-4ubuntu1) ... (Reading database ... 14725 files and directories currently installed.) Preparing to unpack .../archives/init_1.60_s390x.deb ... Unpacking init (1.60) over (1.58) ... Preparing to unpack .../libacl1_2.2.53-9_s390x.deb ... Unpacking libacl1:s390x (2.2.53-9) over (2.2.53-8) ... Setting up libacl1:s390x (2.2.53-9) ... (Reading database ... 14725 files and directories currently installed.) Preparing to unpack .../libattr1_1%3a2.4.48-6_s390x.deb ... Unpacking libattr1:s390x (1:2.4.48-6) over (1:2.4.48-5) ... Setting up libattr1:s390x (1:2.4.48-6) ... (Reading database ... 14725 files and directories currently installed.) Preparing to unpack .../libmount1_2.36.1-1ubuntu2_s390x.deb ... Unpacking libmount1:s390x (2.36.1-1ubuntu2) over (2.36-3ubuntu1) ... Setting up libmount1:s390x (2.36.1-1ubuntu2) ... (Reading database ... 14725 files and directories currently installed.) Preparing to unpack .../libsmartcols1_2.36.1-1ubuntu2_s390x.deb ... Unpacking libsmartcols1:s390x (2.36.1-1ubuntu2) over (2.36-3ubuntu1) ... Setting up libsmartcols1:s390x (2.36.1-1ubuntu2) ... (Reading database ... 14725 files and directories currently installed.) Preparing to unpack .../00-sensible-utils_0.0.14_all.deb ... Unpacking sensible-utils (0.0.14) over (0.0.13) ... Preparing to unpack .../01-libcap2_1%3a2.44-1_s390x.deb ... Unpacking libcap2:s390x (1:2.44-1) over (1:2.43-1) ... Preparing to unpack .../02-readline-common_8.1-1_all.deb ... Unpacking readline-common (8.1-1) over (8.0-4) ... Preparing to unpack .../03-libreadline8_8.1-1_s390x.deb ... Unpacking libreadline8:s390x (8.1-1) over (8.0-4) ... Preparing to unpack .../04-libsqlite3-0_3.34.0-1_s390x.deb ... Unpacking libsqlite3-0:s390x (3.34.0-1) over (3.33.0-1) ... Preparing to unpack .../05-openssl_1.1.1f-1ubuntu5_s390x.deb ... Unpacking openssl (1.1.1f-1ubuntu5) over (1.1.1f-1ubuntu4) ... Preparing to unpack .../06-tzdata_2021a-1ubuntu1_all.deb ... Unpacking tzdata (2021a-1ubuntu1) over (2020b-1ubuntu1) ... Preparing to unpack .../07-dpkg-dev_1.20.7.1ubuntu2_all.deb ... Unpacking dpkg-dev (1.20.7.1ubuntu2) over (1.20.5ubuntu2) ... Preparing to unpack .../08-libdpkg-perl_1.20.7.1ubuntu2_all.deb ... Unpacking libdpkg-perl (1.20.7.1ubuntu2) over (1.20.5ubuntu2) ... Preparing to unpack .../09-patch_2.7.6-7_s390x.deb ... Unpacking patch (2.7.6-7) over (2.7.6-6) ... Selecting previously unselected package lto-disabled-list. Preparing to unpack .../10-lto-disabled-list_1_all.deb ... Unpacking lto-disabled-list (1) ... Preparing to unpack .../11-libfakeroot_1.25.3-1.1_s390x.deb ... Unpacking libfakeroot:s390x (1.25.3-1.1) over (1.25.2-1) ... Preparing to unpack .../12-fakeroot_1.25.3-1.1_s390x.deb ... Unpacking fakeroot (1.25.3-1.1) over (1.25.2-1) ... Preparing to unpack .../13-libnpth0_1.6-3_s390x.deb ... Unpacking libnpth0:s390x (1.6-3) over (1.6-2) ... Preparing to unpack .../14-gpg_2.2.20-1ubuntu2_s390x.deb ... Unpacking gpg (2.2.20-1ubuntu2) over (2.2.20-1ubuntu1) ... Preparing to unpack .../15-gpgconf_2.2.20-1ubuntu2_s390x.deb ... Unpacking gpgconf (2.2.20-1ubuntu2) over (2.2.20-1ubuntu1) ... Preparing to unpack .../16-gpg-agent_2.2.20-1ubuntu2_s390x.deb ... Unpacking gpg-agent (2.2.20-1ubuntu2) over (2.2.20-1ubuntu1) ... Preparing to unpack .../17-liblockfile-bin_1.17-1_s390x.deb ... Unpacking liblockfile-bin (1.17-1) over (1.16-1.1) ... Preparing to unpack .../18-liblockfile1_1.17-1_s390x.deb ... Unpacking liblockfile1:s390x (1.17-1) over (1.16-1.1) ... Setting up lto-disabled-list (1) ... Setting up liblockfile-bin (1.17-1) ... Setting up init (1.60) ... Setting up libtirpc-common (1.3.1-1) ... Setting up perl-modules-5.32 (5.32.0-6) ... Setting up libsqlite3-0:s390x (3.34.0-1) ... Setting up binutils-common:s390x (2.35.50.20210106-1ubuntu2) ... Setting up linux-libc-dev:s390x (5.8.0-36.40+21.04.1) ... Setting up libctf-nobfd0:s390x (2.35.50.20210106-1ubuntu2) ... Setting up libnpth0:s390x (1.6-3) ... Setting up libgomp1:s390x (10.2.1-6ubuntu1) ... Setting up libcap2:s390x (1:2.44-1) ... Setting up libfakeroot:s390x (1.25.3-1.1) ... Setting up libasan6:s390x (10.2.1-6ubuntu1) ... Setting up libkrb5support0:s390x (1.18.3-4) ... Setting up tzdata (2021a-1ubuntu1) ... Current default time zone: 'Etc/UTC' Local time is now: Fri Jan 29 19:28:26 UTC 2021. Universal Time is now: Fri Jan 29 19:28:26 UTC 2021. Run 'dpkg-reconfigure tzdata' if you wish to change it. Setting up fakeroot (1.25.3-1.1) ... Setting up libncurses6:s390x (6.2+20201114-2) ... Setting up libmpc3:s390x (1.2.0-1) ... Setting up libatomic1:s390x (10.2.1-6ubuntu1) ... Setting up patch (2.7.6-7) ... Setting up libncursesw6:s390x (6.2+20201114-2) ... Setting up libk5crypto3:s390x (1.18.3-4) ... Setting up libubsan1:s390x (10.2.1-6ubuntu1) ... Setting up sensible-utils (0.0.14) ... Setting up libcrypt-dev:s390x (1:4.4.17-1ubuntu1) ... Setting up libkrb5-3:s390x (1.18.3-4) ... Setting up liblockfile1:s390x (1.17-1) ... Setting up libbinutils:s390x (2.35.50.20210106-1ubuntu2) ... Setting up libisl23:s390x (0.23-1) ... Setting up libc-dev-bin (2.32-0ubuntu6) ... Setting up openssl (1.1.1f-1ubuntu5) ... Setting up readline-common (8.1-1) ... Setting up libcc1-0:s390x (10.2.1-6ubuntu1) ... Setting up cpp-10 (10.2.1-6ubuntu1) ... Setting up libitm1:s390x (10.2.1-6ubuntu1) ... Setting up libgdbm6:s390x (1.19-2) ... Setting up libctf0:s390x (2.35.50.20210106-1ubuntu2) ... Setting up binutils-s390x-linux-gnu (2.35.50.20210106-1ubuntu2) ... Setting up libgcc-10-dev:s390x (10.2.1-6ubuntu1) ... Setting up libreadline8:s390x (8.1-1) ... Setting up binutils (2.35.50.20210106-1ubuntu2) ... Setting up libgssapi-krb5-2:s390x (1.18.3-4) ... Setting up libgdbm-compat4:s390x (1.19-2) ... Setting up libperl5.32:s390x (5.32.0-6) ... Setting up gcc-10 (10.2.1-6ubuntu1) ... Setting up gpgconf (2.2.20-1ubuntu2) ... Setting up gpg (2.2.20-1ubuntu2) ... Setting up libtirpc3:s390x (1.3.1-1) ... Setting up gpg-agent (2.2.20-1ubuntu2) ... Setting up perl (5.32.0-6) ... Setting up libtirpc-dev:s390x (1.3.1-1) ... Setting up libdpkg-perl (1.20.7.1ubuntu2) ... Setting up libc6-dev:s390x (2.32-0ubuntu6) ... Setting up libstdc++-10-dev:s390x (10.2.1-6ubuntu1) ... Setting up g++-10 (10.2.1-6ubuntu1) ... Setting up dpkg-dev (1.20.7.1ubuntu2) ... Processing triggers for libc-bin (2.32-0ubuntu6) ... RUN: /usr/share/launchpad-buildd/bin/sbuild-package PACKAGEBUILD-20426680 s390x hirsute -c chroot:build-PACKAGEBUILD-20426680 --arch=s390x --dist=hirsute --nolog consul_1.7.4+dfsg1-1.dsc Initiating build PACKAGEBUILD-20426680 with 4 jobs across 4 processor cores. Kernel reported to sbuild: 4.15.0-134-generic #138-Ubuntu SMP Fri Jan 15 10:53:04 UTC 2021 s390x sbuild (Debian sbuild) 0.75.0 (21 Mar 2018) on bos02-s390x-015.buildd +==============================================================================+ | consul 1.7.4+dfsg1-1 (s390x) Fri, 29 Jan 2021 19:28:27 +0000 | +==============================================================================+ Package: consul Version: 1.7.4+dfsg1-1 Source Version: 1.7.4+dfsg1-1 Distribution: hirsute Machine Architecture: s390x Host Architecture: s390x Build Architecture: s390x Build Type: any I: NOTICE: Log filtering will replace 'home/buildd/build-PACKAGEBUILD-20426680/chroot-autobuild' with '<>' +------------------------------------------------------------------------------+ | Fetch source files | +------------------------------------------------------------------------------+ Local sources ------------- consul_1.7.4+dfsg1-1.dsc exists in .; copying to chroot I: NOTICE: Log filtering will replace 'build/consul-00tgJw/consul-1.7.4+dfsg1' with '<>' I: NOTICE: Log filtering will replace 'build/consul-00tgJw' with '<>' +------------------------------------------------------------------------------+ | Install build-essential | +------------------------------------------------------------------------------+ Setup apt archive ----------------- Merged Build-Depends: build-essential, fakeroot Filtered Build-Depends: build-essential, fakeroot dpkg-deb: building package 'sbuild-build-depends-core-dummy' in '/<>/resolver-RZF8jE/apt_archive/sbuild-build-depends-core-dummy.deb'. dpkg-scanpackages: warning: Packages in archive but missing from override file: dpkg-scanpackages: warning: sbuild-build-depends-core-dummy dpkg-scanpackages: info: Wrote 1 entries to output Packages file. Ign:1 copy:/<>/resolver-RZF8jE/apt_archive ./ InRelease Get:2 copy:/<>/resolver-RZF8jE/apt_archive ./ Release [957 B] Ign:3 copy:/<>/resolver-RZF8jE/apt_archive ./ Release.gpg Get:4 copy:/<>/resolver-RZF8jE/apt_archive ./ Sources [349 B] Get:5 copy:/<>/resolver-RZF8jE/apt_archive ./ Packages [433 B] Fetched 1739 B in 0s (155 kB/s) Reading package lists... Reading package lists... Install core build dependencies (apt-based resolver) ---------------------------------------------------- Installing build dependencies Reading package lists... Building dependency tree... Reading state information... The following packages were automatically installed and are no longer required: libisl22 libperl5.30 perl-modules-5.30 Use 'apt autoremove' to remove them. The following NEW packages will be installed: sbuild-build-depends-core-dummy 0 upgraded, 1 newly installed, 0 to remove and 0 not upgraded. Need to get 852 B of archives. After this operation, 0 B of additional disk space will be used. Get:1 copy:/<>/resolver-RZF8jE/apt_archive ./ sbuild-build-depends-core-dummy 0.invalid.0 [852 B] debconf: delaying package configuration, since apt-utils is not installed Fetched 852 B in 0s (0 B/s) Selecting previously unselected package sbuild-build-depends-core-dummy. (Reading database ... 14766 files and directories currently installed.) Preparing to unpack .../sbuild-build-depends-core-dummy_0.invalid.0_s390x.deb ... Unpacking sbuild-build-depends-core-dummy (0.invalid.0) ... Setting up sbuild-build-depends-core-dummy (0.invalid.0) ... +------------------------------------------------------------------------------+ | Check architectures | +------------------------------------------------------------------------------+ Arch check ok (s390x included in any all) +------------------------------------------------------------------------------+ | Install package build dependencies | +------------------------------------------------------------------------------+ Setup apt archive ----------------- Merged Build-Depends: debhelper-compat (= 12), bash-completion, dh-golang (>= 1.42~), golang-any (>= 2:1.13~), golang-github-asaskevich-govalidator-dev, golang-github-armon-circbuf-dev, golang-github-armon-go-metrics-dev (>= 0.0~git20171117~), golang-github-armon-go-radix-dev, golang-github-azure-go-autorest-dev (>= 10.15.5~), golang-github-bgentry-speakeasy-dev, golang-github-circonus-labs-circonus-gometrics-dev (>= 2.3.1~), golang-github-circonus-labs-circonusllhist-dev, golang-github-datadog-datadog-go-dev, golang-github-davecgh-go-spew-dev, golang-github-denverdino-aliyungo-dev, golang-github-digitalocean-godo-dev, golang-github-docker-go-connections-dev, golang-github-elazarl-go-bindata-assetfs-dev (>= 0.0~git20151224~), golang-github-ghodss-yaml-dev, golang-github-gogo-googleapis-dev, golang-github-gogo-protobuf-dev (>= 1.2.1~), golang-github-golang-snappy-dev, golang-github-googleapis-gnostic-dev, golang-github-google-gofuzz-dev, golang-github-gophercloud-gophercloud-dev, golang-github-gregjones-httpcache-dev, golang-github-hashicorp-go-bexpr-dev, golang-github-hashicorp-go-checkpoint-dev, golang-github-hashicorp-go-cleanhttp-dev (>= 0.5.1~), golang-github-hashicorp-go-discover-dev, golang-github-hashicorp-go-hclog-dev (>= 0.9.2~), golang-github-hashicorp-go-immutable-radix-dev (>= 1.1.0~), golang-github-hashicorp-golang-lru-dev (>= 0.0~git20160207~), golang-github-hashicorp-go-memdb-dev (>= 0.0~git20180224~), golang-github-hashicorp-go-msgpack-dev (>= 0.5.5~), golang-github-hashicorp-go-multierror-dev, golang-github-hashicorp-go-plugin-dev (>= 1.0.1~), golang-github-hashicorp-go-raftchunking-dev, golang-github-hashicorp-go-retryablehttp-dev, golang-github-hashicorp-go-rootcerts-dev (>= 1.0.2~), golang-github-hashicorp-go-sockaddr-dev, golang-github-hashicorp-go-syslog-dev, golang-github-hashicorp-go-uuid-dev, golang-github-hashicorp-go-version-dev, golang-github-hashicorp-hcl-dev, golang-github-hashicorp-hil-dev (>= 0.0~git20160711~), golang-github-hashicorp-logutils-dev, golang-github-hashicorp-memberlist-dev (>= 0.1.5~), golang-github-hashicorp-net-rpc-msgpackrpc-dev, golang-github-hashicorp-raft-boltdb-dev, golang-github-hashicorp-raft-dev (>= 1.1.2~), golang-github-hashicorp-scada-client-dev, golang-github-hashicorp-serf-dev (>= 0.8.4~), golang-github-hashicorp-yamux-dev (>= 0.0~git20151129~), golang-github-inconshreveable-muxado-dev, golang-github-imdario-mergo-dev, golang-github-jefferai-jsonx-dev, golang-github-json-iterator-go-dev, golang-github-kr-text-dev, golang-github-mattn-go-isatty-dev, golang-github-miekg-dns-dev (>= 1.1.26~), golang-github-mitchellh-cli-dev (>= 1.0.0~), golang-github-mitchellh-go-testing-interface-dev, golang-github-mitchellh-copystructure-dev, golang-github-mitchellh-hashstructure-dev, golang-github-mitchellh-mapstructure-dev, golang-github-mitchellh-reflectwalk-dev, golang-github-nytimes-gziphandler-dev, golang-github-packethost-packngo-dev, golang-github-pascaldekloe-goe-dev, golang-github-peterbourgon-diskv-dev, golang-github-pierrec-lz4-dev, golang-github-pmezard-go-difflib-dev, golang-github-ryanuber-columnize-dev, golang-github-ryanuber-go-glob-dev, golang-github-shirou-gopsutil-dev, golang-github-spf13-pflag-dev, golang-golang-x-sys-dev (>= 0.0~git20161012~), golang-gopkg-inf.v0-dev, golang-gopkg-square-go-jose.v2-dev, mockery, golang-github-sap-go-hdb-dev Filtered Build-Depends: debhelper-compat (= 12), bash-completion, dh-golang (>= 1.42~), golang-any (>= 2:1.13~), golang-github-asaskevich-govalidator-dev, golang-github-armon-circbuf-dev, golang-github-armon-go-metrics-dev (>= 0.0~git20171117~), golang-github-armon-go-radix-dev, golang-github-azure-go-autorest-dev (>= 10.15.5~), golang-github-bgentry-speakeasy-dev, golang-github-circonus-labs-circonus-gometrics-dev (>= 2.3.1~), golang-github-circonus-labs-circonusllhist-dev, golang-github-datadog-datadog-go-dev, golang-github-davecgh-go-spew-dev, golang-github-denverdino-aliyungo-dev, golang-github-digitalocean-godo-dev, golang-github-docker-go-connections-dev, golang-github-elazarl-go-bindata-assetfs-dev (>= 0.0~git20151224~), golang-github-ghodss-yaml-dev, golang-github-gogo-googleapis-dev, golang-github-gogo-protobuf-dev (>= 1.2.1~), golang-github-golang-snappy-dev, golang-github-googleapis-gnostic-dev, golang-github-google-gofuzz-dev, golang-github-gophercloud-gophercloud-dev, golang-github-gregjones-httpcache-dev, golang-github-hashicorp-go-bexpr-dev, golang-github-hashicorp-go-checkpoint-dev, golang-github-hashicorp-go-cleanhttp-dev (>= 0.5.1~), golang-github-hashicorp-go-discover-dev, golang-github-hashicorp-go-hclog-dev (>= 0.9.2~), golang-github-hashicorp-go-immutable-radix-dev (>= 1.1.0~), golang-github-hashicorp-golang-lru-dev (>= 0.0~git20160207~), golang-github-hashicorp-go-memdb-dev (>= 0.0~git20180224~), golang-github-hashicorp-go-msgpack-dev (>= 0.5.5~), golang-github-hashicorp-go-multierror-dev, golang-github-hashicorp-go-plugin-dev (>= 1.0.1~), golang-github-hashicorp-go-raftchunking-dev, golang-github-hashicorp-go-retryablehttp-dev, golang-github-hashicorp-go-rootcerts-dev (>= 1.0.2~), golang-github-hashicorp-go-sockaddr-dev, golang-github-hashicorp-go-syslog-dev, golang-github-hashicorp-go-uuid-dev, golang-github-hashicorp-go-version-dev, golang-github-hashicorp-hcl-dev, golang-github-hashicorp-hil-dev (>= 0.0~git20160711~), golang-github-hashicorp-logutils-dev, golang-github-hashicorp-memberlist-dev (>= 0.1.5~), golang-github-hashicorp-net-rpc-msgpackrpc-dev, golang-github-hashicorp-raft-boltdb-dev, golang-github-hashicorp-raft-dev (>= 1.1.2~), golang-github-hashicorp-scada-client-dev, golang-github-hashicorp-serf-dev (>= 0.8.4~), golang-github-hashicorp-yamux-dev (>= 0.0~git20151129~), golang-github-inconshreveable-muxado-dev, golang-github-imdario-mergo-dev, golang-github-jefferai-jsonx-dev, golang-github-json-iterator-go-dev, golang-github-kr-text-dev, golang-github-mattn-go-isatty-dev, golang-github-miekg-dns-dev (>= 1.1.26~), golang-github-mitchellh-cli-dev (>= 1.0.0~), golang-github-mitchellh-go-testing-interface-dev, golang-github-mitchellh-copystructure-dev, golang-github-mitchellh-hashstructure-dev, golang-github-mitchellh-mapstructure-dev, golang-github-mitchellh-reflectwalk-dev, golang-github-nytimes-gziphandler-dev, golang-github-packethost-packngo-dev, golang-github-pascaldekloe-goe-dev, golang-github-peterbourgon-diskv-dev, golang-github-pierrec-lz4-dev, golang-github-pmezard-go-difflib-dev, golang-github-ryanuber-columnize-dev, golang-github-ryanuber-go-glob-dev, golang-github-shirou-gopsutil-dev, golang-github-spf13-pflag-dev, golang-golang-x-sys-dev (>= 0.0~git20161012~), golang-gopkg-inf.v0-dev, golang-gopkg-square-go-jose.v2-dev, mockery, golang-github-sap-go-hdb-dev dpkg-deb: building package 'sbuild-build-depends-consul-dummy' in '/<>/resolver-RZF8jE/apt_archive/sbuild-build-depends-consul-dummy.deb'. dpkg-scanpackages: warning: Packages in archive but missing from override file: dpkg-scanpackages: warning: sbuild-build-depends-consul-dummy sbuild-build-depends-core-dummy dpkg-scanpackages: info: Wrote 2 entries to output Packages file. Ign:1 copy:/<>/resolver-RZF8jE/apt_archive ./ InRelease Get:2 copy:/<>/resolver-RZF8jE/apt_archive ./ Release [969 B] Ign:3 copy:/<>/resolver-RZF8jE/apt_archive ./ Release.gpg Get:4 copy:/<>/resolver-RZF8jE/apt_archive ./ Sources [1404 B] Get:5 copy:/<>/resolver-RZF8jE/apt_archive ./ Packages [1496 B] Fetched 3869 B in 0s (363 kB/s) Reading package lists... Reading package lists... Install consul build dependencies (apt-based resolver) ------------------------------------------------------ Installing build dependencies Reading package lists... Building dependency tree... Reading state information... The following packages were automatically installed and are no longer required: libisl22 libperl5.30 perl-modules-5.30 Use 'apt autoremove' to remove them. The following additional packages will be installed: autoconf automake autopoint autotools-dev bash-completion bsdextrautils debhelper dh-autoreconf dh-golang dh-strip-nondeterminism dwz file gettext gettext-base golang-1.15-go golang-1.15-src golang-any golang-ginkgo-dev golang-github-alecthomas-units-dev golang-github-armon-circbuf-dev golang-github-armon-go-metrics-dev golang-github-armon-go-radix-dev golang-github-asaskevich-govalidator-dev golang-github-aws-aws-sdk-go-dev golang-github-azure-go-autorest-dev golang-github-beorn7-perks-dev golang-github-bgentry-speakeasy-dev golang-github-boltdb-bolt-dev golang-github-bradfitz-gomemcache-dev golang-github-cespare-xxhash-dev golang-github-circonus-labs-circonus-gometrics-dev golang-github-circonus-labs-circonusllhist-dev golang-github-creack-pty-dev golang-github-datadog-datadog-go-dev golang-github-davecgh-go-spew-dev golang-github-denverdino-aliyungo-dev golang-github-dgrijalva-jwt-go-dev golang-github-digitalocean-godo-dev golang-github-dimchansky-utfbom-dev golang-github-docker-go-connections-dev golang-github-docopt-docopt-go-dev golang-github-elazarl-go-bindata-assetfs-dev golang-github-fatih-color-dev golang-github-fsnotify-fsnotify-dev golang-github-garyburd-redigo-dev golang-github-ghodss-yaml-dev golang-github-go-kit-kit-dev golang-github-go-logfmt-logfmt-dev golang-github-go-stack-stack-dev golang-github-go-test-deep-dev golang-github-gogo-googleapis-dev golang-github-gogo-protobuf-dev golang-github-golang-mock-dev golang-github-golang-snappy-dev golang-github-google-btree-dev golang-github-google-go-cmp-dev golang-github-google-go-querystring-dev golang-github-google-gofuzz-dev golang-github-google-uuid-dev golang-github-googleapis-gax-go-dev golang-github-googleapis-gnostic-dev golang-github-gophercloud-gophercloud-dev golang-github-gregjones-httpcache-dev golang-github-hashicorp-errwrap-dev golang-github-hashicorp-go-bexpr-dev golang-github-hashicorp-go-checkpoint-dev golang-github-hashicorp-go-cleanhttp-dev golang-github-hashicorp-go-discover-dev golang-github-hashicorp-go-hclog-dev golang-github-hashicorp-go-immutable-radix-dev golang-github-hashicorp-go-memdb-dev golang-github-hashicorp-go-msgpack-dev golang-github-hashicorp-go-multierror-dev golang-github-hashicorp-go-plugin-dev golang-github-hashicorp-go-raftchunking-dev golang-github-hashicorp-go-retryablehttp-dev golang-github-hashicorp-go-rootcerts-dev golang-github-hashicorp-go-sockaddr-dev golang-github-hashicorp-go-syslog-dev golang-github-hashicorp-go-uuid-dev golang-github-hashicorp-go-version-dev golang-github-hashicorp-golang-lru-dev golang-github-hashicorp-hcl-dev golang-github-hashicorp-hil-dev golang-github-hashicorp-logutils-dev golang-github-hashicorp-mdns-dev golang-github-hashicorp-memberlist-dev golang-github-hashicorp-net-rpc-msgpackrpc-dev golang-github-hashicorp-raft-boltdb-dev golang-github-hashicorp-raft-dev golang-github-hashicorp-scada-client-dev golang-github-hashicorp-serf-dev golang-github-hashicorp-yamux-dev golang-github-imdario-mergo-dev golang-github-inconshreveable-muxado-dev golang-github-influxdata-tail-dev golang-github-jeffail-gabs-dev golang-github-jefferai-jsonx-dev golang-github-jmespath-go-jmespath-dev golang-github-jpillora-backoff-dev golang-github-json-iterator-go-dev golang-github-julienschmidt-httprouter-dev golang-github-kr-pretty-dev golang-github-kr-text-dev golang-github-mattn-go-colorable-dev golang-github-mattn-go-isatty-dev golang-github-miekg-dns-dev golang-github-mitchellh-cli-dev golang-github-mitchellh-copystructure-dev golang-github-mitchellh-go-homedir-dev golang-github-mitchellh-go-testing-interface-dev golang-github-mitchellh-hashstructure-dev golang-github-mitchellh-mapstructure-dev golang-github-mitchellh-reflectwalk-dev golang-github-modern-go-concurrent-dev golang-github-modern-go-reflect2-dev golang-github-mwitkow-go-conntrack-dev golang-github-nytimes-gziphandler-dev golang-github-oklog-run-dev golang-github-onsi-ginkgo-dev golang-github-opencontainers-runc-dev golang-github-opentracing-opentracing-go-dev golang-github-packethost-packngo-dev golang-github-pascaldekloe-goe-dev golang-github-peterbourgon-diskv-dev golang-github-pierrec-lz4-dev golang-github-pkg-errors-dev golang-github-pmezard-go-difflib-dev golang-github-posener-complete-dev golang-github-prometheus-client-golang-dev golang-github-prometheus-client-model-dev golang-github-prometheus-common-dev golang-github-ryanuber-columnize-dev golang-github-ryanuber-go-glob-dev golang-github-sap-go-hdb-dev golang-github-shirou-gopsutil-dev golang-github-sirupsen-logrus-dev golang-github-spf13-pflag-dev golang-github-stretchr-objx-dev golang-github-stretchr-testify-dev golang-github-syndtr-goleveldb-dev golang-github-tv42-httpunix-dev golang-github-ugorji-go-codec-dev golang-github-ugorji-go-msgpack-dev golang-github-vmihailenco-tagparser-dev golang-github-vmware-govmomi-dev golang-glog-dev golang-go golang-go.opencensus-dev golang-golang-x-crypto-dev golang-golang-x-net-dev golang-golang-x-oauth2-dev golang-golang-x-oauth2-google-dev golang-golang-x-sync-dev golang-golang-x-sys-dev golang-golang-x-text-dev golang-golang-x-time-dev golang-golang-x-xerrors-dev golang-gomega-dev golang-google-api-dev golang-google-cloud-compute-metadata-dev golang-google-genproto-dev golang-google-grpc-dev golang-gopkg-alecthomas-kingpin.v2-dev golang-gopkg-check.v1-dev golang-gopkg-inf.v0-dev golang-gopkg-mgo.v2-dev golang-gopkg-square-go-jose.v2-dev golang-gopkg-tomb.v1-dev golang-gopkg-tomb.v2-dev golang-gopkg-vmihailenco-msgpack.v2-dev golang-gopkg-yaml.v2-dev golang-goprotobuf-dev golang-procfs-dev golang-protobuf-extensions-dev golang-src golang-x-text-dev groff-base intltool-debian iproute2 libarchive-zip-perl libbpf0 libbsd0 libcap2-bin libdebhelper-perl libelf1 libfile-stripnondeterminism-perl libicu67 libmagic-mgc libmagic1 libmnl0 libpipeline1 libprotobuf-dev libprotobuf-lite23 libprotobuf23 libprotoc23 libsasl2-2 libsasl2-dev libsasl2-modules-db libsigsegv2 libsub-override-perl libtool libuchardet0 libxml2 libxtables12 m4 man-db mockery po-debconf protobuf-compiler zlib1g-dev Suggested packages: autoconf-archive gnu-standards autoconf-doc dh-make gettext-doc libasprintf-dev libgettextpo-dev bzr | brz git mercurial subversion mockgen golang-google-appengine-dev groff iproute2-doc libtool-doc gfortran | fortran95-compiler gcj-jdk m4-doc apparmor less www-browser libmail-box-perl protobuf-mode-el Recommended packages: curl | wget | lynx pkg-config gogoprotobuf libatm1 libpam-cap libarchive-cpio-perl libsasl2-modules libltdl-dev libmail-sendmail-perl The following NEW packages will be installed: autoconf automake autopoint autotools-dev bash-completion bsdextrautils debhelper dh-autoreconf dh-golang dh-strip-nondeterminism dwz file gettext gettext-base golang-1.15-go golang-1.15-src golang-any golang-ginkgo-dev golang-github-alecthomas-units-dev golang-github-armon-circbuf-dev golang-github-armon-go-metrics-dev golang-github-armon-go-radix-dev golang-github-asaskevich-govalidator-dev golang-github-aws-aws-sdk-go-dev golang-github-azure-go-autorest-dev golang-github-beorn7-perks-dev golang-github-bgentry-speakeasy-dev golang-github-boltdb-bolt-dev golang-github-bradfitz-gomemcache-dev golang-github-cespare-xxhash-dev golang-github-circonus-labs-circonus-gometrics-dev golang-github-circonus-labs-circonusllhist-dev golang-github-creack-pty-dev golang-github-datadog-datadog-go-dev golang-github-davecgh-go-spew-dev golang-github-denverdino-aliyungo-dev golang-github-dgrijalva-jwt-go-dev golang-github-digitalocean-godo-dev golang-github-dimchansky-utfbom-dev golang-github-docker-go-connections-dev golang-github-docopt-docopt-go-dev golang-github-elazarl-go-bindata-assetfs-dev golang-github-fatih-color-dev golang-github-fsnotify-fsnotify-dev golang-github-garyburd-redigo-dev golang-github-ghodss-yaml-dev golang-github-go-kit-kit-dev golang-github-go-logfmt-logfmt-dev golang-github-go-stack-stack-dev golang-github-go-test-deep-dev golang-github-gogo-googleapis-dev golang-github-gogo-protobuf-dev golang-github-golang-mock-dev golang-github-golang-snappy-dev golang-github-google-btree-dev golang-github-google-go-cmp-dev golang-github-google-go-querystring-dev golang-github-google-gofuzz-dev golang-github-google-uuid-dev golang-github-googleapis-gax-go-dev golang-github-googleapis-gnostic-dev golang-github-gophercloud-gophercloud-dev golang-github-gregjones-httpcache-dev golang-github-hashicorp-errwrap-dev golang-github-hashicorp-go-bexpr-dev golang-github-hashicorp-go-checkpoint-dev golang-github-hashicorp-go-cleanhttp-dev golang-github-hashicorp-go-discover-dev golang-github-hashicorp-go-hclog-dev golang-github-hashicorp-go-immutable-radix-dev golang-github-hashicorp-go-memdb-dev golang-github-hashicorp-go-msgpack-dev golang-github-hashicorp-go-multierror-dev golang-github-hashicorp-go-plugin-dev golang-github-hashicorp-go-raftchunking-dev golang-github-hashicorp-go-retryablehttp-dev golang-github-hashicorp-go-rootcerts-dev golang-github-hashicorp-go-sockaddr-dev golang-github-hashicorp-go-syslog-dev golang-github-hashicorp-go-uuid-dev golang-github-hashicorp-go-version-dev golang-github-hashicorp-golang-lru-dev golang-github-hashicorp-hcl-dev golang-github-hashicorp-hil-dev golang-github-hashicorp-logutils-dev golang-github-hashicorp-mdns-dev golang-github-hashicorp-memberlist-dev golang-github-hashicorp-net-rpc-msgpackrpc-dev golang-github-hashicorp-raft-boltdb-dev golang-github-hashicorp-raft-dev golang-github-hashicorp-scada-client-dev golang-github-hashicorp-serf-dev golang-github-hashicorp-yamux-dev golang-github-imdario-mergo-dev golang-github-inconshreveable-muxado-dev golang-github-influxdata-tail-dev golang-github-jeffail-gabs-dev golang-github-jefferai-jsonx-dev golang-github-jmespath-go-jmespath-dev golang-github-jpillora-backoff-dev golang-github-json-iterator-go-dev golang-github-julienschmidt-httprouter-dev golang-github-kr-pretty-dev golang-github-kr-text-dev golang-github-mattn-go-colorable-dev golang-github-mattn-go-isatty-dev golang-github-miekg-dns-dev golang-github-mitchellh-cli-dev golang-github-mitchellh-copystructure-dev golang-github-mitchellh-go-homedir-dev golang-github-mitchellh-go-testing-interface-dev golang-github-mitchellh-hashstructure-dev golang-github-mitchellh-mapstructure-dev golang-github-mitchellh-reflectwalk-dev golang-github-modern-go-concurrent-dev golang-github-modern-go-reflect2-dev golang-github-mwitkow-go-conntrack-dev golang-github-nytimes-gziphandler-dev golang-github-oklog-run-dev golang-github-onsi-ginkgo-dev golang-github-opencontainers-runc-dev golang-github-opentracing-opentracing-go-dev golang-github-packethost-packngo-dev golang-github-pascaldekloe-goe-dev golang-github-peterbourgon-diskv-dev golang-github-pierrec-lz4-dev golang-github-pkg-errors-dev golang-github-pmezard-go-difflib-dev golang-github-posener-complete-dev golang-github-prometheus-client-golang-dev golang-github-prometheus-client-model-dev golang-github-prometheus-common-dev golang-github-ryanuber-columnize-dev golang-github-ryanuber-go-glob-dev golang-github-sap-go-hdb-dev golang-github-shirou-gopsutil-dev golang-github-sirupsen-logrus-dev golang-github-spf13-pflag-dev golang-github-stretchr-objx-dev golang-github-stretchr-testify-dev golang-github-syndtr-goleveldb-dev golang-github-tv42-httpunix-dev golang-github-ugorji-go-codec-dev golang-github-ugorji-go-msgpack-dev golang-github-vmihailenco-tagparser-dev golang-github-vmware-govmomi-dev golang-glog-dev golang-go golang-go.opencensus-dev golang-golang-x-crypto-dev golang-golang-x-net-dev golang-golang-x-oauth2-dev golang-golang-x-oauth2-google-dev golang-golang-x-sync-dev golang-golang-x-sys-dev golang-golang-x-text-dev golang-golang-x-time-dev golang-golang-x-xerrors-dev golang-gomega-dev golang-google-api-dev golang-google-cloud-compute-metadata-dev golang-google-genproto-dev golang-google-grpc-dev golang-gopkg-alecthomas-kingpin.v2-dev golang-gopkg-check.v1-dev golang-gopkg-inf.v0-dev golang-gopkg-mgo.v2-dev golang-gopkg-square-go-jose.v2-dev golang-gopkg-tomb.v1-dev golang-gopkg-tomb.v2-dev golang-gopkg-vmihailenco-msgpack.v2-dev golang-gopkg-yaml.v2-dev golang-goprotobuf-dev golang-procfs-dev golang-protobuf-extensions-dev golang-src golang-x-text-dev groff-base intltool-debian iproute2 libarchive-zip-perl libbpf0 libbsd0 libcap2-bin libdebhelper-perl libelf1 libfile-stripnondeterminism-perl libicu67 libmagic-mgc libmagic1 libmnl0 libpipeline1 libprotobuf-dev libprotobuf-lite23 libprotobuf23 libprotoc23 libsasl2-2 libsasl2-dev libsasl2-modules-db libsigsegv2 libsub-override-perl libtool libuchardet0 libxml2 libxtables12 m4 man-db mockery po-debconf protobuf-compiler sbuild-build-depends-consul-dummy zlib1g-dev 0 upgraded, 212 newly installed, 0 to remove and 0 not upgraded. Need to get 117 MB of archives. After this operation, 870 MB of additional disk space will be used. Get:1 copy:/<>/resolver-RZF8jE/apt_archive ./ sbuild-build-depends-consul-dummy 0.invalid.0 [1740 B] Get:2 http://ftpmaster.internal/ubuntu hirsute/main s390x bsdextrautils s390x 2.36.1-1ubuntu2 [77.5 kB] Get:3 http://ftpmaster.internal/ubuntu hirsute/main s390x libuchardet0 s390x 0.0.7-1 [67.5 kB] Get:4 http://ftpmaster.internal/ubuntu hirsute/main s390x groff-base s390x 1.22.4-5 [812 kB] Get:5 http://ftpmaster.internal/ubuntu hirsute/main s390x libpipeline1 s390x 1.5.3-1 [28.7 kB] Get:6 http://ftpmaster.internal/ubuntu hirsute/main s390x man-db s390x 2.9.3-2 [1106 kB] Get:7 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-golang-x-sys-dev all 0.0~git20201223.0d417f6-1 [297 kB] Get:8 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-sirupsen-logrus-dev all 1.7.0-2 [41.5 kB] Get:9 http://ftpmaster.internal/ubuntu hirsute/main s390x libmagic-mgc s390x 1:5.39-3 [229 kB] Get:10 http://ftpmaster.internal/ubuntu hirsute/main s390x libmagic1 s390x 1:5.39-3 [83.6 kB] Get:11 http://ftpmaster.internal/ubuntu hirsute/main s390x file s390x 1:5.39-3 [24.0 kB] Get:12 http://ftpmaster.internal/ubuntu hirsute/main s390x libelf1 s390x 0.182-3 [48.4 kB] Get:13 http://ftpmaster.internal/ubuntu hirsute/main s390x libbpf0 s390x 1:0.3-2 [109 kB] Get:14 http://ftpmaster.internal/ubuntu hirsute/main s390x libbsd0 s390x 0.10.0-1 [43.7 kB] Get:15 http://ftpmaster.internal/ubuntu hirsute/main s390x libmnl0 s390x 1.0.4-3 [12.3 kB] Get:16 http://ftpmaster.internal/ubuntu hirsute/main s390x libxtables12 s390x 1.8.5-3ubuntu4 [27.9 kB] Get:17 http://ftpmaster.internal/ubuntu hirsute/main s390x libcap2-bin s390x 1:2.44-1 [24.5 kB] Get:18 http://ftpmaster.internal/ubuntu hirsute/main s390x iproute2 s390x 5.10.0-2ubuntu1 [888 kB] Get:19 http://ftpmaster.internal/ubuntu hirsute/main s390x libicu67 s390x 67.1-5 [8472 kB] Get:20 http://ftpmaster.internal/ubuntu hirsute/main s390x libxml2 s390x 2.9.10+dfsg-6.3build1 [645 kB] Get:21 http://ftpmaster.internal/ubuntu hirsute/main s390x bash-completion all 1:2.11-2ubuntu1 [179 kB] Get:22 http://ftpmaster.internal/ubuntu hirsute/main s390x gettext-base s390x 0.21-3ubuntu2 [41.0 kB] Get:23 http://ftpmaster.internal/ubuntu hirsute/main s390x libsigsegv2 s390x 2.12-3 [14.0 kB] Get:24 http://ftpmaster.internal/ubuntu hirsute/main s390x m4 s390x 1.4.18-5 [209 kB] Get:25 http://ftpmaster.internal/ubuntu hirsute/main s390x autoconf all 2.69-14 [293 kB] Get:26 http://ftpmaster.internal/ubuntu hirsute/main s390x autotools-dev all 20180224.1+nmu1 [39.4 kB] Get:27 http://ftpmaster.internal/ubuntu hirsute/main s390x automake all 1:1.16.3-2ubuntu1 [552 kB] Get:28 http://ftpmaster.internal/ubuntu hirsute/main s390x autopoint all 0.21-3ubuntu2 [422 kB] Get:29 http://ftpmaster.internal/ubuntu hirsute/main s390x libtool all 2.4.6-15 [161 kB] Get:30 http://ftpmaster.internal/ubuntu hirsute/main s390x dh-autoreconf all 19 [16.1 kB] Get:31 http://ftpmaster.internal/ubuntu hirsute/main s390x libdebhelper-perl all 13.3.1ubuntu1 [62.2 kB] Get:32 http://ftpmaster.internal/ubuntu hirsute/main s390x libarchive-zip-perl all 1.68-1 [90.2 kB] Get:33 http://ftpmaster.internal/ubuntu hirsute/main s390x libsub-override-perl all 0.09-2 [9532 B] Get:34 http://ftpmaster.internal/ubuntu hirsute/main s390x libfile-stripnondeterminism-perl all 1.10.0-1 [17.0 kB] Get:35 http://ftpmaster.internal/ubuntu hirsute/main s390x dh-strip-nondeterminism all 1.10.0-1 [5228 B] Get:36 http://ftpmaster.internal/ubuntu hirsute/main s390x dwz s390x 0.13+20210118-1 [168 kB] Get:37 http://ftpmaster.internal/ubuntu hirsute/main s390x gettext s390x 0.21-3ubuntu2 [869 kB] Get:38 http://ftpmaster.internal/ubuntu hirsute/main s390x intltool-debian all 0.35.0+20060710.5 [24.9 kB] Get:39 http://ftpmaster.internal/ubuntu hirsute/main s390x po-debconf all 1.0.21+nmu1 [233 kB] Get:40 http://ftpmaster.internal/ubuntu hirsute/main s390x debhelper all 13.3.1ubuntu1 [882 kB] Get:41 http://ftpmaster.internal/ubuntu hirsute/main s390x golang-1.15-src s390x 1.15.4-1ubuntu2 [13.8 MB] Get:42 http://ftpmaster.internal/ubuntu hirsute/main s390x golang-1.15-go s390x 1.15.4-1ubuntu2 [44.1 MB] Get:43 http://ftpmaster.internal/ubuntu hirsute/main s390x golang-src s390x 2:1.15~1 [4088 B] Get:44 http://ftpmaster.internal/ubuntu hirsute/main s390x golang-go s390x 2:1.15~1 [22.0 kB] Get:45 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-any s390x 2:1.15~1 [3032 B] Get:46 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-creack-pty-dev all 1.1.9-2 [8048 B] Get:47 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-dimchansky-utfbom-dev all 1.1.1-1 [5520 B] Get:48 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-dgrijalva-jwt-go-dev all 3.2.0-3 [33.5 kB] Get:49 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-go-logfmt-logfmt-dev all 0.5.0-2 [12.8 kB] Get:50 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-go-stack-stack-dev all 1.8.0-1 [8736 B] Get:51 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-davecgh-go-spew-dev all 1.1.1-2 [29.7 kB] Get:52 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-pmezard-go-difflib-dev all 1.0.0-3 [12.4 kB] Get:53 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-stretchr-objx-dev all 0.3.0-1 [25.4 kB] Get:54 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-gopkg-yaml.v2-dev all 2.3.0-1 [60.8 kB] Get:55 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-stretchr-testify-dev all 1.4.0+ds-1 [53.0 kB] Get:56 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-opentracing-opentracing-go-dev all 1.2.0-1 [28.0 kB] Get:57 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-golang-x-text-dev all 0.3.5-1 [3861 kB] Get:58 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-golang-x-net-dev all 1:0.0+git20200226.491c5fc+dfsg-1 [638 kB] Get:59 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-golang-x-time-dev all 0.0+git20200630.3af7569-1 [10.1 kB] Get:60 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-golang-mock-dev all 1.3.1-2build1 [35.2 kB] Get:61 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-golang-x-xerrors-dev all 0.0~git20191204.9bdfabe-1 [14.3 kB] Get:62 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-google-go-cmp-dev all 0.4.0-1 [69.5 kB] Get:63 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-glog-dev all 0.0~git20160126.23def4e-3 [17.3 kB] Get:64 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-golang-x-oauth2-dev all 0.0~git20190604.0f29369-2 [31.8 kB] Get:65 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-google-cloud-compute-metadata-dev all 0.56.0-1 [17.7 kB] Get:66 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-golang-x-oauth2-google-dev all 0.0~git20190604.0f29369-2 [13.3 kB] Get:67 http://ftpmaster.internal/ubuntu hirsute/main s390x zlib1g-dev s390x 1:1.2.11.dfsg-2ubuntu4 [166 kB] Get:68 http://ftpmaster.internal/ubuntu hirsute/main s390x libprotobuf23 s390x 3.12.4-1ubuntu1 [778 kB] Get:69 http://ftpmaster.internal/ubuntu hirsute/main s390x libprotobuf-lite23 s390x 3.12.4-1ubuntu1 [176 kB] Get:70 http://ftpmaster.internal/ubuntu hirsute/main s390x libprotobuf-dev s390x 3.12.4-1ubuntu1 [1214 kB] Get:71 http://ftpmaster.internal/ubuntu hirsute/main s390x libprotoc23 s390x 3.12.4-1ubuntu1 [715 kB] Get:72 http://ftpmaster.internal/ubuntu hirsute/universe s390x protobuf-compiler s390x 3.12.4-1ubuntu1 [27.9 kB] Get:73 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-goprotobuf-dev s390x 1.3.4-2 [1534 kB] Get:74 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-google-genproto-dev all 0.0~git20200413.b5235f6-1 [4034 kB] Get:75 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-google-grpc-dev all 1.29.1-0ubuntu1 [541 kB] Get:76 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-go-kit-kit-dev all 0.6.0-3 [103 kB] Get:77 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-go-test-deep-dev all 1.0.3-1 [9324 B] Get:78 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-gogo-protobuf-dev all 1.3.2-1 [848 kB] Get:79 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-gogo-googleapis-dev all 1.4.0-1 [79.9 kB] Get:80 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-google-uuid-dev all 1.2.0-1 [15.0 kB] Get:81 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-docopt-docopt-go-dev all 0.6.2+git20180111.ee0de3b-1 [17.6 kB] Get:82 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-googleapis-gnostic-dev all 0.2.0-2 [83.7 kB] Get:83 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-bradfitz-gomemcache-dev all 0.0~git20141109-3 [10.4 kB] Get:84 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-garyburd-redigo-dev all 0.0~git20150901.0.d8dbe4d-2.1 [28.2 kB] Get:85 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-google-btree-dev all 1.0.0-1 [13.3 kB] Get:86 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-peterbourgon-diskv-dev all 3.0.0-1 [18.8 kB] Get:87 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-golang-snappy-dev all 0.0+git20160529.d9eb7a3-3 [51.2 kB] Get:88 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-gopkg-tomb.v1-dev all 0.0~git20141024.0.dd63297-7 [5384 B] Get:89 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-fsnotify-fsnotify-dev all 1.4.9-2 [23.4 kB] Get:90 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-influxdata-tail-dev all 1.0.0+git20180327.c434825-3build2 [16.6 kB] Get:91 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-gomega-dev all 1.10.3-1 [81.0 kB] Get:92 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-onsi-ginkgo-dev all 1.14.2-1 [117 kB] Get:93 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-ginkgo-dev all 1.14.2-1 [1948 B] Get:94 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-syndtr-goleveldb-dev all 0.0~git20200815.5c35d60-1 [115 kB] Get:95 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-gregjones-httpcache-dev all 0.0~git20180305.9cad4c3-1.1 [13.8 kB] Get:96 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-bexpr-dev all 0.1.2-2 [33.9 kB] Get:97 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-jmespath-go-jmespath-dev all 0.4.0-1 [91.5 kB] Get:98 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-pkg-errors-dev all 0.9.1-1 [13.1 kB] Get:99 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-aws-aws-sdk-go-dev all 1.36.15-1 [7469 kB] Get:100 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-mitchellh-go-homedir-dev all 1.1.0-1 [5208 B] Get:101 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-golang-x-crypto-dev all 1:0.0~git20200604.70a84ac-2 [1541 kB] Get:102 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-azure-go-autorest-dev all 14.1.1-1 [102 kB] Get:103 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-denverdino-aliyungo-dev all 0.0~git20180921.13fa8aa-2 [125 kB] Get:104 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-google-go-querystring-dev all 1.0.0-1 [7500 B] Get:105 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-digitalocean-godo-dev all 1.37.0-1 [70.6 kB] Get:106 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-gophercloud-gophercloud-dev all 0.12.0-1 [622 kB] Get:107 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-errwrap-dev all 1.0.0-1.1 [10.5 kB] Get:108 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-multierror-dev all 1.0.0-1 [10.6 kB] Get:109 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-golang-x-sync-dev all 0.0~git20190911.cd5d95a-1 [17.4 kB] Get:110 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-miekg-dns-dev all 1.1.26-2 [148 kB] Get:111 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-mdns-dev all 1.0.3-1 [12.4 kB] Get:112 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-packethost-packngo-dev all 0.2.0-2 [39.6 kB] Get:113 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-vmware-govmomi-dev all 0.23.0-1 [703 kB] Get:114 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-googleapis-gax-go-dev all 2.0.5-1 [12.1 kB] Get:115 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-golang-lru-dev all 0.5.4-1 [14.6 kB] Get:116 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-go.opencensus-dev all 0.22.0-1 [120 kB] Get:117 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-google-api-dev all 0.28.0-2 [3815 kB] Get:118 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-discover-dev all 0.0+git20190905.34a6505-2 [26.7 kB] Get:119 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-mattn-go-isatty-dev all 0.0.12-1 [5984 B] Get:120 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-mattn-go-colorable-dev all 0.1.7-1 [9984 B] Get:121 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-fatih-color-dev all 1.7.0-1 [11.4 kB] Get:122 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-hclog-dev all 0.11.0-1 [18.0 kB] Get:123 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-circonus-labs-circonusllhist-dev all 0.0~git20191022.ec08cde-1 [11.2 kB] Get:124 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-cleanhttp-dev all 0.5.1-1 [10.4 kB] Get:125 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-retryablehttp-dev all 0.6.4-1 [17.4 kB] Get:126 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-tv42-httpunix-dev all 0.0~git20150427.b75d861-2 [3784 B] Get:127 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-circonus-labs-circonus-gometrics-dev all 2.3.1-3 [64.8 kB] Get:128 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-datadog-datadog-go-dev all 2.1.0-2 [13.3 kB] Get:129 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-uuid-dev all 1.0.2-1 [8812 B] Get:130 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-immutable-radix-dev all 1.3.0-1 [24.7 kB] Get:131 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-pascaldekloe-goe-dev all 0.1.0-2 [21.7 kB] Get:132 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-beorn7-perks-dev all 1.0.1-1 [12.2 kB] Get:133 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-cespare-xxhash-dev all 2.1.1-1 [8784 B] Get:134 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-modern-go-concurrent-dev all 1.0.3-1.1 [4708 B] Get:135 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-modern-go-reflect2-dev all 1.0.1-1 [10.7 kB] Get:136 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-json-iterator-go-dev all 1.1.10-2 [69.4 kB] Get:137 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-prometheus-client-model-dev all 0.2.0-1 [8252 B] Get:138 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-julienschmidt-httprouter-dev all 1.3.0-1 [19.4 kB] Get:139 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-jpillora-backoff-dev all 1.0.0-1.1 [3744 B] Get:140 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-mwitkow-go-conntrack-dev all 0.0~git20190716.2f06839-2 [14.7 kB] Get:141 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-alecthomas-units-dev all 0.0~git20201120.1786d5e-1 [7552 B] Get:142 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-gopkg-alecthomas-kingpin.v2-dev all 2.2.6-2 [42.8 kB] Get:143 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-protobuf-extensions-dev all 1.0.1-1 [29.6 kB] Get:144 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-prometheus-common-dev all 0.15.0-2 [101 kB] Get:145 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-procfs-dev all 0.1.3-3 [139 kB] Get:146 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-prometheus-client-golang-dev all 1.9.0-2 [127 kB] Get:147 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-armon-go-metrics-dev all 0.3.0-3 [26.5 kB] Get:148 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-kr-text-dev all 0.2.0-1 [11.1 kB] Get:149 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-kr-pretty-dev all 0.2.1+git20200831.59b4212-1 [14.7 kB] Get:150 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-gopkg-check.v1-dev all 0.0+git20200902.038fdea-1 [31.7 kB] Get:151 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-ugorji-go-codec-dev all 1.1.7-2build1 [201 kB] Get:152 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-ugorji-go-msgpack-dev all 0.0~git20130605.792643-5 [20.8 kB] Get:153 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-vmihailenco-tagparser-dev all 0.1.1-2 [4484 B] Get:154 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-gopkg-vmihailenco-msgpack.v2-dev all 4.3.1-1 [26.4 kB] Get:155 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-gopkg-tomb.v2-dev all 0.0~git20161208.d5d1b58-3 [6884 B] Get:156 http://ftpmaster.internal/ubuntu hirsute/main s390x libsasl2-modules-db s390x 2.1.27+dfsg-2ubuntu1 [14.5 kB] Get:157 http://ftpmaster.internal/ubuntu hirsute/main s390x libsasl2-2 s390x 2.1.27+dfsg-2ubuntu1 [48.8 kB] Get:158 http://ftpmaster.internal/ubuntu hirsute/main s390x libsasl2-dev s390x 2.1.27+dfsg-2ubuntu1 [232 kB] Get:159 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-gopkg-mgo.v2-dev all 2016.08.01-6 [316 kB] Get:160 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-msgpack-dev all 0.5.5-1 [43.3 kB] Get:161 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-raft-dev all 1.1.2-1 [88.7 kB] Get:162 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-mitchellh-reflectwalk-dev all 1.0.1-1 [8228 B] Get:163 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-mitchellh-copystructure-dev all 0.0~git20161013.0.5af94ae-2.1 [8944 B] Get:164 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-raftchunking-dev all 0.6.2-2 [12.4 kB] Get:165 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-jeffail-gabs-dev all 2.3.0-1 [16.9 kB] Get:166 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-jefferai-jsonx-dev all 1.0.1-2 [4596 B] Get:167 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-mitchellh-go-testing-interface-dev all 1.14.1-1 [4372 B] Get:168 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-oklog-run-dev all 1.1.0-1 [5624 B] Get:169 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-posener-complete-dev all 1.2.3-1 [15.9 kB] Get:170 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-ryanuber-go-glob-dev all 1.0.0-2 [4640 B] Get:171 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-x-text-dev all 0.3.2-1 [2072 B] Get:172 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-sap-go-hdb-dev all 0.14.1-2 [61.9 kB] Get:173 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-gopkg-inf.v0-dev all 0.9.1-1 [14.3 kB] Get:174 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-gopkg-square-go-jose.v2-dev all 2.3.1-1 [260 kB] Get:175 http://ftpmaster.internal/ubuntu hirsute/universe s390x mockery s390x 0.0~git20181123.e78b021-3build1 [1418 kB] Get:176 http://ftpmaster.internal/ubuntu hirsute/main s390x dh-golang all 1.51 [20.6 kB] Get:177 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-armon-circbuf-dev all 0.0~git20150827.0.bbbad09-2 [3984 B] Get:178 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-armon-go-radix-dev all 1.0.0-1 [7464 B] Get:179 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-asaskevich-govalidator-dev all 9+git20180720.0.f9ffefc3-1.1 [41.4 kB] Get:180 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-bgentry-speakeasy-dev all 0.1.0-2 [5428 B] Get:181 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-boltdb-bolt-dev all 1.3.1-7 [60.6 kB] Get:182 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-opencontainers-runc-dev all 1.0.0~rc92-0ubuntu1 [950 kB] Get:183 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-docker-go-connections-dev all 0.4.0-1 [26.3 kB] Get:184 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-elazarl-go-bindata-assetfs-dev all 1.0.0-1.1 [5624 B] Get:185 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-ghodss-yaml-dev all 1.0.0-1.1 [13.2 kB] Get:186 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-google-gofuzz-dev all 1.2.0-1 [13.0 kB] Get:187 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-checkpoint-dev all 0.0~git20171009.1545e56-2 [8220 B] Get:188 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-memdb-dev all 1.2.1-1 [34.6 kB] Get:189 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-yamux-dev all 0.0+git20190923.df201c7-1 [22.0 kB] Get:190 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-plugin-dev all 1.0.1-2 [48.5 kB] Get:191 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-rootcerts-dev all 1.0.2-1 [8292 B] Get:192 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-sockaddr-dev all 0.0~git20170627.41949a1+ds-2 [62.8 kB] Get:193 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-syslog-dev all 1.0.0-1 [6092 B] Get:194 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-go-version-dev all 1.2.0-1 [13.8 kB] Get:195 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-hcl-dev all 1.0.0-1.1 [58.9 kB] Get:196 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-mitchellh-mapstructure-dev all 1.3.3-1 [25.1 kB] Get:197 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-hil-dev all 0.0~git20160711.1e86c6b-1 [32.6 kB] Get:198 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-logutils-dev all 1.0.0-1 [8864 B] Get:199 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-memberlist-dev all 0.1.7-1ubuntu1 [77.4 kB] Get:200 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-net-rpc-msgpackrpc-dev all 0.0~git20151116.0.a14192a-1.1 [4432 B] Get:201 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-raft-boltdb-dev all 0.0~git20171010.6e5ba93-3 [11.1 kB] Get:202 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-scada-client-dev all 0.0~git20160601.0.6e89678-2.1 [19.6 kB] Get:203 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-mitchellh-cli-dev all 1.1.1-1 [24.4 kB] Get:204 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-ryanuber-columnize-dev all 2.1.1-1 [6644 B] Get:205 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-hashicorp-serf-dev all 0.8.5~ds1-1build1 [120 kB] Get:206 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-imdario-mergo-dev all 0.3.8-2 [18.6 kB] Get:207 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-inconshreveable-muxado-dev all 0.0~git20140312.0.f693c7e-2.1 [26.6 kB] Get:208 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-mitchellh-hashstructure-dev all 1.0.0-1.1 [7632 B] Get:209 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-nytimes-gziphandler-dev all 1.1.1-1 [40.0 kB] Get:210 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-pierrec-lz4-dev all 2.5.2-1 [5102 kB] Get:211 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-shirou-gopsutil-dev all 2.19.11-4 [104 kB] Get:212 http://ftpmaster.internal/ubuntu hirsute/universe s390x golang-github-spf13-pflag-dev all 1.0.5-2 [42.3 kB] debconf: delaying package configuration, since apt-utils is not installed Fetched 117 MB in 6s (20.3 MB/s) Selecting previously unselected package bsdextrautils. (Reading database ... 14766 files and directories currently installed.) Preparing to unpack .../000-bsdextrautils_2.36.1-1ubuntu2_s390x.deb ... Unpacking bsdextrautils (2.36.1-1ubuntu2) ... Selecting previously unselected package libuchardet0:s390x. Preparing to unpack .../001-libuchardet0_0.0.7-1_s390x.deb ... Unpacking libuchardet0:s390x (0.0.7-1) ... Selecting previously unselected package groff-base. Preparing to unpack .../002-groff-base_1.22.4-5_s390x.deb ... Unpacking groff-base (1.22.4-5) ... Selecting previously unselected package libpipeline1:s390x. Preparing to unpack .../003-libpipeline1_1.5.3-1_s390x.deb ... Unpacking libpipeline1:s390x (1.5.3-1) ... Selecting previously unselected package man-db. Preparing to unpack .../004-man-db_2.9.3-2_s390x.deb ... Unpacking man-db (2.9.3-2) ... Selecting previously unselected package golang-golang-x-sys-dev. Preparing to unpack .../005-golang-golang-x-sys-dev_0.0~git20201223.0d417f6-1_all.deb ... Unpacking golang-golang-x-sys-dev (0.0~git20201223.0d417f6-1) ... Selecting previously unselected package golang-github-sirupsen-logrus-dev. Preparing to unpack .../006-golang-github-sirupsen-logrus-dev_1.7.0-2_all.deb ... Unpacking golang-github-sirupsen-logrus-dev (1.7.0-2) ... Selecting previously unselected package libmagic-mgc. Preparing to unpack .../007-libmagic-mgc_1%3a5.39-3_s390x.deb ... Unpacking libmagic-mgc (1:5.39-3) ... Selecting previously unselected package libmagic1:s390x. Preparing to unpack .../008-libmagic1_1%3a5.39-3_s390x.deb ... Unpacking libmagic1:s390x (1:5.39-3) ... Selecting previously unselected package file. Preparing to unpack .../009-file_1%3a5.39-3_s390x.deb ... Unpacking file (1:5.39-3) ... Selecting previously unselected package libelf1:s390x. Preparing to unpack .../010-libelf1_0.182-3_s390x.deb ... Unpacking libelf1:s390x (0.182-3) ... Selecting previously unselected package libbpf0:s390x. Preparing to unpack .../011-libbpf0_1%3a0.3-2_s390x.deb ... Unpacking libbpf0:s390x (1:0.3-2) ... Selecting previously unselected package libbsd0:s390x. Preparing to unpack .../012-libbsd0_0.10.0-1_s390x.deb ... Unpacking libbsd0:s390x (0.10.0-1) ... Selecting previously unselected package libmnl0:s390x. Preparing to unpack .../013-libmnl0_1.0.4-3_s390x.deb ... Unpacking libmnl0:s390x (1.0.4-3) ... Selecting previously unselected package libxtables12:s390x. Preparing to unpack .../014-libxtables12_1.8.5-3ubuntu4_s390x.deb ... Unpacking libxtables12:s390x (1.8.5-3ubuntu4) ... Selecting previously unselected package libcap2-bin. Preparing to unpack .../015-libcap2-bin_1%3a2.44-1_s390x.deb ... Unpacking libcap2-bin (1:2.44-1) ... Selecting previously unselected package iproute2. Preparing to unpack .../016-iproute2_5.10.0-2ubuntu1_s390x.deb ... Unpacking iproute2 (5.10.0-2ubuntu1) ... Selecting previously unselected package libicu67:s390x. Preparing to unpack .../017-libicu67_67.1-5_s390x.deb ... Unpacking libicu67:s390x (67.1-5) ... Selecting previously unselected package libxml2:s390x. Preparing to unpack .../018-libxml2_2.9.10+dfsg-6.3build1_s390x.deb ... Unpacking libxml2:s390x (2.9.10+dfsg-6.3build1) ... Selecting previously unselected package bash-completion. Preparing to unpack .../019-bash-completion_1%3a2.11-2ubuntu1_all.deb ... Unpacking bash-completion (1:2.11-2ubuntu1) ... Selecting previously unselected package gettext-base. Preparing to unpack .../020-gettext-base_0.21-3ubuntu2_s390x.deb ... Unpacking gettext-base (0.21-3ubuntu2) ... Selecting previously unselected package libsigsegv2:s390x. Preparing to unpack .../021-libsigsegv2_2.12-3_s390x.deb ... Unpacking libsigsegv2:s390x (2.12-3) ... Selecting previously unselected package m4. Preparing to unpack .../022-m4_1.4.18-5_s390x.deb ... Unpacking m4 (1.4.18-5) ... Selecting previously unselected package autoconf. Preparing to unpack .../023-autoconf_2.69-14_all.deb ... Unpacking autoconf (2.69-14) ... Selecting previously unselected package autotools-dev. Preparing to unpack .../024-autotools-dev_20180224.1+nmu1_all.deb ... Unpacking autotools-dev (20180224.1+nmu1) ... Selecting previously unselected package automake. Preparing to unpack .../025-automake_1%3a1.16.3-2ubuntu1_all.deb ... Unpacking automake (1:1.16.3-2ubuntu1) ... Selecting previously unselected package autopoint. Preparing to unpack .../026-autopoint_0.21-3ubuntu2_all.deb ... Unpacking autopoint (0.21-3ubuntu2) ... Selecting previously unselected package libtool. Preparing to unpack .../027-libtool_2.4.6-15_all.deb ... Unpacking libtool (2.4.6-15) ... Selecting previously unselected package dh-autoreconf. Preparing to unpack .../028-dh-autoreconf_19_all.deb ... Unpacking dh-autoreconf (19) ... Selecting previously unselected package libdebhelper-perl. Preparing to unpack .../029-libdebhelper-perl_13.3.1ubuntu1_all.deb ... Unpacking libdebhelper-perl (13.3.1ubuntu1) ... Selecting previously unselected package libarchive-zip-perl. Preparing to unpack .../030-libarchive-zip-perl_1.68-1_all.deb ... Unpacking libarchive-zip-perl (1.68-1) ... Selecting previously unselected package libsub-override-perl. Preparing to unpack .../031-libsub-override-perl_0.09-2_all.deb ... Unpacking libsub-override-perl (0.09-2) ... Selecting previously unselected package libfile-stripnondeterminism-perl. Preparing to unpack .../032-libfile-stripnondeterminism-perl_1.10.0-1_all.deb ... Unpacking libfile-stripnondeterminism-perl (1.10.0-1) ... Selecting previously unselected package dh-strip-nondeterminism. Preparing to unpack .../033-dh-strip-nondeterminism_1.10.0-1_all.deb ... Unpacking dh-strip-nondeterminism (1.10.0-1) ... Selecting previously unselected package dwz. Preparing to unpack .../034-dwz_0.13+20210118-1_s390x.deb ... Unpacking dwz (0.13+20210118-1) ... Selecting previously unselected package gettext. Preparing to unpack .../035-gettext_0.21-3ubuntu2_s390x.deb ... Unpacking gettext (0.21-3ubuntu2) ... Selecting previously unselected package intltool-debian. Preparing to unpack .../036-intltool-debian_0.35.0+20060710.5_all.deb ... Unpacking intltool-debian (0.35.0+20060710.5) ... Selecting previously unselected package po-debconf. Preparing to unpack .../037-po-debconf_1.0.21+nmu1_all.deb ... Unpacking po-debconf (1.0.21+nmu1) ... Selecting previously unselected package debhelper. Preparing to unpack .../038-debhelper_13.3.1ubuntu1_all.deb ... Unpacking debhelper (13.3.1ubuntu1) ... Selecting previously unselected package golang-1.15-src. Preparing to unpack .../039-golang-1.15-src_1.15.4-1ubuntu2_s390x.deb ... Unpacking golang-1.15-src (1.15.4-1ubuntu2) ... Selecting previously unselected package golang-1.15-go. Preparing to unpack .../040-golang-1.15-go_1.15.4-1ubuntu2_s390x.deb ... Unpacking golang-1.15-go (1.15.4-1ubuntu2) ... Selecting previously unselected package golang-src:s390x. Preparing to unpack .../041-golang-src_2%3a1.15~1_s390x.deb ... Unpacking golang-src:s390x (2:1.15~1) ... Selecting previously unselected package golang-go. Preparing to unpack .../042-golang-go_2%3a1.15~1_s390x.deb ... Unpacking golang-go (2:1.15~1) ... Selecting previously unselected package golang-any:s390x. Preparing to unpack .../043-golang-any_2%3a1.15~1_s390x.deb ... Unpacking golang-any:s390x (2:1.15~1) ... Selecting previously unselected package golang-github-creack-pty-dev. Preparing to unpack .../044-golang-github-creack-pty-dev_1.1.9-2_all.deb ... Unpacking golang-github-creack-pty-dev (1.1.9-2) ... Selecting previously unselected package golang-github-dimchansky-utfbom-dev. Preparing to unpack .../045-golang-github-dimchansky-utfbom-dev_1.1.1-1_all.deb ... Unpacking golang-github-dimchansky-utfbom-dev (1.1.1-1) ... Selecting previously unselected package golang-github-dgrijalva-jwt-go-dev. Preparing to unpack .../046-golang-github-dgrijalva-jwt-go-dev_3.2.0-3_all.deb ... Unpacking golang-github-dgrijalva-jwt-go-dev (3.2.0-3) ... Selecting previously unselected package golang-github-go-logfmt-logfmt-dev. Preparing to unpack .../047-golang-github-go-logfmt-logfmt-dev_0.5.0-2_all.deb ... Unpacking golang-github-go-logfmt-logfmt-dev (0.5.0-2) ... Selecting previously unselected package golang-github-go-stack-stack-dev. Preparing to unpack .../048-golang-github-go-stack-stack-dev_1.8.0-1_all.deb ... Unpacking golang-github-go-stack-stack-dev (1.8.0-1) ... Selecting previously unselected package golang-github-davecgh-go-spew-dev. Preparing to unpack .../049-golang-github-davecgh-go-spew-dev_1.1.1-2_all.deb ... Unpacking golang-github-davecgh-go-spew-dev (1.1.1-2) ... Selecting previously unselected package golang-github-pmezard-go-difflib-dev. Preparing to unpack .../050-golang-github-pmezard-go-difflib-dev_1.0.0-3_all.deb ... Unpacking golang-github-pmezard-go-difflib-dev (1.0.0-3) ... Selecting previously unselected package golang-github-stretchr-objx-dev. Preparing to unpack .../051-golang-github-stretchr-objx-dev_0.3.0-1_all.deb ... Unpacking golang-github-stretchr-objx-dev (0.3.0-1) ... Selecting previously unselected package golang-gopkg-yaml.v2-dev. Preparing to unpack .../052-golang-gopkg-yaml.v2-dev_2.3.0-1_all.deb ... Unpacking golang-gopkg-yaml.v2-dev (2.3.0-1) ... Selecting previously unselected package golang-github-stretchr-testify-dev. Preparing to unpack .../053-golang-github-stretchr-testify-dev_1.4.0+ds-1_all.deb ... Unpacking golang-github-stretchr-testify-dev (1.4.0+ds-1) ... Selecting previously unselected package golang-github-opentracing-opentracing-go-dev. Preparing to unpack .../054-golang-github-opentracing-opentracing-go-dev_1.2.0-1_all.deb ... Unpacking golang-github-opentracing-opentracing-go-dev (1.2.0-1) ... Selecting previously unselected package golang-golang-x-text-dev. Preparing to unpack .../055-golang-golang-x-text-dev_0.3.5-1_all.deb ... Unpacking golang-golang-x-text-dev (0.3.5-1) ... Selecting previously unselected package golang-golang-x-net-dev. Preparing to unpack .../056-golang-golang-x-net-dev_1%3a0.0+git20200226.491c5fc+dfsg-1_all.deb ... Unpacking golang-golang-x-net-dev (1:0.0+git20200226.491c5fc+dfsg-1) ... Selecting previously unselected package golang-golang-x-time-dev. Preparing to unpack .../057-golang-golang-x-time-dev_0.0+git20200630.3af7569-1_all.deb ... Unpacking golang-golang-x-time-dev (0.0+git20200630.3af7569-1) ... Selecting previously unselected package golang-github-golang-mock-dev. Preparing to unpack .../058-golang-github-golang-mock-dev_1.3.1-2build1_all.deb ... Unpacking golang-github-golang-mock-dev (1.3.1-2build1) ... Selecting previously unselected package golang-golang-x-xerrors-dev. Preparing to unpack .../059-golang-golang-x-xerrors-dev_0.0~git20191204.9bdfabe-1_all.deb ... Unpacking golang-golang-x-xerrors-dev (0.0~git20191204.9bdfabe-1) ... Selecting previously unselected package golang-github-google-go-cmp-dev. Preparing to unpack .../060-golang-github-google-go-cmp-dev_0.4.0-1_all.deb ... Unpacking golang-github-google-go-cmp-dev (0.4.0-1) ... Selecting previously unselected package golang-glog-dev. Preparing to unpack .../061-golang-glog-dev_0.0~git20160126.23def4e-3_all.deb ... Unpacking golang-glog-dev (0.0~git20160126.23def4e-3) ... Selecting previously unselected package golang-golang-x-oauth2-dev. Preparing to unpack .../062-golang-golang-x-oauth2-dev_0.0~git20190604.0f29369-2_all.deb ... Unpacking golang-golang-x-oauth2-dev (0.0~git20190604.0f29369-2) ... Selecting previously unselected package golang-google-cloud-compute-metadata-dev. Preparing to unpack .../063-golang-google-cloud-compute-metadata-dev_0.56.0-1_all.deb ... Unpacking golang-google-cloud-compute-metadata-dev (0.56.0-1) ... Selecting previously unselected package golang-golang-x-oauth2-google-dev. Preparing to unpack .../064-golang-golang-x-oauth2-google-dev_0.0~git20190604.0f29369-2_all.deb ... Unpacking golang-golang-x-oauth2-google-dev (0.0~git20190604.0f29369-2) ... Selecting previously unselected package zlib1g-dev:s390x. Preparing to unpack .../065-zlib1g-dev_1%3a1.2.11.dfsg-2ubuntu4_s390x.deb ... Unpacking zlib1g-dev:s390x (1:1.2.11.dfsg-2ubuntu4) ... Selecting previously unselected package libprotobuf23:s390x. Preparing to unpack .../066-libprotobuf23_3.12.4-1ubuntu1_s390x.deb ... Unpacking libprotobuf23:s390x (3.12.4-1ubuntu1) ... Selecting previously unselected package libprotobuf-lite23:s390x. Preparing to unpack .../067-libprotobuf-lite23_3.12.4-1ubuntu1_s390x.deb ... Unpacking libprotobuf-lite23:s390x (3.12.4-1ubuntu1) ... Selecting previously unselected package libprotobuf-dev:s390x. Preparing to unpack .../068-libprotobuf-dev_3.12.4-1ubuntu1_s390x.deb ... Unpacking libprotobuf-dev:s390x (3.12.4-1ubuntu1) ... Selecting previously unselected package libprotoc23:s390x. Preparing to unpack .../069-libprotoc23_3.12.4-1ubuntu1_s390x.deb ... Unpacking libprotoc23:s390x (3.12.4-1ubuntu1) ... Selecting previously unselected package protobuf-compiler. Preparing to unpack .../070-protobuf-compiler_3.12.4-1ubuntu1_s390x.deb ... Unpacking protobuf-compiler (3.12.4-1ubuntu1) ... Selecting previously unselected package golang-goprotobuf-dev. Preparing to unpack .../071-golang-goprotobuf-dev_1.3.4-2_s390x.deb ... Unpacking golang-goprotobuf-dev (1.3.4-2) ... Selecting previously unselected package golang-google-genproto-dev. Preparing to unpack .../072-golang-google-genproto-dev_0.0~git20200413.b5235f6-1_all.deb ... Unpacking golang-google-genproto-dev (0.0~git20200413.b5235f6-1) ... Selecting previously unselected package golang-google-grpc-dev. Preparing to unpack .../073-golang-google-grpc-dev_1.29.1-0ubuntu1_all.deb ... Unpacking golang-google-grpc-dev (1.29.1-0ubuntu1) ... Selecting previously unselected package golang-github-go-kit-kit-dev. Preparing to unpack .../074-golang-github-go-kit-kit-dev_0.6.0-3_all.deb ... Unpacking golang-github-go-kit-kit-dev (0.6.0-3) ... Selecting previously unselected package golang-github-go-test-deep-dev. Preparing to unpack .../075-golang-github-go-test-deep-dev_1.0.3-1_all.deb ... Unpacking golang-github-go-test-deep-dev (1.0.3-1) ... Selecting previously unselected package golang-github-gogo-protobuf-dev. Preparing to unpack .../076-golang-github-gogo-protobuf-dev_1.3.2-1_all.deb ... Unpacking golang-github-gogo-protobuf-dev (1.3.2-1) ... Selecting previously unselected package golang-github-gogo-googleapis-dev. Preparing to unpack .../077-golang-github-gogo-googleapis-dev_1.4.0-1_all.deb ... Unpacking golang-github-gogo-googleapis-dev (1.4.0-1) ... Selecting previously unselected package golang-github-google-uuid-dev. Preparing to unpack .../078-golang-github-google-uuid-dev_1.2.0-1_all.deb ... Unpacking golang-github-google-uuid-dev (1.2.0-1) ... Selecting previously unselected package golang-github-docopt-docopt-go-dev. Preparing to unpack .../079-golang-github-docopt-docopt-go-dev_0.6.2+git20180111.ee0de3b-1_all.deb ... Unpacking golang-github-docopt-docopt-go-dev (0.6.2+git20180111.ee0de3b-1) ... Selecting previously unselected package golang-github-googleapis-gnostic-dev. Preparing to unpack .../080-golang-github-googleapis-gnostic-dev_0.2.0-2_all.deb ... Unpacking golang-github-googleapis-gnostic-dev (0.2.0-2) ... Selecting previously unselected package golang-github-bradfitz-gomemcache-dev. Preparing to unpack .../081-golang-github-bradfitz-gomemcache-dev_0.0~git20141109-3_all.deb ... Unpacking golang-github-bradfitz-gomemcache-dev (0.0~git20141109-3) ... Selecting previously unselected package golang-github-garyburd-redigo-dev. Preparing to unpack .../082-golang-github-garyburd-redigo-dev_0.0~git20150901.0.d8dbe4d-2.1_all.deb ... Unpacking golang-github-garyburd-redigo-dev (0.0~git20150901.0.d8dbe4d-2.1) ... Selecting previously unselected package golang-github-google-btree-dev. Preparing to unpack .../083-golang-github-google-btree-dev_1.0.0-1_all.deb ... Unpacking golang-github-google-btree-dev (1.0.0-1) ... Selecting previously unselected package golang-github-peterbourgon-diskv-dev. Preparing to unpack .../084-golang-github-peterbourgon-diskv-dev_3.0.0-1_all.deb ... Unpacking golang-github-peterbourgon-diskv-dev (3.0.0-1) ... Selecting previously unselected package golang-github-golang-snappy-dev. Preparing to unpack .../085-golang-github-golang-snappy-dev_0.0+git20160529.d9eb7a3-3_all.deb ... Unpacking golang-github-golang-snappy-dev (0.0+git20160529.d9eb7a3-3) ... Selecting previously unselected package golang-gopkg-tomb.v1-dev. Preparing to unpack .../086-golang-gopkg-tomb.v1-dev_0.0~git20141024.0.dd63297-7_all.deb ... Unpacking golang-gopkg-tomb.v1-dev (0.0~git20141024.0.dd63297-7) ... Selecting previously unselected package golang-github-fsnotify-fsnotify-dev. Preparing to unpack .../087-golang-github-fsnotify-fsnotify-dev_1.4.9-2_all.deb ... Unpacking golang-github-fsnotify-fsnotify-dev (1.4.9-2) ... Selecting previously unselected package golang-github-influxdata-tail-dev. Preparing to unpack .../088-golang-github-influxdata-tail-dev_1.0.0+git20180327.c434825-3build2_all.deb ... Unpacking golang-github-influxdata-tail-dev (1.0.0+git20180327.c434825-3build2) ... Selecting previously unselected package golang-gomega-dev. Preparing to unpack .../089-golang-gomega-dev_1.10.3-1_all.deb ... Unpacking golang-gomega-dev (1.10.3-1) ... Selecting previously unselected package golang-github-onsi-ginkgo-dev. Preparing to unpack .../090-golang-github-onsi-ginkgo-dev_1.14.2-1_all.deb ... Unpacking golang-github-onsi-ginkgo-dev (1.14.2-1) ... Selecting previously unselected package golang-ginkgo-dev. Preparing to unpack .../091-golang-ginkgo-dev_1.14.2-1_all.deb ... Unpacking golang-ginkgo-dev (1.14.2-1) ... Selecting previously unselected package golang-github-syndtr-goleveldb-dev. Preparing to unpack .../092-golang-github-syndtr-goleveldb-dev_0.0~git20200815.5c35d60-1_all.deb ... Unpacking golang-github-syndtr-goleveldb-dev (0.0~git20200815.5c35d60-1) ... Selecting previously unselected package golang-github-gregjones-httpcache-dev. Preparing to unpack .../093-golang-github-gregjones-httpcache-dev_0.0~git20180305.9cad4c3-1.1_all.deb ... Unpacking golang-github-gregjones-httpcache-dev (0.0~git20180305.9cad4c3-1.1) ... Selecting previously unselected package golang-github-hashicorp-go-bexpr-dev. Preparing to unpack .../094-golang-github-hashicorp-go-bexpr-dev_0.1.2-2_all.deb ... Unpacking golang-github-hashicorp-go-bexpr-dev (0.1.2-2) ... Selecting previously unselected package golang-github-jmespath-go-jmespath-dev. Preparing to unpack .../095-golang-github-jmespath-go-jmespath-dev_0.4.0-1_all.deb ... Unpacking golang-github-jmespath-go-jmespath-dev (0.4.0-1) ... Selecting previously unselected package golang-github-pkg-errors-dev. Preparing to unpack .../096-golang-github-pkg-errors-dev_0.9.1-1_all.deb ... Unpacking golang-github-pkg-errors-dev (0.9.1-1) ... Selecting previously unselected package golang-github-aws-aws-sdk-go-dev. Preparing to unpack .../097-golang-github-aws-aws-sdk-go-dev_1.36.15-1_all.deb ... Unpacking golang-github-aws-aws-sdk-go-dev (1.36.15-1) ... Selecting previously unselected package golang-github-mitchellh-go-homedir-dev. Preparing to unpack .../098-golang-github-mitchellh-go-homedir-dev_1.1.0-1_all.deb ... Unpacking golang-github-mitchellh-go-homedir-dev (1.1.0-1) ... Selecting previously unselected package golang-golang-x-crypto-dev. Preparing to unpack .../099-golang-golang-x-crypto-dev_1%3a0.0~git20200604.70a84ac-2_all.deb ... Unpacking golang-golang-x-crypto-dev (1:0.0~git20200604.70a84ac-2) ... Selecting previously unselected package golang-github-azure-go-autorest-dev. Preparing to unpack .../100-golang-github-azure-go-autorest-dev_14.1.1-1_all.deb ... Unpacking golang-github-azure-go-autorest-dev (14.1.1-1) ... Selecting previously unselected package golang-github-denverdino-aliyungo-dev. Preparing to unpack .../101-golang-github-denverdino-aliyungo-dev_0.0~git20180921.13fa8aa-2_all.deb ... Unpacking golang-github-denverdino-aliyungo-dev (0.0~git20180921.13fa8aa-2) ... Selecting previously unselected package golang-github-google-go-querystring-dev. Preparing to unpack .../102-golang-github-google-go-querystring-dev_1.0.0-1_all.deb ... Unpacking golang-github-google-go-querystring-dev (1.0.0-1) ... Selecting previously unselected package golang-github-digitalocean-godo-dev. Preparing to unpack .../103-golang-github-digitalocean-godo-dev_1.37.0-1_all.deb ... Unpacking golang-github-digitalocean-godo-dev (1.37.0-1) ... Selecting previously unselected package golang-github-gophercloud-gophercloud-dev. Preparing to unpack .../104-golang-github-gophercloud-gophercloud-dev_0.12.0-1_all.deb ... Unpacking golang-github-gophercloud-gophercloud-dev (0.12.0-1) ... Selecting previously unselected package golang-github-hashicorp-errwrap-dev. Preparing to unpack .../105-golang-github-hashicorp-errwrap-dev_1.0.0-1.1_all.deb ... Unpacking golang-github-hashicorp-errwrap-dev (1.0.0-1.1) ... Selecting previously unselected package golang-github-hashicorp-go-multierror-dev. Preparing to unpack .../106-golang-github-hashicorp-go-multierror-dev_1.0.0-1_all.deb ... Unpacking golang-github-hashicorp-go-multierror-dev (1.0.0-1) ... Selecting previously unselected package golang-golang-x-sync-dev. Preparing to unpack .../107-golang-golang-x-sync-dev_0.0~git20190911.cd5d95a-1_all.deb ... Unpacking golang-golang-x-sync-dev (0.0~git20190911.cd5d95a-1) ... Selecting previously unselected package golang-github-miekg-dns-dev. Preparing to unpack .../108-golang-github-miekg-dns-dev_1.1.26-2_all.deb ... Unpacking golang-github-miekg-dns-dev (1.1.26-2) ... Selecting previously unselected package golang-github-hashicorp-mdns-dev. Preparing to unpack .../109-golang-github-hashicorp-mdns-dev_1.0.3-1_all.deb ... Unpacking golang-github-hashicorp-mdns-dev (1.0.3-1) ... Selecting previously unselected package golang-github-packethost-packngo-dev. Preparing to unpack .../110-golang-github-packethost-packngo-dev_0.2.0-2_all.deb ... Unpacking golang-github-packethost-packngo-dev (0.2.0-2) ... Selecting previously unselected package golang-github-vmware-govmomi-dev. Preparing to unpack .../111-golang-github-vmware-govmomi-dev_0.23.0-1_all.deb ... Unpacking golang-github-vmware-govmomi-dev (0.23.0-1) ... Selecting previously unselected package golang-github-googleapis-gax-go-dev. Preparing to unpack .../112-golang-github-googleapis-gax-go-dev_2.0.5-1_all.deb ... Unpacking golang-github-googleapis-gax-go-dev (2.0.5-1) ... Selecting previously unselected package golang-github-hashicorp-golang-lru-dev. Preparing to unpack .../113-golang-github-hashicorp-golang-lru-dev_0.5.4-1_all.deb ... Unpacking golang-github-hashicorp-golang-lru-dev (0.5.4-1) ... Selecting previously unselected package golang-go.opencensus-dev. Preparing to unpack .../114-golang-go.opencensus-dev_0.22.0-1_all.deb ... Unpacking golang-go.opencensus-dev (0.22.0-1) ... Selecting previously unselected package golang-google-api-dev. Preparing to unpack .../115-golang-google-api-dev_0.28.0-2_all.deb ... Unpacking golang-google-api-dev (0.28.0-2) ... Selecting previously unselected package golang-github-hashicorp-go-discover-dev. Preparing to unpack .../116-golang-github-hashicorp-go-discover-dev_0.0+git20190905.34a6505-2_all.deb ... Unpacking golang-github-hashicorp-go-discover-dev (0.0+git20190905.34a6505-2) ... Selecting previously unselected package golang-github-mattn-go-isatty-dev. Preparing to unpack .../117-golang-github-mattn-go-isatty-dev_0.0.12-1_all.deb ... Unpacking golang-github-mattn-go-isatty-dev (0.0.12-1) ... Selecting previously unselected package golang-github-mattn-go-colorable-dev. Preparing to unpack .../118-golang-github-mattn-go-colorable-dev_0.1.7-1_all.deb ... Unpacking golang-github-mattn-go-colorable-dev (0.1.7-1) ... Selecting previously unselected package golang-github-fatih-color-dev. Preparing to unpack .../119-golang-github-fatih-color-dev_1.7.0-1_all.deb ... Unpacking golang-github-fatih-color-dev (1.7.0-1) ... Selecting previously unselected package golang-github-hashicorp-go-hclog-dev. Preparing to unpack .../120-golang-github-hashicorp-go-hclog-dev_0.11.0-1_all.deb ... Unpacking golang-github-hashicorp-go-hclog-dev (0.11.0-1) ... Selecting previously unselected package golang-github-circonus-labs-circonusllhist-dev. Preparing to unpack .../121-golang-github-circonus-labs-circonusllhist-dev_0.0~git20191022.ec08cde-1_all.deb ... Unpacking golang-github-circonus-labs-circonusllhist-dev (0.0~git20191022.ec08cde-1) ... Selecting previously unselected package golang-github-hashicorp-go-cleanhttp-dev. Preparing to unpack .../122-golang-github-hashicorp-go-cleanhttp-dev_0.5.1-1_all.deb ... Unpacking golang-github-hashicorp-go-cleanhttp-dev (0.5.1-1) ... Selecting previously unselected package golang-github-hashicorp-go-retryablehttp-dev. Preparing to unpack .../123-golang-github-hashicorp-go-retryablehttp-dev_0.6.4-1_all.deb ... Unpacking golang-github-hashicorp-go-retryablehttp-dev (0.6.4-1) ... Selecting previously unselected package golang-github-tv42-httpunix-dev. Preparing to unpack .../124-golang-github-tv42-httpunix-dev_0.0~git20150427.b75d861-2_all.deb ... Unpacking golang-github-tv42-httpunix-dev (0.0~git20150427.b75d861-2) ... Selecting previously unselected package golang-github-circonus-labs-circonus-gometrics-dev. Preparing to unpack .../125-golang-github-circonus-labs-circonus-gometrics-dev_2.3.1-3_all.deb ... Unpacking golang-github-circonus-labs-circonus-gometrics-dev (2.3.1-3) ... Selecting previously unselected package golang-github-datadog-datadog-go-dev. Preparing to unpack .../126-golang-github-datadog-datadog-go-dev_2.1.0-2_all.deb ... Unpacking golang-github-datadog-datadog-go-dev (2.1.0-2) ... Selecting previously unselected package golang-github-hashicorp-go-uuid-dev. Preparing to unpack .../127-golang-github-hashicorp-go-uuid-dev_1.0.2-1_all.deb ... Unpacking golang-github-hashicorp-go-uuid-dev (1.0.2-1) ... Selecting previously unselected package golang-github-hashicorp-go-immutable-radix-dev. Preparing to unpack .../128-golang-github-hashicorp-go-immutable-radix-dev_1.3.0-1_all.deb ... Unpacking golang-github-hashicorp-go-immutable-radix-dev (1.3.0-1) ... Selecting previously unselected package golang-github-pascaldekloe-goe-dev. Preparing to unpack .../129-golang-github-pascaldekloe-goe-dev_0.1.0-2_all.deb ... Unpacking golang-github-pascaldekloe-goe-dev (0.1.0-2) ... Selecting previously unselected package golang-github-beorn7-perks-dev. Preparing to unpack .../130-golang-github-beorn7-perks-dev_1.0.1-1_all.deb ... Unpacking golang-github-beorn7-perks-dev (1.0.1-1) ... Selecting previously unselected package golang-github-cespare-xxhash-dev. Preparing to unpack .../131-golang-github-cespare-xxhash-dev_2.1.1-1_all.deb ... Unpacking golang-github-cespare-xxhash-dev (2.1.1-1) ... Selecting previously unselected package golang-github-modern-go-concurrent-dev. Preparing to unpack .../132-golang-github-modern-go-concurrent-dev_1.0.3-1.1_all.deb ... Unpacking golang-github-modern-go-concurrent-dev (1.0.3-1.1) ... Selecting previously unselected package golang-github-modern-go-reflect2-dev. Preparing to unpack .../133-golang-github-modern-go-reflect2-dev_1.0.1-1_all.deb ... Unpacking golang-github-modern-go-reflect2-dev (1.0.1-1) ... Selecting previously unselected package golang-github-json-iterator-go-dev. Preparing to unpack .../134-golang-github-json-iterator-go-dev_1.1.10-2_all.deb ... Unpacking golang-github-json-iterator-go-dev (1.1.10-2) ... Selecting previously unselected package golang-github-prometheus-client-model-dev. Preparing to unpack .../135-golang-github-prometheus-client-model-dev_0.2.0-1_all.deb ... Unpacking golang-github-prometheus-client-model-dev (0.2.0-1) ... Selecting previously unselected package golang-github-julienschmidt-httprouter-dev. Preparing to unpack .../136-golang-github-julienschmidt-httprouter-dev_1.3.0-1_all.deb ... Unpacking golang-github-julienschmidt-httprouter-dev (1.3.0-1) ... Selecting previously unselected package golang-github-jpillora-backoff-dev. Preparing to unpack .../137-golang-github-jpillora-backoff-dev_1.0.0-1.1_all.deb ... Unpacking golang-github-jpillora-backoff-dev (1.0.0-1.1) ... Selecting previously unselected package golang-github-mwitkow-go-conntrack-dev. Preparing to unpack .../138-golang-github-mwitkow-go-conntrack-dev_0.0~git20190716.2f06839-2_all.deb ... Unpacking golang-github-mwitkow-go-conntrack-dev (0.0~git20190716.2f06839-2) ... Selecting previously unselected package golang-github-alecthomas-units-dev. Preparing to unpack .../139-golang-github-alecthomas-units-dev_0.0~git20201120.1786d5e-1_all.deb ... Unpacking golang-github-alecthomas-units-dev (0.0~git20201120.1786d5e-1) ... Selecting previously unselected package golang-gopkg-alecthomas-kingpin.v2-dev. Preparing to unpack .../140-golang-gopkg-alecthomas-kingpin.v2-dev_2.2.6-2_all.deb ... Unpacking golang-gopkg-alecthomas-kingpin.v2-dev (2.2.6-2) ... Selecting previously unselected package golang-protobuf-extensions-dev. Preparing to unpack .../141-golang-protobuf-extensions-dev_1.0.1-1_all.deb ... Unpacking golang-protobuf-extensions-dev (1.0.1-1) ... Selecting previously unselected package golang-github-prometheus-common-dev. Preparing to unpack .../142-golang-github-prometheus-common-dev_0.15.0-2_all.deb ... Unpacking golang-github-prometheus-common-dev (0.15.0-2) ... Selecting previously unselected package golang-procfs-dev. Preparing to unpack .../143-golang-procfs-dev_0.1.3-3_all.deb ... Unpacking golang-procfs-dev (0.1.3-3) ... Selecting previously unselected package golang-github-prometheus-client-golang-dev. Preparing to unpack .../144-golang-github-prometheus-client-golang-dev_1.9.0-2_all.deb ... Unpacking golang-github-prometheus-client-golang-dev (1.9.0-2) ... Selecting previously unselected package golang-github-armon-go-metrics-dev. Preparing to unpack .../145-golang-github-armon-go-metrics-dev_0.3.0-3_all.deb ... Unpacking golang-github-armon-go-metrics-dev (0.3.0-3) ... Selecting previously unselected package golang-github-kr-text-dev. Preparing to unpack .../146-golang-github-kr-text-dev_0.2.0-1_all.deb ... Unpacking golang-github-kr-text-dev (0.2.0-1) ... Selecting previously unselected package golang-github-kr-pretty-dev. Preparing to unpack .../147-golang-github-kr-pretty-dev_0.2.1+git20200831.59b4212-1_all.deb ... Unpacking golang-github-kr-pretty-dev (0.2.1+git20200831.59b4212-1) ... Selecting previously unselected package golang-gopkg-check.v1-dev. Preparing to unpack .../148-golang-gopkg-check.v1-dev_0.0+git20200902.038fdea-1_all.deb ... Unpacking golang-gopkg-check.v1-dev (0.0+git20200902.038fdea-1) ... Selecting previously unselected package golang-github-ugorji-go-codec-dev. Preparing to unpack .../149-golang-github-ugorji-go-codec-dev_1.1.7-2build1_all.deb ... Unpacking golang-github-ugorji-go-codec-dev (1.1.7-2build1) ... Selecting previously unselected package golang-github-ugorji-go-msgpack-dev. Preparing to unpack .../150-golang-github-ugorji-go-msgpack-dev_0.0~git20130605.792643-5_all.deb ... Unpacking golang-github-ugorji-go-msgpack-dev (0.0~git20130605.792643-5) ... Selecting previously unselected package golang-github-vmihailenco-tagparser-dev. Preparing to unpack .../151-golang-github-vmihailenco-tagparser-dev_0.1.1-2_all.deb ... Unpacking golang-github-vmihailenco-tagparser-dev (0.1.1-2) ... Selecting previously unselected package golang-gopkg-vmihailenco-msgpack.v2-dev. Preparing to unpack .../152-golang-gopkg-vmihailenco-msgpack.v2-dev_4.3.1-1_all.deb ... Unpacking golang-gopkg-vmihailenco-msgpack.v2-dev (4.3.1-1) ... Selecting previously unselected package golang-gopkg-tomb.v2-dev. Preparing to unpack .../153-golang-gopkg-tomb.v2-dev_0.0~git20161208.d5d1b58-3_all.deb ... Unpacking golang-gopkg-tomb.v2-dev (0.0~git20161208.d5d1b58-3) ... Selecting previously unselected package libsasl2-modules-db:s390x. Preparing to unpack .../154-libsasl2-modules-db_2.1.27+dfsg-2ubuntu1_s390x.deb ... Unpacking libsasl2-modules-db:s390x (2.1.27+dfsg-2ubuntu1) ... Selecting previously unselected package libsasl2-2:s390x. Preparing to unpack .../155-libsasl2-2_2.1.27+dfsg-2ubuntu1_s390x.deb ... Unpacking libsasl2-2:s390x (2.1.27+dfsg-2ubuntu1) ... Selecting previously unselected package libsasl2-dev. Preparing to unpack .../156-libsasl2-dev_2.1.27+dfsg-2ubuntu1_s390x.deb ... Unpacking libsasl2-dev (2.1.27+dfsg-2ubuntu1) ... Selecting previously unselected package golang-gopkg-mgo.v2-dev. Preparing to unpack .../157-golang-gopkg-mgo.v2-dev_2016.08.01-6_all.deb ... Unpacking golang-gopkg-mgo.v2-dev (2016.08.01-6) ... Selecting previously unselected package golang-github-hashicorp-go-msgpack-dev. Preparing to unpack .../158-golang-github-hashicorp-go-msgpack-dev_0.5.5-1_all.deb ... Unpacking golang-github-hashicorp-go-msgpack-dev (0.5.5-1) ... Selecting previously unselected package golang-github-hashicorp-raft-dev. Preparing to unpack .../159-golang-github-hashicorp-raft-dev_1.1.2-1_all.deb ... Unpacking golang-github-hashicorp-raft-dev (1.1.2-1) ... Selecting previously unselected package golang-github-mitchellh-reflectwalk-dev. Preparing to unpack .../160-golang-github-mitchellh-reflectwalk-dev_1.0.1-1_all.deb ... Unpacking golang-github-mitchellh-reflectwalk-dev (1.0.1-1) ... Selecting previously unselected package golang-github-mitchellh-copystructure-dev. Preparing to unpack .../161-golang-github-mitchellh-copystructure-dev_0.0~git20161013.0.5af94ae-2.1_all.deb ... Unpacking golang-github-mitchellh-copystructure-dev (0.0~git20161013.0.5af94ae-2.1) ... Selecting previously unselected package golang-github-hashicorp-go-raftchunking-dev. Preparing to unpack .../162-golang-github-hashicorp-go-raftchunking-dev_0.6.2-2_all.deb ... Unpacking golang-github-hashicorp-go-raftchunking-dev (0.6.2-2) ... Selecting previously unselected package golang-github-jeffail-gabs-dev. Preparing to unpack .../163-golang-github-jeffail-gabs-dev_2.3.0-1_all.deb ... Unpacking golang-github-jeffail-gabs-dev (2.3.0-1) ... Selecting previously unselected package golang-github-jefferai-jsonx-dev. Preparing to unpack .../164-golang-github-jefferai-jsonx-dev_1.0.1-2_all.deb ... Unpacking golang-github-jefferai-jsonx-dev (1.0.1-2) ... Selecting previously unselected package golang-github-mitchellh-go-testing-interface-dev. Preparing to unpack .../165-golang-github-mitchellh-go-testing-interface-dev_1.14.1-1_all.deb ... Unpacking golang-github-mitchellh-go-testing-interface-dev (1.14.1-1) ... Selecting previously unselected package golang-github-oklog-run-dev. Preparing to unpack .../166-golang-github-oklog-run-dev_1.1.0-1_all.deb ... Unpacking golang-github-oklog-run-dev (1.1.0-1) ... Selecting previously unselected package golang-github-posener-complete-dev. Preparing to unpack .../167-golang-github-posener-complete-dev_1.2.3-1_all.deb ... Unpacking golang-github-posener-complete-dev (1.2.3-1) ... Selecting previously unselected package golang-github-ryanuber-go-glob-dev. Preparing to unpack .../168-golang-github-ryanuber-go-glob-dev_1.0.0-2_all.deb ... Unpacking golang-github-ryanuber-go-glob-dev (1.0.0-2) ... Selecting previously unselected package golang-x-text-dev. Preparing to unpack .../169-golang-x-text-dev_0.3.2-1_all.deb ... Unpacking golang-x-text-dev (0.3.2-1) ... Selecting previously unselected package golang-github-sap-go-hdb-dev. Preparing to unpack .../170-golang-github-sap-go-hdb-dev_0.14.1-2_all.deb ... Unpacking golang-github-sap-go-hdb-dev (0.14.1-2) ... Selecting previously unselected package golang-gopkg-inf.v0-dev. Preparing to unpack .../171-golang-gopkg-inf.v0-dev_0.9.1-1_all.deb ... Unpacking golang-gopkg-inf.v0-dev (0.9.1-1) ... Selecting previously unselected package golang-gopkg-square-go-jose.v2-dev. Preparing to unpack .../172-golang-gopkg-square-go-jose.v2-dev_2.3.1-1_all.deb ... Unpacking golang-gopkg-square-go-jose.v2-dev (2.3.1-1) ... Selecting previously unselected package mockery. Preparing to unpack .../173-mockery_0.0~git20181123.e78b021-3build1_s390x.deb ... Unpacking mockery (0.0~git20181123.e78b021-3build1) ... Selecting previously unselected package dh-golang. Preparing to unpack .../174-dh-golang_1.51_all.deb ... Unpacking dh-golang (1.51) ... Selecting previously unselected package golang-github-armon-circbuf-dev. Preparing to unpack .../175-golang-github-armon-circbuf-dev_0.0~git20150827.0.bbbad09-2_all.deb ... Unpacking golang-github-armon-circbuf-dev (0.0~git20150827.0.bbbad09-2) ... Selecting previously unselected package golang-github-armon-go-radix-dev. Preparing to unpack .../176-golang-github-armon-go-radix-dev_1.0.0-1_all.deb ... Unpacking golang-github-armon-go-radix-dev (1.0.0-1) ... Selecting previously unselected package golang-github-asaskevich-govalidator-dev. Preparing to unpack .../177-golang-github-asaskevich-govalidator-dev_9+git20180720.0.f9ffefc3-1.1_all.deb ... Unpacking golang-github-asaskevich-govalidator-dev (9+git20180720.0.f9ffefc3-1.1) ... Selecting previously unselected package golang-github-bgentry-speakeasy-dev. Preparing to unpack .../178-golang-github-bgentry-speakeasy-dev_0.1.0-2_all.deb ... Unpacking golang-github-bgentry-speakeasy-dev (0.1.0-2) ... Selecting previously unselected package golang-github-boltdb-bolt-dev. Preparing to unpack .../179-golang-github-boltdb-bolt-dev_1.3.1-7_all.deb ... Unpacking golang-github-boltdb-bolt-dev (1.3.1-7) ... Selecting previously unselected package golang-github-opencontainers-runc-dev. Preparing to unpack .../180-golang-github-opencontainers-runc-dev_1.0.0~rc92-0ubuntu1_all.deb ... Unpacking golang-github-opencontainers-runc-dev (1.0.0~rc92-0ubuntu1) ... Selecting previously unselected package golang-github-docker-go-connections-dev. Preparing to unpack .../181-golang-github-docker-go-connections-dev_0.4.0-1_all.deb ... Unpacking golang-github-docker-go-connections-dev (0.4.0-1) ... Selecting previously unselected package golang-github-elazarl-go-bindata-assetfs-dev. Preparing to unpack .../182-golang-github-elazarl-go-bindata-assetfs-dev_1.0.0-1.1_all.deb ... Unpacking golang-github-elazarl-go-bindata-assetfs-dev (1.0.0-1.1) ... Selecting previously unselected package golang-github-ghodss-yaml-dev. Preparing to unpack .../183-golang-github-ghodss-yaml-dev_1.0.0-1.1_all.deb ... Unpacking golang-github-ghodss-yaml-dev (1.0.0-1.1) ... Selecting previously unselected package golang-github-google-gofuzz-dev. Preparing to unpack .../184-golang-github-google-gofuzz-dev_1.2.0-1_all.deb ... Unpacking golang-github-google-gofuzz-dev (1.2.0-1) ... Selecting previously unselected package golang-github-hashicorp-go-checkpoint-dev. Preparing to unpack .../185-golang-github-hashicorp-go-checkpoint-dev_0.0~git20171009.1545e56-2_all.deb ... Unpacking golang-github-hashicorp-go-checkpoint-dev (0.0~git20171009.1545e56-2) ... Selecting previously unselected package golang-github-hashicorp-go-memdb-dev. Preparing to unpack .../186-golang-github-hashicorp-go-memdb-dev_1.2.1-1_all.deb ... Unpacking golang-github-hashicorp-go-memdb-dev (1.2.1-1) ... Selecting previously unselected package golang-github-hashicorp-yamux-dev. Preparing to unpack .../187-golang-github-hashicorp-yamux-dev_0.0+git20190923.df201c7-1_all.deb ... Unpacking golang-github-hashicorp-yamux-dev (0.0+git20190923.df201c7-1) ... Selecting previously unselected package golang-github-hashicorp-go-plugin-dev. Preparing to unpack .../188-golang-github-hashicorp-go-plugin-dev_1.0.1-2_all.deb ... Unpacking golang-github-hashicorp-go-plugin-dev (1.0.1-2) ... Selecting previously unselected package golang-github-hashicorp-go-rootcerts-dev. Preparing to unpack .../189-golang-github-hashicorp-go-rootcerts-dev_1.0.2-1_all.deb ... Unpacking golang-github-hashicorp-go-rootcerts-dev (1.0.2-1) ... Selecting previously unselected package golang-github-hashicorp-go-sockaddr-dev. Preparing to unpack .../190-golang-github-hashicorp-go-sockaddr-dev_0.0~git20170627.41949a1+ds-2_all.deb ... Unpacking golang-github-hashicorp-go-sockaddr-dev (0.0~git20170627.41949a1+ds-2) ... Selecting previously unselected package golang-github-hashicorp-go-syslog-dev. Preparing to unpack .../191-golang-github-hashicorp-go-syslog-dev_1.0.0-1_all.deb ... Unpacking golang-github-hashicorp-go-syslog-dev (1.0.0-1) ... Selecting previously unselected package golang-github-hashicorp-go-version-dev. Preparing to unpack .../192-golang-github-hashicorp-go-version-dev_1.2.0-1_all.deb ... Unpacking golang-github-hashicorp-go-version-dev (1.2.0-1) ... Selecting previously unselected package golang-github-hashicorp-hcl-dev. Preparing to unpack .../193-golang-github-hashicorp-hcl-dev_1.0.0-1.1_all.deb ... Unpacking golang-github-hashicorp-hcl-dev (1.0.0-1.1) ... Selecting previously unselected package golang-github-mitchellh-mapstructure-dev. Preparing to unpack .../194-golang-github-mitchellh-mapstructure-dev_1.3.3-1_all.deb ... Unpacking golang-github-mitchellh-mapstructure-dev (1.3.3-1) ... Selecting previously unselected package golang-github-hashicorp-hil-dev. Preparing to unpack .../195-golang-github-hashicorp-hil-dev_0.0~git20160711.1e86c6b-1_all.deb ... Unpacking golang-github-hashicorp-hil-dev (0.0~git20160711.1e86c6b-1) ... Selecting previously unselected package golang-github-hashicorp-logutils-dev. Preparing to unpack .../196-golang-github-hashicorp-logutils-dev_1.0.0-1_all.deb ... Unpacking golang-github-hashicorp-logutils-dev (1.0.0-1) ... Selecting previously unselected package golang-github-hashicorp-memberlist-dev. Preparing to unpack .../197-golang-github-hashicorp-memberlist-dev_0.1.7-1ubuntu1_all.deb ... Unpacking golang-github-hashicorp-memberlist-dev (0.1.7-1ubuntu1) ... Selecting previously unselected package golang-github-hashicorp-net-rpc-msgpackrpc-dev. Preparing to unpack .../198-golang-github-hashicorp-net-rpc-msgpackrpc-dev_0.0~git20151116.0.a14192a-1.1_all.deb ... Unpacking golang-github-hashicorp-net-rpc-msgpackrpc-dev (0.0~git20151116.0.a14192a-1.1) ... Selecting previously unselected package golang-github-hashicorp-raft-boltdb-dev. Preparing to unpack .../199-golang-github-hashicorp-raft-boltdb-dev_0.0~git20171010.6e5ba93-3_all.deb ... Unpacking golang-github-hashicorp-raft-boltdb-dev (0.0~git20171010.6e5ba93-3) ... Selecting previously unselected package golang-github-hashicorp-scada-client-dev. Preparing to unpack .../200-golang-github-hashicorp-scada-client-dev_0.0~git20160601.0.6e89678-2.1_all.deb ... Unpacking golang-github-hashicorp-scada-client-dev (0.0~git20160601.0.6e89678-2.1) ... Selecting previously unselected package golang-github-mitchellh-cli-dev. Preparing to unpack .../201-golang-github-mitchellh-cli-dev_1.1.1-1_all.deb ... Unpacking golang-github-mitchellh-cli-dev (1.1.1-1) ... Selecting previously unselected package golang-github-ryanuber-columnize-dev. Preparing to unpack .../202-golang-github-ryanuber-columnize-dev_2.1.1-1_all.deb ... Unpacking golang-github-ryanuber-columnize-dev (2.1.1-1) ... Selecting previously unselected package golang-github-hashicorp-serf-dev. Preparing to unpack .../203-golang-github-hashicorp-serf-dev_0.8.5~ds1-1build1_all.deb ... Unpacking golang-github-hashicorp-serf-dev (0.8.5~ds1-1build1) ... Selecting previously unselected package golang-github-imdario-mergo-dev. Preparing to unpack .../204-golang-github-imdario-mergo-dev_0.3.8-2_all.deb ... Unpacking golang-github-imdario-mergo-dev (0.3.8-2) ... Selecting previously unselected package golang-github-inconshreveable-muxado-dev. Preparing to unpack .../205-golang-github-inconshreveable-muxado-dev_0.0~git20140312.0.f693c7e-2.1_all.deb ... Unpacking golang-github-inconshreveable-muxado-dev (0.0~git20140312.0.f693c7e-2.1) ... Selecting previously unselected package golang-github-mitchellh-hashstructure-dev. Preparing to unpack .../206-golang-github-mitchellh-hashstructure-dev_1.0.0-1.1_all.deb ... Unpacking golang-github-mitchellh-hashstructure-dev (1.0.0-1.1) ... Selecting previously unselected package golang-github-nytimes-gziphandler-dev. Preparing to unpack .../207-golang-github-nytimes-gziphandler-dev_1.1.1-1_all.deb ... Unpacking golang-github-nytimes-gziphandler-dev (1.1.1-1) ... Selecting previously unselected package golang-github-pierrec-lz4-dev. Preparing to unpack .../208-golang-github-pierrec-lz4-dev_2.5.2-1_all.deb ... Unpacking golang-github-pierrec-lz4-dev (2.5.2-1) ... Selecting previously unselected package golang-github-shirou-gopsutil-dev. Preparing to unpack .../209-golang-github-shirou-gopsutil-dev_2.19.11-4_all.deb ... Unpacking golang-github-shirou-gopsutil-dev (2.19.11-4) ... Selecting previously unselected package golang-github-spf13-pflag-dev. Preparing to unpack .../210-golang-github-spf13-pflag-dev_1.0.5-2_all.deb ... Unpacking golang-github-spf13-pflag-dev (1.0.5-2) ... Selecting previously unselected package sbuild-build-depends-consul-dummy. Preparing to unpack .../211-sbuild-build-depends-consul-dummy_0.invalid.0_s390x.deb ... Unpacking sbuild-build-depends-consul-dummy (0.invalid.0) ... Setting up golang-github-dimchansky-utfbom-dev (1.1.1-1) ... Setting up libpipeline1:s390x (1.5.3-1) ... Setting up golang-github-ryanuber-go-glob-dev (1.0.0-2) ... Setting up golang-github-oklog-run-dev (1.1.0-1) ... Setting up golang-gopkg-yaml.v2-dev (2.3.0-1) ... Setting up golang-github-hashicorp-go-uuid-dev (1.0.2-1) ... Setting up golang-github-mitchellh-go-homedir-dev (1.1.0-1) ... Setting up golang-github-google-go-querystring-dev (1.0.0-1) ... Setting up golang-github-imdario-mergo-dev (0.3.8-2) ... Setting up golang-github-mitchellh-mapstructure-dev (1.3.3-1) ... Setting up bsdextrautils (2.36.1-1ubuntu2) ... update-alternatives: using /usr/bin/write.ul to provide /usr/bin/write (write) in auto mode Setting up golang-github-creack-pty-dev (1.1.9-2) ... Setting up libicu67:s390x (67.1-5) ... Setting up golang-github-golang-mock-dev (1.3.1-2build1) ... Setting up golang-github-stretchr-objx-dev (0.3.0-1) ... Setting up golang-github-mitchellh-hashstructure-dev (1.0.0-1.1) ... Setting up libmagic-mgc (1:5.39-3) ... Setting up golang-github-pkg-errors-dev (0.9.1-1) ... Setting up golang-github-hashicorp-golang-lru-dev (0.5.4-1) ... Setting up golang-1.15-src (1.15.4-1ubuntu2) ... Setting up golang-github-google-gofuzz-dev (1.2.0-1) ... Setting up golang-github-inconshreveable-muxado-dev (0.0~git20140312.0.f693c7e-2.1) ... Setting up libarchive-zip-perl (1.68-1) ... Setting up golang-github-ryanuber-columnize-dev (2.1.1-1) ... Setting up libdebhelper-perl (13.3.1ubuntu1) ... Setting up golang-golang-x-sys-dev (0.0~git20201223.0d417f6-1) ... Setting up libmagic1:s390x (1:5.39-3) ... Setting up golang-github-hashicorp-go-syslog-dev (1.0.0-1) ... Setting up golang-github-golang-snappy-dev (0.0+git20160529.d9eb7a3-3) ... Setting up golang-github-pmezard-go-difflib-dev (1.0.0-3) ... Setting up golang-github-modern-go-concurrent-dev (1.0.3-1.1) ... Setting up gettext-base (0.21-3ubuntu2) ... Setting up golang-github-circonus-labs-circonusllhist-dev (0.0~git20191022.ec08cde-1) ... Setting up golang-github-bradfitz-gomemcache-dev (0.0~git20141109-3) ... Setting up mockery (0.0~git20181123.e78b021-3build1) ... Setting up libprotobuf23:s390x (3.12.4-1ubuntu1) ... Setting up golang-github-mitchellh-go-testing-interface-dev (1.14.1-1) ... Setting up file (1:5.39-3) ... Setting up golang-github-asaskevich-govalidator-dev (9+git20180720.0.f9ffefc3-1.1) ... Setting up golang-github-google-btree-dev (1.0.0-1) ... Setting up golang-github-go-stack-stack-dev (1.8.0-1) ... Setting up golang-github-beorn7-perks-dev (1.0.1-1) ... Setting up golang-github-mitchellh-reflectwalk-dev (1.0.1-1) ... Setting up golang-github-hashicorp-go-cleanhttp-dev (0.5.1-1) ... Setting up golang-github-hashicorp-errwrap-dev (1.0.0-1.1) ... Setting up libsasl2-modules-db:s390x (2.1.27+dfsg-2ubuntu1) ... Setting up libcap2-bin (1:2.44-1) ... Setting up golang-github-cespare-xxhash-dev (2.1.1-1) ... Setting up golang-github-spf13-pflag-dev (1.0.5-2) ... Setting up golang-gopkg-tomb.v2-dev (0.0~git20161208.d5d1b58-3) ... Setting up golang-github-bgentry-speakeasy-dev (0.1.0-2) ... Setting up golang-github-jpillora-backoff-dev (1.0.0-1.1) ... Setting up golang-github-davecgh-go-spew-dev (1.1.1-2) ... Setting up libprotobuf-lite23:s390x (3.12.4-1ubuntu1) ... Setting up golang-github-pierrec-lz4-dev (2.5.2-1) ... Setting up autotools-dev (20180224.1+nmu1) ... Setting up golang-github-pascaldekloe-goe-dev (0.1.0-2) ... Setting up golang-github-go-logfmt-logfmt-dev (0.5.0-2) ... Setting up golang-github-kr-text-dev (0.2.0-1) ... Setting up golang-github-stretchr-testify-dev (1.4.0+ds-1) ... Setting up golang-github-ugorji-go-msgpack-dev (0.0~git20130605.792643-5) ... Setting up golang-github-go-test-deep-dev (1.0.3-1) ... Setting up bash-completion (1:2.11-2ubuntu1) ... Setting up golang-github-hashicorp-go-immutable-radix-dev (1.3.0-1) ... Setting up golang-github-boltdb-bolt-dev (1.3.1-7) ... Setting up libprotoc23:s390x (3.12.4-1ubuntu1) ... Setting up libsigsegv2:s390x (2.12-3) ... Setting up golang-gopkg-tomb.v1-dev (0.0~git20141024.0.dd63297-7) ... Setting up libmnl0:s390x (1.0.4-3) ... Setting up golang-golang-x-sync-dev (0.0~git20190911.cd5d95a-1) ... Setting up autopoint (0.21-3ubuntu2) ... Setting up golang-github-hashicorp-hcl-dev (1.0.0-1.1) ... Setting up libxtables12:s390x (1.8.5-3ubuntu4) ... Setting up golang-glog-dev (0.0~git20160126.23def4e-3) ... Setting up libsasl2-2:s390x (2.1.27+dfsg-2ubuntu1) ... Setting up golang-github-julienschmidt-httprouter-dev (1.3.0-1) ... Setting up golang-github-hashicorp-go-multierror-dev (1.0.0-1) ... Setting up zlib1g-dev:s390x (1:1.2.11.dfsg-2ubuntu4) ... Setting up golang-github-tv42-httpunix-dev (0.0~git20150427.b75d861-2) ... Setting up golang-github-shirou-gopsutil-dev (2.19.11-4) ... Setting up golang-github-hashicorp-go-version-dev (1.2.0-1) ... Setting up golang-github-opencontainers-runc-dev (1.0.0~rc92-0ubuntu1) ... Setting up golang-gopkg-inf.v0-dev (0.9.1-1) ... Setting up golang-golang-x-text-dev (0.3.5-1) ... Setting up golang-github-google-uuid-dev (1.2.0-1) ... Setting up golang-github-hashicorp-hil-dev (0.0~git20160711.1e86c6b-1) ... Setting up libuchardet0:s390x (0.0.7-1) ... Setting up golang-github-fsnotify-fsnotify-dev (1.4.9-2) ... Setting up golang-github-alecthomas-units-dev (0.0~git20201120.1786d5e-1) ... Setting up golang-github-modern-go-reflect2-dev (1.0.1-1) ... Setting up libsub-override-perl (0.09-2) ... Setting up golang-github-dgrijalva-jwt-go-dev (3.2.0-3) ... Setting up golang-github-vmware-govmomi-dev (0.23.0-1) ... Setting up golang-github-armon-go-radix-dev (1.0.0-1) ... Setting up golang-github-datadog-datadog-go-dev (2.1.0-2) ... Setting up golang-github-docopt-docopt-go-dev (0.6.2+git20180111.ee0de3b-1) ... Setting up golang-golang-x-xerrors-dev (0.0~git20191204.9bdfabe-1) ... Setting up golang-github-ghodss-yaml-dev (1.0.0-1.1) ... Setting up golang-procfs-dev (0.1.3-3) ... Setting up golang-github-hashicorp-go-checkpoint-dev (0.0~git20171009.1545e56-2) ... Setting up golang-src:s390x (2:1.15~1) ... Setting up libbsd0:s390x (0.10.0-1) ... Setting up golang-1.15-go (1.15.4-1ubuntu2) ... Setting up libelf1:s390x (0.182-3) ... Setting up golang-github-vmihailenco-tagparser-dev (0.1.1-2) ... Setting up golang-github-armon-circbuf-dev (0.0~git20150827.0.bbbad09-2) ... Setting up golang-github-ugorji-go-codec-dev (1.1.7-2build1) ... Setting up golang-github-jeffail-gabs-dev (2.3.0-1) ... Setting up libxml2:s390x (2.9.10+dfsg-6.3build1) ... Setting up golang-github-jefferai-jsonx-dev (1.0.1-2) ... Setting up golang-github-jmespath-go-jmespath-dev (0.4.0-1) ... Setting up golang-github-hashicorp-yamux-dev (0.0+git20190923.df201c7-1) ... Setting up golang-github-hashicorp-go-rootcerts-dev (1.0.2-1) ... Setting up golang-x-text-dev (0.3.2-1) ... Setting up golang-github-hashicorp-logutils-dev (1.0.0-1) ... Setting up golang-github-kr-pretty-dev (0.2.1+git20200831.59b4212-1) ... Setting up libfile-stripnondeterminism-perl (1.10.0-1) ... Setting up golang-github-google-go-cmp-dev (0.4.0-1) ... Setting up golang-github-mattn-go-isatty-dev (0.0.12-1) ... Setting up golang-github-sirupsen-logrus-dev (1.7.0-2) ... Setting up golang-github-opentracing-opentracing-go-dev (1.2.0-1) ... Setting up golang-github-hashicorp-go-memdb-dev (1.2.1-1) ... Setting up golang-github-hashicorp-go-bexpr-dev (0.1.2-2) ... Setting up gettext (0.21-3ubuntu2) ... Setting up protobuf-compiler (3.12.4-1ubuntu1) ... Setting up golang-gopkg-alecthomas-kingpin.v2-dev (2.2.6-2) ... Setting up libtool (2.4.6-15) ... Setting up golang-go (2:1.15~1) ... Setting up golang-github-mattn-go-colorable-dev (0.1.7-1) ... Setting up golang-github-posener-complete-dev (1.2.3-1) ... Setting up golang-github-nytimes-gziphandler-dev (1.1.1-1) ... Setting up m4 (1.4.18-5) ... Setting up golang-github-json-iterator-go-dev (1.1.10-2) ... Setting up golang-gopkg-check.v1-dev (0.0+git20200902.038fdea-1) ... Setting up golang-github-mitchellh-copystructure-dev (0.0~git20161013.0.5af94ae-2.1) ... Setting up golang-any:s390x (2:1.15~1) ... Setting up intltool-debian (0.35.0+20060710.5) ... Setting up libprotobuf-dev:s390x (3.12.4-1ubuntu1) ... Setting up libsasl2-dev (2.1.27+dfsg-2ubuntu1) ... Setting up golang-gopkg-mgo.v2-dev (2016.08.01-6) ... Setting up golang-goprotobuf-dev (1.3.4-2) ... Setting up golang-github-sap-go-hdb-dev (0.14.1-2) ... Setting up golang-github-elazarl-go-bindata-assetfs-dev (1.0.0-1.1) ... Setting up golang-golang-x-net-dev (1:0.0+git20200226.491c5fc+dfsg-1) ... Setting up autoconf (2.69-14) ... Setting up golang-github-influxdata-tail-dev (1.0.0+git20180327.c434825-3build2) ... Setting up dh-strip-nondeterminism (1.10.0-1) ... Setting up dwz (0.13+20210118-1) ... Setting up groff-base (1.22.4-5) ... Setting up golang-golang-x-crypto-dev (1:0.0~git20200604.70a84ac-2) ... Setting up golang-github-prometheus-client-model-dev (0.2.0-1) ... Setting up golang-golang-x-oauth2-dev (0.0~git20190604.0f29369-2) ... Setting up libbpf0:s390x (1:0.3-2) ... Setting up golang-github-aws-aws-sdk-go-dev (1.36.15-1) ... Setting up automake (1:1.16.3-2ubuntu1) ... update-alternatives: using /usr/bin/automake-1.16 to provide /usr/bin/automake (automake) in auto mode Setting up golang-golang-x-time-dev (0.0+git20200630.3af7569-1) ... Setting up golang-github-digitalocean-godo-dev (1.37.0-1) ... Setting up golang-github-gophercloud-gophercloud-dev (0.12.0-1) ... Setting up golang-github-peterbourgon-diskv-dev (3.0.0-1) ... Setting up golang-github-gogo-protobuf-dev (1.3.2-1) ... Setting up golang-gopkg-vmihailenco-msgpack.v2-dev (4.3.1-1) ... Setting up golang-github-fatih-color-dev (1.7.0-1) ... Setting up golang-github-miekg-dns-dev (1.1.26-2) ... Setting up iproute2 (5.10.0-2ubuntu1) ... Setting up golang-github-hashicorp-go-msgpack-dev (0.5.5-1) ... Setting up po-debconf (1.0.21+nmu1) ... Setting up golang-github-garyburd-redigo-dev (0.0~git20150901.0.d8dbe4d-2.1) ... Setting up golang-protobuf-extensions-dev (1.0.1-1) ... Setting up golang-gomega-dev (1.10.3-1) ... Setting up golang-google-cloud-compute-metadata-dev (0.56.0-1) ... Setting up golang-github-docker-go-connections-dev (0.4.0-1) ... Setting up man-db (2.9.3-2) ... Not building database; man-db/auto-update is not 'true'. Created symlink /etc/systemd/system/timers.target.wants/man-db.timer → /lib/systemd/system/man-db.timer. Setting up golang-github-packethost-packngo-dev (0.2.0-2) ... Setting up golang-github-denverdino-aliyungo-dev (0.0~git20180921.13fa8aa-2) ... Setting up golang-github-onsi-ginkgo-dev (1.14.2-1) ... Setting up golang-golang-x-oauth2-google-dev (0.0~git20190604.0f29369-2) ... Setting up golang-gopkg-square-go-jose.v2-dev (2.3.1-1) ... Setting up golang-github-azure-go-autorest-dev (14.1.1-1) ... Setting up golang-github-mitchellh-cli-dev (1.1.1-1) ... Setting up golang-github-googleapis-gnostic-dev (0.2.0-2) ... Setting up golang-github-gogo-googleapis-dev (1.4.0-1) ... Setting up golang-github-hashicorp-net-rpc-msgpackrpc-dev (0.0~git20151116.0.a14192a-1.1) ... Setting up golang-github-hashicorp-go-hclog-dev (0.11.0-1) ... Setting up golang-github-syndtr-goleveldb-dev (0.0~git20200815.5c35d60-1) ... Setting up golang-ginkgo-dev (1.14.2-1) ... Setting up golang-github-hashicorp-mdns-dev (1.0.3-1) ... Setting up golang-github-hashicorp-go-sockaddr-dev (0.0~git20170627.41949a1+ds-2) ... Setting up golang-github-hashicorp-go-retryablehttp-dev (0.6.4-1) ... Setting up golang-github-hashicorp-go-plugin-dev (1.0.1-2) ... Setting up golang-github-gregjones-httpcache-dev (0.0~git20180305.9cad4c3-1.1) ... Setting up golang-github-circonus-labs-circonus-gometrics-dev (2.3.1-3) ... Setting up golang-github-mwitkow-go-conntrack-dev (0.0~git20190716.2f06839-2) ... Setting up golang-google-genproto-dev (0.0~git20200413.b5235f6-1) ... Setting up dh-autoreconf (19) ... Setting up golang-github-prometheus-client-golang-dev (1.9.0-2) ... Setting up golang-github-armon-go-metrics-dev (0.3.0-3) ... Setting up golang-google-grpc-dev (1.29.1-0ubuntu1) ... Setting up golang-go.opencensus-dev (0.22.0-1) ... Setting up golang-github-hashicorp-raft-dev (1.1.2-1) ... Setting up golang-github-go-kit-kit-dev (0.6.0-3) ... Setting up debhelper (13.3.1ubuntu1) ... Setting up golang-github-googleapis-gax-go-dev (2.0.5-1) ... Setting up golang-github-hashicorp-scada-client-dev (0.0~git20160601.0.6e89678-2.1) ... Setting up golang-github-hashicorp-memberlist-dev (0.1.7-1ubuntu1) ... Setting up golang-github-hashicorp-go-raftchunking-dev (0.6.2-2) ... Setting up golang-github-prometheus-common-dev (0.15.0-2) ... Setting up golang-google-api-dev (0.28.0-2) ... Setting up dh-golang (1.51) ... Setting up golang-github-hashicorp-raft-boltdb-dev (0.0~git20171010.6e5ba93-3) ... Setting up golang-github-hashicorp-serf-dev (0.8.5~ds1-1build1) ... Setting up golang-github-hashicorp-go-discover-dev (0.0+git20190905.34a6505-2) ... Setting up sbuild-build-depends-consul-dummy (0.invalid.0) ... Processing triggers for libc-bin (2.32-0ubuntu6) ... +------------------------------------------------------------------------------+ | Build environment | +------------------------------------------------------------------------------+ Kernel: Linux 4.15.0-134-generic s390x (s390x) Toolchain package versions: binutils_2.35.50.20210106-1ubuntu2 dpkg-dev_1.20.7.1ubuntu2 g++-10_10.2.1-6ubuntu1 gcc-10_10.2.1-6ubuntu1 libc6-dev_2.32-0ubuntu6 libstdc++-10-dev_10.2.1-6ubuntu1 libstdc++6_10.2.1-6ubuntu1 linux-libc-dev_5.8.0-36.40+21.04.1 Package versions: adduser_3.118ubuntu5 advancecomp_2.1-2.1build1 apt_2.1.18 autoconf_2.69-14 automake_1:1.16.3-2ubuntu1 autopoint_0.21-3ubuntu2 autotools-dev_20180224.1+nmu1 base-files_11ubuntu16 base-passwd_3.5.48 bash_5.1-1ubuntu1 bash-completion_1:2.11-2ubuntu1 binutils_2.35.50.20210106-1ubuntu2 binutils-common_2.35.50.20210106-1ubuntu2 binutils-s390x-linux-gnu_2.35.50.20210106-1ubuntu2 bsdextrautils_2.36.1-1ubuntu2 bsdutils_1:2.36.1-1ubuntu2 build-essential_12.8ubuntu3 bzip2_1.0.8-4ubuntu2 ca-certificates_20200601 coreutils_8.32-4ubuntu2 cpp_4:10.2.0-1ubuntu1 cpp-10_10.2.1-6ubuntu1 dash_0.5.11+git20200708+dd9ef66+really0.5.11+git20200708+dd9ef66-5ubuntu1 debconf_1.5.74 debhelper_13.3.1ubuntu1 debianutils_4.11.2 dh-autoreconf_19 dh-golang_1.51 dh-strip-nondeterminism_1.10.0-1 diffutils_1:3.7-3ubuntu1 dpkg_1.20.7.1ubuntu2 dpkg-dev_1.20.7.1ubuntu2 dwz_0.13+20210118-1 e2fsprogs_1.45.6-1ubuntu1 fakeroot_1.25.3-1.1 file_1:5.39-3 findutils_4.7.0-1ubuntu2 g++_4:10.2.0-1ubuntu1 g++-10_10.2.1-6ubuntu1 gcc_4:10.2.0-1ubuntu1 gcc-10_10.2.1-6ubuntu1 gcc-10-base_10.2.1-6ubuntu1 gettext_0.21-3ubuntu2 gettext-base_0.21-3ubuntu2 golang-1.15-go_1.15.4-1ubuntu2 golang-1.15-src_1.15.4-1ubuntu2 golang-any_2:1.15~1 golang-ginkgo-dev_1.14.2-1 golang-github-alecthomas-units-dev_0.0~git20201120.1786d5e-1 golang-github-armon-circbuf-dev_0.0~git20150827.0.bbbad09-2 golang-github-armon-go-metrics-dev_0.3.0-3 golang-github-armon-go-radix-dev_1.0.0-1 golang-github-asaskevich-govalidator-dev_9+git20180720.0.f9ffefc3-1.1 golang-github-aws-aws-sdk-go-dev_1.36.15-1 golang-github-azure-go-autorest-dev_14.1.1-1 golang-github-beorn7-perks-dev_1.0.1-1 golang-github-bgentry-speakeasy-dev_0.1.0-2 golang-github-boltdb-bolt-dev_1.3.1-7 golang-github-bradfitz-gomemcache-dev_0.0~git20141109-3 golang-github-cespare-xxhash-dev_2.1.1-1 golang-github-circonus-labs-circonus-gometrics-dev_2.3.1-3 golang-github-circonus-labs-circonusllhist-dev_0.0~git20191022.ec08cde-1 golang-github-creack-pty-dev_1.1.9-2 golang-github-datadog-datadog-go-dev_2.1.0-2 golang-github-davecgh-go-spew-dev_1.1.1-2 golang-github-denverdino-aliyungo-dev_0.0~git20180921.13fa8aa-2 golang-github-dgrijalva-jwt-go-dev_3.2.0-3 golang-github-digitalocean-godo-dev_1.37.0-1 golang-github-dimchansky-utfbom-dev_1.1.1-1 golang-github-docker-go-connections-dev_0.4.0-1 golang-github-docopt-docopt-go-dev_0.6.2+git20180111.ee0de3b-1 golang-github-elazarl-go-bindata-assetfs-dev_1.0.0-1.1 golang-github-fatih-color-dev_1.7.0-1 golang-github-fsnotify-fsnotify-dev_1.4.9-2 golang-github-garyburd-redigo-dev_0.0~git20150901.0.d8dbe4d-2.1 golang-github-ghodss-yaml-dev_1.0.0-1.1 golang-github-go-kit-kit-dev_0.6.0-3 golang-github-go-logfmt-logfmt-dev_0.5.0-2 golang-github-go-stack-stack-dev_1.8.0-1 golang-github-go-test-deep-dev_1.0.3-1 golang-github-gogo-googleapis-dev_1.4.0-1 golang-github-gogo-protobuf-dev_1.3.2-1 golang-github-golang-mock-dev_1.3.1-2build1 golang-github-golang-snappy-dev_0.0+git20160529.d9eb7a3-3 golang-github-google-btree-dev_1.0.0-1 golang-github-google-go-cmp-dev_0.4.0-1 golang-github-google-go-querystring-dev_1.0.0-1 golang-github-google-gofuzz-dev_1.2.0-1 golang-github-google-uuid-dev_1.2.0-1 golang-github-googleapis-gax-go-dev_2.0.5-1 golang-github-googleapis-gnostic-dev_0.2.0-2 golang-github-gophercloud-gophercloud-dev_0.12.0-1 golang-github-gregjones-httpcache-dev_0.0~git20180305.9cad4c3-1.1 golang-github-hashicorp-errwrap-dev_1.0.0-1.1 golang-github-hashicorp-go-bexpr-dev_0.1.2-2 golang-github-hashicorp-go-checkpoint-dev_0.0~git20171009.1545e56-2 golang-github-hashicorp-go-cleanhttp-dev_0.5.1-1 golang-github-hashicorp-go-discover-dev_0.0+git20190905.34a6505-2 golang-github-hashicorp-go-hclog-dev_0.11.0-1 golang-github-hashicorp-go-immutable-radix-dev_1.3.0-1 golang-github-hashicorp-go-memdb-dev_1.2.1-1 golang-github-hashicorp-go-msgpack-dev_0.5.5-1 golang-github-hashicorp-go-multierror-dev_1.0.0-1 golang-github-hashicorp-go-plugin-dev_1.0.1-2 golang-github-hashicorp-go-raftchunking-dev_0.6.2-2 golang-github-hashicorp-go-retryablehttp-dev_0.6.4-1 golang-github-hashicorp-go-rootcerts-dev_1.0.2-1 golang-github-hashicorp-go-sockaddr-dev_0.0~git20170627.41949a1+ds-2 golang-github-hashicorp-go-syslog-dev_1.0.0-1 golang-github-hashicorp-go-uuid-dev_1.0.2-1 golang-github-hashicorp-go-version-dev_1.2.0-1 golang-github-hashicorp-golang-lru-dev_0.5.4-1 golang-github-hashicorp-hcl-dev_1.0.0-1.1 golang-github-hashicorp-hil-dev_0.0~git20160711.1e86c6b-1 golang-github-hashicorp-logutils-dev_1.0.0-1 golang-github-hashicorp-mdns-dev_1.0.3-1 golang-github-hashicorp-memberlist-dev_0.1.7-1ubuntu1 golang-github-hashicorp-net-rpc-msgpackrpc-dev_0.0~git20151116.0.a14192a-1.1 golang-github-hashicorp-raft-boltdb-dev_0.0~git20171010.6e5ba93-3 golang-github-hashicorp-raft-dev_1.1.2-1 golang-github-hashicorp-scada-client-dev_0.0~git20160601.0.6e89678-2.1 golang-github-hashicorp-serf-dev_0.8.5~ds1-1build1 golang-github-hashicorp-yamux-dev_0.0+git20190923.df201c7-1 golang-github-imdario-mergo-dev_0.3.8-2 golang-github-inconshreveable-muxado-dev_0.0~git20140312.0.f693c7e-2.1 golang-github-influxdata-tail-dev_1.0.0+git20180327.c434825-3build2 golang-github-jeffail-gabs-dev_2.3.0-1 golang-github-jefferai-jsonx-dev_1.0.1-2 golang-github-jmespath-go-jmespath-dev_0.4.0-1 golang-github-jpillora-backoff-dev_1.0.0-1.1 golang-github-json-iterator-go-dev_1.1.10-2 golang-github-julienschmidt-httprouter-dev_1.3.0-1 golang-github-kr-pretty-dev_0.2.1+git20200831.59b4212-1 golang-github-kr-text-dev_0.2.0-1 golang-github-mattn-go-colorable-dev_0.1.7-1 golang-github-mattn-go-isatty-dev_0.0.12-1 golang-github-miekg-dns-dev_1.1.26-2 golang-github-mitchellh-cli-dev_1.1.1-1 golang-github-mitchellh-copystructure-dev_0.0~git20161013.0.5af94ae-2.1 golang-github-mitchellh-go-homedir-dev_1.1.0-1 golang-github-mitchellh-go-testing-interface-dev_1.14.1-1 golang-github-mitchellh-hashstructure-dev_1.0.0-1.1 golang-github-mitchellh-mapstructure-dev_1.3.3-1 golang-github-mitchellh-reflectwalk-dev_1.0.1-1 golang-github-modern-go-concurrent-dev_1.0.3-1.1 golang-github-modern-go-reflect2-dev_1.0.1-1 golang-github-mwitkow-go-conntrack-dev_0.0~git20190716.2f06839-2 golang-github-nytimes-gziphandler-dev_1.1.1-1 golang-github-oklog-run-dev_1.1.0-1 golang-github-onsi-ginkgo-dev_1.14.2-1 golang-github-opencontainers-runc-dev_1.0.0~rc92-0ubuntu1 golang-github-opentracing-opentracing-go-dev_1.2.0-1 golang-github-packethost-packngo-dev_0.2.0-2 golang-github-pascaldekloe-goe-dev_0.1.0-2 golang-github-peterbourgon-diskv-dev_3.0.0-1 golang-github-pierrec-lz4-dev_2.5.2-1 golang-github-pkg-errors-dev_0.9.1-1 golang-github-pmezard-go-difflib-dev_1.0.0-3 golang-github-posener-complete-dev_1.2.3-1 golang-github-prometheus-client-golang-dev_1.9.0-2 golang-github-prometheus-client-model-dev_0.2.0-1 golang-github-prometheus-common-dev_0.15.0-2 golang-github-ryanuber-columnize-dev_2.1.1-1 golang-github-ryanuber-go-glob-dev_1.0.0-2 golang-github-sap-go-hdb-dev_0.14.1-2 golang-github-shirou-gopsutil-dev_2.19.11-4 golang-github-sirupsen-logrus-dev_1.7.0-2 golang-github-spf13-pflag-dev_1.0.5-2 golang-github-stretchr-objx-dev_0.3.0-1 golang-github-stretchr-testify-dev_1.4.0+ds-1 golang-github-syndtr-goleveldb-dev_0.0~git20200815.5c35d60-1 golang-github-tv42-httpunix-dev_0.0~git20150427.b75d861-2 golang-github-ugorji-go-codec-dev_1.1.7-2build1 golang-github-ugorji-go-msgpack-dev_0.0~git20130605.792643-5 golang-github-vmihailenco-tagparser-dev_0.1.1-2 golang-github-vmware-govmomi-dev_0.23.0-1 golang-glog-dev_0.0~git20160126.23def4e-3 golang-go_2:1.15~1 golang-go.opencensus-dev_0.22.0-1 golang-golang-x-crypto-dev_1:0.0~git20200604.70a84ac-2 golang-golang-x-net-dev_1:0.0+git20200226.491c5fc+dfsg-1 golang-golang-x-oauth2-dev_0.0~git20190604.0f29369-2 golang-golang-x-oauth2-google-dev_0.0~git20190604.0f29369-2 golang-golang-x-sync-dev_0.0~git20190911.cd5d95a-1 golang-golang-x-sys-dev_0.0~git20201223.0d417f6-1 golang-golang-x-text-dev_0.3.5-1 golang-golang-x-time-dev_0.0+git20200630.3af7569-1 golang-golang-x-xerrors-dev_0.0~git20191204.9bdfabe-1 golang-gomega-dev_1.10.3-1 golang-google-api-dev_0.28.0-2 golang-google-cloud-compute-metadata-dev_0.56.0-1 golang-google-genproto-dev_0.0~git20200413.b5235f6-1 golang-google-grpc-dev_1.29.1-0ubuntu1 golang-gopkg-alecthomas-kingpin.v2-dev_2.2.6-2 golang-gopkg-check.v1-dev_0.0+git20200902.038fdea-1 golang-gopkg-inf.v0-dev_0.9.1-1 golang-gopkg-mgo.v2-dev_2016.08.01-6 golang-gopkg-square-go-jose.v2-dev_2.3.1-1 golang-gopkg-tomb.v1-dev_0.0~git20141024.0.dd63297-7 golang-gopkg-tomb.v2-dev_0.0~git20161208.d5d1b58-3 golang-gopkg-vmihailenco-msgpack.v2-dev_4.3.1-1 golang-gopkg-yaml.v2-dev_2.3.0-1 golang-goprotobuf-dev_1.3.4-2 golang-procfs-dev_0.1.3-3 golang-protobuf-extensions-dev_1.0.1-1 golang-src_2:1.15~1 golang-x-text-dev_0.3.2-1 gpg_2.2.20-1ubuntu2 gpg-agent_2.2.20-1ubuntu2 gpgconf_2.2.20-1ubuntu2 gpgv_2.2.20-1ubuntu2 grep_3.6-1 groff-base_1.22.4-5 gzip_1.10-2ubuntu3 hostname_3.23 init_1.60 init-system-helpers_1.60 intltool-debian_0.35.0+20060710.5 iproute2_5.10.0-2ubuntu1 libacl1_2.2.53-9 libapparmor1_3.0.0-0ubuntu5 libapt-pkg6.0_2.1.18 libarchive-zip-perl_1.68-1 libargon2-1_0~20171227-0.2build20.10.0 libasan6_10.2.1-6ubuntu1 libassuan0_2.5.3-7.1 libatomic1_10.2.1-6ubuntu1 libattr1_1:2.4.48-6 libaudit-common_1:2.8.5-3ubuntu3 libaudit1_1:2.8.5-3ubuntu3 libbinutils_2.35.50.20210106-1ubuntu2 libblkid1_2.36.1-1ubuntu2 libbpf0_1:0.3-2 libbsd0_0.10.0-1 libbz2-1.0_1.0.8-4ubuntu2 libc-bin_2.32-0ubuntu6 libc-dev-bin_2.32-0ubuntu6 libc6_2.32-0ubuntu6 libc6-dev_2.32-0ubuntu6 libcap-ng0_0.7.9-2.2build1 libcap2_1:2.44-1 libcap2-bin_1:2.44-1 libcc1-0_10.2.1-6ubuntu1 libcom-err2_1.45.6-1ubuntu1 libcrypt-dev_1:4.4.17-1ubuntu1 libcrypt1_1:4.4.17-1ubuntu1 libcryptsetup12_2:2.3.4-1ubuntu1 libctf-nobfd0_2.35.50.20210106-1ubuntu2 libctf0_2.35.50.20210106-1ubuntu2 libdb5.3_5.3.28+dfsg1-0.6ubuntu3 libdebconfclient0_0.256ubuntu1 libdebhelper-perl_13.3.1ubuntu1 libdevmapper1.02.1_2:1.02.167-1ubuntu4 libdpkg-perl_1.20.7.1ubuntu2 libelf1_0.182-3 libext2fs2_1.45.6-1ubuntu1 libfakeroot_1.25.3-1.1 libffi8ubuntu1_3.4~20200819gead65ca871-0ubuntu3 libfile-stripnondeterminism-perl_1.10.0-1 libgcc-10-dev_10.2.1-6ubuntu1 libgcc-s1_10.2.1-6ubuntu1 libgcrypt20_1.8.7-2ubuntu1 libgdbm-compat4_1.19-2 libgdbm6_1.19-2 libgmp10_2:6.2.0+dfsg-6ubuntu1 libgnutls30_3.7.0-5ubuntu1 libgomp1_10.2.1-6ubuntu1 libgpg-error0_1.38-2 libgssapi-krb5-2_1.18.3-4 libhogweed6_3.6-2 libicu67_67.1-5 libidn2-0_2.3.0-5 libip4tc2_1.8.5-3ubuntu4 libisl22_0.22.1-1 libisl23_0.23-1 libitm1_10.2.1-6ubuntu1 libjson-c5_0.15-1 libk5crypto3_1.18.3-4 libkeyutils1_1.6.1-2ubuntu1 libkmod2_27+20200310-2ubuntu1 libkrb5-3_1.18.3-4 libkrb5support0_1.18.3-4 liblockfile-bin_1.17-1 liblockfile1_1.17-1 liblz4-1_1.9.3-0ubuntu1 liblzma5_5.2.4-1ubuntu1 libmagic-mgc_1:5.39-3 libmagic1_1:5.39-3 libmnl0_1.0.4-3 libmount1_2.36.1-1ubuntu2 libmpc3_1.2.0-1 libmpfr6_4.1.0-3 libncurses6_6.2+20201114-2 libncursesw6_6.2+20201114-2 libnettle8_3.6-2 libnpth0_1.6-3 libnsl-dev_1.3.0-0ubuntu3 libnsl2_1.3.0-0ubuntu3 libnss-nis_3.1-0ubuntu4 libnss-nisplus_1.3-0ubuntu4 libp11-kit0_0.23.22-1 libpam-modules_1.3.1-5ubuntu6 libpam-modules-bin_1.3.1-5ubuntu6 libpam-runtime_1.3.1-5ubuntu6 libpam0g_1.3.1-5ubuntu6 libpcre2-8-0_10.35-2ubuntu1 libpcre3_2:8.39-13 libperl5.30_5.30.3-4 libperl5.32_5.32.0-6 libpipeline1_1.5.3-1 libpng16-16_1.6.37-3 libprocps8_2:3.3.16-5ubuntu2 libprotobuf-dev_3.12.4-1ubuntu1 libprotobuf-lite23_3.12.4-1ubuntu1 libprotobuf23_3.12.4-1ubuntu1 libprotoc23_3.12.4-1ubuntu1 libreadline8_8.1-1 libsasl2-2_2.1.27+dfsg-2ubuntu1 libsasl2-dev_2.1.27+dfsg-2ubuntu1 libsasl2-modules-db_2.1.27+dfsg-2ubuntu1 libseccomp2_2.4.3-1ubuntu6 libselinux1_3.1-2build2 libsemanage-common_3.1-1build2 libsemanage1_3.1-1build2 libsepol1_3.1-1 libsigsegv2_2.12-3 libsmartcols1_2.36.1-1ubuntu2 libsqlite3-0_3.34.0-1 libss2_1.45.6-1ubuntu1 libssl1.1_1.1.1f-1ubuntu5 libstdc++-10-dev_10.2.1-6ubuntu1 libstdc++6_10.2.1-6ubuntu1 libsub-override-perl_0.09-2 libsystemd0_247.1-4ubuntu1 libtasn1-6_4.16.0-2 libtinfo6_6.2+20201114-2 libtirpc-common_1.3.1-1 libtirpc-dev_1.3.1-1 libtirpc3_1.3.1-1 libtool_2.4.6-15 libubsan1_10.2.1-6ubuntu1 libuchardet0_0.0.7-1 libudev1_247.1-4ubuntu1 libunistring2_0.9.10-4 libuuid1_2.36.1-1ubuntu2 libxml2_2.9.10+dfsg-6.3build1 libxtables12_1.8.5-3ubuntu4 libxxhash0_0.8.0-2 libzstd1_1.4.8+dfsg-1 linux-libc-dev_5.8.0-36.40+21.04.1 lockfile-progs_0.1.18 login_1:4.8.1-1ubuntu8 logsave_1.45.6-1ubuntu1 lsb-base_11.1.0ubuntu2 lto-disabled-list_1 m4_1.4.18-5 make_4.3-4ubuntu1 man-db_2.9.3-2 mawk_1.3.4.20200120-2 mockery_0.0~git20181123.e78b021-3build1 mount_2.36.1-1ubuntu2 ncurses-base_6.2+20201114-2 ncurses-bin_6.2+20201114-2 openssl_1.1.1f-1ubuntu5 optipng_0.7.7-1 passwd_1:4.8.1-1ubuntu8 patch_2.7.6-7 perl_5.32.0-6 perl-base_5.32.0-6 perl-modules-5.30_5.30.3-4 perl-modules-5.32_5.32.0-6 pinentry-curses_1.1.0-4build1 pkgbinarymangler_146 po-debconf_1.0.21+nmu1 policyrcd-script-zg2_0.1-3 procps_2:3.3.16-5ubuntu2 protobuf-compiler_3.12.4-1ubuntu1 readline-common_8.1-1 rpcsvc-proto_1.4.2-0ubuntu4 sbuild-build-depends-consul-dummy_0.invalid.0 sbuild-build-depends-core-dummy_0.invalid.0 sed_4.7-1ubuntu1 sensible-utils_0.0.14 systemd_247.1-4ubuntu1 systemd-sysv_247.1-4ubuntu1 systemd-timesyncd_247.1-4ubuntu1 sysvinit-utils_2.96-5ubuntu1 tar_1.32+dfsg-1 tzdata_2021a-1ubuntu1 ubuntu-keyring_2020.06.17.1 util-linux_2.36.1-1ubuntu2 xz-utils_5.2.4-1ubuntu1 zlib1g_1:1.2.11.dfsg-2ubuntu4 zlib1g-dev_1:1.2.11.dfsg-2ubuntu4 +------------------------------------------------------------------------------+ | Build | +------------------------------------------------------------------------------+ Unpack source ------------- gpgv: Signature made Thu Jun 11 05:53:17 2020 UTC gpgv: using RSA key 50BC7CF939D20C272A6B065652B6BBD953968D1B gpgv: Can't check signature: No public key dpkg-source: warning: failed to verify signature on ./consul_1.7.4+dfsg1-1.dsc dpkg-source: info: extracting consul in /<>/consul-1.7.4+dfsg1 dpkg-source: info: unpacking consul_1.7.4+dfsg1.orig.tar.xz dpkg-source: info: unpacking consul_1.7.4+dfsg1-1.debian.tar.xz dpkg-source: info: using patch list from debian/patches/series dpkg-source: info: applying provider-no-k8s.patch dpkg-source: info: applying t-skip-unreliable-tests.patch dpkg-source: info: applying vendor-envoyproxy.patch Check disk space ---------------- Sufficient free space for build User Environment ---------------- APT_CONFIG=/var/lib/sbuild/apt.conf DEB_BUILD_OPTIONS=noautodbgsym parallel=4 HOME=/sbuild-nonexistent LANG=C.UTF-8 LC_ALL=C.UTF-8 LOGNAME=buildd PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games SCHROOT_ALIAS_NAME=build-PACKAGEBUILD-20426680 SCHROOT_CHROOT_NAME=build-PACKAGEBUILD-20426680 SCHROOT_COMMAND=env SCHROOT_GID=2501 SCHROOT_GROUP=buildd SCHROOT_SESSION_ID=build-PACKAGEBUILD-20426680 SCHROOT_UID=2001 SCHROOT_USER=buildd SHELL=/bin/sh TERM=unknown USER=buildd V=1 dpkg-buildpackage ----------------- dpkg-buildpackage: info: source package consul dpkg-buildpackage: info: source version 1.7.4+dfsg1-1 dpkg-buildpackage: info: source distribution unstable dpkg-source --before-build . dpkg-buildpackage: info: host architecture s390x fakeroot debian/rules clean dh clean --buildsystem=golang --with=golang,bash-completion --builddirectory=_build dh_auto_clean -O--buildsystem=golang -O--builddirectory=_build dh_autoreconf_clean -O--buildsystem=golang -O--builddirectory=_build debian/rules override_dh_clean make[1]: Entering directory '/<>/consul-1.7.4+dfsg1' dh_clean ## Remove Files-Excluded (when built from checkout or non-DFSG tarball): rm -f -rv `perl -0nE 'say $1 if m{^Files\-Excluded\:\s*(.*?)(?:\n\n|Files:|Comment:)}sm;' debian/copyright` find vendor -type d -empty -delete -print vendor/github.com/dgrijalva vendor/github.com/joyent vendor/github.com/gogo vendor/github.com/fatih vendor/github.com/mattn vendor/github.com/modern-go vendor/github.com/kr vendor/github.com/mitchellh vendor/github.com/DataDog vendor/github.com/spf13 vendor/github.com/imdario vendor/github.com/renier vendor/github.com/pierrec vendor/github.com/ryanuber vendor/github.com/miekg vendor/github.com/StackExchange vendor/github.com/circonus-labs vendor/github.com/googleapis vendor/github.com/boltdb vendor/github.com/peterbourgon vendor/github.com/gregjones vendor/github.com/pascaldekloe vendor/github.com/tv42 vendor/github.com/ghodss vendor/github.com/softlayer vendor/github.com/pmezard vendor/github.com/digitalocean vendor/github.com/golang vendor/github.com/beorn7 vendor/github.com/gophercloud vendor/github.com/armon vendor/github.com/shirou vendor/github.com/dimchansky vendor/github.com/json-iterator vendor/github.com/Azure vendor/github.com/posener vendor/github.com/NYTimes vendor/github.com/nicolai86 vendor/github.com/bgentry vendor/github.com/google vendor/github.com/davecgh vendor/github.com/go-ole vendor/github.com/pkg vendor/github.com/sirupsen vendor/github.com/prometheus vendor/github.com/denverdino vendor/github.com/elazarl vendor/github.com/jmespath vendor/github.com/Microsoft vendor/github.com/stretchr vendor/github.com/docker vendor/github.com/vmware vendor/github.com/matttproud vendor/github.com/packethost vendor/gopkg.in/square vendor/golang.org rm -f -r test/integration make[1]: Leaving directory '/<>/consul-1.7.4+dfsg1' debian/rules build-arch dh build-arch --buildsystem=golang --with=golang,bash-completion --builddirectory=_build dh_update_autotools_config -a -O--buildsystem=golang -O--builddirectory=_build dh_autoreconf -a -O--buildsystem=golang -O--builddirectory=_build debian/rules override_dh_auto_configure make[1]: Entering directory '/<>/consul-1.7.4+dfsg1' dh_auto_configure mkdir -v -p _build/src/github.com/keybase/ mkdir: created directory '_build/src/github.com/keybase/' ln -sv /usr/share/gocode/src/golang.org/x/crypto _build/src/github.com/keybase/go-crypto '_build/src/github.com/keybase/go-crypto' -> '/usr/share/gocode/src/golang.org/x/crypto' mkdir -v -p _build/src/github.com/SermoDigital/ mkdir: created directory '_build/src/github.com/SermoDigital/' ln -sv /usr/share/gocode/src/gopkg.in/square/go-jose.v1 _build/src/github.com/SermoDigital/jose '_build/src/github.com/SermoDigital/jose' -> '/usr/share/gocode/src/gopkg.in/square/go-jose.v1' make[1]: Leaving directory '/<>/consul-1.7.4+dfsg1' debian/rules override_dh_auto_build make[1]: Entering directory '/<>/consul-1.7.4+dfsg1' export GOPATH=/<>/consul-1.7.4+dfsg1/_build \ && /usr/bin/make -C _build/src/github.com/hashicorp/consul --makefile=/<>/consul-1.7.4+dfsg1/GNUmakefile proto make[2]: Entering directory '/<>/consul-1.7.4+dfsg1/_build/src/github.com/hashicorp/consul' bash: line 1: git: command not found bash: line 1: git: command not found bash: line 1: git: command not found bash: line 1: git: command not found bash: line 1: git: command not found bash: line 1: git: command not found bash: line 1: git: command not found Generated all protobuf Go files make[2]: Leaving directory '/<>/consul-1.7.4+dfsg1/_build/src/github.com/hashicorp/consul' dh_auto_build -v cd _build && go version go version go1.15.4 linux/s390x cd _build && go env GO111MODULE="off" GOARCH="s390x" GOBIN="" GOCACHE="/<>/consul-1.7.4+dfsg1/_build/go-build" GOENV="/<>/consul-1.7.4+dfsg1/debian/tmp/.config/go/env" GOEXE="" GOFLAGS="" GOHOSTARCH="s390x" GOHOSTOS="linux" GOINSECURE="" GOMODCACHE="/<>/consul-1.7.4+dfsg1/_build/pkg/mod" GONOPROXY="" GONOSUMDB="" GOOS="linux" GOPATH="/<>/consul-1.7.4+dfsg1/_build" GOPRIVATE="" GOPROXY="off" GOROOT="/usr/lib/go-1.15" GOSUMDB="sum.golang.org" GOTMPDIR="" GOTOOLDIR="/usr/lib/go-1.15/pkg/tool/linux_s390x" GCCGO="gccgo" AR="ar" CC="gcc" CXX="g++" CGO_ENABLED="1" GOMOD="" CGO_CFLAGS="-g -O2 -ffile-prefix-map=/<>/consul-1.7.4+dfsg1=. -fstack-protector-strong -Wformat -Werror=format-security" CGO_CPPFLAGS="-Wdate-time -D_FORTIFY_SOURCE=2" CGO_CXXFLAGS="-g -O2 -ffile-prefix-map=/<>/consul-1.7.4+dfsg1=. -fstack-protector-strong -Wformat -Werror=format-security" CGO_FFLAGS="-g -O2 -ffile-prefix-map=/<>/consul-1.7.4+dfsg1=. -fstack-protector-strong" CGO_LDFLAGS="-Wl,-Bsymbolic-functions -Wl,-z,relro" PKG_CONFIG="pkg-config" GOGCCFLAGS="-fPIC -m64 -march=z196 -pthread -fmessage-length=0 -fdebug-prefix-map=/tmp/go-build302978195=/tmp/go-build -gno-record-gcc-switches" cd _build && go generate -v github.com/hashicorp/consul github.com/hashicorp/consul/acl github.com/hashicorp/consul/agent github.com/hashicorp/consul/agent/ae github.com/hashicorp/consul/agent/agentpb github.com/hashicorp/consul/agent/cache github.com/hashicorp/consul/agent/cache-types github.com/hashicorp/consul/agent/checks github.com/hashicorp/consul/agent/config github.com/hashicorp/consul/agent/connect github.com/hashicorp/consul/agent/connect/ca github.com/hashicorp/consul/agent/consul github.com/hashicorp/consul/agent/consul/authmethod github.com/hashicorp/consul/agent/consul/authmethod/kubeauth github.com/hashicorp/consul/agent/consul/authmethod/testauth github.com/hashicorp/consul/agent/consul/autopilot github.com/hashicorp/consul/agent/consul/discoverychain github.com/hashicorp/consul/agent/consul/fsm github.com/hashicorp/consul/agent/consul/prepared_query github.com/hashicorp/consul/agent/consul/state github.com/hashicorp/consul/agent/debug github.com/hashicorp/consul/agent/exec github.com/hashicorp/consul/agent/local github.com/hashicorp/consul/agent/metadata github.com/hashicorp/consul/agent/mock github.com/hashicorp/consul/agent/pool github.com/hashicorp/consul/agent/proxycfg github.com/hashicorp/consul/agent/router github.com/hashicorp/consul/agent/structs github.com/hashicorp/consul/agent/systemd github.com/hashicorp/consul/agent/token github.com/hashicorp/consul/agent/xds github.com/hashicorp/consul/api github.com/hashicorp/consul/api/watch github.com/hashicorp/consul/command github.com/hashicorp/consul/command/acl github.com/hashicorp/consul/command/acl/agenttokens github.com/hashicorp/consul/command/acl/authmethod github.com/hashicorp/consul/command/acl/authmethod/create github.com/hashicorp/consul/command/acl/authmethod/delete github.com/hashicorp/consul/command/acl/authmethod/list github.com/hashicorp/consul/command/acl/authmethod/read github.com/hashicorp/consul/command/acl/authmethod/update github.com/hashicorp/consul/command/acl/bindingrule github.com/hashicorp/consul/command/acl/bindingrule/create github.com/hashicorp/consul/command/acl/bindingrule/delete github.com/hashicorp/consul/command/acl/bindingrule/list github.com/hashicorp/consul/command/acl/bindingrule/read github.com/hashicorp/consul/command/acl/bindingrule/update github.com/hashicorp/consul/command/acl/bootstrap github.com/hashicorp/consul/command/acl/policy github.com/hashicorp/consul/command/acl/policy/create github.com/hashicorp/consul/command/acl/policy/delete github.com/hashicorp/consul/command/acl/policy/list github.com/hashicorp/consul/command/acl/policy/read github.com/hashicorp/consul/command/acl/policy/update github.com/hashicorp/consul/command/acl/role github.com/hashicorp/consul/command/acl/role/create github.com/hashicorp/consul/command/acl/role/delete github.com/hashicorp/consul/command/acl/role/list github.com/hashicorp/consul/command/acl/role/read github.com/hashicorp/consul/command/acl/role/update github.com/hashicorp/consul/command/acl/rules github.com/hashicorp/consul/command/acl/token github.com/hashicorp/consul/command/acl/token/clone github.com/hashicorp/consul/command/acl/token/create github.com/hashicorp/consul/command/acl/token/delete github.com/hashicorp/consul/command/acl/token/list github.com/hashicorp/consul/command/acl/token/read github.com/hashicorp/consul/command/acl/token/update github.com/hashicorp/consul/command/agent github.com/hashicorp/consul/command/catalog github.com/hashicorp/consul/command/catalog/list/dc github.com/hashicorp/consul/command/catalog/list/nodes github.com/hashicorp/consul/command/catalog/list/services github.com/hashicorp/consul/command/config github.com/hashicorp/consul/command/config/delete github.com/hashicorp/consul/command/config/list github.com/hashicorp/consul/command/config/read github.com/hashicorp/consul/command/config/write github.com/hashicorp/consul/command/connect github.com/hashicorp/consul/command/connect/ca github.com/hashicorp/consul/command/connect/ca/get github.com/hashicorp/consul/command/connect/ca/set github.com/hashicorp/consul/command/connect/envoy github.com/hashicorp/consul/command/connect/envoy/pipe-bootstrap github.com/hashicorp/consul/command/connect/proxy github.com/hashicorp/consul/command/debug github.com/hashicorp/consul/command/event github.com/hashicorp/consul/command/exec github.com/hashicorp/consul/command/flags github.com/hashicorp/consul/command/forceleave github.com/hashicorp/consul/command/helpers github.com/hashicorp/consul/command/info github.com/hashicorp/consul/command/intention github.com/hashicorp/consul/command/intention/check github.com/hashicorp/consul/command/intention/create github.com/hashicorp/consul/command/intention/delete github.com/hashicorp/consul/command/intention/finder github.com/hashicorp/consul/command/intention/get github.com/hashicorp/consul/command/intention/match github.com/hashicorp/consul/command/join github.com/hashicorp/consul/command/keygen github.com/hashicorp/consul/command/keyring github.com/hashicorp/consul/command/kv github.com/hashicorp/consul/command/kv/del github.com/hashicorp/consul/command/kv/exp github.com/hashicorp/consul/command/kv/get github.com/hashicorp/consul/command/kv/imp github.com/hashicorp/consul/command/kv/impexp github.com/hashicorp/consul/command/kv/put github.com/hashicorp/consul/command/leave github.com/hashicorp/consul/command/lock github.com/hashicorp/consul/command/login github.com/hashicorp/consul/command/logout github.com/hashicorp/consul/command/maint github.com/hashicorp/consul/command/members github.com/hashicorp/consul/command/monitor github.com/hashicorp/consul/command/operator github.com/hashicorp/consul/command/operator/autopilot github.com/hashicorp/consul/command/operator/autopilot/get github.com/hashicorp/consul/command/operator/autopilot/set github.com/hashicorp/consul/command/operator/raft github.com/hashicorp/consul/command/operator/raft/listpeers github.com/hashicorp/consul/command/operator/raft/removepeer github.com/hashicorp/consul/command/reload github.com/hashicorp/consul/command/rtt github.com/hashicorp/consul/command/services github.com/hashicorp/consul/command/services/deregister github.com/hashicorp/consul/command/services/register github.com/hashicorp/consul/command/snapshot github.com/hashicorp/consul/command/snapshot/inspect github.com/hashicorp/consul/command/snapshot/restore github.com/hashicorp/consul/command/snapshot/save github.com/hashicorp/consul/command/tls github.com/hashicorp/consul/command/tls/ca github.com/hashicorp/consul/command/tls/ca/create github.com/hashicorp/consul/command/tls/cert github.com/hashicorp/consul/command/tls/cert/create github.com/hashicorp/consul/command/validate github.com/hashicorp/consul/command/version github.com/hashicorp/consul/command/watch github.com/hashicorp/consul/connect github.com/hashicorp/consul/connect/certgen github.com/hashicorp/consul/connect/proxy github.com/hashicorp/consul/ipaddr github.com/hashicorp/consul/lib github.com/hashicorp/consul/lib/file github.com/hashicorp/consul/lib/semaphore github.com/hashicorp/consul/logging github.com/hashicorp/consul/logging/monitor github.com/hashicorp/consul/sdk/freeport github.com/hashicorp/consul/sdk/testutil github.com/hashicorp/consul/sdk/testutil/retry github.com/hashicorp/consul/sentinel github.com/hashicorp/consul/service_os github.com/hashicorp/consul/snapshot github.com/hashicorp/consul/testrpc github.com/hashicorp/consul/tlsutil github.com/hashicorp/consul/types github.com/hashicorp/consul/version src/github.com/hashicorp/consul/main.go src/github.com/hashicorp/consul/main_test.go src/github.com/hashicorp/consul/acl/acl.go src/github.com/hashicorp/consul/acl/acl_oss.go src/github.com/hashicorp/consul/acl/acl_test.go src/github.com/hashicorp/consul/acl/authorizer.go src/github.com/hashicorp/consul/acl/authorizer_oss.go src/github.com/hashicorp/consul/acl/authorizer_test.go src/github.com/hashicorp/consul/acl/chained_authorizer.go src/github.com/hashicorp/consul/acl/chained_authorizer_test.go src/github.com/hashicorp/consul/acl/errors.go src/github.com/hashicorp/consul/acl/policy.go src/github.com/hashicorp/consul/acl/policy_authorizer.go src/github.com/hashicorp/consul/acl/policy_authorizer_oss.go src/github.com/hashicorp/consul/acl/policy_authorizer_test.go src/github.com/hashicorp/consul/acl/policy_merger.go src/github.com/hashicorp/consul/acl/policy_merger_oss.go src/github.com/hashicorp/consul/acl/policy_oss.go src/github.com/hashicorp/consul/acl/policy_test.go src/github.com/hashicorp/consul/acl/static_authorizer.go src/github.com/hashicorp/consul/acl/static_authorizer_test.go src/github.com/hashicorp/consul/agent/acl.go src/github.com/hashicorp/consul/agent/acl_endpoint.go src/github.com/hashicorp/consul/agent/acl_endpoint_legacy.go src/github.com/hashicorp/consul/agent/acl_endpoint_legacy_test.go src/github.com/hashicorp/consul/agent/acl_endpoint_test.go src/github.com/hashicorp/consul/agent/acl_test.go src/github.com/hashicorp/consul/agent/agent.go src/github.com/hashicorp/consul/agent/agent_endpoint.go src/github.com/hashicorp/consul/agent/agent_endpoint_test.go src/github.com/hashicorp/consul/agent/agent_oss.go src/github.com/hashicorp/consul/agent/agent_test.go src/github.com/hashicorp/consul/agent/bindata_assetfs.go src/github.com/hashicorp/consul/agent/blacklist.go src/github.com/hashicorp/consul/agent/blacklist_test.go src/github.com/hashicorp/consul/agent/catalog_endpoint.go src/github.com/hashicorp/consul/agent/catalog_endpoint_test.go src/github.com/hashicorp/consul/agent/check.go src/github.com/hashicorp/consul/agent/config.go src/github.com/hashicorp/consul/agent/config_endpoint.go src/github.com/hashicorp/consul/agent/config_endpoint_test.go src/github.com/hashicorp/consul/agent/connect_auth.go src/github.com/hashicorp/consul/agent/connect_ca_endpoint.go src/github.com/hashicorp/consul/agent/connect_ca_endpoint_test.go src/github.com/hashicorp/consul/agent/coordinate_endpoint.go src/github.com/hashicorp/consul/agent/coordinate_endpoint_test.go src/github.com/hashicorp/consul/agent/discovery_chain_endpoint.go src/github.com/hashicorp/consul/agent/discovery_chain_endpoint_test.go src/github.com/hashicorp/consul/agent/dns.go src/github.com/hashicorp/consul/agent/dns_oss.go src/github.com/hashicorp/consul/agent/dns_test.go src/github.com/hashicorp/consul/agent/enterprise_delegate_oss.go src/github.com/hashicorp/consul/agent/event_endpoint.go src/github.com/hashicorp/consul/agent/event_endpoint_test.go src/github.com/hashicorp/consul/agent/health_endpoint.go src/github.com/hashicorp/consul/agent/health_endpoint_test.go src/github.com/hashicorp/consul/agent/http.go src/github.com/hashicorp/consul/agent/http_decode_test.go src/github.com/hashicorp/consul/agent/http_oss.go src/github.com/hashicorp/consul/agent/http_oss_test.go src/github.com/hashicorp/consul/agent/http_register.go src/github.com/hashicorp/consul/agent/http_test.go src/github.com/hashicorp/consul/agent/intentions_endpoint.go src/github.com/hashicorp/consul/agent/intentions_endpoint_test.go src/github.com/hashicorp/consul/agent/keyring.go src/github.com/hashicorp/consul/agent/keyring_test.go src/github.com/hashicorp/consul/agent/kvs_endpoint.go src/github.com/hashicorp/consul/agent/kvs_endpoint_test.go src/github.com/hashicorp/consul/agent/notify.go src/github.com/hashicorp/consul/agent/notify_test.go src/github.com/hashicorp/consul/agent/operator_endpoint.go src/github.com/hashicorp/consul/agent/operator_endpoint_test.go src/github.com/hashicorp/consul/agent/prepared_query_endpoint.go src/github.com/hashicorp/consul/agent/prepared_query_endpoint_test.go src/github.com/hashicorp/consul/agent/remote_exec.go src/github.com/hashicorp/consul/agent/remote_exec_test.go src/github.com/hashicorp/consul/agent/retry_join.go src/github.com/hashicorp/consul/agent/service_checks_test.go src/github.com/hashicorp/consul/agent/service_manager.go src/github.com/hashicorp/consul/agent/service_manager_test.go src/github.com/hashicorp/consul/agent/session_endpoint.go src/github.com/hashicorp/consul/agent/session_endpoint_test.go src/github.com/hashicorp/consul/agent/sidecar_service.go src/github.com/hashicorp/consul/agent/sidecar_service_test.go src/github.com/hashicorp/consul/agent/signal_unix.go src/github.com/hashicorp/consul/agent/snapshot_endpoint.go src/github.com/hashicorp/consul/agent/snapshot_endpoint_test.go src/github.com/hashicorp/consul/agent/status_endpoint.go src/github.com/hashicorp/consul/agent/status_endpoint_test.go src/github.com/hashicorp/consul/agent/testagent.go src/github.com/hashicorp/consul/agent/testagent_test.go src/github.com/hashicorp/consul/agent/translate_addr.go src/github.com/hashicorp/consul/agent/txn_endpoint.go src/github.com/hashicorp/consul/agent/txn_endpoint_test.go src/github.com/hashicorp/consul/agent/ui_endpoint.go src/github.com/hashicorp/consul/agent/ui_endpoint_test.go src/github.com/hashicorp/consul/agent/user_event.go src/github.com/hashicorp/consul/agent/user_event_test.go src/github.com/hashicorp/consul/agent/util.go src/github.com/hashicorp/consul/agent/util_test.go src/github.com/hashicorp/consul/agent/watch_handler.go src/github.com/hashicorp/consul/agent/watch_handler_test.go src/github.com/hashicorp/consul/agent/ae/ae.go src/github.com/hashicorp/consul/agent/ae/ae_test.go src/github.com/hashicorp/consul/agent/ae/trigger.go src/github.com/hashicorp/consul/agent/agentpb/acl.pb.binary.go src/github.com/hashicorp/consul/agent/agentpb/acl.pb.go src/github.com/hashicorp/consul/agent/agentpb/common.go src/github.com/hashicorp/consul/agent/agentpb/common.pb.binary.go src/github.com/hashicorp/consul/agent/agentpb/common.pb.go src/github.com/hashicorp/consul/agent/cache/cache.go Generating mock for: Request in file: /<>/consul-1.7.4+dfsg1/_build/src/github.com/hashicorp/consul/agent/cache/mock_Request.go Generating mock for: Type in file: /<>/consul-1.7.4+dfsg1/_build/src/github.com/hashicorp/consul/agent/cache/mock_Type.go src/github.com/hashicorp/consul/agent/cache/cache_test.go src/github.com/hashicorp/consul/agent/cache/entry.go src/github.com/hashicorp/consul/agent/cache/entry_test.go src/github.com/hashicorp/consul/agent/cache/mock_Request.go src/github.com/hashicorp/consul/agent/cache/mock_Type.go src/github.com/hashicorp/consul/agent/cache/request.go src/github.com/hashicorp/consul/agent/cache/testing.go src/github.com/hashicorp/consul/agent/cache/type.go src/github.com/hashicorp/consul/agent/cache/watch.go src/github.com/hashicorp/consul/agent/cache/watch_test.go src/github.com/hashicorp/consul/agent/cache-types/catalog_datacenters.go src/github.com/hashicorp/consul/agent/cache-types/catalog_datacenters_test.go src/github.com/hashicorp/consul/agent/cache-types/catalog_list_services.go src/github.com/hashicorp/consul/agent/cache-types/catalog_list_services_test.go src/github.com/hashicorp/consul/agent/cache-types/catalog_service_list.go src/github.com/hashicorp/consul/agent/cache-types/catalog_service_list_test.go src/github.com/hashicorp/consul/agent/cache-types/catalog_services.go src/github.com/hashicorp/consul/agent/cache-types/catalog_services_test.go src/github.com/hashicorp/consul/agent/cache-types/config_entry.go src/github.com/hashicorp/consul/agent/cache-types/config_entry_test.go src/github.com/hashicorp/consul/agent/cache-types/connect_ca_leaf.go src/github.com/hashicorp/consul/agent/cache-types/connect_ca_leaf_oss.go src/github.com/hashicorp/consul/agent/cache-types/connect_ca_leaf_test.go src/github.com/hashicorp/consul/agent/cache-types/connect_ca_root.go src/github.com/hashicorp/consul/agent/cache-types/connect_ca_root_test.go src/github.com/hashicorp/consul/agent/cache-types/discovery_chain.go src/github.com/hashicorp/consul/agent/cache-types/discovery_chain_test.go src/github.com/hashicorp/consul/agent/cache-types/health_services.go src/github.com/hashicorp/consul/agent/cache-types/health_services_test.go src/github.com/hashicorp/consul/agent/cache-types/intention_match.go src/github.com/hashicorp/consul/agent/cache-types/intention_match_test.go src/github.com/hashicorp/consul/agent/cache-types/mock_RPC.go src/github.com/hashicorp/consul/agent/cache-types/node_services.go src/github.com/hashicorp/consul/agent/cache-types/node_services_test.go src/github.com/hashicorp/consul/agent/cache-types/prepared_query.go src/github.com/hashicorp/consul/agent/cache-types/prepared_query_test.go src/github.com/hashicorp/consul/agent/cache-types/resolved_service_config.go src/github.com/hashicorp/consul/agent/cache-types/resolved_service_config_test.go src/github.com/hashicorp/consul/agent/cache-types/rpc.go Generating mock for: Agent in file: /<>/consul-1.7.4+dfsg1/_build/src/github.com/hashicorp/consul/agent/cache-types/mock_Agent.go Generating mock for: RPC in file: /<>/consul-1.7.4+dfsg1/_build/src/github.com/hashicorp/consul/agent/cache-types/mock_RPC.go src/github.com/hashicorp/consul/agent/cache-types/service_checks.go src/github.com/hashicorp/consul/agent/cache-types/service_checks_test.go src/github.com/hashicorp/consul/agent/cache-types/service_dump.go src/github.com/hashicorp/consul/agent/cache-types/service_dump_test.go src/github.com/hashicorp/consul/agent/cache-types/testing.go src/github.com/hashicorp/consul/agent/checks/alias.go src/github.com/hashicorp/consul/agent/checks/alias_test.go src/github.com/hashicorp/consul/agent/checks/check.go src/github.com/hashicorp/consul/agent/checks/check_test.go src/github.com/hashicorp/consul/agent/checks/docker.go src/github.com/hashicorp/consul/agent/checks/docker_unix.go src/github.com/hashicorp/consul/agent/checks/grpc.go src/github.com/hashicorp/consul/agent/checks/grpc_test.go src/github.com/hashicorp/consul/agent/config/builder.go src/github.com/hashicorp/consul/agent/config/builder_oss.go src/github.com/hashicorp/consul/agent/config/builder_oss_test.go src/github.com/hashicorp/consul/agent/config/config.go src/github.com/hashicorp/consul/agent/config/config_oss.go src/github.com/hashicorp/consul/agent/config/default.go src/github.com/hashicorp/consul/agent/config/default_oss.go src/github.com/hashicorp/consul/agent/config/doc.go src/github.com/hashicorp/consul/agent/config/flags.go src/github.com/hashicorp/consul/agent/config/flags_test.go src/github.com/hashicorp/consul/agent/config/flagset.go src/github.com/hashicorp/consul/agent/config/merge.go src/github.com/hashicorp/consul/agent/config/merge_test.go src/github.com/hashicorp/consul/agent/config/runtime.go src/github.com/hashicorp/consul/agent/config/runtime_oss.go src/github.com/hashicorp/consul/agent/config/runtime_oss_test.go src/github.com/hashicorp/consul/agent/config/runtime_test.go src/github.com/hashicorp/consul/agent/config/segment_oss.go src/github.com/hashicorp/consul/agent/config/segment_oss_test.go src/github.com/hashicorp/consul/agent/connect/common_names.go src/github.com/hashicorp/consul/agent/connect/common_names_test.go src/github.com/hashicorp/consul/agent/connect/csr.go src/github.com/hashicorp/consul/agent/connect/generate.go src/github.com/hashicorp/consul/agent/connect/generate_test.go src/github.com/hashicorp/consul/agent/connect/parsing.go src/github.com/hashicorp/consul/agent/connect/sni.go src/github.com/hashicorp/consul/agent/connect/sni_test.go src/github.com/hashicorp/consul/agent/connect/testing_ca.go src/github.com/hashicorp/consul/agent/connect/testing_ca_test.go src/github.com/hashicorp/consul/agent/connect/testing_spiffe.go src/github.com/hashicorp/consul/agent/connect/uri.go src/github.com/hashicorp/consul/agent/connect/uri_agent.go src/github.com/hashicorp/consul/agent/connect/uri_agent_test.go src/github.com/hashicorp/consul/agent/connect/uri_service.go src/github.com/hashicorp/consul/agent/connect/uri_service_oss.go src/github.com/hashicorp/consul/agent/connect/uri_service_test.go src/github.com/hashicorp/consul/agent/connect/uri_signing.go src/github.com/hashicorp/consul/agent/connect/uri_signing_test.go src/github.com/hashicorp/consul/agent/connect/uri_test.go src/github.com/hashicorp/consul/agent/connect/ca/common.go src/github.com/hashicorp/consul/agent/connect/ca/mock_Provider.go src/github.com/hashicorp/consul/agent/connect/ca/provider.go Generating mock for: Provider in file: /<>/consul-1.7.4+dfsg1/_build/src/github.com/hashicorp/consul/agent/connect/ca/mock_Provider.go src/github.com/hashicorp/consul/agent/connect/ca/provider_aws.go src/github.com/hashicorp/consul/agent/connect/ca/provider_aws_test.go src/github.com/hashicorp/consul/agent/connect/ca/provider_consul.go src/github.com/hashicorp/consul/agent/connect/ca/provider_consul_config.go src/github.com/hashicorp/consul/agent/connect/ca/provider_consul_test.go src/github.com/hashicorp/consul/agent/connect/ca/provider_test.go src/github.com/hashicorp/consul/agent/connect/ca/provider_vault.go src/github.com/hashicorp/consul/agent/connect/ca/provider_vault_test.go src/github.com/hashicorp/consul/agent/connect/ca/testing.go src/github.com/hashicorp/consul/agent/consul/acl.go src/github.com/hashicorp/consul/agent/consul/acl_authmethod.go src/github.com/hashicorp/consul/agent/consul/acl_authmethod_test.go src/github.com/hashicorp/consul/agent/consul/acl_client.go src/github.com/hashicorp/consul/agent/consul/acl_endpoint.go src/github.com/hashicorp/consul/agent/consul/acl_endpoint_legacy.go src/github.com/hashicorp/consul/agent/consul/acl_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/acl_oss.go src/github.com/hashicorp/consul/agent/consul/acl_oss_test.go src/github.com/hashicorp/consul/agent/consul/acl_replication.go src/github.com/hashicorp/consul/agent/consul/acl_replication_legacy.go src/github.com/hashicorp/consul/agent/consul/acl_replication_legacy_test.go src/github.com/hashicorp/consul/agent/consul/acl_replication_test.go src/github.com/hashicorp/consul/agent/consul/acl_replication_types.go src/github.com/hashicorp/consul/agent/consul/acl_server.go src/github.com/hashicorp/consul/agent/consul/acl_server_oss.go src/github.com/hashicorp/consul/agent/consul/acl_test.go src/github.com/hashicorp/consul/agent/consul/acl_token_exp.go src/github.com/hashicorp/consul/agent/consul/acl_token_exp_test.go src/github.com/hashicorp/consul/agent/consul/auto_encrypt.go src/github.com/hashicorp/consul/agent/consul/auto_encrypt_endpoint.go src/github.com/hashicorp/consul/agent/consul/auto_encrypt_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/auto_encrypt_test.go src/github.com/hashicorp/consul/agent/consul/autopilot.go src/github.com/hashicorp/consul/agent/consul/autopilot_oss.go src/github.com/hashicorp/consul/agent/consul/autopilot_test.go src/github.com/hashicorp/consul/agent/consul/catalog_endpoint.go src/github.com/hashicorp/consul/agent/consul/catalog_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/client.go src/github.com/hashicorp/consul/agent/consul/client_serf.go src/github.com/hashicorp/consul/agent/consul/client_test.go src/github.com/hashicorp/consul/agent/consul/config.go src/github.com/hashicorp/consul/agent/consul/config_endpoint.go src/github.com/hashicorp/consul/agent/consul/config_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/config_replication.go src/github.com/hashicorp/consul/agent/consul/config_replication_test.go src/github.com/hashicorp/consul/agent/consul/connect_ca_endpoint.go src/github.com/hashicorp/consul/agent/consul/connect_ca_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/consul_ca_delegate.go src/github.com/hashicorp/consul/agent/consul/coordinate_endpoint.go src/github.com/hashicorp/consul/agent/consul/coordinate_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/discovery_chain_endpoint.go src/github.com/hashicorp/consul/agent/consul/discovery_chain_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/enterprise_client_oss.go src/github.com/hashicorp/consul/agent/consul/enterprise_config_oss.go src/github.com/hashicorp/consul/agent/consul/enterprise_server_oss.go src/github.com/hashicorp/consul/agent/consul/filter.go src/github.com/hashicorp/consul/agent/consul/filter_test.go src/github.com/hashicorp/consul/agent/consul/flood.go src/github.com/hashicorp/consul/agent/consul/health_endpoint.go src/github.com/hashicorp/consul/agent/consul/health_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/helper_test.go src/github.com/hashicorp/consul/agent/consul/intention_endpoint.go src/github.com/hashicorp/consul/agent/consul/intention_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/internal_endpoint.go src/github.com/hashicorp/consul/agent/consul/internal_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/issue_test.go src/github.com/hashicorp/consul/agent/consul/kvs_endpoint.go src/github.com/hashicorp/consul/agent/consul/kvs_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/leader.go src/github.com/hashicorp/consul/agent/consul/leader_connect.go src/github.com/hashicorp/consul/agent/consul/leader_connect_test.go src/github.com/hashicorp/consul/agent/consul/leader_routine_manager.go src/github.com/hashicorp/consul/agent/consul/leader_routine_manager_test.go src/github.com/hashicorp/consul/agent/consul/leader_test.go src/github.com/hashicorp/consul/agent/consul/logging.go src/github.com/hashicorp/consul/agent/consul/logging_test.go src/github.com/hashicorp/consul/agent/consul/merge.go src/github.com/hashicorp/consul/agent/consul/merge_test.go src/github.com/hashicorp/consul/agent/consul/operator_autopilot_endpoint.go src/github.com/hashicorp/consul/agent/consul/operator_autopilot_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/operator_endpoint.go src/github.com/hashicorp/consul/agent/consul/operator_raft_endpoint.go src/github.com/hashicorp/consul/agent/consul/operator_raft_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/prepared_query_endpoint.go src/github.com/hashicorp/consul/agent/consul/prepared_query_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/raft_rpc.go src/github.com/hashicorp/consul/agent/consul/replication.go src/github.com/hashicorp/consul/agent/consul/replication_test.go src/github.com/hashicorp/consul/agent/consul/rpc.go src/github.com/hashicorp/consul/agent/consul/rpc_test.go src/github.com/hashicorp/consul/agent/consul/rtt.go src/github.com/hashicorp/consul/agent/consul/rtt_test.go src/github.com/hashicorp/consul/agent/consul/segment_oss.go src/github.com/hashicorp/consul/agent/consul/serf_test.go src/github.com/hashicorp/consul/agent/consul/server.go src/github.com/hashicorp/consul/agent/consul/server_lookup.go src/github.com/hashicorp/consul/agent/consul/server_lookup_test.go src/github.com/hashicorp/consul/agent/consul/server_oss.go src/github.com/hashicorp/consul/agent/consul/server_serf.go src/github.com/hashicorp/consul/agent/consul/server_test.go src/github.com/hashicorp/consul/agent/consul/session_endpoint.go src/github.com/hashicorp/consul/agent/consul/session_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/session_timers.go src/github.com/hashicorp/consul/agent/consul/session_timers_test.go src/github.com/hashicorp/consul/agent/consul/session_ttl.go src/github.com/hashicorp/consul/agent/consul/session_ttl_test.go src/github.com/hashicorp/consul/agent/consul/snapshot_endpoint.go src/github.com/hashicorp/consul/agent/consul/snapshot_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/stats_fetcher.go src/github.com/hashicorp/consul/agent/consul/stats_fetcher_test.go src/github.com/hashicorp/consul/agent/consul/status_endpoint.go src/github.com/hashicorp/consul/agent/consul/status_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/txn_endpoint.go src/github.com/hashicorp/consul/agent/consul/txn_endpoint_test.go src/github.com/hashicorp/consul/agent/consul/util.go src/github.com/hashicorp/consul/agent/consul/util_test.go src/github.com/hashicorp/consul/agent/consul/authmethod/authmethods.go src/github.com/hashicorp/consul/agent/consul/authmethod/authmethods_oss.go src/github.com/hashicorp/consul/agent/consul/authmethod/kubeauth/k8s.go src/github.com/hashicorp/consul/agent/consul/authmethod/kubeauth/k8s_oss.go src/github.com/hashicorp/consul/agent/consul/authmethod/kubeauth/k8s_test.go src/github.com/hashicorp/consul/agent/consul/authmethod/kubeauth/testing.go src/github.com/hashicorp/consul/agent/consul/authmethod/testauth/testing.go src/github.com/hashicorp/consul/agent/consul/authmethod/testauth/testing_oss.go src/github.com/hashicorp/consul/agent/consul/autopilot/autopilot.go src/github.com/hashicorp/consul/agent/consul/autopilot/autopilot_test.go src/github.com/hashicorp/consul/agent/consul/autopilot/promotion.go src/github.com/hashicorp/consul/agent/consul/autopilot/promotion_test.go src/github.com/hashicorp/consul/agent/consul/autopilot/structs.go src/github.com/hashicorp/consul/agent/consul/autopilot/structs_test.go src/github.com/hashicorp/consul/agent/consul/discoverychain/compile.go src/github.com/hashicorp/consul/agent/consul/discoverychain/compile_oss.go src/github.com/hashicorp/consul/agent/consul/discoverychain/compile_test.go src/github.com/hashicorp/consul/agent/consul/discoverychain/string_stack.go src/github.com/hashicorp/consul/agent/consul/discoverychain/string_stack_test.go src/github.com/hashicorp/consul/agent/consul/discoverychain/testing.go src/github.com/hashicorp/consul/agent/consul/fsm/commands_oss.go src/github.com/hashicorp/consul/agent/consul/fsm/commands_oss_test.go src/github.com/hashicorp/consul/agent/consul/fsm/fsm.go src/github.com/hashicorp/consul/agent/consul/fsm/fsm_test.go src/github.com/hashicorp/consul/agent/consul/fsm/snapshot.go src/github.com/hashicorp/consul/agent/consul/fsm/snapshot_oss.go src/github.com/hashicorp/consul/agent/consul/fsm/snapshot_oss_test.go src/github.com/hashicorp/consul/agent/consul/prepared_query/template.go src/github.com/hashicorp/consul/agent/consul/prepared_query/template_test.go src/github.com/hashicorp/consul/agent/consul/prepared_query/walk.go src/github.com/hashicorp/consul/agent/consul/prepared_query/walk_oss_test.go src/github.com/hashicorp/consul/agent/consul/prepared_query/walk_test.go src/github.com/hashicorp/consul/agent/consul/state/acl.go src/github.com/hashicorp/consul/agent/consul/state/acl_oss.go src/github.com/hashicorp/consul/agent/consul/state/acl_test.go src/github.com/hashicorp/consul/agent/consul/state/autopilot.go src/github.com/hashicorp/consul/agent/consul/state/autopilot_test.go src/github.com/hashicorp/consul/agent/consul/state/catalog.go src/github.com/hashicorp/consul/agent/consul/state/catalog_oss.go src/github.com/hashicorp/consul/agent/consul/state/catalog_test.go src/github.com/hashicorp/consul/agent/consul/state/config_entry.go src/github.com/hashicorp/consul/agent/consul/state/config_entry_oss.go src/github.com/hashicorp/consul/agent/consul/state/config_entry_test.go src/github.com/hashicorp/consul/agent/consul/state/connect_ca.go src/github.com/hashicorp/consul/agent/consul/state/connect_ca_test.go src/github.com/hashicorp/consul/agent/consul/state/coordinate.go src/github.com/hashicorp/consul/agent/consul/state/coordinate_test.go src/github.com/hashicorp/consul/agent/consul/state/delay.go src/github.com/hashicorp/consul/agent/consul/state/delay_test.go src/github.com/hashicorp/consul/agent/consul/state/graveyard.go src/github.com/hashicorp/consul/agent/consul/state/graveyard_oss.go src/github.com/hashicorp/consul/agent/consul/state/graveyard_test.go src/github.com/hashicorp/consul/agent/consul/state/index_connect.go src/github.com/hashicorp/consul/agent/consul/state/index_connect_test.go src/github.com/hashicorp/consul/agent/consul/state/index_service_kind.go src/github.com/hashicorp/consul/agent/consul/state/intention.go src/github.com/hashicorp/consul/agent/consul/state/intention_test.go src/github.com/hashicorp/consul/agent/consul/state/kvs.go src/github.com/hashicorp/consul/agent/consul/state/kvs_oss.go src/github.com/hashicorp/consul/agent/consul/state/kvs_test.go src/github.com/hashicorp/consul/agent/consul/state/operations_oss.go src/github.com/hashicorp/consul/agent/consul/state/prepared_query.go src/github.com/hashicorp/consul/agent/consul/state/prepared_query_index.go src/github.com/hashicorp/consul/agent/consul/state/prepared_query_index_test.go src/github.com/hashicorp/consul/agent/consul/state/prepared_query_test.go src/github.com/hashicorp/consul/agent/consul/state/schema.go src/github.com/hashicorp/consul/agent/consul/state/schema_test.go src/github.com/hashicorp/consul/agent/consul/state/session.go src/github.com/hashicorp/consul/agent/consul/state/session_oss.go src/github.com/hashicorp/consul/agent/consul/state/session_test.go src/github.com/hashicorp/consul/agent/consul/state/state_store.go src/github.com/hashicorp/consul/agent/consul/state/state_store_oss_test.go src/github.com/hashicorp/consul/agent/consul/state/state_store_test.go src/github.com/hashicorp/consul/agent/consul/state/tombstone_gc.go src/github.com/hashicorp/consul/agent/consul/state/tombstone_gc_test.go src/github.com/hashicorp/consul/agent/consul/state/txn.go src/github.com/hashicorp/consul/agent/consul/state/txn_test.go src/github.com/hashicorp/consul/agent/debug/host.go src/github.com/hashicorp/consul/agent/debug/host_test.go src/github.com/hashicorp/consul/agent/exec/exec.go src/github.com/hashicorp/consul/agent/exec/exec_unix.go src/github.com/hashicorp/consul/agent/local/state.go src/github.com/hashicorp/consul/agent/local/testing.go src/github.com/hashicorp/consul/agent/local/state_test.go src/github.com/hashicorp/consul/agent/metadata/build.go src/github.com/hashicorp/consul/agent/metadata/build_test.go src/github.com/hashicorp/consul/agent/metadata/server.go src/github.com/hashicorp/consul/agent/metadata/server_internal_test.go src/github.com/hashicorp/consul/agent/metadata/server_test.go src/github.com/hashicorp/consul/agent/mock/notify.go src/github.com/hashicorp/consul/agent/pool/conn.go src/github.com/hashicorp/consul/agent/pool/peek.go src/github.com/hashicorp/consul/agent/pool/peeked_conn.go src/github.com/hashicorp/consul/agent/pool/pool.go src/github.com/hashicorp/consul/agent/proxycfg/manager.go src/github.com/hashicorp/consul/agent/proxycfg/manager_test.go src/github.com/hashicorp/consul/agent/proxycfg/proxycfg.go src/github.com/hashicorp/consul/agent/proxycfg/snapshot.go src/github.com/hashicorp/consul/agent/proxycfg/state.go src/github.com/hashicorp/consul/agent/proxycfg/state_test.go src/github.com/hashicorp/consul/agent/proxycfg/testing.go src/github.com/hashicorp/consul/agent/router/manager.go src/github.com/hashicorp/consul/agent/router/manager_internal_test.go src/github.com/hashicorp/consul/agent/router/router.go src/github.com/hashicorp/consul/agent/router/router_test.go src/github.com/hashicorp/consul/agent/router/serf_adapter.go src/github.com/hashicorp/consul/agent/router/serf_flooder.go src/github.com/hashicorp/consul/agent/router/manager_test.go src/github.com/hashicorp/consul/agent/structs/acl.go src/github.com/hashicorp/consul/agent/structs/acl_cache.go src/github.com/hashicorp/consul/agent/structs/acl_cache_test.go src/github.com/hashicorp/consul/agent/structs/acl_legacy.go src/github.com/hashicorp/consul/agent/structs/acl_legacy_test.go src/github.com/hashicorp/consul/agent/structs/acl_oss.go src/github.com/hashicorp/consul/agent/structs/acl_test.go src/github.com/hashicorp/consul/agent/structs/auto_encrypt.go src/github.com/hashicorp/consul/agent/structs/catalog.go src/github.com/hashicorp/consul/agent/structs/check_definition.go src/github.com/hashicorp/consul/agent/structs/check_definition_test.go src/github.com/hashicorp/consul/agent/structs/check_type.go src/github.com/hashicorp/consul/agent/structs/config_entry.go src/github.com/hashicorp/consul/agent/structs/config_entry_discoverychain.go src/github.com/hashicorp/consul/agent/structs/config_entry_discoverychain_oss.go src/github.com/hashicorp/consul/agent/structs/config_entry_discoverychain_test.go src/github.com/hashicorp/consul/agent/structs/config_entry_oss.go src/github.com/hashicorp/consul/agent/structs/config_entry_test.go src/github.com/hashicorp/consul/agent/structs/connect.go src/github.com/hashicorp/consul/agent/structs/connect_ca.go src/github.com/hashicorp/consul/agent/structs/connect_ca_test.go src/github.com/hashicorp/consul/agent/structs/connect_oss.go src/github.com/hashicorp/consul/agent/structs/connect_proxy_config.go src/github.com/hashicorp/consul/agent/structs/connect_proxy_config_oss.go src/github.com/hashicorp/consul/agent/structs/connect_proxy_config_test.go src/github.com/hashicorp/consul/agent/structs/discovery_chain.go src/github.com/hashicorp/consul/agent/structs/discovery_chain_oss.go src/github.com/hashicorp/consul/agent/structs/errors.go src/github.com/hashicorp/consul/agent/structs/intention.go src/github.com/hashicorp/consul/agent/structs/intention_oss.go src/github.com/hashicorp/consul/agent/structs/intention_test.go src/github.com/hashicorp/consul/agent/structs/operator.go src/github.com/hashicorp/consul/agent/structs/prepared_query.go src/github.com/hashicorp/consul/agent/structs/prepared_query_test.go src/github.com/hashicorp/consul/agent/structs/protobuf_compat.go src/github.com/hashicorp/consul/agent/structs/sanitize_oss.go src/github.com/hashicorp/consul/agent/structs/service_definition.go src/github.com/hashicorp/consul/agent/structs/service_definition_test.go src/github.com/hashicorp/consul/agent/structs/snapshot.go src/github.com/hashicorp/consul/agent/structs/structs.go src/github.com/hashicorp/consul/agent/structs/structs_filtering_test.go src/github.com/hashicorp/consul/agent/structs/structs_oss.go src/github.com/hashicorp/consul/agent/structs/structs_test.go src/github.com/hashicorp/consul/agent/structs/testing.go src/github.com/hashicorp/consul/agent/structs/testing_catalog.go src/github.com/hashicorp/consul/agent/structs/testing_connect_proxy_config.go src/github.com/hashicorp/consul/agent/structs/testing_intention.go src/github.com/hashicorp/consul/agent/structs/testing_service_definition.go src/github.com/hashicorp/consul/agent/structs/txn.go src/github.com/hashicorp/consul/agent/systemd/notify.go src/github.com/hashicorp/consul/agent/token/store.go src/github.com/hashicorp/consul/agent/token/store_oss.go src/github.com/hashicorp/consul/agent/token/store_test.go src/github.com/hashicorp/consul/agent/xds/clusters.go src/github.com/hashicorp/consul/agent/xds/clusters_test.go src/github.com/hashicorp/consul/agent/xds/config.go src/github.com/hashicorp/consul/agent/xds/config_test.go src/github.com/hashicorp/consul/agent/xds/endpoints.go src/github.com/hashicorp/consul/agent/xds/endpoints_test.go src/github.com/hashicorp/consul/agent/xds/failover_math.go src/github.com/hashicorp/consul/agent/xds/failover_math_test.go src/github.com/hashicorp/consul/agent/xds/golden_test.go src/github.com/hashicorp/consul/agent/xds/listeners.go src/github.com/hashicorp/consul/agent/xds/listeners_test.go src/github.com/hashicorp/consul/agent/xds/naming.go src/github.com/hashicorp/consul/agent/xds/response.go src/github.com/hashicorp/consul/agent/xds/routes.go src/github.com/hashicorp/consul/agent/xds/routes_test.go src/github.com/hashicorp/consul/agent/xds/server.go src/github.com/hashicorp/consul/agent/xds/server_oss.go src/github.com/hashicorp/consul/agent/xds/server_test.go src/github.com/hashicorp/consul/agent/xds/testing.go src/github.com/hashicorp/consul/agent/xds/xds.go src/github.com/hashicorp/consul/api/acl.go src/github.com/hashicorp/consul/api/acl_test.go src/github.com/hashicorp/consul/api/agent.go src/github.com/hashicorp/consul/api/agent_test.go src/github.com/hashicorp/consul/api/api.go src/github.com/hashicorp/consul/api/api_test.go src/github.com/hashicorp/consul/api/catalog.go src/github.com/hashicorp/consul/api/catalog_test.go src/github.com/hashicorp/consul/api/config_entry.go src/github.com/hashicorp/consul/api/config_entry_discoverychain.go src/github.com/hashicorp/consul/api/config_entry_discoverychain_test.go src/github.com/hashicorp/consul/api/config_entry_test.go src/github.com/hashicorp/consul/api/connect.go src/github.com/hashicorp/consul/api/connect_ca.go src/github.com/hashicorp/consul/api/connect_ca_test.go src/github.com/hashicorp/consul/api/connect_intention.go src/github.com/hashicorp/consul/api/connect_intention_test.go src/github.com/hashicorp/consul/api/coordinate.go src/github.com/hashicorp/consul/api/coordinate_test.go src/github.com/hashicorp/consul/api/debug.go src/github.com/hashicorp/consul/api/debug_test.go src/github.com/hashicorp/consul/api/discovery_chain.go src/github.com/hashicorp/consul/api/discovery_chain_test.go src/github.com/hashicorp/consul/api/event.go src/github.com/hashicorp/consul/api/event_test.go src/github.com/hashicorp/consul/api/health.go src/github.com/hashicorp/consul/api/health_test.go src/github.com/hashicorp/consul/api/kv.go src/github.com/hashicorp/consul/api/kv_test.go src/github.com/hashicorp/consul/api/lock.go src/github.com/hashicorp/consul/api/lock_test.go src/github.com/hashicorp/consul/api/namespace.go src/github.com/hashicorp/consul/api/operator.go src/github.com/hashicorp/consul/api/operator_area.go src/github.com/hashicorp/consul/api/operator_autopilot.go src/github.com/hashicorp/consul/api/operator_autopilot_test.go src/github.com/hashicorp/consul/api/operator_keyring.go src/github.com/hashicorp/consul/api/operator_keyring_test.go src/github.com/hashicorp/consul/api/operator_license.go src/github.com/hashicorp/consul/api/operator_raft.go src/github.com/hashicorp/consul/api/operator_raft_test.go src/github.com/hashicorp/consul/api/operator_segment.go src/github.com/hashicorp/consul/api/oss_test.go src/github.com/hashicorp/consul/api/prepared_query.go src/github.com/hashicorp/consul/api/prepared_query_test.go src/github.com/hashicorp/consul/api/raw.go src/github.com/hashicorp/consul/api/semaphore.go src/github.com/hashicorp/consul/api/semaphore_test.go src/github.com/hashicorp/consul/api/session.go src/github.com/hashicorp/consul/api/session_test.go src/github.com/hashicorp/consul/api/snapshot.go src/github.com/hashicorp/consul/api/snapshot_test.go src/github.com/hashicorp/consul/api/status.go src/github.com/hashicorp/consul/api/status_test.go src/github.com/hashicorp/consul/api/txn.go src/github.com/hashicorp/consul/api/txn_test.go src/github.com/hashicorp/consul/api/watch/funcs.go src/github.com/hashicorp/consul/api/watch/plan.go src/github.com/hashicorp/consul/api/watch/plan_test.go src/github.com/hashicorp/consul/api/watch/watch.go src/github.com/hashicorp/consul/api/watch/watch_test.go src/github.com/hashicorp/consul/api/watch/funcs_test.go src/github.com/hashicorp/consul/command/commands_oss.go src/github.com/hashicorp/consul/command/registry.go src/github.com/hashicorp/consul/command/acl/acl.go src/github.com/hashicorp/consul/command/acl/acl_helpers.go src/github.com/hashicorp/consul/command/acl/agenttokens/agent_tokens.go src/github.com/hashicorp/consul/command/acl/agenttokens/agent_tokens_test.go src/github.com/hashicorp/consul/command/acl/authmethod/authmethod.go src/github.com/hashicorp/consul/command/acl/authmethod/formatter.go src/github.com/hashicorp/consul/command/acl/authmethod/create/authmethod_create.go src/github.com/hashicorp/consul/command/acl/authmethod/create/authmethod_create_test.go src/github.com/hashicorp/consul/command/acl/authmethod/delete/authmethod_delete.go src/github.com/hashicorp/consul/command/acl/authmethod/delete/authmethod_delete_test.go src/github.com/hashicorp/consul/command/acl/authmethod/list/authmethod_list.go src/github.com/hashicorp/consul/command/acl/authmethod/list/authmethod_list_test.go src/github.com/hashicorp/consul/command/acl/authmethod/read/authmethod_read.go src/github.com/hashicorp/consul/command/acl/authmethod/read/authmethod_read_test.go src/github.com/hashicorp/consul/command/acl/authmethod/update/authmethod_update.go src/github.com/hashicorp/consul/command/acl/authmethod/update/authmethod_update_test.go src/github.com/hashicorp/consul/command/acl/bindingrule/bindingrule.go src/github.com/hashicorp/consul/command/acl/bindingrule/formatter.go src/github.com/hashicorp/consul/command/acl/bindingrule/create/bindingrule_create.go src/github.com/hashicorp/consul/command/acl/bindingrule/create/bindingrule_create_test.go src/github.com/hashicorp/consul/command/acl/bindingrule/delete/bindingrule_delete.go src/github.com/hashicorp/consul/command/acl/bindingrule/delete/bindingrule_delete_test.go src/github.com/hashicorp/consul/command/acl/bindingrule/list/bindingrule_list.go src/github.com/hashicorp/consul/command/acl/bindingrule/list/bindingrule_list_test.go src/github.com/hashicorp/consul/command/acl/bindingrule/read/bindingrule_read.go src/github.com/hashicorp/consul/command/acl/bindingrule/read/bindingrule_read_test.go src/github.com/hashicorp/consul/command/acl/bindingrule/update/bindingrule_update.go src/github.com/hashicorp/consul/command/acl/bindingrule/update/bindingrule_update_test.go src/github.com/hashicorp/consul/command/acl/bootstrap/bootstrap.go src/github.com/hashicorp/consul/command/acl/bootstrap/bootstrap_test.go src/github.com/hashicorp/consul/command/acl/policy/formatter.go src/github.com/hashicorp/consul/command/acl/policy/policy.go src/github.com/hashicorp/consul/command/acl/policy/create/policy_create.go src/github.com/hashicorp/consul/command/acl/policy/create/policy_create_test.go src/github.com/hashicorp/consul/command/acl/policy/delete/policy_delete.go src/github.com/hashicorp/consul/command/acl/policy/delete/policy_delete_test.go src/github.com/hashicorp/consul/command/acl/policy/list/policy_list.go src/github.com/hashicorp/consul/command/acl/policy/list/policy_list_test.go src/github.com/hashicorp/consul/command/acl/policy/read/policy_read.go src/github.com/hashicorp/consul/command/acl/policy/read/policy_read_test.go src/github.com/hashicorp/consul/command/acl/policy/update/policy_update.go src/github.com/hashicorp/consul/command/acl/policy/update/policy_update_test.go src/github.com/hashicorp/consul/command/acl/role/formatter.go src/github.com/hashicorp/consul/command/acl/role/role.go src/github.com/hashicorp/consul/command/acl/role/create/role_create.go src/github.com/hashicorp/consul/command/acl/role/create/role_create_test.go src/github.com/hashicorp/consul/command/acl/role/delete/role_delete.go src/github.com/hashicorp/consul/command/acl/role/delete/role_delete_test.go src/github.com/hashicorp/consul/command/acl/role/list/role_list.go src/github.com/hashicorp/consul/command/acl/role/list/role_list_test.go src/github.com/hashicorp/consul/command/acl/role/read/role_read.go src/github.com/hashicorp/consul/command/acl/role/read/role_read_test.go src/github.com/hashicorp/consul/command/acl/role/update/role_update.go src/github.com/hashicorp/consul/command/acl/role/update/role_update_test.go src/github.com/hashicorp/consul/command/acl/rules/translate.go src/github.com/hashicorp/consul/command/acl/rules/translate_test.go src/github.com/hashicorp/consul/command/acl/token/formatter.go src/github.com/hashicorp/consul/command/acl/token/token.go src/github.com/hashicorp/consul/command/acl/token/clone/token_clone.go src/github.com/hashicorp/consul/command/acl/token/clone/token_clone_test.go src/github.com/hashicorp/consul/command/acl/token/create/token_create.go src/github.com/hashicorp/consul/command/acl/token/create/token_create_test.go src/github.com/hashicorp/consul/command/acl/token/delete/token_delete.go src/github.com/hashicorp/consul/command/acl/token/delete/token_delete_test.go src/github.com/hashicorp/consul/command/acl/token/list/token_list.go src/github.com/hashicorp/consul/command/acl/token/list/token_list_test.go src/github.com/hashicorp/consul/command/acl/token/read/token_read.go src/github.com/hashicorp/consul/command/acl/token/read/token_read_test.go src/github.com/hashicorp/consul/command/acl/token/update/token_update.go src/github.com/hashicorp/consul/command/acl/token/update/token_update_test.go src/github.com/hashicorp/consul/command/agent/agent.go src/github.com/hashicorp/consul/command/agent/agent_test.go src/github.com/hashicorp/consul/command/catalog/catalog.go src/github.com/hashicorp/consul/command/catalog/catalog_test.go src/github.com/hashicorp/consul/command/catalog/list/dc/catalog_list_datacenters.go src/github.com/hashicorp/consul/command/catalog/list/dc/catalog_list_datacenters_test.go src/github.com/hashicorp/consul/command/catalog/list/nodes/catalog_list_nodes.go src/github.com/hashicorp/consul/command/catalog/list/nodes/catalog_list_nodes_test.go src/github.com/hashicorp/consul/command/catalog/list/services/catalog_list_services.go src/github.com/hashicorp/consul/command/catalog/list/services/catalog_list_services_test.go src/github.com/hashicorp/consul/command/config/config.go src/github.com/hashicorp/consul/command/config/delete/config_delete.go src/github.com/hashicorp/consul/command/config/delete/config_delete_test.go src/github.com/hashicorp/consul/command/config/list/config_list.go src/github.com/hashicorp/consul/command/config/list/config_list_test.go src/github.com/hashicorp/consul/command/config/read/config_read.go src/github.com/hashicorp/consul/command/config/read/config_read_test.go src/github.com/hashicorp/consul/command/config/write/config_write.go src/github.com/hashicorp/consul/command/config/write/config_write_test.go src/github.com/hashicorp/consul/command/config/write/decode_shim.go src/github.com/hashicorp/consul/command/connect/connect.go src/github.com/hashicorp/consul/command/connect/connect_test.go src/github.com/hashicorp/consul/command/connect/ca/ca.go src/github.com/hashicorp/consul/command/connect/ca/ca_test.go src/github.com/hashicorp/consul/command/connect/ca/get/connect_ca_get.go src/github.com/hashicorp/consul/command/connect/ca/get/connect_ca_get_test.go src/github.com/hashicorp/consul/command/connect/ca/set/connect_ca_set.go src/github.com/hashicorp/consul/command/connect/ca/set/connect_ca_set_test.go src/github.com/hashicorp/consul/command/connect/envoy/bootstrap_config.go src/github.com/hashicorp/consul/command/connect/envoy/bootstrap_config_test.go src/github.com/hashicorp/consul/command/connect/envoy/bootstrap_tpl.go src/github.com/hashicorp/consul/command/connect/envoy/envoy.go src/github.com/hashicorp/consul/command/connect/envoy/envoy_test.go src/github.com/hashicorp/consul/command/connect/envoy/exec_test.go src/github.com/hashicorp/consul/command/connect/envoy/exec_unix.go src/github.com/hashicorp/consul/command/connect/envoy/pipe-bootstrap/connect_envoy_pipe-bootstrap.go src/github.com/hashicorp/consul/command/connect/envoy/pipe-bootstrap/connect_envoy_pipe-bootstrap_test.go src/github.com/hashicorp/consul/command/connect/proxy/flag_upstreams.go src/github.com/hashicorp/consul/command/connect/proxy/flag_upstreams_test.go src/github.com/hashicorp/consul/command/connect/proxy/proxy.go src/github.com/hashicorp/consul/command/connect/proxy/proxy_test.go src/github.com/hashicorp/consul/command/connect/proxy/register.go src/github.com/hashicorp/consul/command/connect/proxy/register_test.go src/github.com/hashicorp/consul/command/debug/debug.go src/github.com/hashicorp/consul/command/debug/debug_test.go src/github.com/hashicorp/consul/command/event/event.go src/github.com/hashicorp/consul/command/event/event_test.go src/github.com/hashicorp/consul/command/exec/exec.go src/github.com/hashicorp/consul/command/exec/exec_test.go src/github.com/hashicorp/consul/command/flags/config.go src/github.com/hashicorp/consul/command/flags/config_test.go src/github.com/hashicorp/consul/command/flags/flag_map_value.go src/github.com/hashicorp/consul/command/flags/flag_map_value_test.go src/github.com/hashicorp/consul/command/flags/flag_slice_value.go src/github.com/hashicorp/consul/command/flags/flag_slice_value_test.go src/github.com/hashicorp/consul/command/flags/http.go src/github.com/hashicorp/consul/command/flags/http_test.go src/github.com/hashicorp/consul/command/flags/merge.go src/github.com/hashicorp/consul/command/flags/usage.go src/github.com/hashicorp/consul/command/forceleave/forceleave.go src/github.com/hashicorp/consul/command/forceleave/forceleave_test.go src/github.com/hashicorp/consul/command/helpers/helpers.go src/github.com/hashicorp/consul/command/info/info.go src/github.com/hashicorp/consul/command/info/info_test.go src/github.com/hashicorp/consul/command/intention/intention.go src/github.com/hashicorp/consul/command/intention/intention_test.go src/github.com/hashicorp/consul/command/intention/check/check.go src/github.com/hashicorp/consul/command/intention/check/check_test.go src/github.com/hashicorp/consul/command/intention/create/create.go src/github.com/hashicorp/consul/command/intention/create/create_test.go src/github.com/hashicorp/consul/command/intention/delete/delete.go src/github.com/hashicorp/consul/command/intention/delete/delete_test.go src/github.com/hashicorp/consul/command/intention/finder/finder.go src/github.com/hashicorp/consul/command/intention/finder/finder_test.go src/github.com/hashicorp/consul/command/intention/get/get.go src/github.com/hashicorp/consul/command/intention/get/get_test.go src/github.com/hashicorp/consul/command/intention/match/match.go src/github.com/hashicorp/consul/command/intention/match/match_test.go src/github.com/hashicorp/consul/command/join/join.go src/github.com/hashicorp/consul/command/join/join_test.go src/github.com/hashicorp/consul/command/keygen/keygen.go src/github.com/hashicorp/consul/command/keygen/keygen_test.go src/github.com/hashicorp/consul/command/keyring/keyring.go src/github.com/hashicorp/consul/command/keyring/keyring_test.go src/github.com/hashicorp/consul/command/kv/kv.go src/github.com/hashicorp/consul/command/kv/kv_test.go src/github.com/hashicorp/consul/command/kv/del/kv_delete.go src/github.com/hashicorp/consul/command/kv/del/kv_delete_test.go src/github.com/hashicorp/consul/command/kv/exp/kv_export.go src/github.com/hashicorp/consul/command/kv/exp/kv_export_test.go src/github.com/hashicorp/consul/command/kv/get/kv_get.go src/github.com/hashicorp/consul/command/kv/get/kv_get_test.go src/github.com/hashicorp/consul/command/kv/imp/kv_import.go src/github.com/hashicorp/consul/command/kv/imp/kv_import_test.go src/github.com/hashicorp/consul/command/kv/impexp/kvimpexp.go src/github.com/hashicorp/consul/command/kv/put/kv_put.go src/github.com/hashicorp/consul/command/kv/put/kv_put_test.go src/github.com/hashicorp/consul/command/leave/leave.go src/github.com/hashicorp/consul/command/leave/leave_test.go src/github.com/hashicorp/consul/command/lock/lock.go src/github.com/hashicorp/consul/command/lock/lock_test.go src/github.com/hashicorp/consul/command/lock/util_unix.go src/github.com/hashicorp/consul/command/login/login.go src/github.com/hashicorp/consul/command/login/login_test.go src/github.com/hashicorp/consul/command/logout/logout.go src/github.com/hashicorp/consul/command/logout/logout_test.go src/github.com/hashicorp/consul/command/maint/maint.go src/github.com/hashicorp/consul/command/maint/maint_test.go src/github.com/hashicorp/consul/command/members/members.go src/github.com/hashicorp/consul/command/members/members_test.go src/github.com/hashicorp/consul/command/monitor/monitor.go src/github.com/hashicorp/consul/command/monitor/monitor_test.go src/github.com/hashicorp/consul/command/operator/operator.go src/github.com/hashicorp/consul/command/operator/operator_test.go src/github.com/hashicorp/consul/command/operator/autopilot/operator_autopilot.go src/github.com/hashicorp/consul/command/operator/autopilot/operator_autopilot_test.go src/github.com/hashicorp/consul/command/operator/autopilot/get/operator_autopilot_get.go src/github.com/hashicorp/consul/command/operator/autopilot/get/operator_autopilot_get_test.go src/github.com/hashicorp/consul/command/operator/autopilot/set/operator_autopilot_set.go src/github.com/hashicorp/consul/command/operator/autopilot/set/operator_autopilot_set_test.go src/github.com/hashicorp/consul/command/operator/raft/operator_raft.go src/github.com/hashicorp/consul/command/operator/raft/operator_raft_test.go src/github.com/hashicorp/consul/command/operator/raft/listpeers/operator_raft_list.go src/github.com/hashicorp/consul/command/operator/raft/listpeers/operator_raft_list_test.go src/github.com/hashicorp/consul/command/operator/raft/removepeer/operator_raft_remove.go src/github.com/hashicorp/consul/command/operator/raft/removepeer/operator_raft_remove_test.go src/github.com/hashicorp/consul/command/reload/reload.go src/github.com/hashicorp/consul/command/reload/reload_test.go src/github.com/hashicorp/consul/command/rtt/rtt.go src/github.com/hashicorp/consul/command/rtt/rtt_test.go src/github.com/hashicorp/consul/command/services/config.go src/github.com/hashicorp/consul/command/services/config_test.go src/github.com/hashicorp/consul/command/services/services.go src/github.com/hashicorp/consul/command/services/services_test.go src/github.com/hashicorp/consul/command/services/deregister/deregister.go src/github.com/hashicorp/consul/command/services/deregister/deregister_test.go src/github.com/hashicorp/consul/command/services/register/register.go src/github.com/hashicorp/consul/command/services/register/register_test.go src/github.com/hashicorp/consul/command/snapshot/snapshot_command.go src/github.com/hashicorp/consul/command/snapshot/snapshot_command_test.go src/github.com/hashicorp/consul/command/snapshot/inspect/snapshot_inspect.go src/github.com/hashicorp/consul/command/snapshot/inspect/snapshot_inspect_test.go src/github.com/hashicorp/consul/command/snapshot/restore/snapshot_restore.go src/github.com/hashicorp/consul/command/snapshot/restore/snapshot_restore_test.go src/github.com/hashicorp/consul/command/snapshot/save/snapshot_save.go src/github.com/hashicorp/consul/command/snapshot/save/snapshot_save_test.go src/github.com/hashicorp/consul/command/tls/tls.go src/github.com/hashicorp/consul/command/tls/tls_test.go src/github.com/hashicorp/consul/command/tls/ca/tls_ca.go src/github.com/hashicorp/consul/command/tls/ca/tls_ca_test.go src/github.com/hashicorp/consul/command/tls/ca/create/tls_ca_create.go src/github.com/hashicorp/consul/command/tls/ca/create/tls_ca_create_test.go src/github.com/hashicorp/consul/command/tls/cert/tls_cert.go src/github.com/hashicorp/consul/command/tls/cert/tls_cert_test.go src/github.com/hashicorp/consul/command/tls/cert/create/tls_cert_create.go src/github.com/hashicorp/consul/command/tls/cert/create/tls_cert_create_test.go src/github.com/hashicorp/consul/command/validate/validate.go src/github.com/hashicorp/consul/command/validate/validate_test.go src/github.com/hashicorp/consul/command/version/version.go src/github.com/hashicorp/consul/command/version/version_test.go src/github.com/hashicorp/consul/command/watch/watch.go src/github.com/hashicorp/consul/command/watch/watch_test.go src/github.com/hashicorp/consul/connect/example_test.go src/github.com/hashicorp/consul/connect/resolver.go src/github.com/hashicorp/consul/connect/resolver_test.go src/github.com/hashicorp/consul/connect/service.go src/github.com/hashicorp/consul/connect/service_test.go src/github.com/hashicorp/consul/connect/testing.go src/github.com/hashicorp/consul/connect/tls.go src/github.com/hashicorp/consul/connect/tls_test.go src/github.com/hashicorp/consul/connect/certgen/certgen.go src/github.com/hashicorp/consul/connect/proxy/config.go src/github.com/hashicorp/consul/connect/proxy/config_test.go src/github.com/hashicorp/consul/connect/proxy/conn.go src/github.com/hashicorp/consul/connect/proxy/conn_test.go src/github.com/hashicorp/consul/connect/proxy/listener.go src/github.com/hashicorp/consul/connect/proxy/listener_test.go src/github.com/hashicorp/consul/connect/proxy/proxy.go src/github.com/hashicorp/consul/connect/proxy/proxy_test.go src/github.com/hashicorp/consul/connect/proxy/testing.go src/github.com/hashicorp/consul/ipaddr/detect.go src/github.com/hashicorp/consul/ipaddr/detect_test.go src/github.com/hashicorp/consul/ipaddr/ipaddr.go src/github.com/hashicorp/consul/ipaddr/ipaddr_test.go src/github.com/hashicorp/consul/lib/cluster.go src/github.com/hashicorp/consul/lib/cluster_test.go src/github.com/hashicorp/consul/lib/eof.go src/github.com/hashicorp/consul/lib/json.go src/github.com/hashicorp/consul/lib/map_walker.go src/github.com/hashicorp/consul/lib/map_walker_test.go src/github.com/hashicorp/consul/lib/math.go src/github.com/hashicorp/consul/lib/patch_hcl.go src/github.com/hashicorp/consul/lib/patch_hcl_test.go src/github.com/hashicorp/consul/lib/path.go src/github.com/hashicorp/consul/lib/rand.go src/github.com/hashicorp/consul/lib/retry.go src/github.com/hashicorp/consul/lib/retry_test.go src/github.com/hashicorp/consul/lib/rtt.go src/github.com/hashicorp/consul/lib/rtt_test.go src/github.com/hashicorp/consul/lib/serf.go src/github.com/hashicorp/consul/lib/stop_context.go src/github.com/hashicorp/consul/lib/string.go src/github.com/hashicorp/consul/lib/string_test.go src/github.com/hashicorp/consul/lib/telemetry.go src/github.com/hashicorp/consul/lib/telemetry_test.go src/github.com/hashicorp/consul/lib/testing_httpserver.go src/github.com/hashicorp/consul/lib/translate.go src/github.com/hashicorp/consul/lib/translate_test.go src/github.com/hashicorp/consul/lib/useragent.go src/github.com/hashicorp/consul/lib/useragent_test.go src/github.com/hashicorp/consul/lib/uuid.go src/github.com/hashicorp/consul/lib/math_test.go src/github.com/hashicorp/consul/lib/file/atomic.go src/github.com/hashicorp/consul/lib/file/atomic_test.go src/github.com/hashicorp/consul/lib/semaphore/semaphore.go src/github.com/hashicorp/consul/lib/semaphore/semaphore_test.go src/github.com/hashicorp/consul/logging/gated_writer.go src/github.com/hashicorp/consul/logging/gated_writer_test.go src/github.com/hashicorp/consul/logging/grpc.go src/github.com/hashicorp/consul/logging/grpc_test.go src/github.com/hashicorp/consul/logging/log_levels.go src/github.com/hashicorp/consul/logging/logfile.go src/github.com/hashicorp/consul/logging/logfile_test.go src/github.com/hashicorp/consul/logging/logger.go src/github.com/hashicorp/consul/logging/logger_test.go src/github.com/hashicorp/consul/logging/names.go src/github.com/hashicorp/consul/logging/syslog.go src/github.com/hashicorp/consul/logging/monitor/monitor.go src/github.com/hashicorp/consul/logging/monitor/monitor_test.go src/github.com/hashicorp/consul/sdk/freeport/ephemeral_linux.go src/github.com/hashicorp/consul/sdk/freeport/ephemeral_linux_test.go src/github.com/hashicorp/consul/sdk/freeport/freeport.go src/github.com/hashicorp/consul/sdk/freeport/freeport_test.go src/github.com/hashicorp/consul/sdk/freeport/systemlimit.go src/github.com/hashicorp/consul/sdk/testutil/io.go src/github.com/hashicorp/consul/sdk/testutil/server.go src/github.com/hashicorp/consul/sdk/testutil/server_methods.go src/github.com/hashicorp/consul/sdk/testutil/server_wrapper.go src/github.com/hashicorp/consul/sdk/testutil/testlog.go src/github.com/hashicorp/consul/sdk/testutil/retry/retry.go src/github.com/hashicorp/consul/sdk/testutil/retry/retry_test.go src/github.com/hashicorp/consul/sentinel/evaluator.go src/github.com/hashicorp/consul/sentinel/scope.go src/github.com/hashicorp/consul/sentinel/sentinel_oss.go src/github.com/hashicorp/consul/service_os/service.go src/github.com/hashicorp/consul/snapshot/archive.go src/github.com/hashicorp/consul/snapshot/archive_test.go src/github.com/hashicorp/consul/snapshot/snapshot.go src/github.com/hashicorp/consul/snapshot/snapshot_test.go src/github.com/hashicorp/consul/testrpc/wait.go src/github.com/hashicorp/consul/tlsutil/config.go src/github.com/hashicorp/consul/tlsutil/config_test.go src/github.com/hashicorp/consul/tlsutil/generate.go src/github.com/hashicorp/consul/tlsutil/generate_test.go src/github.com/hashicorp/consul/types/area.go src/github.com/hashicorp/consul/types/checks.go src/github.com/hashicorp/consul/types/node_id.go src/github.com/hashicorp/consul/version/version.go cd _build && go install -trimpath -v -p 4 github.com/hashicorp/consul github.com/hashicorp/consul/acl github.com/hashicorp/consul/agent github.com/hashicorp/consul/agent/ae github.com/hashicorp/consul/agent/agentpb github.com/hashicorp/consul/agent/cache github.com/hashicorp/consul/agent/cache-types github.com/hashicorp/consul/agent/checks github.com/hashicorp/consul/agent/config github.com/hashicorp/consul/agent/connect github.com/hashicorp/consul/agent/connect/ca github.com/hashicorp/consul/agent/consul github.com/hashicorp/consul/agent/consul/authmethod github.com/hashicorp/consul/agent/consul/authmethod/kubeauth github.com/hashicorp/consul/agent/consul/authmethod/testauth github.com/hashicorp/consul/agent/consul/autopilot github.com/hashicorp/consul/agent/consul/discoverychain github.com/hashicorp/consul/agent/consul/fsm github.com/hashicorp/consul/agent/consul/prepared_query github.com/hashicorp/consul/agent/consul/state github.com/hashicorp/consul/agent/debug github.com/hashicorp/consul/agent/exec github.com/hashicorp/consul/agent/local github.com/hashicorp/consul/agent/metadata github.com/hashicorp/consul/agent/mock github.com/hashicorp/consul/agent/pool github.com/hashicorp/consul/agent/proxycfg github.com/hashicorp/consul/agent/router github.com/hashicorp/consul/agent/structs github.com/hashicorp/consul/agent/systemd github.com/hashicorp/consul/agent/token github.com/hashicorp/consul/agent/xds github.com/hashicorp/consul/api github.com/hashicorp/consul/api/watch github.com/hashicorp/consul/command github.com/hashicorp/consul/command/acl github.com/hashicorp/consul/command/acl/agenttokens github.com/hashicorp/consul/command/acl/authmethod github.com/hashicorp/consul/command/acl/authmethod/create github.com/hashicorp/consul/command/acl/authmethod/delete github.com/hashicorp/consul/command/acl/authmethod/list github.com/hashicorp/consul/command/acl/authmethod/read github.com/hashicorp/consul/command/acl/authmethod/update github.com/hashicorp/consul/command/acl/bindingrule github.com/hashicorp/consul/command/acl/bindingrule/create github.com/hashicorp/consul/command/acl/bindingrule/delete github.com/hashicorp/consul/command/acl/bindingrule/list github.com/hashicorp/consul/command/acl/bindingrule/read github.com/hashicorp/consul/command/acl/bindingrule/update github.com/hashicorp/consul/command/acl/bootstrap github.com/hashicorp/consul/command/acl/policy github.com/hashicorp/consul/command/acl/policy/create github.com/hashicorp/consul/command/acl/policy/delete github.com/hashicorp/consul/command/acl/policy/list github.com/hashicorp/consul/command/acl/policy/read github.com/hashicorp/consul/command/acl/policy/update github.com/hashicorp/consul/command/acl/role github.com/hashicorp/consul/command/acl/role/create github.com/hashicorp/consul/command/acl/role/delete github.com/hashicorp/consul/command/acl/role/list github.com/hashicorp/consul/command/acl/role/read github.com/hashicorp/consul/command/acl/role/update github.com/hashicorp/consul/command/acl/rules github.com/hashicorp/consul/command/acl/token github.com/hashicorp/consul/command/acl/token/clone github.com/hashicorp/consul/command/acl/token/create github.com/hashicorp/consul/command/acl/token/delete github.com/hashicorp/consul/command/acl/token/list github.com/hashicorp/consul/command/acl/token/read github.com/hashicorp/consul/command/acl/token/update github.com/hashicorp/consul/command/agent github.com/hashicorp/consul/command/catalog github.com/hashicorp/consul/command/catalog/list/dc github.com/hashicorp/consul/command/catalog/list/nodes github.com/hashicorp/consul/command/catalog/list/services github.com/hashicorp/consul/command/config github.com/hashicorp/consul/command/config/delete github.com/hashicorp/consul/command/config/list github.com/hashicorp/consul/command/config/read github.com/hashicorp/consul/command/config/write github.com/hashicorp/consul/command/connect github.com/hashicorp/consul/command/connect/ca github.com/hashicorp/consul/command/connect/ca/get github.com/hashicorp/consul/command/connect/ca/set github.com/hashicorp/consul/command/connect/envoy github.com/hashicorp/consul/command/connect/envoy/pipe-bootstrap github.com/hashicorp/consul/command/connect/proxy github.com/hashicorp/consul/command/debug github.com/hashicorp/consul/command/event github.com/hashicorp/consul/command/exec github.com/hashicorp/consul/command/flags github.com/hashicorp/consul/command/forceleave github.com/hashicorp/consul/command/helpers github.com/hashicorp/consul/command/info github.com/hashicorp/consul/command/intention github.com/hashicorp/consul/command/intention/check github.com/hashicorp/consul/command/intention/create github.com/hashicorp/consul/command/intention/delete github.com/hashicorp/consul/command/intention/finder github.com/hashicorp/consul/command/intention/get github.com/hashicorp/consul/command/intention/match github.com/hashicorp/consul/command/join github.com/hashicorp/consul/command/keygen github.com/hashicorp/consul/command/keyring github.com/hashicorp/consul/command/kv github.com/hashicorp/consul/command/kv/del github.com/hashicorp/consul/command/kv/exp github.com/hashicorp/consul/command/kv/get github.com/hashicorp/consul/command/kv/imp github.com/hashicorp/consul/command/kv/impexp github.com/hashicorp/consul/command/kv/put github.com/hashicorp/consul/command/leave github.com/hashicorp/consul/command/lock github.com/hashicorp/consul/command/login github.com/hashicorp/consul/command/logout github.com/hashicorp/consul/command/maint github.com/hashicorp/consul/command/members github.com/hashicorp/consul/command/monitor github.com/hashicorp/consul/command/operator github.com/hashicorp/consul/command/operator/autopilot github.com/hashicorp/consul/command/operator/autopilot/get github.com/hashicorp/consul/command/operator/autopilot/set github.com/hashicorp/consul/command/operator/raft github.com/hashicorp/consul/command/operator/raft/listpeers github.com/hashicorp/consul/command/operator/raft/removepeer github.com/hashicorp/consul/command/reload github.com/hashicorp/consul/command/rtt github.com/hashicorp/consul/command/services github.com/hashicorp/consul/command/services/deregister github.com/hashicorp/consul/command/services/register github.com/hashicorp/consul/command/snapshot github.com/hashicorp/consul/command/snapshot/inspect github.com/hashicorp/consul/command/snapshot/restore github.com/hashicorp/consul/command/snapshot/save github.com/hashicorp/consul/command/tls github.com/hashicorp/consul/command/tls/ca github.com/hashicorp/consul/command/tls/ca/create github.com/hashicorp/consul/command/tls/cert github.com/hashicorp/consul/command/tls/cert/create github.com/hashicorp/consul/command/validate github.com/hashicorp/consul/command/version github.com/hashicorp/consul/command/watch github.com/hashicorp/consul/connect github.com/hashicorp/consul/connect/certgen github.com/hashicorp/consul/connect/proxy github.com/hashicorp/consul/ipaddr github.com/hashicorp/consul/lib github.com/hashicorp/consul/lib/file github.com/hashicorp/consul/lib/semaphore github.com/hashicorp/consul/logging github.com/hashicorp/consul/logging/monitor github.com/hashicorp/consul/sdk/freeport github.com/hashicorp/consul/sdk/testutil github.com/hashicorp/consul/sdk/testutil/retry github.com/hashicorp/consul/sentinel github.com/hashicorp/consul/service_os github.com/hashicorp/consul/snapshot github.com/hashicorp/consul/testrpc github.com/hashicorp/consul/tlsutil github.com/hashicorp/consul/types github.com/hashicorp/consul/version internal/unsafeheader runtime/internal/sys runtime/internal/atomic internal/cpu math/bits runtime/internal/math unicode/utf8 internal/race sync/atomic unicode internal/bytealg math internal/testlog encoding unicode/utf16 runtime container/list internal/nettrace crypto/internal/subtle runtime/cgo crypto/subtle vendor/golang.org/x/crypto/cryptobyte/asn1 vendor/golang.org/x/crypto/internal/subtle github.com/circonus-labs/circonus-gometrics/api/config golang.org/x/sys/internal/unsafeheader golang.org/x/net/internal/iana github.com/hashicorp/consul/types google.golang.org/grpc/internal/grpclog google.golang.org/grpc/serviceconfig github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/client/metadata github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/selection github.com/hashicorp/consul/vendor/k8s.io/client-go/util/integer github.com/aws/aws-sdk-go/aws/client/metadata go.opencensus.io go.opencensus.io/trace/internal go.opencensus.io/internal/tagencoding github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/internal/sdkmath github.com/aws/aws-sdk-go/internal/sdkmath github.com/hashicorp/consul/service_os internal/reflectlite sync internal/singleflight google.golang.org/grpc/internal/buffer math/rand google.golang.org/grpc/internal/grpcsync github.com/hashicorp/consul/agent/token golang.org/x/sync/singleflight github.com/aws/aws-sdk-go/internal/sync/singleflight errors sort io internal/oserror strconv github.com/hashicorp/hcl/hcl/strconv syscall container/heap bytes hash strings reflect crypto text/tabwriter github.com/armon/go-radix regexp/syntax bufio internal/syscall/unix time internal/syscall/execenv github.com/hashicorp/golang-lru/simplelru github.com/hashicorp/go-immutable-radix regexp vendor/golang.org/x/net/dns/dnsmessage internal/poll internal/fmtsort encoding/binary context os hash/crc32 crypto/internal/randutil crypto/md5 encoding/base64 golang.org/x/crypto/blake2b crypto/cipher crypto/sha512 crypto/ed25519/internal/edwards25519 crypto/aes fmt net crypto/des crypto/hmac crypto/rc4 crypto/sha1 crypto/sha256 encoding/pem encoding/json github.com/hashicorp/hcl/hcl/token log github.com/hashicorp/hcl/hcl/ast github.com/hashicorp/hcl/hcl/scanner github.com/hashicorp/hcl/json/token github.com/hashicorp/hcl/json/scanner github.com/hashicorp/hcl/hcl/parser github.com/hashicorp/hcl/json/parser github.com/hashicorp/hcl/hcl/printer github.com/hashicorp/hcl compress/flate math/big github.com/hashicorp/consul/acl compress/gzip encoding/hex path/filepath net/url io/ioutil vendor/golang.org/x/sys/cpu vendor/golang.org/x/crypto/curve25519 vendor/golang.org/x/crypto/hkdf vendor/golang.org/x/crypto/chacha20 vendor/golang.org/x/crypto/poly1305 vendor/golang.org/x/text/transform vendor/golang.org/x/crypto/chacha20poly1305 vendor/golang.org/x/text/unicode/bidi vendor/golang.org/x/text/unicode/norm crypto/rand crypto/elliptic encoding/asn1 crypto/ed25519 crypto/rsa crypto/dsa vendor/golang.org/x/crypto/cryptobyte crypto/x509/pkix vendor/golang.org/x/text/secure/bidirule net/textproto vendor/golang.org/x/net/http2/hpack vendor/golang.org/x/net/idna mime crypto/ecdsa mime/quotedprintable net/http/internal vendor/golang.org/x/net/http/httpguts vendor/golang.org/x/net/http/httpproxy mime/multipart path crypto/x509 os/signal github.com/pkg/errors github.com/circonus-labs/circonusllhist github.com/DataDog/datadog-go/statsd github.com/beorn7/perks/quantile github.com/cespare/xxhash github.com/golang/protobuf/proto github.com/prometheus/common/internal/bitbucket.org/ww/goautoneg github.com/prometheus/common/model github.com/prometheus/procfs/internal/fs github.com/prometheus/procfs/internal/util crypto/tls golang.org/x/sys/unix runtime/debug github.com/hashicorp/consul/ipaddr github.com/mitchellh/go-testing-interface os/exec github.com/hashicorp/consul/version github.com/hashicorp/go-uuid encoding/gob github.com/prometheus/procfs github.com/golang/protobuf/ptypes/any github.com/golang/protobuf/ptypes/duration github.com/golang/protobuf/ptypes/timestamp github.com/golang/protobuf/ptypes github.com/prometheus/client_model/go net/http/httptrace net/http github.com/prometheus/client_golang/prometheus/internal github.com/matttproud/golang_protobuf_extensions/pbutil github.com/hashicorp/consul/sdk/freeport go/token html text/template/parse compress/lzw github.com/google/btree github.com/hashicorp/errwrap github.com/hashicorp/go-multierror github.com/hashicorp/go-sockaddr encoding/base32 golang.org/x/crypto/ed25519 golang.org/x/net/bpf text/template golang.org/x/net/internal/socket github.com/hashicorp/memberlist/vendor/github.com/sean-/seed golang.org/x/net/ipv4 golang.org/x/net/ipv6 html/template github.com/hashicorp/yamux github.com/miekg/dns github.com/mitchellh/reflectwalk github.com/mitchellh/mapstructure github.com/mitchellh/copystructure github.com/davecgh/go-spew/spew github.com/pmezard/go-difflib/difflib github.com/stretchr/objx gopkg.in/yaml.v2 github.com/armon/go-metrics github.com/hashicorp/go-cleanhttp github.com/hashicorp/go-retryablehttp github.com/tv42/httpunix github.com/armon/go-metrics/datadog expvar github.com/circonus-labs/circonus-gometrics/api github.com/prometheus/common/expfmt github.com/hashicorp/serf/coordinate net/rpc github.com/prometheus/client_golang/prometheus github.com/circonus-labs/circonus-gometrics/checkmgr github.com/hashicorp/go-msgpack/codec github.com/circonus-labs/circonus-gometrics flag github.com/armon/go-metrics/circonus github.com/prometheus/client_golang/prometheus/push net/http/httptest github.com/mattn/go-isatty github.com/mattn/go-colorable github.com/armon/go-metrics/prometheus github.com/fatih/color github.com/hashicorp/memberlist github.com/stretchr/testify/assert github.com/hashicorp/go-hclog log/syslog github.com/hashicorp/go-syslog github.com/bgentry/speakeasy os/user github.com/hashicorp/go-version hash/crc64 github.com/stretchr/testify/mock github.com/stretchr/testify/require github.com/hashicorp/serf/serf runtime/trace testing github.com/hashicorp/go-rootcerts github.com/hashicorp/consul/api github.com/hashicorp/consul/lib github.com/hashicorp/raft github.com/hashicorp/consul/agent/cache github.com/hashicorp/golang-lru hash/fnv github.com/kr/text github.com/mitchellh/hashstructure github.com/hashicorp/consul/command/flags github.com/hashicorp/consul/command/helpers github.com/NYTimes/gziphandler github.com/armon/circbuf github.com/hashicorp/consul/vendor/github.com/coredns/coredns/plugin/pkg/dnsutil github.com/elazarl/go-bindata-assetfs golang.org/x/net/internal/socks github.com/hashicorp/consul/agent/exec github.com/hashicorp/consul/sdk/testutil/retry github.com/hashicorp/consul/sdk/testutil golang.org/x/net/proxy github.com/posener/complete/cmd/install golang.org/x/net/internal/timeseries golang.org/x/net/trace google.golang.org/grpc/backoff github.com/docker/go-connections/sockets github.com/posener/complete/cmd google.golang.org/grpc/grpclog github.com/posener/complete google.golang.org/grpc/credentials/internal google.golang.org/grpc/connectivity google.golang.org/grpc/internal google.golang.org/grpc/metadata google.golang.org/grpc/credentials github.com/mitchellh/cli google.golang.org/grpc/attributes google.golang.org/grpc/internal/grpcrand google.golang.org/grpc/codes google.golang.org/grpc/encoding google.golang.org/grpc/resolver google.golang.org/grpc/encoding/proto google.golang.org/grpc/internal/backoff google.golang.org/grpc/balancer google.golang.org/grpc/internal/balancerload google.golang.org/grpc/binarylog/grpc_binarylog_v1 google.golang.org/genproto/googleapis/rpc/status google.golang.org/grpc/balancer/base google.golang.org/grpc/internal/status github.com/hashicorp/consul/logging github.com/hashicorp/consul/command/acl/agenttokens github.com/hashicorp/consul/command/acl/authmethod github.com/hashicorp/consul/command/acl/authmethod/delete github.com/hashicorp/consul/agent/consul/autopilot github.com/hashicorp/consul/command/acl/bindingrule github.com/hashicorp/consul/command/acl/authmethod/create github.com/hashicorp/consul/command/acl/authmethod/list github.com/hashicorp/consul/command/acl/authmethod/read github.com/hashicorp/consul/command/acl/authmethod/update github.com/hashicorp/consul/command/acl/bindingrule/create github.com/hashicorp/consul/agent/structs github.com/hashicorp/consul/command/acl/bindingrule/list github.com/hashicorp/consul/command/acl/token github.com/hashicorp/consul/command/acl/policy github.com/hashicorp/consul/command/acl/role github.com/hashicorp/consul/command/acl/policy/list github.com/hashicorp/consul/command/acl/bootstrap github.com/hashicorp/consul/command/acl/role/list github.com/hashicorp/consul/command/acl/token/list github.com/hashicorp/consul/agent/ae google.golang.org/grpc/balancer/roundrobin google.golang.org/grpc/status google.golang.org/grpc/internal/channelz google.golang.org/grpc/internal/envconfig google.golang.org/grpc/internal/binarylog google.golang.org/grpc/internal/grpcutil google.golang.org/grpc/internal/resolver/dns google.golang.org/grpc/internal/resolver/passthrough golang.org/x/text/transform golang.org/x/text/unicode/bidi golang.org/x/net/http2/hpack golang.org/x/text/unicode/norm google.golang.org/grpc/internal/syscall golang.org/x/text/secure/bidirule google.golang.org/grpc/keepalive google.golang.org/grpc/peer google.golang.org/grpc/stats google.golang.org/grpc/tap google.golang.org/grpc/naming net/http/httputil github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/awserr github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/internal/ini github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/internal/shareddefaults golang.org/x/net/idna github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/endpoints github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/credentials github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/internal/sdkio golang.org/x/net/http/httpguts github.com/jmespath/go-jmespath golang.org/x/net/http2 github.com/hashicorp/consul/command/acl github.com/hashicorp/consul/command/acl/bindingrule/delete github.com/hashicorp/consul/command/acl/bindingrule/read github.com/hashicorp/consul/command/acl/bindingrule/update github.com/hashicorp/consul/command/acl/policy/create github.com/hashicorp/consul/command/acl/policy/delete github.com/hashicorp/consul/command/acl/policy/read github.com/hashicorp/consul/command/acl/policy/update github.com/hashicorp/consul/command/acl/role/create github.com/hashicorp/consul/command/acl/role/delete github.com/hashicorp/consul/command/acl/role/read github.com/hashicorp/consul/command/acl/role/update github.com/hashicorp/consul/command/acl/rules github.com/hashicorp/consul/command/acl/token/clone github.com/hashicorp/consul/command/acl/token/create github.com/hashicorp/consul/command/acl/token/delete github.com/hashicorp/consul/command/acl/token/read github.com/hashicorp/consul/command/acl/token/update github.com/hashicorp/consul/agent/connect github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/awsutil google.golang.org/grpc/internal/transport github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/internal/sdkrand github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/credentials/processcreds encoding/xml github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/internal/sdkuri github.com/gogo/protobuf/proto github.com/gogo/protobuf/sortkeys github.com/hashicorp/consul/agent/consul/discoverychain github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws github.com/hashicorp/hil/ast google.golang.org/grpc github.com/hashicorp/hil github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/request github.com/hashicorp/consul/agent/consul/prepared_query github.com/hashicorp/go-memdb github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/client github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/corehandlers github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/private/protocol github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/csm github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/private/protocol/rest github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/private/protocol/query/queryutil google.golang.org/grpc/health/grpc_health_v1 github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/private/protocol/xml/xmlutil github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/signer/v4 github.com/hashicorp/consul/agent/checks github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/private/protocol/query github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/ec2metadata github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/private/protocol/json/jsonutil github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/service/sts github.com/gogo/protobuf/protoc-gen-gogo/descriptor github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/credentials/ec2rolecreds github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/credentials/endpointcreds github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/private/protocol/jsonrpc github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/defaults github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/service/sts/stsiface github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/service/acmpca github.com/gogo/protobuf/types github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/credentials/stscreds github.com/hashicorp/consul/vendor/github.com/aws/aws-sdk-go/aws/session github.com/gogo/protobuf/gogoproto github.com/hashicorp/consul/vendor/github.com/hashicorp/vault/sdk/helper/consts github.com/hashicorp/consul/vendor/github.com/hashicorp/vault/sdk/helper/hclutil github.com/golang/snappy github.com/pierrec/lz4/internal/xxh32 github.com/ryanuber/go-glob github.com/hashicorp/consul/vendor/github.com/hashicorp/vault/sdk/helper/strutil github.com/pierrec/lz4 golang.org/x/time/rate github.com/hashicorp/consul/vendor/github.com/hashicorp/vault/sdk/helper/parseutil golang.org/x/crypto/pbkdf2 gopkg.in/square/go-jose.v2/cipher github.com/hashicorp/consul/vendor/github.com/hashicorp/vault/sdk/helper/compressutil gopkg.in/square/go-jose.v2/json github.com/hashicorp/consul/agent/consul/authmethod github.com/hashicorp/consul/vendor/github.com/hashicorp/vault/sdk/helper/jsonutil github.com/google/gofuzz/bytesource gopkg.in/inf.v0 github.com/google/gofuzz github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/api/resource github.com/hashicorp/consul/vendor/k8s.io/apimachinery/third_party/forked/golang/reflect github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/conversion github.com/hashicorp/consul/agent/agentpb github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/fields gopkg.in/square/go-jose.v2 github.com/golang/glog github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/sets github.com/hashicorp/consul/agent/consul/state github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/errors github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/validation/field go/scanner go/ast github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/validation gopkg.in/square/go-jose.v2/jwt github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/labels github.com/hashicorp/consul/vendor/github.com/hashicorp/vault/api internal/lazyregexp github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/conversion/queryparams go/doc go/parser github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/runtime/schema github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/json github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/runtime github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/types github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/intstr github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/runtime github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/net github.com/hashicorp/consul/agent/connect/ca github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/wait github.com/googleapis/gnostic/extensions github.com/gregjones/httpcache github.com/googleapis/gnostic/compiler hash/adler32 github.com/googleapis/gnostic/OpenAPIv2 compress/zlib github.com/ghodss/yaml github.com/peterbourgon/diskv github.com/modern-go/concurrent github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/watch github.com/gregjones/httpcache/diskcache github.com/modern-go/reflect2 github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/runtime/serializer/recognizer github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/framer github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/apis/meta/v1 github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/yaml github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/runtime/serializer/protobuf github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/version github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/runtime/serializer/streaming github.com/hashicorp/consul/vendor/k8s.io/client-go/pkg/version golang.org/x/crypto/ssh/terminal github.com/hashicorp/consul/vendor/k8s.io/client-go/tools/clientcmd/api github.com/hashicorp/consul/vendor/k8s.io/client-go/transport github.com/json-iterator/go github.com/hashicorp/consul/vendor/k8s.io/client-go/util/connrotation github.com/hashicorp/consul/vendor/k8s.io/client-go/tools/metrics github.com/hashicorp/consul/vendor/k8s.io/client-go/util/cert github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/util/clock github.com/hashicorp/consul/vendor/k8s.io/client-go/util/flowcontrol github.com/hashicorp/consul/vendor/k8s.io/api/authentication/v1 github.com/hashicorp/consul/vendor/k8s.io/api/core/v1 github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/api/errors github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/apis/meta/v1/unstructured github.com/hashicorp/consul/vendor/k8s.io/api/admissionregistration/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/runtime/serializer/versioning github.com/hashicorp/consul/vendor/k8s.io/api/admissionregistration/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/runtime/serializer/json github.com/hashicorp/consul/vendor/k8s.io/api/authentication/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/runtime/serializer github.com/hashicorp/consul/vendor/k8s.io/api/authorization/v1 github.com/hashicorp/consul/vendor/k8s.io/api/authorization/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/api/certificates/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/api/rbac/v1 github.com/hashicorp/consul/vendor/k8s.io/api/rbac/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/api/rbac/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/api/scheduling/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/api/scheduling/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/api/storage/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/client-go/pkg/apis/clientauthentication github.com/hashicorp/consul/vendor/k8s.io/client-go/rest/watch github.com/hashicorp/consul/vendor/k8s.io/client-go/pkg/apis/clientauthentication/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/client-go/pkg/apis/clientauthentication/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/apis/meta/v1beta1 github.com/hashicorp/go-raftchunking/types github.com/hashicorp/consul/vendor/k8s.io/client-go/plugin/pkg/client/auth/exec github.com/hashicorp/go-raftchunking github.com/hashicorp/consul/vendor/k8s.io/apimachinery/pkg/api/meta github.com/hashicorp/consul/agent/consul/fsm github.com/hashicorp/consul/vendor/k8s.io/client-go/rest github.com/hashicorp/consul/agent/metadata github.com/hashicorp/consul/tlsutil github.com/hashicorp/net-rpc-msgpackrpc github.com/hashicorp/consul/agent/router github.com/hashicorp/consul/lib/semaphore archive/tar github.com/hashicorp/consul/agent/pool github.com/hashicorp/go-bexpr github.com/hashicorp/consul/vendor/github.com/hashicorp/go-connlimit github.com/boltdb/bolt github.com/hashicorp/consul/snapshot github.com/hashicorp/consul/agent/local github.com/hashicorp/go-sockaddr/template github.com/shirou/gopsutil/internal/common github.com/hashicorp/raft-boltdb github.com/hashicorp/consul/agent/systemd github.com/golang/protobuf/protoc-gen-go/descriptor net/mail github.com/shirou/gopsutil/cpu github.com/shirou/gopsutil/disk github.com/shirou/gopsutil/host github.com/shirou/gopsutil/mem github.com/gogo/googleapis/google/api github.com/hashicorp/consul/vendor/github.com/envoyproxy/protoc-gen-validate/validate github.com/hashicorp/consul/agent/debug github.com/gogo/googleapis/google/rpc github.com/gogo/protobuf/jsonpb github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/type github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/api/v2/core github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/pkg/util github.com/hashicorp/consul/api/watch github.com/hashicorp/consul/lib/file github.com/hashicorp/consul/vendor/k8s.io/api/apps/v1 github.com/hashicorp/consul/vendor/k8s.io/api/apps/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/api/apps/v1beta2 github.com/hashicorp/consul/vendor/k8s.io/api/autoscaling/v1 github.com/hashicorp/consul/vendor/k8s.io/api/autoscaling/v2beta1 github.com/hashicorp/consul/vendor/k8s.io/api/batch/v1 github.com/hashicorp/consul/vendor/k8s.io/api/batch/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/api/batch/v2alpha1 github.com/hashicorp/consul/vendor/k8s.io/api/events/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/api/extensions/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/api/networking/v1 github.com/hashicorp/consul/vendor/k8s.io/api/policy/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/api/settings/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/api/storage/v1 github.com/hashicorp/consul/vendor/k8s.io/api/storage/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/tools/reference github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/api/v2/auth github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/api/v2/cluster github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/api/v2/endpoint github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/api/v2/route github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/scheme github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/config/filter/network/ext_authz/v2 github.com/hashicorp/consul/vendor/k8s.io/client-go/discovery github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/admissionregistration/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/admissionregistration/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/apps/v1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/apps/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/apps/v1beta2 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/authentication/v1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/authentication/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/authorization/v1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/authorization/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/autoscaling/v1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/autoscaling/v2beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/batch/v1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/batch/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/batch/v2alpha1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/certificates/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/core/v1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/events/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/extensions/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/networking/v1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/policy/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/rbac/v1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/rbac/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/rbac/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/scheduling/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/scheduling/v1beta1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/settings/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/storage/v1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/storage/v1alpha1 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes/typed/storage/v1beta1 github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/api/v2/listener github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/config/filter/accesslog/v2 github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/service/auth/v2 github.com/hashicorp/consul/vendor/k8s.io/client-go/kubernetes github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/api/v2 github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/config/filter/network/tcp_proxy/v2 github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/service/auth/v2alpha github.com/hashicorp/consul/agent/consul/authmethod/kubeauth github.com/hashicorp/consul/logging/monitor github.com/denverdino/aliyungo/util github.com/aws/aws-sdk-go/aws/awserr github.com/denverdino/aliyungo/common github.com/aws/aws-sdk-go/internal/ini github.com/hashicorp/consul/agent/consul github.com/denverdino/aliyungo/ecs github.com/aws/aws-sdk-go/internal/shareddefaults github.com/aws/aws-sdk-go/aws/credentials github.com/aws/aws-sdk-go/aws/endpoints github.com/hashicorp/go-discover/provider/aliyun github.com/aws/aws-sdk-go/internal/sdkio github.com/aws/aws-sdk-go/aws/awsutil github.com/aws/aws-sdk-go/internal/sdkrand github.com/aws/aws-sdk-go/internal/sdkuri github.com/aws/aws-sdk-go/aws/credentials/processcreds github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/config/filter/network/http_connection_manager/v2 github.com/hashicorp/consul/vendor/github.com/envoyproxy/go-control-plane/envoy/service/discovery/v2 github.com/aws/aws-sdk-go/internal/strings golang.org/x/net/context/ctxhttp cloud.google.com/go/compute/metadata golang.org/x/oauth2/internal golang.org/x/oauth2 golang.org/x/oauth2/jws google.golang.org/api/internal/third_party/uritemplates golang.org/x/oauth2/jwt google.golang.org/api/googleapi golang.org/x/oauth2/google github.com/googleapis/gax-go/v2 google.golang.org/api/internal google.golang.org/api/internal/gensupport google.golang.org/api/option github.com/hashicorp/consul/agent/cache-types github.com/hashicorp/consul/agent/config google.golang.org/api/option/internaloption go.opencensus.io/internal go.opencensus.io/trace/tracestate go.opencensus.io/trace github.com/hashicorp/consul/agent/proxycfg go.opencensus.io/trace/propagation go.opencensus.io/plugin/ochttp/propagation/b3 go.opencensus.io/resource runtime/pprof go.opencensus.io/metric/metricdata github.com/hashicorp/consul/agent/xds go.opencensus.io/metric/metricproducer google.golang.org/api/googleapi/transport google.golang.org/api/transport/cert go.opencensus.io/tag google.golang.org/api/transport/http/internal/propagation github.com/aws/aws-sdk-go/aws go.opencensus.io/stats/internal github.com/hashicorp/mdns go.opencensus.io/stats go.opencensus.io/stats/view github.com/hashicorp/go-discover/provider/mdns github.com/aws/aws-sdk-go/aws/request github.com/gophercloud/gophercloud go.opencensus.io/plugin/ochttp github.com/packethost/packngo github.com/aws/aws-sdk-go/aws/corehandlers github.com/aws/aws-sdk-go/aws/client github.com/aws/aws-sdk-go/private/protocol github.com/aws/aws-sdk-go/aws/csm github.com/aws/aws-sdk-go/private/protocol/json/jsonutil github.com/aws/aws-sdk-go/aws/ec2metadata github.com/aws/aws-sdk-go/private/protocol/rest github.com/aws/aws-sdk-go/private/protocol/query/queryutil github.com/aws/aws-sdk-go/aws/credentials/ec2rolecreds github.com/aws/aws-sdk-go/aws/credentials/endpointcreds github.com/aws/aws-sdk-go/private/protocol/xml/xmlutil github.com/aws/aws-sdk-go/aws/signer/v4 google.golang.org/api/transport/http github.com/aws/aws-sdk-go/aws/defaults google.golang.org/api/compute/v1 github.com/aws/aws-sdk-go/private/protocol/query github.com/aws/aws-sdk-go/private/protocol/ec2query github.com/gophercloud/gophercloud/pagination github.com/aws/aws-sdk-go/service/ec2 github.com/aws/aws-sdk-go/service/sts github.com/gophercloud/gophercloud/openstack/identity/v2/tenants github.com/aws/aws-sdk-go/service/sts/stsiface github.com/gophercloud/gophercloud/openstack/identity/v2/tokens github.com/aws/aws-sdk-go/aws/credentials/stscreds github.com/gophercloud/gophercloud/openstack/identity/v3/tokens github.com/aws/aws-sdk-go/aws/session github.com/gophercloud/gophercloud/openstack/identity/v3/extensions/ec2tokens github.com/gophercloud/gophercloud/openstack/identity/v3/extensions/oauth1 github.com/gophercloud/gophercloud/openstack/utils github.com/gophercloud/gophercloud/openstack/compute/v2/servers github.com/gophercloud/gophercloud/openstack github.com/hashicorp/go-discover/provider/packet github.com/hashicorp/go-discover/provider/os github.com/imdario/mergo github.com/prometheus/client_golang/prometheus/promhttp golang.org/x/net/context internal/profile github.com/hashicorp/go-checkpoint github.com/hashicorp/consul/command/catalog net/http/pprof github.com/hashicorp/consul/command/catalog/list/dc github.com/ryanuber/columnize github.com/hashicorp/consul/command/catalog/list/services github.com/hashicorp/consul/command/catalog/list/nodes github.com/hashicorp/consul/command/config github.com/hashicorp/consul/command/config/delete github.com/hashicorp/consul/command/config/list github.com/hashicorp/consul/command/config/read github.com/hashicorp/consul/command/config/write github.com/hashicorp/consul/command/connect github.com/hashicorp/consul/command/connect/ca github.com/hashicorp/consul/command/connect/ca/get github.com/hashicorp/consul/command/connect/ca/set github.com/hashicorp/consul/connect github.com/hashicorp/consul/command/connect/envoy/pipe-bootstrap github.com/hashicorp/consul/command/debug github.com/hashicorp/consul/connect/proxy github.com/hashicorp/consul/command/event github.com/hashicorp/consul/command/exec github.com/hashicorp/consul/command/connect/proxy github.com/hashicorp/consul/command/forceleave github.com/hashicorp/consul/command/connect/envoy github.com/hashicorp/consul/command/info github.com/hashicorp/consul/command/intention github.com/hashicorp/consul/command/intention/check github.com/hashicorp/consul/command/intention/finder github.com/hashicorp/consul/command/intention/match github.com/hashicorp/consul/command/intention/create github.com/hashicorp/consul/command/intention/delete github.com/hashicorp/consul/command/intention/get github.com/hashicorp/consul/command/join github.com/hashicorp/consul/command/keygen github.com/hashicorp/consul/command/kv github.com/hashicorp/consul/command/kv/del github.com/hashicorp/consul/command/kv/impexp github.com/hashicorp/consul/command/kv/exp github.com/hashicorp/consul/command/kv/get github.com/hashicorp/consul/command/kv/imp github.com/hashicorp/consul/command/kv/put github.com/hashicorp/consul/command/leave github.com/hashicorp/consul/command/login github.com/hashicorp/consul/command/logout github.com/hashicorp/consul/command/maint github.com/hashicorp/consul/command/members github.com/hashicorp/consul/command/monitor github.com/hashicorp/consul/command/operator github.com/hashicorp/consul/command/operator/autopilot github.com/hashicorp/consul/command/operator/autopilot/get github.com/hashicorp/consul/command/operator/autopilot/set github.com/hashicorp/consul/command/operator/raft github.com/hashicorp/consul/command/operator/raft/listpeers github.com/hashicorp/consul/command/operator/raft/removepeer github.com/hashicorp/consul/command/reload github.com/hashicorp/consul/command/rtt github.com/hashicorp/consul/command/services github.com/hashicorp/consul/command/snapshot github.com/hashicorp/consul/command/services/deregister github.com/hashicorp/consul/command/services/register github.com/hashicorp/consul/command/snapshot/inspect github.com/hashicorp/consul/command/snapshot/restore github.com/hashicorp/consul/vendor/github.com/rboyer/safeio github.com/hashicorp/consul/command/tls github.com/hashicorp/consul/command/snapshot/save github.com/hashicorp/consul/command/tls/ca github.com/hashicorp/consul/command/tls/ca/create github.com/hashicorp/consul/command/tls/cert github.com/hashicorp/consul/command/tls/cert/create github.com/hashicorp/consul/command/validate github.com/hashicorp/consul/command/version github.com/hashicorp/consul/agent/consul/authmethod/testauth github.com/hashicorp/consul/agent/mock github.com/hashicorp/consul/connect/certgen github.com/hashicorp/consul/sentinel github.com/hashicorp/consul/testrpc github.com/hashicorp/go-discover/provider/gce github.com/hashicorp/go-discover/provider/aws github.com/hashicorp/go-discover github.com/hashicorp/consul/agent github.com/hashicorp/consul/command/agent github.com/hashicorp/consul/command/lock github.com/hashicorp/consul/command/watch github.com/hashicorp/consul/command/keyring github.com/hashicorp/consul/command github.com/hashicorp/consul make[1]: Leaving directory '/<>/consul-1.7.4+dfsg1' debian/rules override_dh_auto_test make[1]: Entering directory '/<>/consul-1.7.4+dfsg1' PATH="/<>/consul-1.7.4+dfsg1/_build/bin:${PATH}" \ DH_GOLANG_EXCLUDES="test/integration api agent/cache agent/checks agent/connect agent/consul agent/proxycfg command/debug command/tls" \ dh_auto_test -v --max-parallel=4 -- -short -failfast -timeout 8m cd _build && go test -vet=off -v -p 4 -short -failfast -timeout 8m github.com/hashicorp/consul github.com/hashicorp/consul/acl github.com/hashicorp/consul/agent github.com/hashicorp/consul/agent/ae github.com/hashicorp/consul/agent/agentpb github.com/hashicorp/consul/agent/config github.com/hashicorp/consul/agent/debug github.com/hashicorp/consul/agent/exec github.com/hashicorp/consul/agent/local github.com/hashicorp/consul/agent/metadata github.com/hashicorp/consul/agent/mock github.com/hashicorp/consul/agent/pool github.com/hashicorp/consul/agent/router github.com/hashicorp/consul/agent/structs github.com/hashicorp/consul/agent/systemd github.com/hashicorp/consul/agent/token github.com/hashicorp/consul/agent/xds github.com/hashicorp/consul/command github.com/hashicorp/consul/command/acl github.com/hashicorp/consul/command/acl/agenttokens github.com/hashicorp/consul/command/acl/authmethod github.com/hashicorp/consul/command/acl/authmethod/create github.com/hashicorp/consul/command/acl/authmethod/delete github.com/hashicorp/consul/command/acl/authmethod/list github.com/hashicorp/consul/command/acl/authmethod/read github.com/hashicorp/consul/command/acl/authmethod/update github.com/hashicorp/consul/command/acl/bindingrule github.com/hashicorp/consul/command/acl/bindingrule/create github.com/hashicorp/consul/command/acl/bindingrule/delete github.com/hashicorp/consul/command/acl/bindingrule/list github.com/hashicorp/consul/command/acl/bindingrule/read github.com/hashicorp/consul/command/acl/bindingrule/update github.com/hashicorp/consul/command/acl/bootstrap github.com/hashicorp/consul/command/acl/policy github.com/hashicorp/consul/command/acl/policy/create github.com/hashicorp/consul/command/acl/policy/delete github.com/hashicorp/consul/command/acl/policy/list github.com/hashicorp/consul/command/acl/policy/read github.com/hashicorp/consul/command/acl/policy/update github.com/hashicorp/consul/command/acl/role github.com/hashicorp/consul/command/acl/role/create github.com/hashicorp/consul/command/acl/role/delete github.com/hashicorp/consul/command/acl/role/list github.com/hashicorp/consul/command/acl/role/read github.com/hashicorp/consul/command/acl/role/update github.com/hashicorp/consul/command/acl/rules github.com/hashicorp/consul/command/acl/token github.com/hashicorp/consul/command/acl/token/clone github.com/hashicorp/consul/command/acl/token/create github.com/hashicorp/consul/command/acl/token/delete github.com/hashicorp/consul/command/acl/token/list github.com/hashicorp/consul/command/acl/token/read github.com/hashicorp/consul/command/acl/token/update github.com/hashicorp/consul/command/agent github.com/hashicorp/consul/command/catalog github.com/hashicorp/consul/command/catalog/list/dc github.com/hashicorp/consul/command/catalog/list/nodes github.com/hashicorp/consul/command/catalog/list/services github.com/hashicorp/consul/command/config github.com/hashicorp/consul/command/config/delete github.com/hashicorp/consul/command/config/list github.com/hashicorp/consul/command/config/read github.com/hashicorp/consul/command/config/write github.com/hashicorp/consul/command/connect github.com/hashicorp/consul/command/connect/ca github.com/hashicorp/consul/command/connect/ca/get github.com/hashicorp/consul/command/connect/ca/set github.com/hashicorp/consul/command/connect/envoy github.com/hashicorp/consul/command/connect/envoy/pipe-bootstrap github.com/hashicorp/consul/command/connect/proxy github.com/hashicorp/consul/command/event github.com/hashicorp/consul/command/exec github.com/hashicorp/consul/command/flags github.com/hashicorp/consul/command/forceleave github.com/hashicorp/consul/command/helpers github.com/hashicorp/consul/command/info github.com/hashicorp/consul/command/intention github.com/hashicorp/consul/command/intention/check github.com/hashicorp/consul/command/intention/create github.com/hashicorp/consul/command/intention/delete github.com/hashicorp/consul/command/intention/finder github.com/hashicorp/consul/command/intention/get github.com/hashicorp/consul/command/intention/match github.com/hashicorp/consul/command/join github.com/hashicorp/consul/command/keygen github.com/hashicorp/consul/command/keyring github.com/hashicorp/consul/command/kv github.com/hashicorp/consul/command/kv/del github.com/hashicorp/consul/command/kv/exp github.com/hashicorp/consul/command/kv/get github.com/hashicorp/consul/command/kv/imp github.com/hashicorp/consul/command/kv/impexp github.com/hashicorp/consul/command/kv/put github.com/hashicorp/consul/command/leave github.com/hashicorp/consul/command/lock github.com/hashicorp/consul/command/login github.com/hashicorp/consul/command/logout github.com/hashicorp/consul/command/maint github.com/hashicorp/consul/command/members github.com/hashicorp/consul/command/monitor github.com/hashicorp/consul/command/operator github.com/hashicorp/consul/command/operator/autopilot github.com/hashicorp/consul/command/operator/autopilot/get github.com/hashicorp/consul/command/operator/autopilot/set github.com/hashicorp/consul/command/operator/raft github.com/hashicorp/consul/command/operator/raft/listpeers github.com/hashicorp/consul/command/operator/raft/removepeer github.com/hashicorp/consul/command/reload github.com/hashicorp/consul/command/rtt github.com/hashicorp/consul/command/services github.com/hashicorp/consul/command/services/deregister github.com/hashicorp/consul/command/services/register github.com/hashicorp/consul/command/snapshot github.com/hashicorp/consul/command/snapshot/inspect github.com/hashicorp/consul/command/snapshot/restore github.com/hashicorp/consul/command/snapshot/save github.com/hashicorp/consul/command/validate github.com/hashicorp/consul/command/version github.com/hashicorp/consul/command/watch github.com/hashicorp/consul/connect github.com/hashicorp/consul/connect/certgen github.com/hashicorp/consul/connect/proxy github.com/hashicorp/consul/ipaddr github.com/hashicorp/consul/lib github.com/hashicorp/consul/lib/file github.com/hashicorp/consul/lib/semaphore github.com/hashicorp/consul/logging github.com/hashicorp/consul/logging/monitor github.com/hashicorp/consul/sdk/freeport github.com/hashicorp/consul/sdk/testutil github.com/hashicorp/consul/sdk/testutil/retry github.com/hashicorp/consul/sentinel github.com/hashicorp/consul/service_os github.com/hashicorp/consul/snapshot github.com/hashicorp/consul/testrpc github.com/hashicorp/consul/tlsutil github.com/hashicorp/consul/types github.com/hashicorp/consul/version testing: warning: no tests to run PASS ok github.com/hashicorp/consul 0.103s [no tests to run] === RUN TestACL === RUN TestACL/DenyAll === RUN TestACL/DenyAll/DenyACLRead === RUN TestACL/DenyAll/DenyACLWrite === RUN TestACL/DenyAll/DenyAgentRead === RUN TestACL/DenyAll/DenyAgentWrite === RUN TestACL/DenyAll/DenyEventRead === RUN TestACL/DenyAll/DenyEventWrite === RUN TestACL/DenyAll/DenyIntentionDefaultAllow === RUN TestACL/DenyAll/DenyIntentionRead === RUN TestACL/DenyAll/DenyIntentionWrite === RUN TestACL/DenyAll/DenyKeyRead === RUN TestACL/DenyAll/DenyKeyringRead === RUN TestACL/DenyAll/DenyKeyringWrite === RUN TestACL/DenyAll/DenyKeyWrite === RUN TestACL/DenyAll/DenyNodeRead === RUN TestACL/DenyAll/DenyNodeWrite === RUN TestACL/DenyAll/DenyOperatorRead === RUN TestACL/DenyAll/DenyOperatorWrite === RUN TestACL/DenyAll/DenyPreparedQueryRead === RUN TestACL/DenyAll/DenyPreparedQueryWrite === RUN TestACL/DenyAll/DenyServiceRead === RUN TestACL/DenyAll/DenyServiceWrite === RUN TestACL/DenyAll/DenySessionRead === RUN TestACL/DenyAll/DenySessionWrite === RUN TestACL/DenyAll/DenySnapshot === RUN TestACL/AllowAll === RUN TestACL/AllowAll/DenyACLRead === RUN TestACL/AllowAll/DenyACLWrite === RUN TestACL/AllowAll/AllowAgentRead === RUN TestACL/AllowAll/AllowAgentWrite === RUN TestACL/AllowAll/AllowEventRead === RUN TestACL/AllowAll/AllowEventWrite === RUN TestACL/AllowAll/AllowIntentionDefaultAllow === RUN TestACL/AllowAll/AllowIntentionRead === RUN TestACL/AllowAll/AllowIntentionWrite === RUN TestACL/AllowAll/AllowKeyRead === RUN TestACL/AllowAll/AllowKeyringRead === RUN TestACL/AllowAll/AllowKeyringWrite === RUN TestACL/AllowAll/AllowKeyWrite === RUN TestACL/AllowAll/AllowNodeRead === RUN TestACL/AllowAll/AllowNodeWrite === RUN TestACL/AllowAll/AllowOperatorRead === RUN TestACL/AllowAll/AllowOperatorWrite === RUN TestACL/AllowAll/AllowPreparedQueryRead === RUN TestACL/AllowAll/AllowPreparedQueryWrite === RUN TestACL/AllowAll/AllowServiceRead === RUN TestACL/AllowAll/AllowServiceWrite === RUN TestACL/AllowAll/AllowSessionRead === RUN TestACL/AllowAll/AllowSessionWrite === RUN TestACL/AllowAll/DenySnapshot === RUN TestACL/ManageAll === RUN TestACL/ManageAll/AllowACLRead === RUN TestACL/ManageAll/AllowACLWrite === RUN TestACL/ManageAll/AllowAgentRead === RUN TestACL/ManageAll/AllowAgentWrite === RUN TestACL/ManageAll/AllowEventRead === RUN TestACL/ManageAll/AllowEventWrite === RUN TestACL/ManageAll/AllowIntentionDefaultAllow === RUN TestACL/ManageAll/AllowIntentionRead === RUN TestACL/ManageAll/AllowIntentionWrite === RUN TestACL/ManageAll/AllowKeyRead === RUN TestACL/ManageAll/AllowKeyringRead === RUN TestACL/ManageAll/AllowKeyringWrite === RUN TestACL/ManageAll/AllowKeyWrite === RUN TestACL/ManageAll/AllowNodeRead === RUN TestACL/ManageAll/AllowNodeWrite === RUN TestACL/ManageAll/AllowOperatorRead === RUN TestACL/ManageAll/AllowOperatorWrite === RUN TestACL/ManageAll/AllowPreparedQueryRead === RUN TestACL/ManageAll/AllowPreparedQueryWrite === RUN TestACL/ManageAll/AllowServiceRead === RUN TestACL/ManageAll/AllowServiceWrite === RUN TestACL/ManageAll/AllowSessionRead === RUN TestACL/ManageAll/AllowSessionWrite === RUN TestACL/ManageAll/AllowSnapshot === RUN TestACL/AgentBasicDefaultDeny === RUN TestACL/AgentBasicDefaultDeny/DefaultReadDenied.Prefix(ro) === RUN TestACL/AgentBasicDefaultDeny/DefaultWriteDenied.Prefix(ro) === RUN TestACL/AgentBasicDefaultDeny/ROReadAllowed.Prefix(root) === RUN TestACL/AgentBasicDefaultDeny/ROWriteDenied.Prefix(root) === RUN TestACL/AgentBasicDefaultDeny/ROSuffixReadAllowed.Prefix(root-ro) === RUN TestACL/AgentBasicDefaultDeny/ROSuffixWriteDenied.Prefix(root-ro) === RUN TestACL/AgentBasicDefaultDeny/RWReadAllowed.Prefix(root-rw) === RUN TestACL/AgentBasicDefaultDeny/RWWriteDenied.Prefix(root-rw) === RUN TestACL/AgentBasicDefaultDeny/RWSuffixReadAllowed.Prefix(root-rw-sub) === RUN TestACL/AgentBasicDefaultDeny/RWSuffixWriteAllowed.Prefix(root-rw-sub) === RUN TestACL/AgentBasicDefaultDeny/DenyReadDenied.Prefix(root-nope) === RUN TestACL/AgentBasicDefaultDeny/DenyWriteDenied.Prefix(root-nope) === RUN TestACL/AgentBasicDefaultDeny/DenySuffixReadDenied.Prefix(root-nope-sub) === RUN TestACL/AgentBasicDefaultDeny/DenySuffixWriteDenied.Prefix(root-nope-sub) === RUN TestACL/AgentBasicDefaultAllow === RUN TestACL/AgentBasicDefaultAllow/DefaultReadDenied.Prefix(ro) === RUN TestACL/AgentBasicDefaultAllow/DefaultWriteDenied.Prefix(ro) === RUN TestACL/AgentBasicDefaultAllow/ROReadAllowed.Prefix(root) === RUN TestACL/AgentBasicDefaultAllow/ROWriteDenied.Prefix(root) === RUN TestACL/AgentBasicDefaultAllow/ROSuffixReadAllowed.Prefix(root-ro) === RUN TestACL/AgentBasicDefaultAllow/ROSuffixWriteDenied.Prefix(root-ro) === RUN TestACL/AgentBasicDefaultAllow/RWReadAllowed.Prefix(root-rw) === RUN TestACL/AgentBasicDefaultAllow/RWWriteDenied.Prefix(root-rw) === RUN TestACL/AgentBasicDefaultAllow/RWSuffixReadAllowed.Prefix(root-rw-sub) === RUN TestACL/AgentBasicDefaultAllow/RWSuffixWriteAllowed.Prefix(root-rw-sub) === RUN TestACL/AgentBasicDefaultAllow/DenyReadDenied.Prefix(root-nope) === RUN TestACL/AgentBasicDefaultAllow/DenyWriteDenied.Prefix(root-nope) === RUN TestACL/AgentBasicDefaultAllow/DenySuffixReadDenied.Prefix(root-nope-sub) === RUN TestACL/AgentBasicDefaultAllow/DenySuffixWriteDenied.Prefix(root-nope-sub) === RUN TestACL/PreparedQueryDefaultAllow === RUN TestACL/PreparedQueryDefaultAllow/ReadAllowed.Prefix(foo) === RUN TestACL/PreparedQueryDefaultAllow/WriteAllowed.Prefix(foo) === RUN TestACL/PreparedQueryDefaultAllow/ReadDenied.Prefix(other) === RUN TestACL/PreparedQueryDefaultAllow/WriteDenied.Prefix(other) === RUN TestACL/AgentNestedDefaultDeny === RUN TestACL/AgentNestedDefaultDeny/DefaultReadDenied.Prefix(nope) === RUN TestACL/AgentNestedDefaultDeny/DefaultWriteDenied.Prefix(nope) === RUN TestACL/AgentNestedDefaultDeny/DenyReadDenied.Prefix(root-nope) === RUN TestACL/AgentNestedDefaultDeny/DenyWriteDenied.Prefix(root-nope) === RUN TestACL/AgentNestedDefaultDeny/ROReadAllowed.Prefix(root-ro) === RUN TestACL/AgentNestedDefaultDeny/ROWriteDenied.Prefix(root-ro) === RUN TestACL/AgentNestedDefaultDeny/RWReadAllowed.Prefix(root-rw) === RUN TestACL/AgentNestedDefaultDeny/RWWriteAllowed.Prefix(root-rw) === RUN TestACL/AgentNestedDefaultDeny/DenySuffixReadDenied.Prefix(root-nope-prefix) === RUN TestACL/AgentNestedDefaultDeny/DenySuffixWriteDenied.Prefix(root-nope-prefix) === RUN TestACL/AgentNestedDefaultDeny/ROSuffixReadAllowed.Prefix(root-ro-prefix) === RUN TestACL/AgentNestedDefaultDeny/ROSuffixWriteDenied.Prefix(root-ro-prefix) === RUN TestACL/AgentNestedDefaultDeny/RWSuffixReadAllowed.Prefix(root-rw-prefix) === RUN TestACL/AgentNestedDefaultDeny/RWSuffixWriteAllowed.Prefix(root-rw-prefix) === RUN TestACL/AgentNestedDefaultDeny/ChildDenyReadDenied.Prefix(child-nope) === RUN TestACL/AgentNestedDefaultDeny/ChildDenyWriteDenied.Prefix(child-nope) === RUN TestACL/AgentNestedDefaultDeny/ChildROReadAllowed.Prefix(child-ro) === RUN TestACL/AgentNestedDefaultDeny/ChildROWriteDenied.Prefix(child-ro) === RUN TestACL/AgentNestedDefaultDeny/ChildRWReadAllowed.Prefix(child-rw) === RUN TestACL/AgentNestedDefaultDeny/ChildRWWriteAllowed.Prefix(child-rw) === RUN TestACL/AgentNestedDefaultDeny/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) === RUN TestACL/AgentNestedDefaultDeny/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) === RUN TestACL/AgentNestedDefaultDeny/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) === RUN TestACL/AgentNestedDefaultDeny/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) === RUN TestACL/AgentNestedDefaultDeny/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) === RUN TestACL/AgentNestedDefaultDeny/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) === RUN TestACL/AgentNestedDefaultDeny/ChildOverrideReadAllowed.Prefix(override) === RUN TestACL/AgentNestedDefaultDeny/ChildOverrideWriteAllowed.Prefix(override) === RUN TestACL/AgentNestedDefaultAllow === RUN TestACL/AgentNestedDefaultAllow/DefaultReadAllowed.Prefix(nope) === RUN TestACL/AgentNestedDefaultAllow/DefaultWriteAllowed.Prefix(nope) === RUN TestACL/AgentNestedDefaultAllow/DenyReadDenied.Prefix(root-nope) === RUN TestACL/AgentNestedDefaultAllow/DenyWriteDenied.Prefix(root-nope) === RUN TestACL/AgentNestedDefaultAllow/ROReadAllowed.Prefix(root-ro) === RUN TestACL/AgentNestedDefaultAllow/ROWriteDenied.Prefix(root-ro) === RUN TestACL/AgentNestedDefaultAllow/RWReadAllowed.Prefix(root-rw) === RUN TestACL/AgentNestedDefaultAllow/RWWriteAllowed.Prefix(root-rw) === RUN TestACL/AgentNestedDefaultAllow/DenySuffixReadDenied.Prefix(root-nope-prefix) === RUN TestACL/AgentNestedDefaultAllow/DenySuffixWriteDenied.Prefix(root-nope-prefix) === RUN TestACL/AgentNestedDefaultAllow/ROSuffixReadAllowed.Prefix(root-ro-prefix) === RUN TestACL/AgentNestedDefaultAllow/ROSuffixWriteDenied.Prefix(root-ro-prefix) === RUN TestACL/AgentNestedDefaultAllow/RWSuffixReadAllowed.Prefix(root-rw-prefix) === RUN TestACL/AgentNestedDefaultAllow/RWSuffixWriteAllowed.Prefix(root-rw-prefix) === RUN TestACL/AgentNestedDefaultAllow/ChildDenyReadDenied.Prefix(child-nope) === RUN TestACL/AgentNestedDefaultAllow/ChildDenyWriteDenied.Prefix(child-nope) === RUN TestACL/AgentNestedDefaultAllow/ChildROReadAllowed.Prefix(child-ro) === RUN TestACL/AgentNestedDefaultAllow/ChildROWriteDenied.Prefix(child-ro) === RUN TestACL/AgentNestedDefaultAllow/ChildRWReadAllowed.Prefix(child-rw) === RUN TestACL/AgentNestedDefaultAllow/ChildRWWriteAllowed.Prefix(child-rw) === RUN TestACL/AgentNestedDefaultAllow/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) === RUN TestACL/AgentNestedDefaultAllow/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) === RUN TestACL/AgentNestedDefaultAllow/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) === RUN TestACL/AgentNestedDefaultAllow/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) === RUN TestACL/AgentNestedDefaultAllow/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) === RUN TestACL/AgentNestedDefaultAllow/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) === RUN TestACL/AgentNestedDefaultAllow/ChildOverrideReadAllowed.Prefix(override) === RUN TestACL/AgentNestedDefaultAllow/ChildOverrideWriteAllowed.Prefix(override) === RUN TestACL/KeyringDefaultAllowPolicyDeny === RUN TestACL/KeyringDefaultAllowPolicyDeny/ReadDenied === RUN TestACL/KeyringDefaultAllowPolicyDeny/WriteDenied === RUN TestACL/KeyringDefaultAllowPolicyRead === RUN TestACL/KeyringDefaultAllowPolicyRead/ReadAllowed === RUN TestACL/KeyringDefaultAllowPolicyRead/WriteDenied === RUN TestACL/KeyringDefaultAllowPolicyWrite === RUN TestACL/KeyringDefaultAllowPolicyWrite/ReadAllowed === RUN TestACL/KeyringDefaultAllowPolicyWrite/WriteAllowed === RUN TestACL/KeyringDefaultAllowPolicyNone === RUN TestACL/KeyringDefaultAllowPolicyNone/ReadAllowed === RUN TestACL/KeyringDefaultAllowPolicyNone/WriteAllowed === RUN TestACL/KeyringDefaultDenyPolicyDeny === RUN TestACL/KeyringDefaultDenyPolicyDeny/ReadDenied === RUN TestACL/KeyringDefaultDenyPolicyDeny/WriteDenied === RUN TestACL/KeyringDefaultDenyPolicyRead === RUN TestACL/KeyringDefaultDenyPolicyRead/ReadAllowed === RUN TestACL/KeyringDefaultDenyPolicyRead/WriteDenied === RUN TestACL/KeyringDefaultDenyPolicyWrite === RUN TestACL/KeyringDefaultDenyPolicyWrite/ReadAllowed === RUN TestACL/KeyringDefaultDenyPolicyWrite/WriteAllowed === RUN TestACL/KeyringDefaultDenyPolicyNone === RUN TestACL/KeyringDefaultDenyPolicyNone/ReadDenied === RUN TestACL/KeyringDefaultDenyPolicyNone/WriteDenied === RUN TestACL/OperatorDefaultAllowPolicyDeny === RUN TestACL/OperatorDefaultAllowPolicyDeny/ReadDenied === RUN TestACL/OperatorDefaultAllowPolicyDeny/WriteDenied === RUN TestACL/OperatorDefaultAllowPolicyRead === RUN TestACL/OperatorDefaultAllowPolicyRead/ReadAllowed === RUN TestACL/OperatorDefaultAllowPolicyRead/WriteDenied === RUN TestACL/OperatorDefaultAllowPolicyWrite === RUN TestACL/OperatorDefaultAllowPolicyWrite/ReadAllowed === RUN TestACL/OperatorDefaultAllowPolicyWrite/WriteAllowed === RUN TestACL/OperatorDefaultAllowPolicyNone === RUN TestACL/OperatorDefaultAllowPolicyNone/ReadAllowed === RUN TestACL/OperatorDefaultAllowPolicyNone/WriteAllowed === RUN TestACL/OperatorDefaultDenyPolicyDeny === RUN TestACL/OperatorDefaultDenyPolicyDeny/ReadDenied === RUN TestACL/OperatorDefaultDenyPolicyDeny/WriteDenied === RUN TestACL/OperatorDefaultDenyPolicyRead === RUN TestACL/OperatorDefaultDenyPolicyRead/ReadAllowed === RUN TestACL/OperatorDefaultDenyPolicyRead/WriteDenied === RUN TestACL/OperatorDefaultDenyPolicyWrite === RUN TestACL/OperatorDefaultDenyPolicyWrite/ReadAllowed === RUN TestACL/OperatorDefaultDenyPolicyWrite/WriteAllowed === RUN TestACL/OperatorDefaultDenyPolicyNone === RUN TestACL/OperatorDefaultDenyPolicyNone/ReadDenied === RUN TestACL/OperatorDefaultDenyPolicyNone/WriteDenied === RUN TestACL/NodeDefaultDeny === RUN TestACL/NodeDefaultDeny/DefaultReadDenied.Prefix(nope) === RUN TestACL/NodeDefaultDeny/DefaultWriteDenied.Prefix(nope) === RUN TestACL/NodeDefaultDeny/DenyReadDenied.Prefix(root-nope) === RUN TestACL/NodeDefaultDeny/DenyWriteDenied.Prefix(root-nope) === RUN TestACL/NodeDefaultDeny/ROReadAllowed.Prefix(root-ro) === RUN TestACL/NodeDefaultDeny/ROWriteDenied.Prefix(root-ro) === RUN TestACL/NodeDefaultDeny/RWReadAllowed.Prefix(root-rw) === RUN TestACL/NodeDefaultDeny/RWWriteAllowed.Prefix(root-rw) === RUN TestACL/NodeDefaultDeny/DenySuffixReadDenied.Prefix(root-nope-prefix) === RUN TestACL/NodeDefaultDeny/DenySuffixWriteDenied.Prefix(root-nope-prefix) === RUN TestACL/NodeDefaultDeny/ROSuffixReadAllowed.Prefix(root-ro-prefix) === RUN TestACL/NodeDefaultDeny/ROSuffixWriteDenied.Prefix(root-ro-prefix) === RUN TestACL/NodeDefaultDeny/RWSuffixReadAllowed.Prefix(root-rw-prefix) === RUN TestACL/NodeDefaultDeny/RWSuffixWriteAllowed.Prefix(root-rw-prefix) === RUN TestACL/NodeDefaultDeny/ChildDenyReadDenied.Prefix(child-nope) === RUN TestACL/NodeDefaultDeny/ChildDenyWriteDenied.Prefix(child-nope) === RUN TestACL/NodeDefaultDeny/ChildROReadAllowed.Prefix(child-ro) === RUN TestACL/NodeDefaultDeny/ChildROWriteDenied.Prefix(child-ro) === RUN TestACL/NodeDefaultDeny/ChildRWReadAllowed.Prefix(child-rw) === RUN TestACL/NodeDefaultDeny/ChildRWWriteAllowed.Prefix(child-rw) === RUN TestACL/NodeDefaultDeny/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) === RUN TestACL/NodeDefaultDeny/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) === RUN TestACL/NodeDefaultDeny/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) === RUN TestACL/NodeDefaultDeny/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) === RUN TestACL/NodeDefaultDeny/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) === RUN TestACL/NodeDefaultDeny/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) === RUN TestACL/NodeDefaultDeny/ChildOverrideReadAllowed.Prefix(override) === RUN TestACL/NodeDefaultDeny/ChildOverrideWriteAllowed.Prefix(override) === RUN TestACL/NodeDefaultAllow === RUN TestACL/NodeDefaultAllow/DefaultReadAllowed.Prefix(nope) === RUN TestACL/NodeDefaultAllow/DefaultWriteAllowed.Prefix(nope) === RUN TestACL/NodeDefaultAllow/DenyReadDenied.Prefix(root-nope) === RUN TestACL/NodeDefaultAllow/DenyWriteDenied.Prefix(root-nope) === RUN TestACL/NodeDefaultAllow/ROReadAllowed.Prefix(root-ro) === RUN TestACL/NodeDefaultAllow/ROWriteDenied.Prefix(root-ro) === RUN TestACL/NodeDefaultAllow/RWReadAllowed.Prefix(root-rw) === RUN TestACL/NodeDefaultAllow/RWWriteAllowed.Prefix(root-rw) === RUN TestACL/NodeDefaultAllow/DenySuffixReadDenied.Prefix(root-nope-prefix) === RUN TestACL/NodeDefaultAllow/DenySuffixWriteDenied.Prefix(root-nope-prefix) === RUN TestACL/NodeDefaultAllow/ROSuffixReadAllowed.Prefix(root-ro-prefix) === RUN TestACL/NodeDefaultAllow/ROSuffixWriteDenied.Prefix(root-ro-prefix) === RUN TestACL/NodeDefaultAllow/RWSuffixReadAllowed.Prefix(root-rw-prefix) === RUN TestACL/NodeDefaultAllow/RWSuffixWriteAllowed.Prefix(root-rw-prefix) === RUN TestACL/NodeDefaultAllow/ChildDenyReadDenied.Prefix(child-nope) === RUN TestACL/NodeDefaultAllow/ChildDenyWriteDenied.Prefix(child-nope) === RUN TestACL/NodeDefaultAllow/ChildROReadAllowed.Prefix(child-ro) === RUN TestACL/NodeDefaultAllow/ChildROWriteDenied.Prefix(child-ro) === RUN TestACL/NodeDefaultAllow/ChildRWReadAllowed.Prefix(child-rw) === RUN TestACL/NodeDefaultAllow/ChildRWWriteAllowed.Prefix(child-rw) === RUN TestACL/NodeDefaultAllow/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) === RUN TestACL/NodeDefaultAllow/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) === RUN TestACL/NodeDefaultAllow/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) === RUN TestACL/NodeDefaultAllow/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) === RUN TestACL/NodeDefaultAllow/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) === RUN TestACL/NodeDefaultAllow/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) === RUN TestACL/NodeDefaultAllow/ChildOverrideReadAllowed.Prefix(override) === RUN TestACL/NodeDefaultAllow/ChildOverrideWriteAllowed.Prefix(override) === RUN TestACL/SessionDefaultDeny === RUN TestACL/SessionDefaultDeny/DefaultReadDenied.Prefix(nope) === RUN TestACL/SessionDefaultDeny/DefaultWriteDenied.Prefix(nope) === RUN TestACL/SessionDefaultDeny/DenyReadDenied.Prefix(root-nope) === RUN TestACL/SessionDefaultDeny/DenyWriteDenied.Prefix(root-nope) === RUN TestACL/SessionDefaultDeny/ROReadAllowed.Prefix(root-ro) === RUN TestACL/SessionDefaultDeny/ROWriteDenied.Prefix(root-ro) === RUN TestACL/SessionDefaultDeny/RWReadAllowed.Prefix(root-rw) === RUN TestACL/SessionDefaultDeny/RWWriteAllowed.Prefix(root-rw) === RUN TestACL/SessionDefaultDeny/DenySuffixReadDenied.Prefix(root-nope-prefix) === RUN TestACL/SessionDefaultDeny/DenySuffixWriteDenied.Prefix(root-nope-prefix) === RUN TestACL/SessionDefaultDeny/ROSuffixReadAllowed.Prefix(root-ro-prefix) === RUN TestACL/SessionDefaultDeny/ROSuffixWriteDenied.Prefix(root-ro-prefix) === RUN TestACL/SessionDefaultDeny/RWSuffixReadAllowed.Prefix(root-rw-prefix) === RUN TestACL/SessionDefaultDeny/RWSuffixWriteAllowed.Prefix(root-rw-prefix) === RUN TestACL/SessionDefaultDeny/ChildDenyReadDenied.Prefix(child-nope) === RUN TestACL/SessionDefaultDeny/ChildDenyWriteDenied.Prefix(child-nope) === RUN TestACL/SessionDefaultDeny/ChildROReadAllowed.Prefix(child-ro) === RUN TestACL/SessionDefaultDeny/ChildROWriteDenied.Prefix(child-ro) === RUN TestACL/SessionDefaultDeny/ChildRWReadAllowed.Prefix(child-rw) === RUN TestACL/SessionDefaultDeny/ChildRWWriteAllowed.Prefix(child-rw) === RUN TestACL/SessionDefaultDeny/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) === RUN TestACL/SessionDefaultDeny/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) === RUN TestACL/SessionDefaultDeny/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) === RUN TestACL/SessionDefaultDeny/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) === RUN TestACL/SessionDefaultDeny/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) === RUN TestACL/SessionDefaultDeny/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) === RUN TestACL/SessionDefaultDeny/ChildOverrideReadAllowed.Prefix(override) === RUN TestACL/SessionDefaultDeny/ChildOverrideWriteAllowed.Prefix(override) === RUN TestACL/SessionDefaultAllow === RUN TestACL/SessionDefaultAllow/DefaultReadAllowed.Prefix(nope) === RUN TestACL/SessionDefaultAllow/DefaultWriteAllowed.Prefix(nope) === RUN TestACL/SessionDefaultAllow/DenyReadDenied.Prefix(root-nope) === RUN TestACL/SessionDefaultAllow/DenyWriteDenied.Prefix(root-nope) === RUN TestACL/SessionDefaultAllow/ROReadAllowed.Prefix(root-ro) === RUN TestACL/SessionDefaultAllow/ROWriteDenied.Prefix(root-ro) === RUN TestACL/SessionDefaultAllow/RWReadAllowed.Prefix(root-rw) === RUN TestACL/SessionDefaultAllow/RWWriteAllowed.Prefix(root-rw) === RUN TestACL/SessionDefaultAllow/DenySuffixReadDenied.Prefix(root-nope-prefix) === RUN TestACL/SessionDefaultAllow/DenySuffixWriteDenied.Prefix(root-nope-prefix) === RUN TestACL/SessionDefaultAllow/ROSuffixReadAllowed.Prefix(root-ro-prefix) === RUN TestACL/SessionDefaultAllow/ROSuffixWriteDenied.Prefix(root-ro-prefix) === RUN TestACL/SessionDefaultAllow/RWSuffixReadAllowed.Prefix(root-rw-prefix) === RUN TestACL/SessionDefaultAllow/RWSuffixWriteAllowed.Prefix(root-rw-prefix) === RUN TestACL/SessionDefaultAllow/ChildDenyReadDenied.Prefix(child-nope) === RUN TestACL/SessionDefaultAllow/ChildDenyWriteDenied.Prefix(child-nope) === RUN TestACL/SessionDefaultAllow/ChildROReadAllowed.Prefix(child-ro) === RUN TestACL/SessionDefaultAllow/ChildROWriteDenied.Prefix(child-ro) === RUN TestACL/SessionDefaultAllow/ChildRWReadAllowed.Prefix(child-rw) === RUN TestACL/SessionDefaultAllow/ChildRWWriteAllowed.Prefix(child-rw) === RUN TestACL/SessionDefaultAllow/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) === RUN TestACL/SessionDefaultAllow/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) === RUN TestACL/SessionDefaultAllow/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) === RUN TestACL/SessionDefaultAllow/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) === RUN TestACL/SessionDefaultAllow/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) === RUN TestACL/SessionDefaultAllow/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) === RUN TestACL/SessionDefaultAllow/ChildOverrideReadAllowed.Prefix(override) === RUN TestACL/SessionDefaultAllow/ChildOverrideWriteAllowed.Prefix(override) === RUN TestACL/Parent === RUN TestACL/Parent/KeyReadDenied.Prefix(other) === RUN TestACL/Parent/KeyWriteDenied.Prefix(other) === RUN TestACL/Parent/KeyWritePrefixDenied.Prefix(other) === RUN TestACL/Parent/KeyReadAllowed.Prefix(foo/test) === RUN TestACL/Parent/KeyWriteAllowed.Prefix(foo/test) === RUN TestACL/Parent/KeyWritePrefixAllowed.Prefix(foo/test) === RUN TestACL/Parent/KeyReadAllowed.Prefix(foo/priv/test) === RUN TestACL/Parent/KeyWriteDenied.Prefix(foo/priv/test) === RUN TestACL/Parent/KeyWritePrefixDenied.Prefix(foo/priv/test) === RUN TestACL/Parent/KeyReadDenied.Prefix(bar/any) === RUN TestACL/Parent/KeyWriteDenied.Prefix(bar/any) === RUN TestACL/Parent/KeyWritePrefixDenied.Prefix(bar/any) === RUN TestACL/Parent/KeyReadAllowed.Prefix(zip/test) === RUN TestACL/Parent/KeyWriteDenied.Prefix(zip/test) === RUN TestACL/Parent/KeyWritePrefixDenied.Prefix(zip/test) === RUN TestACL/Parent/ServiceReadDenied.Prefix(fail) === RUN TestACL/Parent/ServiceWriteDenied.Prefix(fail) === RUN TestACL/Parent/ServiceReadAllowed.Prefix(other) === RUN TestACL/Parent/ServiceWriteAllowed.Prefix(other) === RUN TestACL/Parent/ServiceReadAllowed.Prefix(foo) === RUN TestACL/Parent/ServiceWriteDenied.Prefix(foo) === RUN TestACL/Parent/ServiceReadDenied.Prefix(bar) === RUN TestACL/Parent/ServiceWriteDenied.Prefix(bar) === RUN TestACL/Parent/PreparedQueryReadAllowed.Prefix(foo) === RUN TestACL/Parent/PreparedQueryWriteDenied.Prefix(foo) === RUN TestACL/Parent/PreparedQueryReadAllowed.Prefix(foobar) === RUN TestACL/Parent/PreparedQueryWriteDenied.Prefix(foobar) === RUN TestACL/Parent/PreparedQueryReadDenied.Prefix(bar) === RUN TestACL/Parent/PreparedQueryWriteDenied.Prefix(bar) === RUN TestACL/Parent/PreparedQueryReadDenied.Prefix(barbaz) === RUN TestACL/Parent/PreparedQueryWriteDenied.Prefix(barbaz) === RUN TestACL/Parent/PreparedQueryReadDenied.Prefix(baz) === RUN TestACL/Parent/PreparedQueryWriteDenied.Prefix(baz) === RUN TestACL/Parent/PreparedQueryReadDenied.Prefix(nope) === RUN TestACL/Parent/PreparedQueryWriteDenied.Prefix(nope) === RUN TestACL/Parent/ACLReadDenied === RUN TestACL/Parent/ACLWriteDenied === RUN TestACL/Parent/SnapshotDenied === RUN TestACL/Parent/IntentionDefaultAllowDenied === RUN TestACL/ComplexDefaultAllow === RUN TestACL/ComplexDefaultAllow/KeyReadAllowed.Prefix(other) === RUN TestACL/ComplexDefaultAllow/KeyWriteAllowed.Prefix(other) === RUN TestACL/ComplexDefaultAllow/KeyWritePrefixAllowed.Prefix(other) === RUN TestACL/ComplexDefaultAllow/KeyListAllowed.Prefix(other) === RUN TestACL/ComplexDefaultAllow/KeyReadAllowed.Prefix(foo/test) === RUN TestACL/ComplexDefaultAllow/KeyWriteAllowed.Prefix(foo/test) === RUN TestACL/ComplexDefaultAllow/KeyWritePrefixAllowed.Prefix(foo/test) === RUN TestACL/ComplexDefaultAllow/KeyListAllowed.Prefix(foo/test) === RUN TestACL/ComplexDefaultAllow/KeyReadDenied.Prefix(foo/priv/test) === RUN TestACL/ComplexDefaultAllow/KeyWriteDenied.Prefix(foo/priv/test) === RUN TestACL/ComplexDefaultAllow/KeyWritePrefixDenied.Prefix(foo/priv/test) === RUN TestACL/ComplexDefaultAllow/KeyListDenied.Prefix(foo/priv/test) === RUN TestACL/ComplexDefaultAllow/KeyReadDenied.Prefix(bar/any) === RUN TestACL/ComplexDefaultAllow/KeyWriteDenied.Prefix(bar/any) === RUN TestACL/ComplexDefaultAllow/KeyWritePrefixDenied.Prefix(bar/any) === RUN TestACL/ComplexDefaultAllow/KeyListDenied.Prefix(bar/any) === RUN TestACL/ComplexDefaultAllow/KeyReadAllowed.Prefix(zip/test) === RUN TestACL/ComplexDefaultAllow/KeyWriteDenied.Prefix(zip/test) === RUN TestACL/ComplexDefaultAllow/KeyWritePrefixDenied.Prefix(zip/test) === RUN TestACL/ComplexDefaultAllow/KeyListDenied.Prefix(zip/test) === RUN TestACL/ComplexDefaultAllow/KeyReadAllowed.Prefix(foo/) === RUN TestACL/ComplexDefaultAllow/KeyWriteAllowed.Prefix(foo/) === RUN TestACL/ComplexDefaultAllow/KeyWritePrefixDenied.Prefix(foo/) === RUN TestACL/ComplexDefaultAllow/KeyListAllowed.Prefix(foo/) === RUN TestACL/ComplexDefaultAllow/KeyReadAllowed === RUN TestACL/ComplexDefaultAllow/KeyWriteAllowed === RUN TestACL/ComplexDefaultAllow/KeyWritePrefixDenied === RUN TestACL/ComplexDefaultAllow/KeyListAllowed === RUN TestACL/ComplexDefaultAllow/KeyReadAllowed.Prefix(zap/test) === RUN TestACL/ComplexDefaultAllow/KeyWriteDenied.Prefix(zap/test) === RUN TestACL/ComplexDefaultAllow/KeyWritePrefixDenied.Prefix(zap/test) === RUN TestACL/ComplexDefaultAllow/KeyListAllowed.Prefix(zap/test) === RUN TestACL/ComplexDefaultAllow/IntentionReadAllowed.Prefix(other) === RUN TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(other) === RUN TestACL/ComplexDefaultAllow/IntentionReadAllowed.Prefix(foo) === RUN TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(foo) === RUN TestACL/ComplexDefaultAllow/IntentionReadDenied.Prefix(bar) === RUN TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(bar) === RUN TestACL/ComplexDefaultAllow/IntentionReadAllowed.Prefix(foobar) === RUN TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(foobar) === RUN TestACL/ComplexDefaultAllow/IntentionReadDenied.Prefix(barfo) === RUN TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(barfo) === RUN TestACL/ComplexDefaultAllow/IntentionReadAllowed.Prefix(barfoo) === RUN TestACL/ComplexDefaultAllow/IntentionWriteAllowed.Prefix(barfoo) === RUN TestACL/ComplexDefaultAllow/IntentionReadAllowed.Prefix(barfoo2) === RUN TestACL/ComplexDefaultAllow/IntentionWriteAllowed.Prefix(barfoo2) === RUN TestACL/ComplexDefaultAllow/IntentionReadDenied.Prefix(intbaz) === RUN TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(intbaz) === RUN TestACL/ComplexDefaultAllow/IntentionDefaultAllowAllowed === RUN TestACL/ComplexDefaultAllow/ServiceReadAllowed.Prefix(other) === RUN TestACL/ComplexDefaultAllow/ServiceWriteAllowed.Prefix(other) === RUN TestACL/ComplexDefaultAllow/ServiceReadAllowed.Prefix(foo) === RUN TestACL/ComplexDefaultAllow/ServiceWriteDenied.Prefix(foo) === RUN TestACL/ComplexDefaultAllow/ServiceReadDenied.Prefix(bar) === RUN TestACL/ComplexDefaultAllow/ServiceWriteDenied.Prefix(bar) === RUN TestACL/ComplexDefaultAllow/ServiceReadAllowed.Prefix(foobar) === RUN TestACL/ComplexDefaultAllow/ServiceWriteDenied.Prefix(foobar) === RUN TestACL/ComplexDefaultAllow/ServiceReadDenied.Prefix(barfo) === RUN TestACL/ComplexDefaultAllow/ServiceWriteDenied.Prefix(barfo) === RUN TestACL/ComplexDefaultAllow/ServiceReadAllowed.Prefix(barfoo) === RUN TestACL/ComplexDefaultAllow/ServiceWriteAllowed.Prefix(barfoo) === RUN TestACL/ComplexDefaultAllow/ServiceReadAllowed.Prefix(barfoo2) === RUN TestACL/ComplexDefaultAllow/ServiceWriteAllowed.Prefix(barfoo2) === RUN TestACL/ComplexDefaultAllow/EventReadAllowed.Prefix(foo) === RUN TestACL/ComplexDefaultAllow/EventWriteAllowed.Prefix(foo) === RUN TestACL/ComplexDefaultAllow/EventReadAllowed.Prefix(foobar) === RUN TestACL/ComplexDefaultAllow/EventWriteAllowed.Prefix(foobar) === RUN TestACL/ComplexDefaultAllow/EventReadDenied.Prefix(bar) === RUN TestACL/ComplexDefaultAllow/EventWriteDenied.Prefix(bar) === RUN TestACL/ComplexDefaultAllow/EventReadDenied.Prefix(barbaz) === RUN TestACL/ComplexDefaultAllow/EventWriteDenied.Prefix(barbaz) === RUN TestACL/ComplexDefaultAllow/EventReadAllowed.Prefix(baz) === RUN TestACL/ComplexDefaultAllow/EventWriteDenied.Prefix(baz) === RUN TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(foo) === RUN TestACL/ComplexDefaultAllow/PreparedQueryWriteAllowed.Prefix(foo) === RUN TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(foobar) === RUN TestACL/ComplexDefaultAllow/PreparedQueryWriteAllowed.Prefix(foobar) === RUN TestACL/ComplexDefaultAllow/PreparedQueryReadDenied.Prefix(bar) === RUN TestACL/ComplexDefaultAllow/PreparedQueryWriteDenied.Prefix(bar) === RUN TestACL/ComplexDefaultAllow/PreparedQueryReadDenied.Prefix(barbaz) === RUN TestACL/ComplexDefaultAllow/PreparedQueryWriteDenied.Prefix(barbaz) === RUN TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(baz) === RUN TestACL/ComplexDefaultAllow/PreparedQueryWriteDenied.Prefix(baz) === RUN TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(nope) === RUN TestACL/ComplexDefaultAllow/PreparedQueryWriteDenied.Prefix(nope) === RUN TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(zoo) === RUN TestACL/ComplexDefaultAllow/PreparedQueryWriteAllowed.Prefix(zoo) === RUN TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(zookeeper) === RUN TestACL/ComplexDefaultAllow/PreparedQueryWriteAllowed.Prefix(zookeeper) === RUN TestACL/ExactMatchPrecedence === RUN TestACL/ExactMatchPrecedence/AgentReadPrefixAllowed.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/AgentWritePrefixDenied.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/AgentReadPrefixAllowed.Prefix(for) === RUN TestACL/ExactMatchPrecedence/AgentWritePrefixDenied.Prefix(for) === RUN TestACL/ExactMatchPrecedence/AgentReadAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/AgentWriteAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/AgentReadPrefixAllowed.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/AgentWritePrefixDenied.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/AgentReadPrefixAllowed.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/AgentWritePrefixDenied.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/AgentReadPrefixAllowed.Prefix(food) === RUN TestACL/ExactMatchPrecedence/AgentWritePrefixDenied.Prefix(food) === RUN TestACL/ExactMatchPrecedence/AgentReadDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/AgentWriteDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/KeyReadPrefixAllowed.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/KeyWritePrefixDenied.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/KeyReadPrefixAllowed.Prefix(for) === RUN TestACL/ExactMatchPrecedence/KeyWritePrefixDenied.Prefix(for) === RUN TestACL/ExactMatchPrecedence/KeyReadAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/KeyWriteAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/KeyReadPrefixAllowed.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/KeyWritePrefixDenied.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/KeyReadPrefixAllowed.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/KeyWritePrefixDenied.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/KeyReadPrefixAllowed.Prefix(food) === RUN TestACL/ExactMatchPrecedence/KeyWritePrefixDenied.Prefix(food) === RUN TestACL/ExactMatchPrecedence/KeyReadDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/KeyWriteDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(for) === RUN TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(for) === RUN TestACL/ExactMatchPrecedence/NodeReadAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/NodeWriteAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(food) === RUN TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(food) === RUN TestACL/ExactMatchPrecedence/NodeReadDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/NodeWriteDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/ServiceReadPrefixAllowed.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/ServiceWritePrefixDenied.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/ServiceReadPrefixAllowed.Prefix(for) === RUN TestACL/ExactMatchPrecedence/ServiceWritePrefixDenied.Prefix(for) === RUN TestACL/ExactMatchPrecedence/ServiceReadAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/ServiceWriteAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/ServiceReadPrefixAllowed.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/ServiceWritePrefixDenied.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/ServiceReadPrefixAllowed.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/ServiceWritePrefixDenied.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/ServiceReadPrefixAllowed.Prefix(food) === RUN TestACL/ExactMatchPrecedence/ServiceWritePrefixDenied.Prefix(food) === RUN TestACL/ExactMatchPrecedence/ServiceReadDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/ServiceWriteDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(fo)#01 === RUN TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(fo)#01 === RUN TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(for)#01 === RUN TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(for)#01 === RUN TestACL/ExactMatchPrecedence/NodeReadAllowed.Prefix(foo)#01 === RUN TestACL/ExactMatchPrecedence/NodeWriteAllowed.Prefix(foo)#01 === RUN TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(foot)#01 === RUN TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(foot)#01 === RUN TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(foot2)#01 === RUN TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(foot2)#01 === RUN TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(food)#01 === RUN TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(food)#01 === RUN TestACL/ExactMatchPrecedence/NodeReadDenied.Prefix(football)#01 === RUN TestACL/ExactMatchPrecedence/NodeWriteDenied.Prefix(football)#01 === RUN TestACL/ExactMatchPrecedence/IntentionReadPrefixAllowed.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/IntentionWritePrefixDenied.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/IntentionReadPrefixAllowed.Prefix(for) === RUN TestACL/ExactMatchPrecedence/IntentionWritePrefixDenied.Prefix(for) === RUN TestACL/ExactMatchPrecedence/IntentionReadAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/IntentionWriteAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/IntentionReadPrefixAllowed.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/IntentionWritePrefixDenied.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/IntentionReadPrefixAllowed.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/IntentionWritePrefixDenied.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/IntentionReadPrefixAllowed.Prefix(food) === RUN TestACL/ExactMatchPrecedence/IntentionWritePrefixDenied.Prefix(food) === RUN TestACL/ExactMatchPrecedence/IntentionReadDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/IntentionWriteDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/SessionReadPrefixAllowed.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/SessionWritePrefixDenied.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/SessionReadPrefixAllowed.Prefix(for) === RUN TestACL/ExactMatchPrecedence/SessionWritePrefixDenied.Prefix(for) === RUN TestACL/ExactMatchPrecedence/SessionReadAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/SessionWriteAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/SessionReadPrefixAllowed.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/SessionWritePrefixDenied.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/SessionReadPrefixAllowed.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/SessionWritePrefixDenied.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/SessionReadPrefixAllowed.Prefix(food) === RUN TestACL/ExactMatchPrecedence/SessionWritePrefixDenied.Prefix(food) === RUN TestACL/ExactMatchPrecedence/SessionReadDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/SessionWriteDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/EventReadPrefixAllowed.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/EventWritePrefixDenied.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/EventReadPrefixAllowed.Prefix(for) === RUN TestACL/ExactMatchPrecedence/EventWritePrefixDenied.Prefix(for) === RUN TestACL/ExactMatchPrecedence/EventReadAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/EventWriteAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/EventReadPrefixAllowed.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/EventWritePrefixDenied.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/EventReadPrefixAllowed.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/EventWritePrefixDenied.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/EventReadPrefixAllowed.Prefix(food) === RUN TestACL/ExactMatchPrecedence/EventWritePrefixDenied.Prefix(food) === RUN TestACL/ExactMatchPrecedence/EventReadDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/EventWriteDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/PreparedQueryReadPrefixAllowed.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/PreparedQueryWritePrefixDenied.Prefix(fo) === RUN TestACL/ExactMatchPrecedence/PreparedQueryReadPrefixAllowed.Prefix(for) === RUN TestACL/ExactMatchPrecedence/PreparedQueryWritePrefixDenied.Prefix(for) === RUN TestACL/ExactMatchPrecedence/PreparedQueryReadAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/PreparedQueryWriteAllowed.Prefix(foo) === RUN TestACL/ExactMatchPrecedence/PreparedQueryReadPrefixAllowed.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/PreparedQueryWritePrefixDenied.Prefix(foot) === RUN TestACL/ExactMatchPrecedence/PreparedQueryReadPrefixAllowed.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/PreparedQueryWritePrefixDenied.Prefix(foot2) === RUN TestACL/ExactMatchPrecedence/PreparedQueryReadPrefixAllowed.Prefix(food) === RUN TestACL/ExactMatchPrecedence/PreparedQueryWritePrefixDenied.Prefix(food) === RUN TestACL/ExactMatchPrecedence/PreparedQueryReadDenied.Prefix(football) === RUN TestACL/ExactMatchPrecedence/PreparedQueryWriteDenied.Prefix(football) === RUN TestACL/ACLRead === RUN TestACL/ACLRead/ReadAllowed === RUN TestACL/ACLRead/WriteDenied === RUN TestACL/ACLRead#01 === RUN TestACL/ACLRead#01/ReadAllowed === RUN TestACL/ACLRead#01/WriteAllowed === RUN TestACL/KeyWritePrefixDefaultDeny === RUN TestACL/KeyWritePrefixDefaultDeny/DeniedTopLevelPrefix.Prefix(foo) === RUN TestACL/KeyWritePrefixDefaultDeny/AllowedTopLevelPrefix.Prefix(baz/) === RUN TestACL/KeyWritePrefixDefaultDeny/AllowedPrefixWithNestedWrite.Prefix(foo/) === RUN TestACL/KeyWritePrefixDefaultDeny/DenyPrefixWithNestedRead.Prefix(bar/) === RUN TestACL/KeyWritePrefixDefaultDeny/DenyNoPrefixMatch.Prefix(te) === RUN TestACL/KeyWritePrefixDefaultAllow === RUN TestACL/KeyWritePrefixDefaultAllow/KeyWritePrefixDenied.Prefix(foo) === RUN TestACL/KeyWritePrefixDefaultAllow/KeyWritePrefixAllowed.Prefix(bar) --- PASS: TestACL (0.05s) --- PASS: TestACL/DenyAll (0.00s) --- PASS: TestACL/DenyAll/DenyACLRead (0.00s) --- PASS: TestACL/DenyAll/DenyACLWrite (0.00s) --- PASS: TestACL/DenyAll/DenyAgentRead (0.00s) --- PASS: TestACL/DenyAll/DenyAgentWrite (0.00s) --- PASS: TestACL/DenyAll/DenyEventRead (0.00s) --- PASS: TestACL/DenyAll/DenyEventWrite (0.00s) --- PASS: TestACL/DenyAll/DenyIntentionDefaultAllow (0.00s) --- PASS: TestACL/DenyAll/DenyIntentionRead (0.00s) --- PASS: TestACL/DenyAll/DenyIntentionWrite (0.00s) --- PASS: TestACL/DenyAll/DenyKeyRead (0.00s) --- PASS: TestACL/DenyAll/DenyKeyringRead (0.00s) --- PASS: TestACL/DenyAll/DenyKeyringWrite (0.00s) --- PASS: TestACL/DenyAll/DenyKeyWrite (0.00s) --- PASS: TestACL/DenyAll/DenyNodeRead (0.00s) --- PASS: TestACL/DenyAll/DenyNodeWrite (0.00s) --- PASS: TestACL/DenyAll/DenyOperatorRead (0.00s) --- PASS: TestACL/DenyAll/DenyOperatorWrite (0.00s) --- PASS: TestACL/DenyAll/DenyPreparedQueryRead (0.00s) --- PASS: TestACL/DenyAll/DenyPreparedQueryWrite (0.00s) --- PASS: TestACL/DenyAll/DenyServiceRead (0.00s) --- PASS: TestACL/DenyAll/DenyServiceWrite (0.00s) --- PASS: TestACL/DenyAll/DenySessionRead (0.00s) --- PASS: TestACL/DenyAll/DenySessionWrite (0.00s) --- PASS: TestACL/DenyAll/DenySnapshot (0.00s) --- PASS: TestACL/AllowAll (0.00s) --- PASS: TestACL/AllowAll/DenyACLRead (0.00s) --- PASS: TestACL/AllowAll/DenyACLWrite (0.00s) --- PASS: TestACL/AllowAll/AllowAgentRead (0.00s) --- PASS: TestACL/AllowAll/AllowAgentWrite (0.00s) --- PASS: TestACL/AllowAll/AllowEventRead (0.00s) --- PASS: TestACL/AllowAll/AllowEventWrite (0.00s) --- PASS: TestACL/AllowAll/AllowIntentionDefaultAllow (0.00s) --- PASS: TestACL/AllowAll/AllowIntentionRead (0.00s) --- PASS: TestACL/AllowAll/AllowIntentionWrite (0.00s) --- PASS: TestACL/AllowAll/AllowKeyRead (0.00s) --- PASS: TestACL/AllowAll/AllowKeyringRead (0.00s) --- PASS: TestACL/AllowAll/AllowKeyringWrite (0.00s) --- PASS: TestACL/AllowAll/AllowKeyWrite (0.00s) --- PASS: TestACL/AllowAll/AllowNodeRead (0.00s) --- PASS: TestACL/AllowAll/AllowNodeWrite (0.00s) --- PASS: TestACL/AllowAll/AllowOperatorRead (0.00s) --- PASS: TestACL/AllowAll/AllowOperatorWrite (0.00s) --- PASS: TestACL/AllowAll/AllowPreparedQueryRead (0.00s) --- PASS: TestACL/AllowAll/AllowPreparedQueryWrite (0.00s) --- PASS: TestACL/AllowAll/AllowServiceRead (0.00s) --- PASS: TestACL/AllowAll/AllowServiceWrite (0.00s) --- PASS: TestACL/AllowAll/AllowSessionRead (0.00s) --- PASS: TestACL/AllowAll/AllowSessionWrite (0.00s) --- PASS: TestACL/AllowAll/DenySnapshot (0.00s) --- PASS: TestACL/ManageAll (0.00s) --- PASS: TestACL/ManageAll/AllowACLRead (0.00s) --- PASS: TestACL/ManageAll/AllowACLWrite (0.00s) --- PASS: TestACL/ManageAll/AllowAgentRead (0.00s) --- PASS: TestACL/ManageAll/AllowAgentWrite (0.00s) --- PASS: TestACL/ManageAll/AllowEventRead (0.00s) --- PASS: TestACL/ManageAll/AllowEventWrite (0.00s) --- PASS: TestACL/ManageAll/AllowIntentionDefaultAllow (0.00s) --- PASS: TestACL/ManageAll/AllowIntentionRead (0.00s) --- PASS: TestACL/ManageAll/AllowIntentionWrite (0.00s) --- PASS: TestACL/ManageAll/AllowKeyRead (0.00s) --- PASS: TestACL/ManageAll/AllowKeyringRead (0.00s) --- PASS: TestACL/ManageAll/AllowKeyringWrite (0.00s) --- PASS: TestACL/ManageAll/AllowKeyWrite (0.00s) --- PASS: TestACL/ManageAll/AllowNodeRead (0.00s) --- PASS: TestACL/ManageAll/AllowNodeWrite (0.00s) --- PASS: TestACL/ManageAll/AllowOperatorRead (0.00s) --- PASS: TestACL/ManageAll/AllowOperatorWrite (0.00s) --- PASS: TestACL/ManageAll/AllowPreparedQueryRead (0.00s) --- PASS: TestACL/ManageAll/AllowPreparedQueryWrite (0.00s) --- PASS: TestACL/ManageAll/AllowServiceRead (0.00s) --- PASS: TestACL/ManageAll/AllowServiceWrite (0.00s) --- PASS: TestACL/ManageAll/AllowSessionRead (0.00s) --- PASS: TestACL/ManageAll/AllowSessionWrite (0.00s) --- PASS: TestACL/ManageAll/AllowSnapshot (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/DefaultReadDenied.Prefix(ro) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/DefaultWriteDenied.Prefix(ro) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/ROReadAllowed.Prefix(root) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/ROWriteDenied.Prefix(root) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/ROSuffixReadAllowed.Prefix(root-ro) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/ROSuffixWriteDenied.Prefix(root-ro) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/RWReadAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/RWWriteDenied.Prefix(root-rw) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/RWSuffixReadAllowed.Prefix(root-rw-sub) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/RWSuffixWriteAllowed.Prefix(root-rw-sub) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/DenyReadDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/DenyWriteDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/DenySuffixReadDenied.Prefix(root-nope-sub) (0.00s) --- PASS: TestACL/AgentBasicDefaultDeny/DenySuffixWriteDenied.Prefix(root-nope-sub) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/DefaultReadDenied.Prefix(ro) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/DefaultWriteDenied.Prefix(ro) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/ROReadAllowed.Prefix(root) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/ROWriteDenied.Prefix(root) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/ROSuffixReadAllowed.Prefix(root-ro) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/ROSuffixWriteDenied.Prefix(root-ro) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/RWReadAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/RWWriteDenied.Prefix(root-rw) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/RWSuffixReadAllowed.Prefix(root-rw-sub) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/RWSuffixWriteAllowed.Prefix(root-rw-sub) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/DenyReadDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/DenyWriteDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/DenySuffixReadDenied.Prefix(root-nope-sub) (0.00s) --- PASS: TestACL/AgentBasicDefaultAllow/DenySuffixWriteDenied.Prefix(root-nope-sub) (0.00s) --- PASS: TestACL/PreparedQueryDefaultAllow (0.00s) --- PASS: TestACL/PreparedQueryDefaultAllow/ReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/PreparedQueryDefaultAllow/WriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/PreparedQueryDefaultAllow/ReadDenied.Prefix(other) (0.00s) --- PASS: TestACL/PreparedQueryDefaultAllow/WriteDenied.Prefix(other) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/DefaultReadDenied.Prefix(nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/DefaultWriteDenied.Prefix(nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/DenyReadDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/DenyWriteDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ROReadAllowed.Prefix(root-ro) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ROWriteDenied.Prefix(root-ro) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/RWReadAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/RWWriteAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/DenySuffixReadDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/DenySuffixWriteDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ROSuffixReadAllowed.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ROSuffixWriteDenied.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/RWSuffixReadAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/RWSuffixWriteAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildDenyReadDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildDenyWriteDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildROReadAllowed.Prefix(child-ro) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildROWriteDenied.Prefix(child-ro) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildRWReadAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildRWWriteAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildOverrideReadAllowed.Prefix(override) (0.00s) --- PASS: TestACL/AgentNestedDefaultDeny/ChildOverrideWriteAllowed.Prefix(override) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/DefaultReadAllowed.Prefix(nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/DefaultWriteAllowed.Prefix(nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/DenyReadDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/DenyWriteDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ROReadAllowed.Prefix(root-ro) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ROWriteDenied.Prefix(root-ro) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/RWReadAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/RWWriteAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/DenySuffixReadDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/DenySuffixWriteDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ROSuffixReadAllowed.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ROSuffixWriteDenied.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/RWSuffixReadAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/RWSuffixWriteAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildDenyReadDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildDenyWriteDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildROReadAllowed.Prefix(child-ro) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildROWriteDenied.Prefix(child-ro) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildRWReadAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildRWWriteAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildOverrideReadAllowed.Prefix(override) (0.00s) --- PASS: TestACL/AgentNestedDefaultAllow/ChildOverrideWriteAllowed.Prefix(override) (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyDeny (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyDeny/ReadDenied (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyDeny/WriteDenied (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyRead (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyRead/ReadAllowed (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyRead/WriteDenied (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyWrite (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyWrite/ReadAllowed (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyWrite/WriteAllowed (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyNone (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyNone/ReadAllowed (0.00s) --- PASS: TestACL/KeyringDefaultAllowPolicyNone/WriteAllowed (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyDeny (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyDeny/ReadDenied (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyDeny/WriteDenied (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyRead (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyRead/ReadAllowed (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyRead/WriteDenied (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyWrite (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyWrite/ReadAllowed (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyWrite/WriteAllowed (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyNone (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyNone/ReadDenied (0.00s) --- PASS: TestACL/KeyringDefaultDenyPolicyNone/WriteDenied (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyDeny (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyDeny/ReadDenied (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyDeny/WriteDenied (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyRead (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyRead/ReadAllowed (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyRead/WriteDenied (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyWrite (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyWrite/ReadAllowed (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyWrite/WriteAllowed (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyNone (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyNone/ReadAllowed (0.00s) --- PASS: TestACL/OperatorDefaultAllowPolicyNone/WriteAllowed (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyDeny (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyDeny/ReadDenied (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyDeny/WriteDenied (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyRead (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyRead/ReadAllowed (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyRead/WriteDenied (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyWrite (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyWrite/ReadAllowed (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyWrite/WriteAllowed (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyNone (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyNone/ReadDenied (0.00s) --- PASS: TestACL/OperatorDefaultDenyPolicyNone/WriteDenied (0.00s) --- PASS: TestACL/NodeDefaultDeny (0.00s) --- PASS: TestACL/NodeDefaultDeny/DefaultReadDenied.Prefix(nope) (0.00s) --- PASS: TestACL/NodeDefaultDeny/DefaultWriteDenied.Prefix(nope) (0.00s) --- PASS: TestACL/NodeDefaultDeny/DenyReadDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/NodeDefaultDeny/DenyWriteDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ROReadAllowed.Prefix(root-ro) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ROWriteDenied.Prefix(root-ro) (0.00s) --- PASS: TestACL/NodeDefaultDeny/RWReadAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/NodeDefaultDeny/RWWriteAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/NodeDefaultDeny/DenySuffixReadDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/DenySuffixWriteDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ROSuffixReadAllowed.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ROSuffixWriteDenied.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/RWSuffixReadAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/RWSuffixWriteAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildDenyReadDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildDenyWriteDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildROReadAllowed.Prefix(child-ro) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildROWriteDenied.Prefix(child-ro) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildRWReadAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildRWWriteAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildOverrideReadAllowed.Prefix(override) (0.00s) --- PASS: TestACL/NodeDefaultDeny/ChildOverrideWriteAllowed.Prefix(override) (0.00s) --- PASS: TestACL/NodeDefaultAllow (0.00s) --- PASS: TestACL/NodeDefaultAllow/DefaultReadAllowed.Prefix(nope) (0.00s) --- PASS: TestACL/NodeDefaultAllow/DefaultWriteAllowed.Prefix(nope) (0.00s) --- PASS: TestACL/NodeDefaultAllow/DenyReadDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/NodeDefaultAllow/DenyWriteDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ROReadAllowed.Prefix(root-ro) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ROWriteDenied.Prefix(root-ro) (0.00s) --- PASS: TestACL/NodeDefaultAllow/RWReadAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/NodeDefaultAllow/RWWriteAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/NodeDefaultAllow/DenySuffixReadDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/DenySuffixWriteDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ROSuffixReadAllowed.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ROSuffixWriteDenied.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/RWSuffixReadAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/RWSuffixWriteAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildDenyReadDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildDenyWriteDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildROReadAllowed.Prefix(child-ro) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildROWriteDenied.Prefix(child-ro) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildRWReadAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildRWWriteAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildOverrideReadAllowed.Prefix(override) (0.00s) --- PASS: TestACL/NodeDefaultAllow/ChildOverrideWriteAllowed.Prefix(override) (0.00s) --- PASS: TestACL/SessionDefaultDeny (0.00s) --- PASS: TestACL/SessionDefaultDeny/DefaultReadDenied.Prefix(nope) (0.00s) --- PASS: TestACL/SessionDefaultDeny/DefaultWriteDenied.Prefix(nope) (0.00s) --- PASS: TestACL/SessionDefaultDeny/DenyReadDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/SessionDefaultDeny/DenyWriteDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ROReadAllowed.Prefix(root-ro) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ROWriteDenied.Prefix(root-ro) (0.00s) --- PASS: TestACL/SessionDefaultDeny/RWReadAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/SessionDefaultDeny/RWWriteAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/SessionDefaultDeny/DenySuffixReadDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/DenySuffixWriteDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ROSuffixReadAllowed.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ROSuffixWriteDenied.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/RWSuffixReadAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/RWSuffixWriteAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildDenyReadDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildDenyWriteDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildROReadAllowed.Prefix(child-ro) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildROWriteDenied.Prefix(child-ro) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildRWReadAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildRWWriteAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildOverrideReadAllowed.Prefix(override) (0.00s) --- PASS: TestACL/SessionDefaultDeny/ChildOverrideWriteAllowed.Prefix(override) (0.00s) --- PASS: TestACL/SessionDefaultAllow (0.00s) --- PASS: TestACL/SessionDefaultAllow/DefaultReadAllowed.Prefix(nope) (0.00s) --- PASS: TestACL/SessionDefaultAllow/DefaultWriteAllowed.Prefix(nope) (0.00s) --- PASS: TestACL/SessionDefaultAllow/DenyReadDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/SessionDefaultAllow/DenyWriteDenied.Prefix(root-nope) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ROReadAllowed.Prefix(root-ro) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ROWriteDenied.Prefix(root-ro) (0.00s) --- PASS: TestACL/SessionDefaultAllow/RWReadAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/SessionDefaultAllow/RWWriteAllowed.Prefix(root-rw) (0.00s) --- PASS: TestACL/SessionDefaultAllow/DenySuffixReadDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/DenySuffixWriteDenied.Prefix(root-nope-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ROSuffixReadAllowed.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ROSuffixWriteDenied.Prefix(root-ro-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/RWSuffixReadAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/RWSuffixWriteAllowed.Prefix(root-rw-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildDenyReadDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildDenyWriteDenied.Prefix(child-nope) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildROReadAllowed.Prefix(child-ro) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildROWriteDenied.Prefix(child-ro) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildRWReadAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildRWWriteAllowed.Prefix(child-rw) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildDenySuffixReadDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildDenySuffixWriteDenied.Prefix(child-nope-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildROSuffixReadAllowed.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildROSuffixWriteDenied.Prefix(child-ro-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildRWSuffixReadAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildRWSuffixWriteAllowed.Prefix(child-rw-prefix) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildOverrideReadAllowed.Prefix(override) (0.00s) --- PASS: TestACL/SessionDefaultAllow/ChildOverrideWriteAllowed.Prefix(override) (0.00s) --- PASS: TestACL/Parent (0.00s) --- PASS: TestACL/Parent/KeyReadDenied.Prefix(other) (0.00s) --- PASS: TestACL/Parent/KeyWriteDenied.Prefix(other) (0.00s) --- PASS: TestACL/Parent/KeyWritePrefixDenied.Prefix(other) (0.00s) --- PASS: TestACL/Parent/KeyReadAllowed.Prefix(foo/test) (0.00s) --- PASS: TestACL/Parent/KeyWriteAllowed.Prefix(foo/test) (0.00s) --- PASS: TestACL/Parent/KeyWritePrefixAllowed.Prefix(foo/test) (0.00s) --- PASS: TestACL/Parent/KeyReadAllowed.Prefix(foo/priv/test) (0.00s) --- PASS: TestACL/Parent/KeyWriteDenied.Prefix(foo/priv/test) (0.00s) --- PASS: TestACL/Parent/KeyWritePrefixDenied.Prefix(foo/priv/test) (0.00s) --- PASS: TestACL/Parent/KeyReadDenied.Prefix(bar/any) (0.00s) --- PASS: TestACL/Parent/KeyWriteDenied.Prefix(bar/any) (0.00s) --- PASS: TestACL/Parent/KeyWritePrefixDenied.Prefix(bar/any) (0.00s) --- PASS: TestACL/Parent/KeyReadAllowed.Prefix(zip/test) (0.00s) --- PASS: TestACL/Parent/KeyWriteDenied.Prefix(zip/test) (0.00s) --- PASS: TestACL/Parent/KeyWritePrefixDenied.Prefix(zip/test) (0.00s) --- PASS: TestACL/Parent/ServiceReadDenied.Prefix(fail) (0.00s) --- PASS: TestACL/Parent/ServiceWriteDenied.Prefix(fail) (0.00s) --- PASS: TestACL/Parent/ServiceReadAllowed.Prefix(other) (0.00s) --- PASS: TestACL/Parent/ServiceWriteAllowed.Prefix(other) (0.00s) --- PASS: TestACL/Parent/ServiceReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/Parent/ServiceWriteDenied.Prefix(foo) (0.00s) --- PASS: TestACL/Parent/ServiceReadDenied.Prefix(bar) (0.00s) --- PASS: TestACL/Parent/ServiceWriteDenied.Prefix(bar) (0.00s) --- PASS: TestACL/Parent/PreparedQueryReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/Parent/PreparedQueryWriteDenied.Prefix(foo) (0.00s) --- PASS: TestACL/Parent/PreparedQueryReadAllowed.Prefix(foobar) (0.00s) --- PASS: TestACL/Parent/PreparedQueryWriteDenied.Prefix(foobar) (0.00s) --- PASS: TestACL/Parent/PreparedQueryReadDenied.Prefix(bar) (0.00s) --- PASS: TestACL/Parent/PreparedQueryWriteDenied.Prefix(bar) (0.00s) --- PASS: TestACL/Parent/PreparedQueryReadDenied.Prefix(barbaz) (0.00s) --- PASS: TestACL/Parent/PreparedQueryWriteDenied.Prefix(barbaz) (0.00s) --- PASS: TestACL/Parent/PreparedQueryReadDenied.Prefix(baz) (0.00s) --- PASS: TestACL/Parent/PreparedQueryWriteDenied.Prefix(baz) (0.00s) --- PASS: TestACL/Parent/PreparedQueryReadDenied.Prefix(nope) (0.00s) --- PASS: TestACL/Parent/PreparedQueryWriteDenied.Prefix(nope) (0.00s) --- PASS: TestACL/Parent/ACLReadDenied (0.00s) --- PASS: TestACL/Parent/ACLWriteDenied (0.00s) --- PASS: TestACL/Parent/SnapshotDenied (0.00s) --- PASS: TestACL/Parent/IntentionDefaultAllowDenied (0.00s) --- PASS: TestACL/ComplexDefaultAllow (0.01s) --- PASS: TestACL/ComplexDefaultAllow/KeyReadAllowed.Prefix(other) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWriteAllowed.Prefix(other) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWritePrefixAllowed.Prefix(other) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyListAllowed.Prefix(other) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyReadAllowed.Prefix(foo/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWriteAllowed.Prefix(foo/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWritePrefixAllowed.Prefix(foo/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyListAllowed.Prefix(foo/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyReadDenied.Prefix(foo/priv/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWriteDenied.Prefix(foo/priv/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWritePrefixDenied.Prefix(foo/priv/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyListDenied.Prefix(foo/priv/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyReadDenied.Prefix(bar/any) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWriteDenied.Prefix(bar/any) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWritePrefixDenied.Prefix(bar/any) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyListDenied.Prefix(bar/any) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyReadAllowed.Prefix(zip/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWriteDenied.Prefix(zip/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWritePrefixDenied.Prefix(zip/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyListDenied.Prefix(zip/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyReadAllowed.Prefix(foo/) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWriteAllowed.Prefix(foo/) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWritePrefixDenied.Prefix(foo/) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyListAllowed.Prefix(foo/) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyReadAllowed (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWriteAllowed (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWritePrefixDenied (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyListAllowed (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyReadAllowed.Prefix(zap/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWriteDenied.Prefix(zap/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyWritePrefixDenied.Prefix(zap/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/KeyListAllowed.Prefix(zap/test) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionReadAllowed.Prefix(other) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(other) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(foo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionReadDenied.Prefix(bar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(bar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionReadAllowed.Prefix(foobar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(foobar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionReadDenied.Prefix(barfo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(barfo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionReadAllowed.Prefix(barfoo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionWriteAllowed.Prefix(barfoo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionReadAllowed.Prefix(barfoo2) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionWriteAllowed.Prefix(barfoo2) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionReadDenied.Prefix(intbaz) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionWriteDenied.Prefix(intbaz) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/IntentionDefaultAllowAllowed (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceReadAllowed.Prefix(other) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceWriteAllowed.Prefix(other) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceWriteDenied.Prefix(foo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceReadDenied.Prefix(bar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceWriteDenied.Prefix(bar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceReadAllowed.Prefix(foobar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceWriteDenied.Prefix(foobar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceReadDenied.Prefix(barfo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceWriteDenied.Prefix(barfo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceReadAllowed.Prefix(barfoo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceWriteAllowed.Prefix(barfoo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceReadAllowed.Prefix(barfoo2) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/ServiceWriteAllowed.Prefix(barfoo2) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/EventReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/EventWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/EventReadAllowed.Prefix(foobar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/EventWriteAllowed.Prefix(foobar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/EventReadDenied.Prefix(bar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/EventWriteDenied.Prefix(bar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/EventReadDenied.Prefix(barbaz) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/EventWriteDenied.Prefix(barbaz) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/EventReadAllowed.Prefix(baz) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/EventWriteDenied.Prefix(baz) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(foobar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryWriteAllowed.Prefix(foobar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryReadDenied.Prefix(bar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryWriteDenied.Prefix(bar) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryReadDenied.Prefix(barbaz) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryWriteDenied.Prefix(barbaz) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(baz) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryWriteDenied.Prefix(baz) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(nope) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryWriteDenied.Prefix(nope) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(zoo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryWriteAllowed.Prefix(zoo) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryReadAllowed.Prefix(zookeeper) (0.00s) --- PASS: TestACL/ComplexDefaultAllow/PreparedQueryWriteAllowed.Prefix(zookeeper) (0.00s) --- PASS: TestACL/ExactMatchPrecedence (0.01s) --- PASS: TestACL/ExactMatchPrecedence/AgentReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentReadDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/AgentWriteDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyReadDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/KeyWriteDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWriteDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceReadDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/ServiceWriteDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(fo)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(fo)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(for)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(for)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadAllowed.Prefix(foo)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWriteAllowed.Prefix(foo)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(foot)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(foot)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(foot2)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(foot2)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadPrefixAllowed.Prefix(food)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWritePrefixDenied.Prefix(food)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeReadDenied.Prefix(football)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/NodeWriteDenied.Prefix(football)#01 (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionReadDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/IntentionWriteDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionReadDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/SessionWriteDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventReadDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/EventWriteDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryReadAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryReadDenied.Prefix(football) (0.00s) --- PASS: TestACL/ExactMatchPrecedence/PreparedQueryWriteDenied.Prefix(football) (0.00s) --- PASS: TestACL/ACLRead (0.00s) --- PASS: TestACL/ACLRead/ReadAllowed (0.00s) --- PASS: TestACL/ACLRead/WriteDenied (0.00s) --- PASS: TestACL/ACLRead#01 (0.00s) --- PASS: TestACL/ACLRead#01/ReadAllowed (0.00s) --- PASS: TestACL/ACLRead#01/WriteAllowed (0.00s) --- PASS: TestACL/KeyWritePrefixDefaultDeny (0.00s) --- PASS: TestACL/KeyWritePrefixDefaultDeny/DeniedTopLevelPrefix.Prefix(foo) (0.00s) --- PASS: TestACL/KeyWritePrefixDefaultDeny/AllowedTopLevelPrefix.Prefix(baz/) (0.00s) --- PASS: TestACL/KeyWritePrefixDefaultDeny/AllowedPrefixWithNestedWrite.Prefix(foo/) (0.00s) --- PASS: TestACL/KeyWritePrefixDefaultDeny/DenyPrefixWithNestedRead.Prefix(bar/) (0.00s) --- PASS: TestACL/KeyWritePrefixDefaultDeny/DenyNoPrefixMatch.Prefix(te) (0.00s) --- PASS: TestACL/KeyWritePrefixDefaultAllow (0.00s) --- PASS: TestACL/KeyWritePrefixDefaultAllow/KeyWritePrefixDenied.Prefix(foo) (0.00s) --- PASS: TestACL/KeyWritePrefixDefaultAllow/KeyWritePrefixAllowed.Prefix(bar) (0.00s) === RUN TestRootAuthorizer --- PASS: TestRootAuthorizer (0.00s) === RUN TestACLEnforce === RUN TestACLEnforce/RuleNoneRequireRead === RUN TestACLEnforce/RuleNoneRequireWrite === RUN TestACLEnforce/RuleNoneRequireList === RUN TestACLEnforce/RuleReadRequireRead === RUN TestACLEnforce/RuleReadRequireWrite === RUN TestACLEnforce/RuleReadRequireList === RUN TestACLEnforce/RuleListRequireRead === RUN TestACLEnforce/RuleListRequireWrite === RUN TestACLEnforce/RuleListRequireList === RUN TestACLEnforce/RuleWritetRequireRead === RUN TestACLEnforce/RuleWritetRequireWrite === RUN TestACLEnforce/RuleWritetRequireList === RUN TestACLEnforce/RuleDenyRequireRead === RUN TestACLEnforce/RuleDenyRequireWrite === RUN TestACLEnforce/RuleDenyRequireList --- PASS: TestACLEnforce (0.00s) --- PASS: TestACLEnforce/RuleNoneRequireRead (0.00s) --- PASS: TestACLEnforce/RuleNoneRequireWrite (0.00s) --- PASS: TestACLEnforce/RuleNoneRequireList (0.00s) --- PASS: TestACLEnforce/RuleReadRequireRead (0.00s) --- PASS: TestACLEnforce/RuleReadRequireWrite (0.00s) --- PASS: TestACLEnforce/RuleReadRequireList (0.00s) --- PASS: TestACLEnforce/RuleListRequireRead (0.00s) --- PASS: TestACLEnforce/RuleListRequireWrite (0.00s) --- PASS: TestACLEnforce/RuleListRequireList (0.00s) --- PASS: TestACLEnforce/RuleWritetRequireRead (0.00s) --- PASS: TestACLEnforce/RuleWritetRequireWrite (0.00s) --- PASS: TestACLEnforce/RuleWritetRequireList (0.00s) --- PASS: TestACLEnforce/RuleDenyRequireRead (0.00s) --- PASS: TestACLEnforce/RuleDenyRequireWrite (0.00s) --- PASS: TestACLEnforce/RuleDenyRequireList (0.00s) === RUN TestACL_Enforce === PAUSE TestACL_Enforce === RUN TestChainedAuthorizer === PAUSE TestChainedAuthorizer === RUN TestPolicyAuthorizer === PAUSE TestPolicyAuthorizer === RUN TestAnyAllowed === PAUSE TestAnyAllowed === RUN TestAllAllowed === PAUSE TestAllAllowed === RUN TestPolicySourceParse === RUN TestPolicySourceParse/Legacy_Basic === RUN TestPolicySourceParse/Legacy_(JSON) === RUN TestPolicySourceParse/Service_No_Intentions_(Legacy) === RUN TestPolicySourceParse/Service_Intentions_(Legacy) === RUN TestPolicySourceParse/Service_Intention:_invalid_value_(Legacy) === RUN TestPolicySourceParse/Bad_Policy_-_ACL === RUN TestPolicySourceParse/Bad_Policy_-_Agent === RUN TestPolicySourceParse/Bad_Policy_-_Agent_Prefix === RUN TestPolicySourceParse/Bad_Policy_-_Key === RUN TestPolicySourceParse/Bad_Policy_-_Key_Prefix === RUN TestPolicySourceParse/Bad_Policy_-_Node === RUN TestPolicySourceParse/Bad_Policy_-_Node_Prefix === RUN TestPolicySourceParse/Bad_Policy_-_Service === RUN TestPolicySourceParse/Bad_Policy_-_Service_Prefix === RUN TestPolicySourceParse/Bad_Policy_-_Session === RUN TestPolicySourceParse/Bad_Policy_-_Session_Prefix === RUN TestPolicySourceParse/Bad_Policy_-_Event === RUN TestPolicySourceParse/Bad_Policy_-_Event_Prefix === RUN TestPolicySourceParse/Bad_Policy_-_Prepared_Query === RUN TestPolicySourceParse/Bad_Policy_-_Prepared_Query_Prefix === RUN TestPolicySourceParse/Bad_Policy_-_Keyring === RUN TestPolicySourceParse/Bad_Policy_-_Operator === RUN TestPolicySourceParse/Keyring_Empty === RUN TestPolicySourceParse/Operator_Empty --- PASS: TestPolicySourceParse (0.01s) --- PASS: TestPolicySourceParse/Legacy_Basic (0.00s) --- PASS: TestPolicySourceParse/Legacy_(JSON) (0.00s) --- PASS: TestPolicySourceParse/Service_No_Intentions_(Legacy) (0.00s) --- PASS: TestPolicySourceParse/Service_Intentions_(Legacy) (0.00s) --- PASS: TestPolicySourceParse/Service_Intention:_invalid_value_(Legacy) (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_ACL (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Agent (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Agent_Prefix (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Key (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Key_Prefix (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Node (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Node_Prefix (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Service (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Service_Prefix (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Session (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Session_Prefix (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Event (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Event_Prefix (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Prepared_Query (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Prepared_Query_Prefix (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Keyring (0.00s) --- PASS: TestPolicySourceParse/Bad_Policy_-_Operator (0.00s) --- PASS: TestPolicySourceParse/Keyring_Empty (0.00s) --- PASS: TestPolicySourceParse/Operator_Empty (0.00s) === RUN TestMergePolicies === RUN TestMergePolicies/Agents === RUN TestMergePolicies/Events === RUN TestMergePolicies/Node === RUN TestMergePolicies/Keys === RUN TestMergePolicies/Services === RUN TestMergePolicies/Sessions === RUN TestMergePolicies/Prepared_Queries === RUN TestMergePolicies/Write_Precedence === RUN TestMergePolicies/Deny_Precedence === RUN TestMergePolicies/Read_Precedence --- PASS: TestMergePolicies (0.01s) --- PASS: TestMergePolicies/Agents (0.00s) --- PASS: TestMergePolicies/Events (0.00s) --- PASS: TestMergePolicies/Node (0.00s) --- PASS: TestMergePolicies/Keys (0.01s) --- PASS: TestMergePolicies/Services (0.00s) --- PASS: TestMergePolicies/Sessions (0.00s) --- PASS: TestMergePolicies/Prepared_Queries (0.00s) --- PASS: TestMergePolicies/Write_Precedence (0.00s) --- PASS: TestMergePolicies/Deny_Precedence (0.00s) --- PASS: TestMergePolicies/Read_Precedence (0.00s) === RUN TestRulesTranslate --- PASS: TestRulesTranslate (0.00s) === RUN TestRulesTranslate_GH5493 --- PASS: TestRulesTranslate_GH5493 (0.00s) === RUN TestPrecedence === RUN TestPrecedence/Deny_Over_Write === RUN TestPrecedence/Deny_Over_List === RUN TestPrecedence/Deny_Over_Read === RUN TestPrecedence/Deny_Over_Unknown === RUN TestPrecedence/Write_Over_List === RUN TestPrecedence/Write_Over_Read === RUN TestPrecedence/Write_Over_Unknown === RUN TestPrecedence/List_Over_Read === RUN TestPrecedence/List_Over_Unknown === RUN TestPrecedence/Read_Over_Unknown === RUN TestPrecedence/Write_Over_Deny === RUN TestPrecedence/List_Over_Deny === RUN TestPrecedence/Read_Over_Deny === RUN TestPrecedence/Deny_Over_Unknown#01 === RUN TestPrecedence/List_Over_Write === RUN TestPrecedence/Read_Over_Write === RUN TestPrecedence/Unknown_Over_Write === RUN TestPrecedence/Read_Over_List === RUN TestPrecedence/Unknown_Over_List === RUN TestPrecedence/Unknown_Over_Read --- PASS: TestPrecedence (0.00s) --- PASS: TestPrecedence/Deny_Over_Write (0.00s) --- PASS: TestPrecedence/Deny_Over_List (0.00s) --- PASS: TestPrecedence/Deny_Over_Read (0.00s) --- PASS: TestPrecedence/Deny_Over_Unknown (0.00s) --- PASS: TestPrecedence/Write_Over_List (0.00s) --- PASS: TestPrecedence/Write_Over_Read (0.00s) --- PASS: TestPrecedence/Write_Over_Unknown (0.00s) --- PASS: TestPrecedence/List_Over_Read (0.00s) --- PASS: TestPrecedence/List_Over_Unknown (0.00s) --- PASS: TestPrecedence/Read_Over_Unknown (0.00s) --- PASS: TestPrecedence/Write_Over_Deny (0.00s) --- PASS: TestPrecedence/List_Over_Deny (0.00s) --- PASS: TestPrecedence/Read_Over_Deny (0.00s) --- PASS: TestPrecedence/Deny_Over_Unknown#01 (0.00s) --- PASS: TestPrecedence/List_Over_Write (0.00s) --- PASS: TestPrecedence/Read_Over_Write (0.00s) --- PASS: TestPrecedence/Unknown_Over_Write (0.00s) --- PASS: TestPrecedence/Read_Over_List (0.00s) --- PASS: TestPrecedence/Unknown_Over_List (0.00s) --- PASS: TestPrecedence/Unknown_Over_Read (0.00s) === RUN TestStaticAuthorizer === PAUSE TestStaticAuthorizer === CONT TestACL_Enforce === RUN TestACL_Enforce/acl/read/Deny === CONT TestAllAllowed === RUN TestAllAllowed/no-rules-default === CONT TestStaticAuthorizer === RUN TestStaticAuthorizer/AllowAll === RUN TestAllAllowed/prefix-write-allowed === CONT TestPolicyAuthorizer === CONT TestACL_Enforce/acl/read/Deny authorizer_test.go:618: PASS: ACLRead(*acl.AuthorizerContext) === PAUSE TestStaticAuthorizer/AllowAll === RUN TestStaticAuthorizer/DenyAll === RUN TestAllAllowed/prefix-deny === RUN TestAllAllowed/prefix-allow-other-write-prefix === RUN TestAllAllowed/prefix-allow-other-write-exact === RUN TestAllAllowed/prefix-allow-other-list-prefix === RUN TestPolicyAuthorizer/Defaults === PAUSE TestPolicyAuthorizer/Defaults === RUN TestAllAllowed/prefix-allow-other-list-exact === RUN TestPolicyAuthorizer/Prefer_Exact_Matches === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches === RUN TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied === RUN TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed === PAUSE TestStaticAuthorizer/DenyAll === RUN TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed === RUN TestStaticAuthorizer/ManageAll === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed === RUN TestPolicyAuthorizer/Intention_Wildcards_-_all_default === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_all_default === RUN TestPolicyAuthorizer/Intention_Wildcards_-_any_default === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_any_default === RUN TestAllAllowed/prefix-allow-other-read-exact === CONT TestAnyAllowed === RUN TestAllAllowed/prefix-list-allowed === RUN TestAnyAllowed/prefix-deny-other-list-exact === PAUSE TestStaticAuthorizer/ManageAll === RUN TestACL_Enforce/acl/read/Allow === CONT TestChainedAuthorizer === RUN TestAllAllowed/prefix-read-allowed === RUN TestChainedAuthorizer/No_Authorizers === PAUSE TestChainedAuthorizer/No_Authorizers === RUN TestChainedAuthorizer/Authorizer_Defaults === RUN TestAllAllowed/prefix-allow-other-read-prefix === PAUSE TestChainedAuthorizer/Authorizer_Defaults === RUN TestChainedAuthorizer/Authorizer_No_Defaults === PAUSE TestChainedAuthorizer/Authorizer_No_Defaults === RUN TestChainedAuthorizer/First_Found === PAUSE TestChainedAuthorizer/First_Found === CONT TestPolicyAuthorizer/Defaults === RUN TestAllAllowed/prefix-allow-other-deny-prefix === RUN TestPolicyAuthorizer/Defaults/DefaultACLRead.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultACLRead.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultACLWrite.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultACLWrite.Prefix(foo) === RUN TestAllAllowed/prefix-allow-other-deny-exact === RUN TestPolicyAuthorizer/Defaults/DefaultAgentRead.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultAgentRead.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultAgentWrite.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultAgentWrite.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultEventRead.Prefix(foo) === CONT TestACL_Enforce/acl/read/Allow authorizer_test.go:618: PASS: ACLRead(*acl.AuthorizerContext) --- PASS: TestAllAllowed (0.00s) --- PASS: TestAllAllowed/no-rules-default (0.00s) --- PASS: TestAllAllowed/prefix-write-allowed (0.00s) --- PASS: TestAllAllowed/prefix-deny (0.00s) --- PASS: TestAllAllowed/prefix-allow-other-write-prefix (0.00s) --- PASS: TestAllAllowed/prefix-allow-other-write-exact (0.00s) --- PASS: TestAllAllowed/prefix-allow-other-list-prefix (0.00s) --- PASS: TestAllAllowed/prefix-allow-other-list-exact (0.00s) --- PASS: TestAllAllowed/prefix-allow-other-read-exact (0.00s) --- PASS: TestAllAllowed/prefix-list-allowed (0.00s) --- PASS: TestAllAllowed/prefix-read-allowed (0.00s) --- PASS: TestAllAllowed/prefix-allow-other-read-prefix (0.00s) --- PASS: TestAllAllowed/prefix-allow-other-deny-prefix (0.00s) --- PASS: TestAllAllowed/prefix-allow-other-deny-exact (0.00s) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed === RUN TestACL_Enforce/acl/write/Deny === PAUSE TestPolicyAuthorizer/Defaults/DefaultEventRead.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultEventWrite.Prefix(foo) === RUN TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed/AnyAllowed.Prefix(*) === PAUSE TestPolicyAuthorizer/Defaults/DefaultEventWrite.Prefix(foo) === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed/AnyAllowed.Prefix(*) === RUN TestPolicyAuthorizer/Defaults/DefaultIntentionDefaultAllow.Prefix(foo) === RUN TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed/AllAllowed.Prefix(*) === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed/AllAllowed.Prefix(*) === PAUSE TestPolicyAuthorizer/Defaults/DefaultIntentionDefaultAllow.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultIntentionRead.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultIntentionRead.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultIntentionWrite.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultIntentionWrite.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultKeyRead.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultKeyRead.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultKeyList.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultKeyList.Prefix(foo) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied === RUN TestPolicyAuthorizer/Defaults/DefaultKeyringRead.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultKeyringRead.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultKeyringWrite.Prefix(foo) === RUN TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied/AnyAllowed.Prefix(*) === PAUSE TestPolicyAuthorizer/Defaults/DefaultKeyringWrite.Prefix(foo) === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied/AnyAllowed.Prefix(*) === CONT TestACL_Enforce/acl/write/Deny authorizer_test.go:618: PASS: ACLWrite(*acl.AuthorizerContext) === RUN TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied/AllDenied.Prefix(*) === RUN TestACL_Enforce/acl/write/Allow === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied/AllDenied.Prefix(*) === RUN TestPolicyAuthorizer/Defaults/DefaultKeyWrite.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultKeyWrite.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultKeyWritePrefix.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultKeyWritePrefix.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultNodeRead.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultNodeRead.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultNodeWrite.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultNodeWrite.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultOperatorRead.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultOperatorRead.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultOperatorWrite.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultOperatorWrite.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultPreparedQueryRead.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultPreparedQueryRead.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultPreparedQueryWrite.Prefix(foo) === CONT TestACL_Enforce/acl/write/Allow authorizer_test.go:618: PASS: ACLWrite(*acl.AuthorizerContext) === PAUSE TestPolicyAuthorizer/Defaults/DefaultPreparedQueryWrite.Prefix(foo) === RUN TestACL_Enforce/acl/list/Deny === RUN TestPolicyAuthorizer/Defaults/DefaultServiceRead.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultServiceRead.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultServiceWrite.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultServiceWrite.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultSessionRead.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultSessionRead.Prefix(foo) === RUN TestACL_Enforce/operator/read/Deny === RUN TestPolicyAuthorizer/Defaults/DefaultSessionWrite.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultSessionWrite.Prefix(foo) === RUN TestPolicyAuthorizer/Defaults/DefaultSnapshot.Prefix(foo) === PAUSE TestPolicyAuthorizer/Defaults/DefaultSnapshot.Prefix(foo) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed === RUN TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed/AnyAllowed.Prefix(*) === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed/AnyAllowed.Prefix(*) === RUN TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed/AllDenied.Prefix(*) === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed/AllDenied.Prefix(*) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches === CONT TestACL_Enforce/operator/read/Deny authorizer_test.go:618: PASS: OperatorRead(*acl.AuthorizerContext) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_any_default === RUN TestACL_Enforce/operator/read/Allow === RUN TestPolicyAuthorizer/Intention_Wildcards_-_any_default/AnyDefault.Prefix(*) === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_any_default/AnyDefault.Prefix(*) === RUN TestPolicyAuthorizer/Intention_Wildcards_-_any_default/AllDenied.Prefix(*) === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_any_default/AllDenied.Prefix(*) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_all_default === RUN TestPolicyAuthorizer/Intention_Wildcards_-_all_default/AnyAllowed.Prefix(*) === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_all_default/AnyAllowed.Prefix(*) === RUN TestPolicyAuthorizer/Intention_Wildcards_-_all_default/AllDefault.Prefix(*) === PAUSE TestPolicyAuthorizer/Intention_Wildcards_-_all_default/AllDefault.Prefix(*) === CONT TestStaticAuthorizer/AllowAll === CONT TestACL_Enforce/operator/read/Allow authorizer_test.go:618: PASS: OperatorRead(*acl.AuthorizerContext) === RUN TestACL_Enforce/operator/write/Deny authorizer_test.go:618: PASS: OperatorWrite(*acl.AuthorizerContext) === RUN TestACL_Enforce/operator/write/Allow === CONT TestStaticAuthorizer/ManageAll === CONT TestACL_Enforce/operator/write/Allow authorizer_test.go:618: PASS: OperatorWrite(*acl.AuthorizerContext) === RUN TestACL_Enforce/operator/list/Deny === RUN TestACL_Enforce/keyring/read/Deny === CONT TestChainedAuthorizer/No_Authorizers === CONT TestACL_Enforce/keyring/read/Deny authorizer_test.go:618: PASS: KeyringRead(*acl.AuthorizerContext) === RUN TestACL_Enforce/keyring/read/Allow authorizer_test.go:618: PASS: KeyringRead(*acl.AuthorizerContext) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed/AnyAllowed.Prefix(*) === RUN TestACL_Enforce/keyring/write/Deny === CONT TestStaticAuthorizer/DenyAll === CONT TestACL_Enforce/keyring/write/Deny authorizer_test.go:618: PASS: KeyringWrite(*acl.AuthorizerContext) === RUN TestACL_Enforce/keyring/write/Allow --- PASS: TestStaticAuthorizer (0.00s) --- PASS: TestStaticAuthorizer/AllowAll (0.00s) --- PASS: TestStaticAuthorizer/ManageAll (0.00s) --- PASS: TestStaticAuthorizer/DenyAll (0.00s) === CONT TestChainedAuthorizer/Authorizer_No_Defaults === CONT TestACL_Enforce/keyring/write/Allow authorizer_test.go:618: PASS: KeyringWrite(*acl.AuthorizerContext) === RUN TestACL_Enforce/keyring/list/Deny === RUN TestACL_Enforce/agent/foo/read/Deny authorizer_test.go:618: PASS: AgentRead(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/agent/foo/read/Allow === CONT TestChainedAuthorizer/First_Found === CONT TestACL_Enforce/agent/foo/read/Allow authorizer_test.go:618: PASS: AgentRead(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/agent/foo/write/Deny authorizer_test.go:618: PASS: AgentWrite(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/agent/foo/write/Allow authorizer_test.go:618: PASS: AgentWrite(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/agent/foo/list/Deny === CONT TestChainedAuthorizer/Authorizer_Defaults === RUN TestACL_Enforce/event/foo/read/Deny authorizer_test.go:618: PASS: EventRead(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/event/foo/read/Allow --- PASS: TestChainedAuthorizer (0.00s) --- PASS: TestChainedAuthorizer/No_Authorizers (0.00s) --- PASS: TestChainedAuthorizer/Authorizer_No_Defaults (0.00s) --- PASS: TestChainedAuthorizer/First_Found (0.00s) --- PASS: TestChainedAuthorizer/Authorizer_Defaults (0.00s) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied/AnyAllowed.Prefix(*) === CONT TestPolicyAuthorizer/Defaults/DefaultACLRead.Prefix(foo) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed/AnyAllowed.Prefix(*) === CONT TestACL_Enforce/event/foo/read/Allow authorizer_test.go:618: PASS: EventRead(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/event/foo/write/Deny === CONT TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied/AllDenied.Prefix(*) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed/AllAllowed.Prefix(*) === CONT TestPolicyAuthorizer/Defaults/DefaultKeyringWrite.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultSnapshot.Prefix(foo) === CONT TestACL_Enforce/event/foo/write/Deny authorizer_test.go:618: PASS: EventWrite(string,*acl.AuthorizerContext) === CONT TestPolicyAuthorizer/Defaults/DefaultEventWrite.Prefix(foo) === RUN TestACL_Enforce/event/foo/write/Allow === CONT TestPolicyAuthorizer/Defaults/DefaultSessionWrite.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultKeyringRead.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultSessionRead.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultKeyList.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultServiceWrite.Prefix(foo) === CONT TestACL_Enforce/event/foo/write/Allow authorizer_test.go:618: PASS: EventWrite(string,*acl.AuthorizerContext) === CONT TestPolicyAuthorizer/Defaults/DefaultKeyRead.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultServiceRead.Prefix(foo) === RUN TestACL_Enforce/event/foo/list/Deny === CONT TestPolicyAuthorizer/Defaults/DefaultIntentionWrite.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultPreparedQueryWrite.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultIntentionRead.Prefix(foo) === RUN TestACL_Enforce/intention/foo/read/Deny === CONT TestPolicyAuthorizer/Defaults/DefaultPreparedQueryRead.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultOperatorWrite.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultIntentionDefaultAllow.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultOperatorRead.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultNodeWrite.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultAgentRead.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultNodeRead.Prefix(foo) === CONT TestACL_Enforce/intention/foo/read/Deny authorizer_test.go:618: PASS: IntentionRead(string,*acl.AuthorizerContext) === CONT TestPolicyAuthorizer/Defaults/DefaultEventRead.Prefix(foo) === RUN TestACL_Enforce/intention/foo/read/Allow === CONT TestPolicyAuthorizer/Defaults/DefaultKeyWritePrefix.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultKeyWrite.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultAgentWrite.Prefix(foo) === CONT TestPolicyAuthorizer/Defaults/DefaultACLWrite.Prefix(foo) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed/AllDenied.Prefix(*) === CONT TestACL_Enforce/intention/foo/read/Allow authorizer_test.go:618: PASS: IntentionRead(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/intention/foo/write/Deny === CONT TestPolicyAuthorizer/Intention_Wildcards_-_any_default/AnyDefault.Prefix(*) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_all_default/AnyAllowed.Prefix(*) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_any_default/AllDenied.Prefix(*) === CONT TestPolicyAuthorizer/Intention_Wildcards_-_all_default/AllDefault.Prefix(*) === RUN TestAnyAllowed/prefix-deny-other-deny-prefix === CONT TestACL_Enforce/intention/foo/write/Deny authorizer_test.go:618: PASS: IntentionWrite(string,*acl.AuthorizerContext) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(fo) === RUN TestACL_Enforce/intention/foo/write/Allow === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(fo) === RUN TestAnyAllowed/prefix-write-allowed === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(for) === CONT TestACL_Enforce/intention/foo/write/Allow authorizer_test.go:618: PASS: IntentionWrite(string,*acl.AuthorizerContext) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(for) === RUN TestACL_Enforce/intention/foo/list/Deny === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadAllowed.Prefix(foo) === RUN TestACL_Enforce/node/foo/read/Deny === RUN TestAnyAllowed/prefix-list-allowed === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWriteAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWriteAllowed.Prefix(foo) === RUN TestAnyAllowed/prefix-read-allowed === CONT TestACL_Enforce/node/foo/read/Deny authorizer_test.go:618: PASS: NodeRead(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/node/foo/read/Allow === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(foot) === RUN TestAnyAllowed/prefix-deny-other-write-prefix === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(foot2) === CONT TestACL_Enforce/node/foo/read/Allow authorizer_test.go:618: PASS: NodeRead(string,*acl.AuthorizerContext) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(foot2) === RUN TestACL_Enforce/node/foo/write/Deny === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(food) === RUN TestAnyAllowed/prefix-deny-other-write-exact === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWriteDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWriteDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(fo) === CONT TestACL_Enforce/node/foo/write/Deny authorizer_test.go:618: PASS: NodeWrite(string,*acl.AuthorizerContext) === RUN TestAnyAllowed/prefix-deny-other-list-prefix === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(fo) === RUN TestACL_Enforce/node/foo/write/Allow === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(for) === RUN TestAnyAllowed/no-rules-default === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWriteAllowed.Prefix(foo) === RUN TestAnyAllowed/prefix-deny === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWriteAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(foot) === CONT TestACL_Enforce/node/foo/write/Allow authorizer_test.go:618: PASS: NodeWrite(string,*acl.AuthorizerContext) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(foot) === RUN TestACL_Enforce/node/foo/list/Deny === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(foot2) === RUN TestAnyAllowed/prefix-deny-other-read-prefix === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(food) === RUN TestACL_Enforce/query/foo/read/Deny === RUN TestAnyAllowed/prefix-deny-other-read-exact === RUN TestAnyAllowed/prefix-deny-other-deny-exact --- PASS: TestAnyAllowed (0.01s) --- PASS: TestAnyAllowed/prefix-deny-other-list-exact (0.01s) --- PASS: TestAnyAllowed/prefix-deny-other-deny-prefix (0.00s) --- PASS: TestAnyAllowed/prefix-write-allowed (0.00s) --- PASS: TestAnyAllowed/prefix-list-allowed (0.00s) --- PASS: TestAnyAllowed/prefix-read-allowed (0.00s) --- PASS: TestAnyAllowed/prefix-deny-other-write-prefix (0.00s) --- PASS: TestAnyAllowed/prefix-deny-other-write-exact (0.00s) --- PASS: TestAnyAllowed/prefix-deny-other-list-prefix (0.00s) --- PASS: TestAnyAllowed/no-rules-default (0.00s) --- PASS: TestAnyAllowed/prefix-deny (0.00s) --- PASS: TestAnyAllowed/prefix-deny-other-read-prefix (0.00s) --- PASS: TestAnyAllowed/prefix-deny-other-read-exact (0.00s) --- PASS: TestAnyAllowed/prefix-deny-other-deny-exact (0.00s) === CONT TestACL_Enforce/query/foo/read/Deny authorizer_test.go:618: PASS: PreparedQueryRead(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/query/foo/read/Allow === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadDenied.Prefix(football) === CONT TestACL_Enforce/query/foo/read/Allow authorizer_test.go:618: PASS: PreparedQueryRead(string,*acl.AuthorizerContext) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWriteDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWriteDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(fo) === RUN TestACL_Enforce/query/foo/write/Deny === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot2) === CONT TestACL_Enforce/query/foo/write/Deny authorizer_test.go:618: PASS: PreparedQueryWrite(string,*acl.AuthorizerContext) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot2) === RUN TestACL_Enforce/query/foo/write/Allow === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(fo) === CONT TestACL_Enforce/query/foo/write/Allow authorizer_test.go:618: PASS: PreparedQueryWrite(string,*acl.AuthorizerContext) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(fo) === RUN TestACL_Enforce/query/foo/list/Deny === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWriteAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWriteAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(food) === RUN TestACL_Enforce/service/foo/read/Deny === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWriteDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWriteDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(fo)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(fo)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(fo)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(fo)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(for)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(for)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(for)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(for)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadAllowed.Prefix(foo)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadAllowed.Prefix(foo)#01 === CONT TestACL_Enforce/service/foo/read/Deny authorizer_test.go:618: PASS: ServiceRead(string,*acl.AuthorizerContext) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteAllowed.Prefix(foo)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteAllowed.Prefix(foo)#01 === RUN TestACL_Enforce/service/foo/read/Allow === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot2)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot2)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot2)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot2)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(food)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(food)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(food)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(food)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadDenied.Prefix(football)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadDenied.Prefix(football)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteDenied.Prefix(football)#01 === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteDenied.Prefix(football)#01 === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(fo) === CONT TestACL_Enforce/service/foo/read/Allow authorizer_test.go:618: PASS: ServiceRead(string,*acl.AuthorizerContext) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(fo) === RUN TestACL_Enforce/service/foo/write/Deny === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWriteAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWriteAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(foot2) === CONT TestACL_Enforce/service/foo/write/Deny authorizer_test.go:618: PASS: ServiceWrite(string,*acl.AuthorizerContext) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(food) === RUN TestACL_Enforce/service/foo/write/Allow === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWriteDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWriteDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWriteAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWriteAllowed.Prefix(foo) === CONT TestACL_Enforce/service/foo/write/Allow authorizer_test.go:618: PASS: ServiceWrite(string,*acl.AuthorizerContext) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(foot) === RUN TestACL_Enforce/session/foo/list/Deny === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(food) === RUN TestACL_Enforce/session/foo/read/Deny === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWriteDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWriteDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventWriteAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventWriteAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(foot) === CONT TestACL_Enforce/session/foo/read/Deny authorizer_test.go:618: PASS: SessionRead(string,*acl.AuthorizerContext) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(foot) === RUN TestACL_Enforce/session/foo/read/Allow === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/EventWriteDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/EventWriteDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(fo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(fo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(fo) === CONT TestACL_Enforce/session/foo/read/Allow authorizer_test.go:618: PASS: SessionRead(string,*acl.AuthorizerContext) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(for) === RUN TestACL_Enforce/session/foo/write/Deny === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(for) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(for) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWriteAllowed.Prefix(foo) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWriteAllowed.Prefix(foo) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(foot) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(foot) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(foot2) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(foot2) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(foot2) === CONT TestACL_Enforce/session/foo/write/Deny authorizer_test.go:618: PASS: SessionWrite(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/session/foo/write/Allow === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(food) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(food) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadDenied.Prefix(football) === RUN TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWriteDenied.Prefix(football) === PAUSE TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWriteDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventWriteDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWriteDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWriteAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWriteDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventWriteAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWriteAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteDenied.Prefix(football)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadDenied.Prefix(football)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWriteDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(food)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(food)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot2)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot2)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteAllowed.Prefix(foo)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadAllowed.Prefix(foo)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWriteDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(for)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(for)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(foot2) === CONT TestACL_Enforce/session/foo/write/Allow authorizer_test.go:618: PASS: SessionWrite(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/session/foo/list/Deny#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(fo)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(fo)#01 === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWriteAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWriteDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(fo) === RUN TestACL_Enforce/key/foo/read/Deny === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWriteDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWriteAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(foot2) === CONT TestACL_Enforce/key/foo/read/Deny authorizer_test.go:618: PASS: KeyRead(string,*acl.AuthorizerContext) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(for) === RUN TestACL_Enforce/key/foo/read/Allow === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWriteAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWriteAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(for) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot2) === CONT TestACL_Enforce/key/foo/read/Allow authorizer_test.go:618: PASS: KeyRead(string,*acl.AuthorizerContext) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(fo) === RUN TestACL_Enforce/key/foo/write/Deny === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadDenied.Prefix(football) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(food) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot2) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteAllowed.Prefix(foo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(fo) === CONT TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadAllowed.Prefix(foo) === CONT TestACL_Enforce/key/foo/write/Deny authorizer_test.go:618: PASS: KeyWrite(string,*acl.AuthorizerContext) --- PASS: TestPolicyAuthorizer (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied/AnyAllowed.Prefix(*) (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_prefix_denied/AllDenied.Prefix(*) (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed/AnyAllowed.Prefix(*) (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_all_allowed/AllAllowed.Prefix(*) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultACLRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultKeyringWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultSnapshot.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultEventWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultSessionWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultKeyringRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultSessionRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultKeyList.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultServiceWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultKeyRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultServiceRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultIntentionWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultPreparedQueryWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultIntentionRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultPreparedQueryRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultOperatorWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultIntentionDefaultAllow.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultOperatorRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultNodeWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultAgentRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultNodeRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultEventRead.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultKeyWritePrefix.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultKeyWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultAgentWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Defaults/DefaultACLWrite.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed/AnyAllowed.Prefix(*) (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_prefix_allowed/AllDenied.Prefix(*) (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_any_default (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_any_default/AnyDefault.Prefix(*) (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_any_default/AllDenied.Prefix(*) (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_all_default (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_all_default/AnyAllowed.Prefix(*) (0.00s) --- PASS: TestPolicyAuthorizer/Intention_Wildcards_-_all_default/AllDefault.Prefix(*) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches (0.01s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventWriteDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWriteDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWriteDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/EventReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteDenied.Prefix(football)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadDenied.Prefix(football)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/IntentionReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionWriteDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(food)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/SessionReadDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot2)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot2)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteAllowed.Prefix(foo)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadAllowed.Prefix(foo)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWriteDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(for)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(food)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(for)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(fo)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(fo)#01 (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWriteDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWriteDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/KeyWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/AgentWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryWritePrefixDenied.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(for) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadDenied.Prefix(football) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(food) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/ServiceWritePrefixDenied.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot2) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadPrefixAllowed.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWritePrefixDenied.Prefix(foot) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeWriteAllowed.Prefix(foo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/PreparedQueryReadPrefixAllowed.Prefix(fo) (0.00s) --- PASS: TestPolicyAuthorizer/Prefer_Exact_Matches/NodeReadAllowed.Prefix(foo) (0.00s) === RUN TestACL_Enforce/key/foo/write/Allow authorizer_test.go:618: PASS: KeyWrite(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/key/foo/list/Deny authorizer_test.go:618: PASS: KeyList(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/key/foo/list/Allow authorizer_test.go:618: PASS: KeyList(string,*acl.AuthorizerContext) === RUN TestACL_Enforce/key/foo/deny/Deny === RUN TestACL_Enforce/not-a-real-resource/read/Deny --- PASS: TestACL_Enforce (0.02s) --- PASS: TestACL_Enforce/acl/read/Deny (0.00s) --- PASS: TestACL_Enforce/acl/read/Allow (0.00s) --- PASS: TestACL_Enforce/acl/write/Deny (0.00s) --- PASS: TestACL_Enforce/acl/write/Allow (0.00s) --- PASS: TestACL_Enforce/acl/list/Deny (0.00s) --- PASS: TestACL_Enforce/operator/read/Deny (0.00s) --- PASS: TestACL_Enforce/operator/read/Allow (0.00s) --- PASS: TestACL_Enforce/operator/write/Deny (0.00s) --- PASS: TestACL_Enforce/operator/write/Allow (0.00s) --- PASS: TestACL_Enforce/operator/list/Deny (0.00s) --- PASS: TestACL_Enforce/keyring/read/Deny (0.00s) --- PASS: TestACL_Enforce/keyring/read/Allow (0.00s) --- PASS: TestACL_Enforce/keyring/write/Deny (0.00s) --- PASS: TestACL_Enforce/keyring/write/Allow (0.00s) --- PASS: TestACL_Enforce/keyring/list/Deny (0.00s) --- PASS: TestACL_Enforce/agent/foo/read/Deny (0.00s) --- PASS: TestACL_Enforce/agent/foo/read/Allow (0.00s) --- PASS: TestACL_Enforce/agent/foo/write/Deny (0.00s) --- PASS: TestACL_Enforce/agent/foo/write/Allow (0.00s) --- PASS: TestACL_Enforce/agent/foo/list/Deny (0.00s) --- PASS: TestACL_Enforce/event/foo/read/Deny (0.00s) --- PASS: TestACL_Enforce/event/foo/read/Allow (0.00s) --- PASS: TestACL_Enforce/event/foo/write/Deny (0.00s) --- PASS: TestACL_Enforce/event/foo/write/Allow (0.00s) --- PASS: TestACL_Enforce/event/foo/list/Deny (0.00s) --- PASS: TestACL_Enforce/intention/foo/read/Deny (0.00s) --- PASS: TestACL_Enforce/intention/foo/read/Allow (0.00s) --- PASS: TestACL_Enforce/intention/foo/write/Deny (0.00s) --- PASS: TestACL_Enforce/intention/foo/write/Allow (0.00s) --- PASS: TestACL_Enforce/intention/foo/list/Deny (0.00s) --- PASS: TestACL_Enforce/node/foo/read/Deny (0.00s) --- PASS: TestACL_Enforce/node/foo/read/Allow (0.00s) --- PASS: TestACL_Enforce/node/foo/write/Deny (0.00s) --- PASS: TestACL_Enforce/node/foo/write/Allow (0.00s) --- PASS: TestACL_Enforce/node/foo/list/Deny (0.00s) --- PASS: TestACL_Enforce/query/foo/read/Deny (0.00s) --- PASS: TestACL_Enforce/query/foo/read/Allow (0.00s) --- PASS: TestACL_Enforce/query/foo/write/Deny (0.00s) --- PASS: TestACL_Enforce/query/foo/write/Allow (0.00s) --- PASS: TestACL_Enforce/query/foo/list/Deny (0.00s) --- PASS: TestACL_Enforce/service/foo/read/Deny (0.00s) --- PASS: TestACL_Enforce/service/foo/read/Allow (0.00s) --- PASS: TestACL_Enforce/service/foo/write/Deny (0.00s) --- PASS: TestACL_Enforce/service/foo/write/Allow (0.00s) --- PASS: TestACL_Enforce/session/foo/list/Deny (0.00s) --- PASS: TestACL_Enforce/session/foo/read/Deny (0.00s) --- PASS: TestACL_Enforce/session/foo/read/Allow (0.00s) --- PASS: TestACL_Enforce/session/foo/write/Deny (0.00s) --- PASS: TestACL_Enforce/session/foo/write/Allow (0.00s) --- PASS: TestACL_Enforce/session/foo/list/Deny#01 (0.00s) --- PASS: TestACL_Enforce/key/foo/read/Deny (0.00s) --- PASS: TestACL_Enforce/key/foo/read/Allow (0.00s) --- PASS: TestACL_Enforce/key/foo/write/Deny (0.00s) --- PASS: TestACL_Enforce/key/foo/write/Allow (0.00s) --- PASS: TestACL_Enforce/key/foo/list/Deny (0.00s) --- PASS: TestACL_Enforce/key/foo/list/Allow (0.00s) --- PASS: TestACL_Enforce/key/foo/deny/Deny (0.00s) --- PASS: TestACL_Enforce/not-a-real-resource/read/Deny (0.00s) PASS ok github.com/hashicorp/consul/acl 0.101s === RUN TestACL_Legacy_Disabled_Response === PAUSE TestACL_Legacy_Disabled_Response === RUN TestACL_Legacy_Update === PAUSE TestACL_Legacy_Update === RUN TestACL_Legacy_UpdateUpsert === PAUSE TestACL_Legacy_UpdateUpsert === RUN TestACL_Legacy_Destroy acl_endpoint_legacy_test.go:132: DM-skipped --- SKIP: TestACL_Legacy_Destroy (0.00s) === RUN TestACL_Legacy_Clone === PAUSE TestACL_Legacy_Clone === RUN TestACL_Legacy_Get === PAUSE TestACL_Legacy_Get === RUN TestACL_Legacy_List acl_endpoint_legacy_test.go:254: DM-skipped --- SKIP: TestACL_Legacy_List (0.00s) === RUN TestACLReplicationStatus === PAUSE TestACLReplicationStatus === RUN TestACL_Disabled_Response === PAUSE TestACL_Disabled_Response === RUN TestACL_Bootstrap === PAUSE TestACL_Bootstrap === RUN TestACL_HTTP === PAUSE TestACL_HTTP === RUN TestACL_LoginProcedure_HTTP === PAUSE TestACL_LoginProcedure_HTTP === RUN TestACL_Authorize === PAUSE TestACL_Authorize === RUN TestACL_Version8 === PAUSE TestACL_Version8 === RUN TestACL_AgentMasterToken === PAUSE TestACL_AgentMasterToken === RUN TestACL_RootAuthorizersDenied === PAUSE TestACL_RootAuthorizersDenied === RUN TestACL_vetServiceRegister === PAUSE TestACL_vetServiceRegister === RUN TestACL_vetServiceUpdate === PAUSE TestACL_vetServiceUpdate === RUN TestACL_vetCheckRegister === PAUSE TestACL_vetCheckRegister === RUN TestACL_vetCheckUpdate === PAUSE TestACL_vetCheckUpdate === RUN TestACL_filterMembers === PAUSE TestACL_filterMembers === RUN TestACL_filterServices === PAUSE TestACL_filterServices === RUN TestACL_filterChecks === PAUSE TestACL_filterChecks === RUN TestAgent_Services === PAUSE TestAgent_Services === RUN TestAgent_ServicesFiltered === PAUSE TestAgent_ServicesFiltered === RUN TestAgent_Services_ExternalConnectProxy === PAUSE TestAgent_Services_ExternalConnectProxy === RUN TestAgent_Services_Sidecar === PAUSE TestAgent_Services_Sidecar === RUN TestAgent_Services_MeshGateway === PAUSE TestAgent_Services_MeshGateway === RUN TestAgent_Services_ACLFilter === PAUSE TestAgent_Services_ACLFilter === RUN TestAgent_Service agent_endpoint_test.go:276: DM-skipped --- SKIP: TestAgent_Service (0.00s) === RUN TestAgent_Checks === PAUSE TestAgent_Checks === RUN TestAgent_ChecksWithFilter === PAUSE TestAgent_ChecksWithFilter === RUN TestAgent_HealthServiceByID === PAUSE TestAgent_HealthServiceByID === RUN TestAgent_HealthServiceByName === PAUSE TestAgent_HealthServiceByName === RUN TestAgent_HealthServicesACLEnforcement === PAUSE TestAgent_HealthServicesACLEnforcement === RUN TestAgent_Checks_ACLFilter === PAUSE TestAgent_Checks_ACLFilter === RUN TestAgent_Self === PAUSE TestAgent_Self === RUN TestAgent_Self_ACLDeny === PAUSE TestAgent_Self_ACLDeny === RUN TestAgent_Metrics_ACLDeny === PAUSE TestAgent_Metrics_ACLDeny === RUN TestAgent_Reload === PAUSE TestAgent_Reload === RUN TestAgent_Reload_ACLDeny === PAUSE TestAgent_Reload_ACLDeny === RUN TestAgent_Members === PAUSE TestAgent_Members === RUN TestAgent_Members_WAN === PAUSE TestAgent_Members_WAN === RUN TestAgent_Members_ACLFilter === PAUSE TestAgent_Members_ACLFilter === RUN TestAgent_Join === PAUSE TestAgent_Join === RUN TestAgent_Join_WAN === PAUSE TestAgent_Join_WAN === RUN TestAgent_Join_ACLDeny === PAUSE TestAgent_Join_ACLDeny === RUN TestAgent_JoinLANNotify === PAUSE TestAgent_JoinLANNotify === RUN TestAgent_Leave agent_endpoint_test.go:1575: DM-skipped --- SKIP: TestAgent_Leave (0.01s) === RUN TestAgent_Leave_ACLDeny === PAUSE TestAgent_Leave_ACLDeny === RUN TestAgent_ForceLeave agent_endpoint_test.go:1643: DM-skipped --- SKIP: TestAgent_ForceLeave (0.00s) === RUN TestOpenMetricsMimeTypeHeaders === PAUSE TestOpenMetricsMimeTypeHeaders === RUN TestAgent_ForceLeave_ACLDeny === PAUSE TestAgent_ForceLeave_ACLDeny === RUN TestAgent_ForceLeavePrune === PAUSE TestAgent_ForceLeavePrune === RUN TestAgent_RegisterCheck === PAUSE TestAgent_RegisterCheck === RUN TestAgent_RegisterCheck_Scripts agent_endpoint_test.go:1837: DM-skipped --- SKIP: TestAgent_RegisterCheck_Scripts (0.01s) === RUN TestAgent_RegisterCheckScriptsExecDisable === PAUSE TestAgent_RegisterCheckScriptsExecDisable === RUN TestAgent_RegisterCheckScriptsExecRemoteDisable === PAUSE TestAgent_RegisterCheckScriptsExecRemoteDisable === RUN TestAgent_RegisterCheck_Passing === PAUSE TestAgent_RegisterCheck_Passing === RUN TestAgent_RegisterCheck_BadStatus === PAUSE TestAgent_RegisterCheck_BadStatus === RUN TestAgent_RegisterCheck_ACLDeny === PAUSE TestAgent_RegisterCheck_ACLDeny === RUN TestAgent_DeregisterCheck === PAUSE TestAgent_DeregisterCheck === RUN TestAgent_DeregisterCheckACLDeny === PAUSE TestAgent_DeregisterCheckACLDeny === RUN TestAgent_PassCheck === PAUSE TestAgent_PassCheck === RUN TestAgent_PassCheck_ACLDeny === PAUSE TestAgent_PassCheck_ACLDeny === RUN TestAgent_WarnCheck === PAUSE TestAgent_WarnCheck === RUN TestAgent_WarnCheck_ACLDeny === PAUSE TestAgent_WarnCheck_ACLDeny === RUN TestAgent_FailCheck === PAUSE TestAgent_FailCheck === RUN TestAgent_FailCheck_ACLDeny === PAUSE TestAgent_FailCheck_ACLDeny === RUN TestAgent_UpdateCheck agent_endpoint_test.go:2382: DM-skipped --- SKIP: TestAgent_UpdateCheck (0.00s) === RUN TestAgent_UpdateCheck_ACLDeny === PAUSE TestAgent_UpdateCheck_ACLDeny === RUN TestAgent_RegisterService === RUN TestAgent_RegisterService/normal === PAUSE TestAgent_RegisterService/normal === RUN TestAgent_RegisterService/service_manager === PAUSE TestAgent_RegisterService/service_manager === CONT TestAgent_RegisterService/normal === CONT TestAgent_RegisterService/service_manager [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:31:18.182Z [WARN] TestAgent_RegisterService/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:18.183Z [DEBUG] TestAgent_RegisterService/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:18.185Z [DEBUG] TestAgent_RegisterService/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.187Z [WARN] TestAgent_RegisterService/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:18.188Z [DEBUG] TestAgent_RegisterService/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:18.189Z [DEBUG] TestAgent_RegisterService/normal.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:18.193Z [INFO] TestAgent_RegisterService/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0c261fc5-511f-bd10-e649-6639dc031b0d Address:127.0.0.1:29411}]" === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.194Z [INFO] TestAgent_RegisterService/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2c87e472-a03d-3a60-c6c7-aa5bb3e28495 Address:127.0.0.1:29405}]" writer.go:29: 2021-01-29T19:31:18.198Z [INFO] TestAgent_RegisterService/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29405 [Follower]" leader= writer.go:29: 2021-01-29T19:31:18.200Z [INFO] TestAgent_RegisterService/service_manager.server.serf.wan: serf: EventMemberJoin: Node-2c87e472-a03d-3a60-c6c7-aa5bb3e28495.dc1 127.0.0.1 === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.195Z [INFO] TestAgent_RegisterService/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29411 [Follower]" leader= === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.205Z [INFO] TestAgent_RegisterService/service_manager.server.serf.lan: serf: EventMemberJoin: Node-2c87e472-a03d-3a60-c6c7-aa5bb3e28495 127.0.0.1 === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.201Z [INFO] TestAgent_RegisterService/normal.server.serf.wan: serf: EventMemberJoin: Node-0c261fc5-511f-bd10-e649-6639dc031b0d.dc1 127.0.0.1 === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.209Z [INFO] TestAgent_RegisterService/service_manager: Started DNS server: address=127.0.0.1:29400 network=udp === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.209Z [INFO] TestAgent_RegisterService/normal.server.serf.lan: serf: EventMemberJoin: Node-0c261fc5-511f-bd10-e649-6639dc031b0d 127.0.0.1 writer.go:29: 2021-01-29T19:31:18.214Z [INFO] TestAgent_RegisterService/normal: Started DNS server: address=127.0.0.1:29406 network=udp === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.236Z [INFO] TestAgent_RegisterService/service_manager.server: Adding LAN server: server="Node-2c87e472-a03d-3a60-c6c7-aa5bb3e28495 (Addr: tcp/127.0.0.1:29405) (DC: dc1)" === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.236Z [INFO] TestAgent_RegisterService/normal.server: Adding LAN server: server="Node-0c261fc5-511f-bd10-e649-6639dc031b0d (Addr: tcp/127.0.0.1:29411) (DC: dc1)" writer.go:29: 2021-01-29T19:31:18.239Z [INFO] TestAgent_RegisterService/normal.server: Handled event for server in area: event=member-join server=Node-0c261fc5-511f-bd10-e649-6639dc031b0d.dc1 area=wan writer.go:29: 2021-01-29T19:31:18.242Z [WARN] TestAgent_RegisterService/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:18.243Z [INFO] TestAgent_RegisterService/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29411 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:18.252Z [INFO] TestAgent_RegisterService/normal: Started DNS server: address=127.0.0.1:29406 network=tcp writer.go:29: 2021-01-29T19:31:18.255Z [DEBUG] TestAgent_RegisterService/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:18.257Z [DEBUG] TestAgent_RegisterService/normal.server.raft: vote granted: from=0c261fc5-511f-bd10-e649-6639dc031b0d term=2 tally=1 writer.go:29: 2021-01-29T19:31:18.259Z [INFO] TestAgent_RegisterService/normal.server.raft: election won: tally=1 === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.243Z [INFO] TestAgent_RegisterService/service_manager: Started DNS server: address=127.0.0.1:29400 network=tcp === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.261Z [INFO] TestAgent_RegisterService/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29411 [Leader]" === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.252Z [INFO] TestAgent_RegisterService/service_manager.server: Handled event for server in area: event=member-join server=Node-2c87e472-a03d-3a60-c6c7-aa5bb3e28495.dc1 area=wan === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.256Z [INFO] TestAgent_RegisterService/normal: Started HTTP server: address=127.0.0.1:29407 network=tcp === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.256Z [WARN] TestAgent_RegisterService/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.272Z [INFO] TestAgent_RegisterService/normal: started state syncer writer.go:29: 2021-01-29T19:31:18.270Z [INFO] TestAgent_RegisterService/normal.server: cluster leadership acquired === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.273Z [INFO] TestAgent_RegisterService/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29405 [Candidate]" term=2 === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.270Z [INFO] TestAgent_RegisterService/normal.server: New leader elected: payload=Node-0c261fc5-511f-bd10-e649-6639dc031b0d === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.263Z [INFO] TestAgent_RegisterService/service_manager: Started HTTP server: address=127.0.0.1:29401 network=tcp writer.go:29: 2021-01-29T19:31:18.279Z [DEBUG] TestAgent_RegisterService/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:18.282Z [DEBUG] TestAgent_RegisterService/service_manager.server.raft: vote granted: from=2c87e472-a03d-3a60-c6c7-aa5bb3e28495 term=2 tally=1 writer.go:29: 2021-01-29T19:31:18.282Z [INFO] TestAgent_RegisterService/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:18.283Z [INFO] TestAgent_RegisterService/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29405 [Leader]" writer.go:29: 2021-01-29T19:31:18.281Z [INFO] TestAgent_RegisterService/service_manager: started state syncer === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.270Z [DEBUG] TestAgent_RegisterService/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29411 writer.go:29: 2021-01-29T19:31:18.275Z [INFO] TestAgent_RegisterService/normal: Synced node info === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.284Z [INFO] TestAgent_RegisterService/service_manager.server: cluster leadership acquired === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.297Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.285Z [INFO] TestAgent_RegisterService/service_manager.server: New leader elected: payload=Node-2c87e472-a03d-3a60-c6c7-aa5bb3e28495 writer.go:29: 2021-01-29T19:31:18.295Z [DEBUG] TestAgent_RegisterService/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29405 writer.go:29: 2021-01-29T19:31:18.295Z [INFO] TestAgent_RegisterService/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:18.333Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:18.344Z [INFO] TestAgent_RegisterService/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:18.345Z [INFO] TestAgent_RegisterService/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:18.346Z [DEBUG] TestAgent_RegisterService/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-2c87e472-a03d-3a60-c6c7-aa5bb3e28495 writer.go:29: 2021-01-29T19:31:18.347Z [INFO] TestAgent_RegisterService/service_manager.server: member joined, marking health alive: member=Node-2c87e472-a03d-3a60-c6c7-aa5bb3e28495 === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.347Z [INFO] TestAgent_RegisterService/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:18.349Z [INFO] TestAgent_RegisterService/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:18.351Z [DEBUG] TestAgent_RegisterService/normal.server: Skipping self join check for node since the cluster is too small: node=Node-0c261fc5-511f-bd10-e649-6639dc031b0d writer.go:29: 2021-01-29T19:31:18.370Z [INFO] TestAgent_RegisterService/normal.server: member joined, marking health alive: member=Node-0c261fc5-511f-bd10-e649-6639dc031b0d === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.375Z [DEBUG] TestAgent_RegisterService/service_manager: Node info in sync writer.go:29: 2021-01-29T19:31:18.378Z [INFO] TestAgent_RegisterService/service_manager: Synced service: service=test writer.go:29: 2021-01-29T19:31:18.378Z [DEBUG] TestAgent_RegisterService/service_manager: Check in sync: check=service:test:2 writer.go:29: 2021-01-29T19:31:18.379Z [DEBUG] TestAgent_RegisterService/service_manager: Check in sync: check=service:test:3 writer.go:29: 2021-01-29T19:31:18.380Z [DEBUG] TestAgent_RegisterService/service_manager: Check in sync: check=service:test:1 writer.go:29: 2021-01-29T19:31:18.381Z [INFO] TestAgent_RegisterService/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:18.381Z [INFO] TestAgent_RegisterService/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:18.382Z [DEBUG] TestAgent_RegisterService/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:18.383Z [WARN] TestAgent_RegisterService/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:18.383Z [DEBUG] TestAgent_RegisterService/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:18.385Z [WARN] TestAgent_RegisterService/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:18.387Z [INFO] TestAgent_RegisterService/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:18.387Z [INFO] TestAgent_RegisterService/service_manager: consul server down writer.go:29: 2021-01-29T19:31:18.389Z [INFO] TestAgent_RegisterService/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:18.390Z [INFO] TestAgent_RegisterService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29400 network=tcp writer.go:29: 2021-01-29T19:31:18.390Z [INFO] TestAgent_RegisterService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29400 network=udp writer.go:29: 2021-01-29T19:31:18.392Z [INFO] TestAgent_RegisterService/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29401 network=tcp === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.393Z [DEBUG] TestAgent_RegisterService/normal: Node info in sync writer.go:29: 2021-01-29T19:31:18.404Z [INFO] TestAgent_RegisterService/normal: Synced service: service=test writer.go:29: 2021-01-29T19:31:18.405Z [DEBUG] TestAgent_RegisterService/normal: Check in sync: check=service:test:3 writer.go:29: 2021-01-29T19:31:18.407Z [DEBUG] TestAgent_RegisterService/normal: Check in sync: check=service:test:1 writer.go:29: 2021-01-29T19:31:18.408Z [DEBUG] TestAgent_RegisterService/normal: Check in sync: check=service:test:2 writer.go:29: 2021-01-29T19:31:18.409Z [INFO] TestAgent_RegisterService/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:18.410Z [INFO] TestAgent_RegisterService/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:18.411Z [DEBUG] TestAgent_RegisterService/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:18.412Z [WARN] TestAgent_RegisterService/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:18.412Z [DEBUG] TestAgent_RegisterService/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:18.419Z [WARN] TestAgent_RegisterService/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:18.423Z [INFO] TestAgent_RegisterService/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:18.423Z [INFO] TestAgent_RegisterService/normal: consul server down writer.go:29: 2021-01-29T19:31:18.426Z [INFO] TestAgent_RegisterService/normal: shutdown complete writer.go:29: 2021-01-29T19:31:18.427Z [INFO] TestAgent_RegisterService/normal: Stopping server: protocol=DNS address=127.0.0.1:29406 network=tcp writer.go:29: 2021-01-29T19:31:18.428Z [INFO] TestAgent_RegisterService/normal: Stopping server: protocol=DNS address=127.0.0.1:29406 network=udp writer.go:29: 2021-01-29T19:31:18.430Z [INFO] TestAgent_RegisterService/normal: Stopping server: protocol=HTTP address=127.0.0.1:29407 network=tcp === CONT TestAgent_RegisterService/service_manager writer.go:29: 2021-01-29T19:31:18.893Z [INFO] TestAgent_RegisterService/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:18.894Z [INFO] TestAgent_RegisterService/service_manager: Endpoints down === CONT TestAgent_RegisterService/normal writer.go:29: 2021-01-29T19:31:18.931Z [INFO] TestAgent_RegisterService/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:18.931Z [INFO] TestAgent_RegisterService/normal: Endpoints down --- PASS: TestAgent_RegisterService (0.00s) --- PASS: TestAgent_RegisterService/service_manager (0.81s) --- PASS: TestAgent_RegisterService/normal (0.85s) === RUN TestAgent_RegisterService_ReRegister === RUN TestAgent_RegisterService_ReRegister/normal === PAUSE TestAgent_RegisterService_ReRegister/normal === RUN TestAgent_RegisterService_ReRegister/service_manager === PAUSE TestAgent_RegisterService_ReRegister/service_manager === CONT TestAgent_RegisterService_ReRegister/normal === CONT TestAgent_RegisterService_ReRegister/service_manager === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:18.986Z [WARN] TestAgent_RegisterService_ReRegister/normal: bootstrap = true: do not enable unless necessary === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:18.988Z [WARN] TestAgent_RegisterService_ReRegister/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:18.989Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:18.990Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:18.988Z [DEBUG] TestAgent_RegisterService_ReRegister/normal.tlsutil: Update: version=1 === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:18.994Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c499b726-776b-bc82-9189-fc29b911b6ad Address:127.0.0.1:29423}]" writer.go:29: 2021-01-29T19:31:18.996Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server.serf.wan: serf: EventMemberJoin: Node-c499b726-776b-bc82-9189-fc29b911b6ad.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:18.996Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29423 [Follower]" leader= === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:18.994Z [DEBUG] TestAgent_RegisterService_ReRegister/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:18.998Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server.serf.lan: serf: EventMemberJoin: Node-c499b726-776b-bc82-9189-fc29b911b6ad 127.0.0.1 writer.go:29: 2021-01-29T19:31:19.008Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server: Adding LAN server: server="Node-c499b726-776b-bc82-9189-fc29b911b6ad (Addr: tcp/127.0.0.1:29423) (DC: dc1)" writer.go:29: 2021-01-29T19:31:19.008Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server: Handled event for server in area: event=member-join server=Node-c499b726-776b-bc82-9189-fc29b911b6ad.dc1 area=wan === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:19.006Z [INFO] TestAgent_RegisterService_ReRegister/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cceca772-1086-13f3-766e-dc35eb14b0dd Address:127.0.0.1:29417}]" === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:19.009Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Started DNS server: address=127.0.0.1:29418 network=udp writer.go:29: 2021-01-29T19:31:19.012Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Started DNS server: address=127.0.0.1:29418 network=tcp === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:19.011Z [INFO] TestAgent_RegisterService_ReRegister/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29417 [Follower]" leader= === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:19.015Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Started HTTP server: address=127.0.0.1:29419 network=tcp writer.go:29: 2021-01-29T19:31:19.016Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: started state syncer === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:19.017Z [INFO] TestAgent_RegisterService_ReRegister/normal.server.serf.wan: serf: EventMemberJoin: Node-cceca772-1086-13f3-766e-dc35eb14b0dd.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:19.028Z [INFO] TestAgent_RegisterService_ReRegister/normal.server.serf.lan: serf: EventMemberJoin: Node-cceca772-1086-13f3-766e-dc35eb14b0dd 127.0.0.1 writer.go:29: 2021-01-29T19:31:19.036Z [INFO] TestAgent_RegisterService_ReRegister/normal.server: Handled event for server in area: event=member-join server=Node-cceca772-1086-13f3-766e-dc35eb14b0dd.dc1 area=wan writer.go:29: 2021-01-29T19:31:19.036Z [INFO] TestAgent_RegisterService_ReRegister/normal.server: Adding LAN server: server="Node-cceca772-1086-13f3-766e-dc35eb14b0dd (Addr: tcp/127.0.0.1:29417) (DC: dc1)" writer.go:29: 2021-01-29T19:31:19.038Z [INFO] TestAgent_RegisterService_ReRegister/normal: Started DNS server: address=127.0.0.1:29412 network=tcp writer.go:29: 2021-01-29T19:31:19.041Z [INFO] TestAgent_RegisterService_ReRegister/normal: Started DNS server: address=127.0.0.1:29412 network=udp writer.go:29: 2021-01-29T19:31:19.046Z [INFO] TestAgent_RegisterService_ReRegister/normal: Started HTTP server: address=127.0.0.1:29413 network=tcp writer.go:29: 2021-01-29T19:31:19.047Z [INFO] TestAgent_RegisterService_ReRegister/normal: started state syncer === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:19.053Z [WARN] TestAgent_RegisterService_ReRegister/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:19.053Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29423 [Candidate]" term=2 === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:19.064Z [WARN] TestAgent_RegisterService_ReRegister/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:19.066Z [INFO] TestAgent_RegisterService_ReRegister/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29417 [Candidate]" term=2 === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:19.068Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:19.069Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager.server.raft: vote granted: from=c499b726-776b-bc82-9189-fc29b911b6ad term=2 tally=1 === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:19.068Z [DEBUG] TestAgent_RegisterService_ReRegister/normal.server.raft: votes: needed=1 === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:19.070Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:19.070Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29423 [Leader]" writer.go:29: 2021-01-29T19:31:19.071Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:19.072Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server: New leader elected: payload=Node-c499b726-776b-bc82-9189-fc29b911b6ad writer.go:29: 2021-01-29T19:31:19.073Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29423 === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:19.070Z [DEBUG] TestAgent_RegisterService_ReRegister/normal.server.raft: vote granted: from=cceca772-1086-13f3-766e-dc35eb14b0dd term=2 tally=1 writer.go:29: 2021-01-29T19:31:19.077Z [INFO] TestAgent_RegisterService_ReRegister/normal.server.raft: election won: tally=1 === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:19.080Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:19.080Z [INFO] TestAgent_RegisterService_ReRegister/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29417 [Leader]" writer.go:29: 2021-01-29T19:31:19.084Z [INFO] TestAgent_RegisterService_ReRegister/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:19.085Z [INFO] TestAgent_RegisterService_ReRegister/normal.server: New leader elected: payload=Node-cceca772-1086-13f3-766e-dc35eb14b0dd writer.go:29: 2021-01-29T19:31:19.085Z [DEBUG] TestAgent_RegisterService_ReRegister/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29417 === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:19.088Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:19.090Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:19.092Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-c499b726-776b-bc82-9189-fc29b911b6ad writer.go:29: 2021-01-29T19:31:19.092Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server: member joined, marking health alive: member=Node-c499b726-776b-bc82-9189-fc29b911b6ad === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:19.103Z [INFO] TestAgent_RegisterService_ReRegister/normal: Synced node info writer.go:29: 2021-01-29T19:31:19.104Z [DEBUG] TestAgent_RegisterService_ReRegister/normal: Node info in sync writer.go:29: 2021-01-29T19:31:19.105Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:19.134Z [INFO] TestAgent_RegisterService_ReRegister/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:19.137Z [INFO] TestAgent_RegisterService_ReRegister/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:19.141Z [DEBUG] TestAgent_RegisterService_ReRegister/normal.server: Skipping self join check for node since the cluster is too small: node=Node-cceca772-1086-13f3-766e-dc35eb14b0dd writer.go:29: 2021-01-29T19:31:19.142Z [INFO] TestAgent_RegisterService_ReRegister/normal.server: member joined, marking health alive: member=Node-cceca772-1086-13f3-766e-dc35eb14b0dd writer.go:29: 2021-01-29T19:31:19.238Z [DEBUG] TestAgent_RegisterService_ReRegister/normal: Node info in sync writer.go:29: 2021-01-29T19:31:19.241Z [INFO] TestAgent_RegisterService_ReRegister/normal: Synced service: service=test writer.go:29: 2021-01-29T19:31:19.242Z [DEBUG] TestAgent_RegisterService_ReRegister/normal: Check in sync: check=check_1 writer.go:29: 2021-01-29T19:31:19.244Z [DEBUG] TestAgent_RegisterService_ReRegister/normal: Check in sync: check=check_2 writer.go:29: 2021-01-29T19:31:19.247Z [DEBUG] TestAgent_RegisterService_ReRegister/normal: Node info in sync writer.go:29: 2021-01-29T19:31:19.250Z [INFO] TestAgent_RegisterService_ReRegister/normal: Synced service: service=test writer.go:29: 2021-01-29T19:31:19.251Z [DEBUG] TestAgent_RegisterService_ReRegister/normal: Check in sync: check=check_1 writer.go:29: 2021-01-29T19:31:19.252Z [DEBUG] TestAgent_RegisterService_ReRegister/normal: Check in sync: check=check_2 writer.go:29: 2021-01-29T19:31:19.253Z [DEBUG] TestAgent_RegisterService_ReRegister/normal: Check in sync: check=check_3 writer.go:29: 2021-01-29T19:31:19.254Z [INFO] TestAgent_RegisterService_ReRegister/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:19.255Z [INFO] TestAgent_RegisterService_ReRegister/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:19.256Z [DEBUG] TestAgent_RegisterService_ReRegister/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:19.257Z [WARN] TestAgent_RegisterService_ReRegister/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:19.257Z [DEBUG] TestAgent_RegisterService_ReRegister/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:19.261Z [WARN] TestAgent_RegisterService_ReRegister/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:19.264Z [INFO] TestAgent_RegisterService_ReRegister/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:19.264Z [INFO] TestAgent_RegisterService_ReRegister/normal: consul server down writer.go:29: 2021-01-29T19:31:19.266Z [INFO] TestAgent_RegisterService_ReRegister/normal: shutdown complete writer.go:29: 2021-01-29T19:31:19.267Z [INFO] TestAgent_RegisterService_ReRegister/normal: Stopping server: protocol=DNS address=127.0.0.1:29412 network=tcp writer.go:29: 2021-01-29T19:31:19.267Z [INFO] TestAgent_RegisterService_ReRegister/normal: Stopping server: protocol=DNS address=127.0.0.1:29412 network=udp writer.go:29: 2021-01-29T19:31:19.268Z [INFO] TestAgent_RegisterService_ReRegister/normal: Stopping server: protocol=HTTP address=127.0.0.1:29413 network=tcp === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:19.325Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:19.327Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Synced service: service=test writer.go:29: 2021-01-29T19:31:19.328Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager: Check in sync: check=check_2 writer.go:29: 2021-01-29T19:31:19.328Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager: Check in sync: check=check_1 writer.go:29: 2021-01-29T19:31:19.339Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager: Node info in sync writer.go:29: 2021-01-29T19:31:19.341Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Synced service: service=test writer.go:29: 2021-01-29T19:31:19.342Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager: Check in sync: check=check_1 writer.go:29: 2021-01-29T19:31:19.343Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager: Check in sync: check=check_2 writer.go:29: 2021-01-29T19:31:19.344Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager: Check in sync: check=check_3 writer.go:29: 2021-01-29T19:31:19.346Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:19.347Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:19.348Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:19.349Z [WARN] TestAgent_RegisterService_ReRegister/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:19.348Z [ERROR] TestAgent_RegisterService_ReRegister/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:19.349Z [DEBUG] TestAgent_RegisterService_ReRegister/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:19.356Z [WARN] TestAgent_RegisterService_ReRegister/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:19.358Z [INFO] TestAgent_RegisterService_ReRegister/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:19.358Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: consul server down writer.go:29: 2021-01-29T19:31:19.361Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:19.362Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29418 network=tcp writer.go:29: 2021-01-29T19:31:19.365Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29418 network=udp writer.go:29: 2021-01-29T19:31:19.366Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29419 network=tcp === CONT TestAgent_RegisterService_ReRegister/normal writer.go:29: 2021-01-29T19:31:19.769Z [INFO] TestAgent_RegisterService_ReRegister/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:19.771Z [INFO] TestAgent_RegisterService_ReRegister/normal: Endpoints down === CONT TestAgent_RegisterService_ReRegister/service_manager writer.go:29: 2021-01-29T19:31:19.867Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:19.869Z [INFO] TestAgent_RegisterService_ReRegister/service_manager: Endpoints down --- PASS: TestAgent_RegisterService_ReRegister (0.00s) --- PASS: TestAgent_RegisterService_ReRegister/normal (0.84s) --- PASS: TestAgent_RegisterService_ReRegister/service_manager (0.94s) === RUN TestAgent_RegisterService_ReRegister_ReplaceExistingChecks === RUN TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal === PAUSE TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal === RUN TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager === PAUSE TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:19.902Z [WARN] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:19.906Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:19.922Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:19.942Z [WARN] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:19.961Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:19.967Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:19.985Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a8d6d657-35f7-921a-dcff-ff7126dea64e Address:127.0.0.1:29435}]" writer.go:29: 2021-01-29T19:31:19.990Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29435 [Follower]" leader= writer.go:29: 2021-01-29T19:31:19.996Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.serf.wan: serf: EventMemberJoin: Node-a8d6d657-35f7-921a-dcff-ff7126dea64e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:20.004Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.serf.lan: serf: EventMemberJoin: Node-a8d6d657-35f7-921a-dcff-ff7126dea64e 127.0.0.1 writer.go:29: 2021-01-29T19:31:20.010Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server: Adding LAN server: server="Node-a8d6d657-35f7-921a-dcff-ff7126dea64e (Addr: tcp/127.0.0.1:29435) (DC: dc1)" === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.010Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c209b5da-31b7-f663-2e6a-58929fd4d9d6 Address:127.0.0.1:29429}]" === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.012Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server: Handled event for server in area: event=member-join server=Node-a8d6d657-35f7-921a-dcff-ff7126dea64e.dc1 area=wan === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.014Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.serf.wan: serf: EventMemberJoin: Node-c209b5da-31b7-f663-2e6a-58929fd4d9d6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:20.014Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29429 [Follower]" leader= writer.go:29: 2021-01-29T19:31:20.018Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.serf.lan: serf: EventMemberJoin: Node-c209b5da-31b7-f663-2e6a-58929fd4d9d6 127.0.0.1 writer.go:29: 2021-01-29T19:31:20.025Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server: Adding LAN server: server="Node-c209b5da-31b7-f663-2e6a-58929fd4d9d6 (Addr: tcp/127.0.0.1:29429) (DC: dc1)" writer.go:29: 2021-01-29T19:31:20.030Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server: Handled event for server in area: event=member-join server=Node-c209b5da-31b7-f663-2e6a-58929fd4d9d6.dc1 area=wan writer.go:29: 2021-01-29T19:31:20.033Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Started DNS server: address=127.0.0.1:29424 network=udp writer.go:29: 2021-01-29T19:31:20.035Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Started DNS server: address=127.0.0.1:29424 network=tcp === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.041Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Started DNS server: address=127.0.0.1:29430 network=tcp === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.042Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Started HTTP server: address=127.0.0.1:29425 network=tcp writer.go:29: 2021-01-29T19:31:20.046Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: started state syncer === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.046Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Started DNS server: address=127.0.0.1:29430 network=udp writer.go:29: 2021-01-29T19:31:20.043Z [WARN] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:20.055Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29435 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:20.051Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Started HTTP server: address=127.0.0.1:29431 network=tcp writer.go:29: 2021-01-29T19:31:20.059Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:20.061Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.raft: vote granted: from=a8d6d657-35f7-921a-dcff-ff7126dea64e term=2 tally=1 writer.go:29: 2021-01-29T19:31:20.063Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:20.061Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: started state syncer writer.go:29: 2021-01-29T19:31:20.065Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29435 [Leader]" writer.go:29: 2021-01-29T19:31:20.072Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server: cluster leadership acquired === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.081Z [WARN] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:20.083Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29429 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:20.087Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:20.090Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.raft: vote granted: from=c209b5da-31b7-f663-2e6a-58929fd4d9d6 term=2 tally=1 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.077Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server: New leader elected: payload=Node-a8d6d657-35f7-921a-dcff-ff7126dea64e === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.092Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:20.098Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29429 [Leader]" === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.094Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29435 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.106Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:20.107Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server: New leader elected: payload=Node-c209b5da-31b7-f663-2e6a-58929fd4d9d6 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.112Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.109Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29429 writer.go:29: 2021-01-29T19:31:20.129Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.132Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:20.134Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.140Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.139Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server: Skipping self join check for node since the cluster is too small: node=Node-a8d6d657-35f7-921a-dcff-ff7126dea64e === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.144Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.147Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server: member joined, marking health alive: member=Node-a8d6d657-35f7-921a-dcff-ff7126dea64e === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.147Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-c209b5da-31b7-f663-2e6a-58929fd4d9d6 writer.go:29: 2021-01-29T19:31:20.160Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server: member joined, marking health alive: member=Node-c209b5da-31b7-f663-2e6a-58929fd4d9d6 writer.go:29: 2021-01-29T19:31:20.173Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:20.175Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Synced service: service=test writer.go:29: 2021-01-29T19:31:20.177Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Check in sync: check=service:test:1 writer.go:29: 2021-01-29T19:31:20.181Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Check in sync: check=check_2 writer.go:29: 2021-01-29T19:31:20.193Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: removed check: check=check_2 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.194Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Synced node info === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.195Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Node info in sync writer.go:29: 2021-01-29T19:31:20.198Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Synced service: service=test === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.198Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Synced service: service=test === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.199Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Deregistered check: check=check_2 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.199Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Check in sync: check=check_2 writer.go:29: 2021-01-29T19:31:20.200Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Check in sync: check=service:test:1 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.200Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Check in sync: check=check_3 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.204Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: removed check: check=check_2 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.203Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Check in sync: check=service:test:1 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.205Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Node info in sync === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.209Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:20.210Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:20.212Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:20.213Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.208Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Synced service: service=test === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.213Z [WARN] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:20.212Z [ERROR] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:20.218Z [WARN] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.217Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Deregistered check: check=check_2 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.220Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:20.221Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: consul server down === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.221Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Check in sync: check=check_3 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.222Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: shutdown complete === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.223Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Check in sync: check=service:test:1 === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.224Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29424 network=tcp writer.go:29: 2021-01-29T19:31:20.225Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29424 network=udp writer.go:29: 2021-01-29T19:31:20.226Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29425 network=tcp === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.228Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:20.230Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:20.231Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:20.234Z [WARN] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:20.231Z [ERROR] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:20.234Z [DEBUG] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:20.260Z [WARN] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:20.279Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:20.279Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: consul server down writer.go:29: 2021-01-29T19:31:20.282Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: shutdown complete writer.go:29: 2021-01-29T19:31:20.284Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Stopping server: protocol=DNS address=127.0.0.1:29430 network=tcp writer.go:29: 2021-01-29T19:31:20.286Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Stopping server: protocol=DNS address=127.0.0.1:29430 network=udp writer.go:29: 2021-01-29T19:31:20.287Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Stopping server: protocol=HTTP address=127.0.0.1:29431 network=tcp === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager writer.go:29: 2021-01-29T19:31:20.728Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:20.732Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager: Endpoints down === CONT TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal writer.go:29: 2021-01-29T19:31:20.789Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:20.790Z [INFO] TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal: Endpoints down --- PASS: TestAgent_RegisterService_ReRegister_ReplaceExistingChecks (0.00s) --- PASS: TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/service_manager (0.86s) --- PASS: TestAgent_RegisterService_ReRegister_ReplaceExistingChecks/normal (0.91s) === RUN TestAgent_RegisterService_TranslateKeys === RUN TestAgent_RegisterService_TranslateKeys/normal === PAUSE TestAgent_RegisterService_TranslateKeys/normal === RUN TestAgent_RegisterService_TranslateKeys/service_manager === PAUSE TestAgent_RegisterService_TranslateKeys/service_manager === CONT TestAgent_RegisterService_TranslateKeys/normal === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.835Z [WARN] TestAgent_RegisterService_TranslateKeys/service_manager: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:20.836Z [WARN] TestAgent_RegisterService_TranslateKeys/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:20.837Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:20.842Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.846Z [WARN] TestAgent_RegisterService_TranslateKeys/normal: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.850Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fd06f590-dd4d-f483-dda5-b53904af22e2 Address:127.0.0.1:29447}]" writer.go:29: 2021-01-29T19:31:20.852Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29447 [Follower]" leader= === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.848Z [WARN] TestAgent_RegisterService_TranslateKeys/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:20.855Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.tlsutil: Update: version=1 === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.856Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.serf.wan: serf: EventMemberJoin: Node-fd06f590-dd4d-f483-dda5-b53904af22e2.dc1 127.0.0.1 === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.858Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:20.865Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1932ccac-05cc-dc57-693c-5632e301f7f5 Address:127.0.0.1:29441}]" writer.go:29: 2021-01-29T19:31:20.867Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29441 [Follower]" leader= === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.862Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.serf.lan: serf: EventMemberJoin: Node-fd06f590-dd4d-f483-dda5-b53904af22e2 127.0.0.1 writer.go:29: 2021-01-29T19:31:20.871Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: Adding LAN server: server="Node-fd06f590-dd4d-f483-dda5-b53904af22e2 (Addr: tcp/127.0.0.1:29447) (DC: dc1)" writer.go:29: 2021-01-29T19:31:20.872Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: Handled event for server in area: event=member-join server=Node-fd06f590-dd4d-f483-dda5-b53904af22e2.dc1 area=wan writer.go:29: 2021-01-29T19:31:20.877Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Started DNS server: address=127.0.0.1:29442 network=tcp === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.871Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.serf.wan: serf: EventMemberJoin: Node-1932ccac-05cc-dc57-693c-5632e301f7f5.dc1 127.0.0.1 === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.878Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Started DNS server: address=127.0.0.1:29442 network=udp writer.go:29: 2021-01-29T19:31:20.881Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Started HTTP server: address=127.0.0.1:29443 network=tcp writer.go:29: 2021-01-29T19:31:20.882Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: started state syncer === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.880Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.serf.lan: serf: EventMemberJoin: Node-1932ccac-05cc-dc57-693c-5632e301f7f5 127.0.0.1 writer.go:29: 2021-01-29T19:31:20.886Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: Adding LAN server: server="Node-1932ccac-05cc-dc57-693c-5632e301f7f5 (Addr: tcp/127.0.0.1:29441) (DC: dc1)" writer.go:29: 2021-01-29T19:31:20.886Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: Handled event for server in area: event=member-join server=Node-1932ccac-05cc-dc57-693c-5632e301f7f5.dc1 area=wan writer.go:29: 2021-01-29T19:31:20.889Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Started DNS server: address=127.0.0.1:29436 network=tcp writer.go:29: 2021-01-29T19:31:20.891Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Started DNS server: address=127.0.0.1:29436 network=udp writer.go:29: 2021-01-29T19:31:20.892Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Started HTTP server: address=127.0.0.1:29437 network=tcp writer.go:29: 2021-01-29T19:31:20.894Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: started state syncer writer.go:29: 2021-01-29T19:31:20.905Z [WARN] TestAgent_RegisterService_TranslateKeys/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:20.906Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29441 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:20.913Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:20.915Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.server.raft: vote granted: from=1932ccac-05cc-dc57-693c-5632e301f7f5 term=2 tally=1 writer.go:29: 2021-01-29T19:31:20.917Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.raft: election won: tally=1 === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.919Z [WARN] TestAgent_RegisterService_TranslateKeys/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:20.921Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29447 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:20.923Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:20.924Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.server.raft: vote granted: from=fd06f590-dd4d-f483-dda5-b53904af22e2 term=2 tally=1 writer.go:29: 2021-01-29T19:31:20.925Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:20.926Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29447 [Leader]" === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.918Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29441 [Leader]" === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.926Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:20.928Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: cluster leadership acquired === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.928Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:20.929Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: New leader elected: payload=Node-1932ccac-05cc-dc57-693c-5632e301f7f5 === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.930Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: initializing acls writer.go:29: 2021-01-29T19:31:20.930Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: New leader elected: payload=Node-fd06f590-dd4d-f483-dda5-b53904af22e2 writer.go:29: 2021-01-29T19:31:20.932Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:20.933Z [WARN] TestAgent_RegisterService_TranslateKeys/service_manager.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.932Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: initializing acls === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.934Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: Bootstrapped ACL master token from configuration === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.937Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: initializing acls === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.939Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:20.940Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:20.941Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:20.942Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.serf.lan: serf: EventMemberUpdate: Node-fd06f590-dd4d-f483-dda5-b53904af22e2 === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.939Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: Created ACL 'global-management' policy === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.943Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.serf.wan: serf: EventMemberUpdate: Node-fd06f590-dd4d-f483-dda5-b53904af22e2.dc1 writer.go:29: 2021-01-29T19:31:20.943Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: Updating LAN server: server="Node-fd06f590-dd4d-f483-dda5-b53904af22e2 (Addr: tcp/127.0.0.1:29447) (DC: dc1)" writer.go:29: 2021-01-29T19:31:20.944Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: Handled event for server in area: event=member-update server=Node-fd06f590-dd4d-f483-dda5-b53904af22e2.dc1 area=wan === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.943Z [WARN] TestAgent_RegisterService_TranslateKeys/normal.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:20.939Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:20.952Z [WARN] TestAgent_RegisterService_TranslateKeys/normal.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:20.950Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:20.954Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:20.957Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:20.958Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:20.959Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.server: transitioning out of legacy ACL mode === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.963Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.960Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.serf.lan: serf: EventMemberUpdate: Node-1932ccac-05cc-dc57-693c-5632e301f7f5 writer.go:29: 2021-01-29T19:31:20.969Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.serf.wan: serf: EventMemberUpdate: Node-1932ccac-05cc-dc57-693c-5632e301f7f5.dc1 === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.972Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:20.973Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:20.974Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-fd06f590-dd4d-f483-dda5-b53904af22e2 === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.970Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.serf.lan: serf: EventMemberUpdate: Node-1932ccac-05cc-dc57-693c-5632e301f7f5 === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:20.975Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: member joined, marking health alive: member=Node-fd06f590-dd4d-f483-dda5-b53904af22e2 === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:20.970Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: Handled event for server in area: event=member-update server=Node-1932ccac-05cc-dc57-693c-5632e301f7f5.dc1 area=wan writer.go:29: 2021-01-29T19:31:20.971Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: Updating LAN server: server="Node-1932ccac-05cc-dc57-693c-5632e301f7f5 (Addr: tcp/127.0.0.1:29441) (DC: dc1)" writer.go:29: 2021-01-29T19:31:20.979Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: Updating LAN server: server="Node-1932ccac-05cc-dc57-693c-5632e301f7f5 (Addr: tcp/127.0.0.1:29441) (DC: dc1)" writer.go:29: 2021-01-29T19:31:20.986Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.serf.wan: serf: EventMemberUpdate: Node-1932ccac-05cc-dc57-693c-5632e301f7f5.dc1 writer.go:29: 2021-01-29T19:31:21.001Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: Handled event for server in area: event=member-update server=Node-1932ccac-05cc-dc57-693c-5632e301f7f5.dc1 area=wan === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:21.007Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-fd06f590-dd4d-f483-dda5-b53904af22e2 === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:21.031Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:21.078Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Synced node info writer.go:29: 2021-01-29T19:31:21.081Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal: Node info in sync === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:21.082Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:21.078Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:21.096Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Synced node info === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:21.093Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:21.097Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager: Node info in sync === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:21.097Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.server: Skipping self join check for node since the cluster is too small: node=Node-1932ccac-05cc-dc57-693c-5632e301f7f5 writer.go:29: 2021-01-29T19:31:21.101Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: member joined, marking health alive: member=Node-1932ccac-05cc-dc57-693c-5632e301f7f5 writer.go:29: 2021-01-29T19:31:21.113Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.acl: dropping node from result due to ACLs: node=Node-1932ccac-05cc-dc57-693c-5632e301f7f5 === RUN TestAgent_RegisterService_TranslateKeys/normal/no_token === RUN TestAgent_RegisterService_TranslateKeys/normal/root_token === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:21.116Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.server: Skipping self join check for node since the cluster is too small: node=Node-1932ccac-05cc-dc57-693c-5632e301f7f5 writer.go:29: 2021-01-29T19:31:21.123Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.server: Skipping self join check for node since the cluster is too small: node=Node-1932ccac-05cc-dc57-693c-5632e301f7f5 writer.go:29: 2021-01-29T19:31:21.129Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal: Node info in sync writer.go:29: 2021-01-29T19:31:21.131Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Synced service: service=test writer.go:29: 2021-01-29T19:31:21.132Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal: Check in sync: check=service:test:1 writer.go:29: 2021-01-29T19:31:21.133Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal: Check in sync: check=service:test:2 writer.go:29: 2021-01-29T19:31:21.134Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal: Check in sync: check=service:test:3 writer.go:29: 2021-01-29T19:31:21.135Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:21.138Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:21.138Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:21.139Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:21.140Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:21.141Z [WARN] TestAgent_RegisterService_TranslateKeys/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:21.142Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:21.142Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:21.142Z [DEBUG] TestAgent_RegisterService_TranslateKeys/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:21.144Z [WARN] TestAgent_RegisterService_TranslateKeys/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:21.148Z [INFO] TestAgent_RegisterService_TranslateKeys/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:21.148Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: consul server down writer.go:29: 2021-01-29T19:31:21.150Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: shutdown complete writer.go:29: 2021-01-29T19:31:21.151Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Stopping server: protocol=DNS address=127.0.0.1:29436 network=tcp writer.go:29: 2021-01-29T19:31:21.152Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Stopping server: protocol=DNS address=127.0.0.1:29436 network=udp writer.go:29: 2021-01-29T19:31:21.152Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Stopping server: protocol=HTTP address=127.0.0.1:29437 network=tcp === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:21.168Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.acl: dropping node from result due to ACLs: node=Node-fd06f590-dd4d-f483-dda5-b53904af22e2 writer.go:29: 2021-01-29T19:31:21.170Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.acl: dropping node from result due to ACLs: node=Node-fd06f590-dd4d-f483-dda5-b53904af22e2 === RUN TestAgent_RegisterService_TranslateKeys/service_manager/no_token === RUN TestAgent_RegisterService_TranslateKeys/service_manager/root_token === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:21.183Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager: Node info in sync writer.go:29: 2021-01-29T19:31:21.186Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Synced service: service=test writer.go:29: 2021-01-29T19:31:21.188Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager: Check in sync: check=service:test:2 writer.go:29: 2021-01-29T19:31:21.190Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager: Check in sync: check=service:test:3 writer.go:29: 2021-01-29T19:31:21.192Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager: Check in sync: check=service:test:1 writer.go:29: 2021-01-29T19:31:21.194Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:21.196Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:21.198Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:21.199Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:21.201Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:21.202Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:21.202Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:21.207Z [WARN] TestAgent_RegisterService_TranslateKeys/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:21.207Z [DEBUG] TestAgent_RegisterService_TranslateKeys/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:21.218Z [WARN] TestAgent_RegisterService_TranslateKeys/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:21.224Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:21.227Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: consul server down writer.go:29: 2021-01-29T19:31:21.229Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:21.231Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29442 network=tcp writer.go:29: 2021-01-29T19:31:21.234Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29442 network=udp writer.go:29: 2021-01-29T19:31:21.236Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29443 network=tcp === CONT TestAgent_RegisterService_TranslateKeys/normal writer.go:29: 2021-01-29T19:31:21.653Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:21.656Z [INFO] TestAgent_RegisterService_TranslateKeys/normal: Endpoints down === CONT TestAgent_RegisterService_TranslateKeys/service_manager writer.go:29: 2021-01-29T19:31:21.738Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:21.740Z [INFO] TestAgent_RegisterService_TranslateKeys/service_manager: Endpoints down --- PASS: TestAgent_RegisterService_TranslateKeys (0.00s) --- PASS: TestAgent_RegisterService_TranslateKeys/normal (0.86s) --- PASS: TestAgent_RegisterService_TranslateKeys/normal/no_token (0.00s) --- PASS: TestAgent_RegisterService_TranslateKeys/normal/root_token (0.01s) --- PASS: TestAgent_RegisterService_TranslateKeys/service_manager (0.95s) --- PASS: TestAgent_RegisterService_TranslateKeys/service_manager/no_token (0.00s) --- PASS: TestAgent_RegisterService_TranslateKeys/service_manager/root_token (0.02s) === RUN TestAgent_RegisterService_ACLDeny === RUN TestAgent_RegisterService_ACLDeny/normal === PAUSE TestAgent_RegisterService_ACLDeny/normal === RUN TestAgent_RegisterService_ACLDeny/service_manager === PAUSE TestAgent_RegisterService_ACLDeny/service_manager === CONT TestAgent_RegisterService_ACLDeny/normal === CONT TestAgent_RegisterService_ACLDeny/service_manager === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.780Z [WARN] TestAgent_RegisterService_ACLDeny/normal: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:21.782Z [WARN] TestAgent_RegisterService_ACLDeny/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:21.784Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:21.802Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.811Z [WARN] TestAgent_RegisterService_ACLDeny/service_manager: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:21.812Z [WARN] TestAgent_RegisterService_ACLDeny/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:21.816Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:21.817Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.820Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6a90fc70-d3ea-6e5e-8a93-638423643c0a Address:127.0.0.1:29453}]" === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.823Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f2448dee-efe3-5ea0-c327-da8e265c7037 Address:127.0.0.1:29459}]" writer.go:29: 2021-01-29T19:31:21.827Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29459 [Follower]" leader= === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.823Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29453 [Follower]" leader= === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.832Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.serf.wan: serf: EventMemberJoin: Node-f2448dee-efe3-5ea0-c327-da8e265c7037.dc1 127.0.0.1 === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.830Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.serf.wan: serf: EventMemberJoin: Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:21.840Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.serf.lan: serf: EventMemberJoin: Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a 127.0.0.1 === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.841Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.serf.lan: serf: EventMemberJoin: Node-f2448dee-efe3-5ea0-c327-da8e265c7037 127.0.0.1 === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.844Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: Handled event for server in area: event=member-join server=Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a.dc1 area=wan writer.go:29: 2021-01-29T19:31:21.844Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: Adding LAN server: server="Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a (Addr: tcp/127.0.0.1:29453) (DC: dc1)" === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.846Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: Handled event for server in area: event=member-join server=Node-f2448dee-efe3-5ea0-c327-da8e265c7037.dc1 area=wan writer.go:29: 2021-01-29T19:31:21.847Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: Adding LAN server: server="Node-f2448dee-efe3-5ea0-c327-da8e265c7037 (Addr: tcp/127.0.0.1:29459) (DC: dc1)" === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.849Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Started DNS server: address=127.0.0.1:29448 network=tcp === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.852Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Started DNS server: address=127.0.0.1:29454 network=tcp === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.854Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Started DNS server: address=127.0.0.1:29448 network=udp === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.855Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Started DNS server: address=127.0.0.1:29454 network=udp === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.858Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Started HTTP server: address=127.0.0.1:29449 network=tcp writer.go:29: 2021-01-29T19:31:21.860Z [INFO] TestAgent_RegisterService_ACLDeny/normal: started state syncer === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.863Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Started HTTP server: address=127.0.0.1:29455 network=tcp writer.go:29: 2021-01-29T19:31:21.865Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: started state syncer writer.go:29: 2021-01-29T19:31:21.880Z [WARN] TestAgent_RegisterService_ACLDeny/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:21.881Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29459 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:21.883Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:21.884Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.server.raft: vote granted: from=f2448dee-efe3-5ea0-c327-da8e265c7037 term=2 tally=1 writer.go:29: 2021-01-29T19:31:21.885Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:21.885Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29459 [Leader]" writer.go:29: 2021-01-29T19:31:21.886Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:21.887Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: New leader elected: payload=Node-f2448dee-efe3-5ea0-c327-da8e265c7037 writer.go:29: 2021-01-29T19:31:21.888Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: initializing acls writer.go:29: 2021-01-29T19:31:21.891Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:21.893Z [WARN] TestAgent_RegisterService_ACLDeny/service_manager.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.894Z [WARN] TestAgent_RegisterService_ACLDeny/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:21.895Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29453 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:21.896Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:21.897Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:21.899Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.server.raft: vote granted: from=6a90fc70-d3ea-6e5e-8a93-638423643c0a term=2 tally=1 writer.go:29: 2021-01-29T19:31:21.900Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:21.901Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29453 [Leader]" writer.go:29: 2021-01-29T19:31:21.902Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:21.903Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: New leader elected: payload=Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.895Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:21.901Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: initializing acls writer.go:29: 2021-01-29T19:31:21.907Z [WARN] TestAgent_RegisterService_ACLDeny/service_manager.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:21.907Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:21.909Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Node info in sync writer.go:29: 2021-01-29T19:31:21.910Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: Created ACL anonymous token from configuration === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.910Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: initializing acls === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.911Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:21.912Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:21.913Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.serf.lan: serf: EventMemberUpdate: Node-f2448dee-efe3-5ea0-c327-da8e265c7037 === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.913Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:21.914Z [WARN] TestAgent_RegisterService_ACLDeny/normal.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.914Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.serf.wan: serf: EventMemberUpdate: Node-f2448dee-efe3-5ea0-c327-da8e265c7037.dc1 writer.go:29: 2021-01-29T19:31:21.909Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:21.916Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:21.916Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: Updating LAN server: server="Node-f2448dee-efe3-5ea0-c327-da8e265c7037 (Addr: tcp/127.0.0.1:29459) (DC: dc1)" writer.go:29: 2021-01-29T19:31:21.916Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: Handled event for server in area: event=member-update server=Node-f2448dee-efe3-5ea0-c327-da8e265c7037.dc1 area=wan === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.917Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: Bootstrapped ACL master token from configuration === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.917Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.serf.lan: serf: EventMemberUpdate: Node-f2448dee-efe3-5ea0-c327-da8e265c7037 writer.go:29: 2021-01-29T19:31:21.919Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.serf.wan: serf: EventMemberUpdate: Node-f2448dee-efe3-5ea0-c327-da8e265c7037.dc1 === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.920Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: Created ACL anonymous token from configuration === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.921Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: Handled event for server in area: event=member-update server=Node-f2448dee-efe3-5ea0-c327-da8e265c7037.dc1 area=wan writer.go:29: 2021-01-29T19:31:21.921Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: Updating LAN server: server="Node-f2448dee-efe3-5ea0-c327-da8e265c7037 (Addr: tcp/127.0.0.1:29459) (DC: dc1)" === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.921Z [INFO] TestAgent_RegisterService_ACLDeny/normal.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:21.923Z [INFO] TestAgent_RegisterService_ACLDeny/normal.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:21.925Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.serf.lan: serf: EventMemberUpdate: Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a writer.go:29: 2021-01-29T19:31:21.927Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.serf.wan: serf: EventMemberUpdate: Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a.dc1 writer.go:29: 2021-01-29T19:31:21.931Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: Updating LAN server: server="Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a (Addr: tcp/127.0.0.1:29453) (DC: dc1)" writer.go:29: 2021-01-29T19:31:21.931Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: Handled event for server in area: event=member-update server=Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a.dc1 area=wan === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.934Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:21.940Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:21.941Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:21.943Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-f2448dee-efe3-5ea0-c327-da8e265c7037 === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.943Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.946Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: member joined, marking health alive: member=Node-f2448dee-efe3-5ea0-c327-da8e265c7037 === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.953Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.953Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-f2448dee-efe3-5ea0-c327-da8e265c7037 === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.955Z [INFO] TestAgent_RegisterService_ACLDeny/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:21.955Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-f2448dee-efe3-5ea0-c327-da8e265c7037 === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.956Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.server: Skipping self join check for node since the cluster is too small: node=Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a writer.go:29: 2021-01-29T19:31:21.957Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: member joined, marking health alive: member=Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a writer.go:29: 2021-01-29T19:31:21.959Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.server: Skipping self join check for node since the cluster is too small: node=Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a writer.go:29: 2021-01-29T19:31:21.977Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.acl: dropping node from result due to ACLs: node=Node-6a90fc70-d3ea-6e5e-8a93-638423643c0a === RUN TestAgent_RegisterService_ACLDeny/normal/no_token === RUN TestAgent_RegisterService_ACLDeny/normal/root_token === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:21.989Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Synced node info writer.go:29: 2021-01-29T19:31:21.990Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Synced service: service=test writer.go:29: 2021-01-29T19:31:21.991Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal: Check in sync: check=service:test:1 writer.go:29: 2021-01-29T19:31:21.992Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal: Check in sync: check=service:test:2 writer.go:29: 2021-01-29T19:31:21.993Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal: Check in sync: check=service:test:3 writer.go:29: 2021-01-29T19:31:21.994Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:21.995Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:21.996Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:21.997Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:21.998Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:21.998Z [WARN] TestAgent_RegisterService_ACLDeny/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:21.996Z [ERROR] TestAgent_RegisterService_ACLDeny/normal.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:22.000Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:22.001Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:22.001Z [DEBUG] TestAgent_RegisterService_ACLDeny/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:22.002Z [WARN] TestAgent_RegisterService_ACLDeny/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:22.006Z [INFO] TestAgent_RegisterService_ACLDeny/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:22.006Z [INFO] TestAgent_RegisterService_ACLDeny/normal: consul server down writer.go:29: 2021-01-29T19:31:22.010Z [INFO] TestAgent_RegisterService_ACLDeny/normal: shutdown complete writer.go:29: 2021-01-29T19:31:22.012Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Stopping server: protocol=DNS address=127.0.0.1:29448 network=tcp writer.go:29: 2021-01-29T19:31:22.013Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Stopping server: protocol=DNS address=127.0.0.1:29448 network=udp writer.go:29: 2021-01-29T19:31:22.015Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Stopping server: protocol=HTTP address=127.0.0.1:29449 network=tcp === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:22.112Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.acl: dropping node from result due to ACLs: node=Node-f2448dee-efe3-5ea0-c327-da8e265c7037 writer.go:29: 2021-01-29T19:31:22.115Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.acl: dropping node from result due to ACLs: node=Node-f2448dee-efe3-5ea0-c327-da8e265c7037 === RUN TestAgent_RegisterService_ACLDeny/service_manager/no_token === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:22.116Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:22.117Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Node info in sync === RUN TestAgent_RegisterService_ACLDeny/service_manager/root_token === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:22.123Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Node info in sync writer.go:29: 2021-01-29T19:31:22.125Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Synced service: service=test writer.go:29: 2021-01-29T19:31:22.127Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Check in sync: check=service:test:1 writer.go:29: 2021-01-29T19:31:22.129Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Check in sync: check=service:test:2 writer.go:29: 2021-01-29T19:31:22.130Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Check in sync: check=service:test:3 writer.go:29: 2021-01-29T19:31:22.132Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Node info in sync writer.go:29: 2021-01-29T19:31:22.134Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Service in sync: service=test writer.go:29: 2021-01-29T19:31:22.135Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Check in sync: check=service:test:1 writer.go:29: 2021-01-29T19:31:22.137Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Check in sync: check=service:test:2 writer.go:29: 2021-01-29T19:31:22.138Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager: Check in sync: check=service:test:3 writer.go:29: 2021-01-29T19:31:22.140Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:22.142Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:22.145Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:22.146Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:22.147Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:22.149Z [WARN] TestAgent_RegisterService_ACLDeny/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:22.148Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:22.148Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:22.151Z [DEBUG] TestAgent_RegisterService_ACLDeny/service_manager.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:22.153Z [WARN] TestAgent_RegisterService_ACLDeny/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:22.166Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:22.168Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: consul server down writer.go:29: 2021-01-29T19:31:22.170Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:22.172Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29454 network=tcp writer.go:29: 2021-01-29T19:31:22.175Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29454 network=udp writer.go:29: 2021-01-29T19:31:22.178Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29455 network=tcp === CONT TestAgent_RegisterService_ACLDeny/normal writer.go:29: 2021-01-29T19:31:22.517Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:22.518Z [INFO] TestAgent_RegisterService_ACLDeny/normal: Endpoints down === CONT TestAgent_RegisterService_ACLDeny/service_manager writer.go:29: 2021-01-29T19:31:22.680Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:22.681Z [INFO] TestAgent_RegisterService_ACLDeny/service_manager: Endpoints down --- PASS: TestAgent_RegisterService_ACLDeny (0.00s) --- PASS: TestAgent_RegisterService_ACLDeny/normal (0.77s) --- PASS: TestAgent_RegisterService_ACLDeny/normal/no_token (0.00s) --- PASS: TestAgent_RegisterService_ACLDeny/normal/root_token (0.01s) --- PASS: TestAgent_RegisterService_ACLDeny/service_manager (0.93s) --- PASS: TestAgent_RegisterService_ACLDeny/service_manager/no_token (0.00s) --- PASS: TestAgent_RegisterService_ACLDeny/service_manager/root_token (0.01s) === RUN TestAgent_RegisterService_InvalidAddress === RUN TestAgent_RegisterService_InvalidAddress/normal === PAUSE TestAgent_RegisterService_InvalidAddress/normal === RUN TestAgent_RegisterService_InvalidAddress/service_manager === PAUSE TestAgent_RegisterService_InvalidAddress/service_manager === CONT TestAgent_RegisterService_InvalidAddress/normal === CONT TestAgent_RegisterService_InvalidAddress/service_manager === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.740Z [WARN] TestAgent_RegisterService_InvalidAddress/normal: bootstrap = true: do not enable unless necessary === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.746Z [WARN] TestAgent_RegisterService_InvalidAddress/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:22.747Z [DEBUG] TestAgent_RegisterService_InvalidAddress/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:22.749Z [DEBUG] TestAgent_RegisterService_InvalidAddress/service_manager.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:22.758Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:dfea3f3f-e16c-8728-8bbb-0ef73dbdfda1 Address:127.0.0.1:29471}]" writer.go:29: 2021-01-29T19:31:22.760Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29471 [Follower]" leader= writer.go:29: 2021-01-29T19:31:22.760Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server.serf.wan: serf: EventMemberJoin: Node-dfea3f3f-e16c-8728-8bbb-0ef73dbdfda1.dc1 127.0.0.1 === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.742Z [DEBUG] TestAgent_RegisterService_InvalidAddress/normal.tlsutil: Update: version=1 === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.766Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server.serf.lan: serf: EventMemberJoin: Node-dfea3f3f-e16c-8728-8bbb-0ef73dbdfda1 127.0.0.1 writer.go:29: 2021-01-29T19:31:22.774Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server: Adding LAN server: server="Node-dfea3f3f-e16c-8728-8bbb-0ef73dbdfda1 (Addr: tcp/127.0.0.1:29471) (DC: dc1)" === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.768Z [DEBUG] TestAgent_RegisterService_InvalidAddress/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.776Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server: Handled event for server in area: event=member-join server=Node-dfea3f3f-e16c-8728-8bbb-0ef73dbdfda1.dc1 area=wan writer.go:29: 2021-01-29T19:31:22.785Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: Started DNS server: address=127.0.0.1:29466 network=tcp writer.go:29: 2021-01-29T19:31:22.792Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: Started DNS server: address=127.0.0.1:29466 network=udp === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.789Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ff4a4651-4692-a6c6-fe60-b7cc598077d2 Address:127.0.0.1:29465}]" === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.798Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: Started HTTP server: address=127.0.0.1:29467 network=tcp writer.go:29: 2021-01-29T19:31:22.800Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: started state syncer === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.799Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29465 [Follower]" leader= writer.go:29: 2021-01-29T19:31:22.803Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server.serf.wan: serf: EventMemberJoin: Node-ff4a4651-4692-a6c6-fe60-b7cc598077d2.dc1 127.0.0.1 === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.804Z [WARN] TestAgent_RegisterService_InvalidAddress/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:22.806Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29471 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:22.811Z [DEBUG] TestAgent_RegisterService_InvalidAddress/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:22.814Z [DEBUG] TestAgent_RegisterService_InvalidAddress/service_manager.server.raft: vote granted: from=dfea3f3f-e16c-8728-8bbb-0ef73dbdfda1 term=2 tally=1 === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.820Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server.serf.lan: serf: EventMemberJoin: Node-ff4a4651-4692-a6c6-fe60-b7cc598077d2 127.0.0.1 === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.822Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server.raft: election won: tally=1 === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.825Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server: Adding LAN server: server="Node-ff4a4651-4692-a6c6-fe60-b7cc598077d2 (Addr: tcp/127.0.0.1:29465) (DC: dc1)" writer.go:29: 2021-01-29T19:31:22.826Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server: Handled event for server in area: event=member-join server=Node-ff4a4651-4692-a6c6-fe60-b7cc598077d2.dc1 area=wan === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.825Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29471 [Leader]" writer.go:29: 2021-01-29T19:31:22.829Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:22.829Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server: New leader elected: payload=Node-dfea3f3f-e16c-8728-8bbb-0ef73dbdfda1 === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.835Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Started DNS server: address=127.0.0.1:29460 network=tcp writer.go:29: 2021-01-29T19:31:22.839Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Started DNS server: address=127.0.0.1:29460 network=udp === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.843Z [DEBUG] TestAgent_RegisterService_InvalidAddress/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29471 === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.845Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Started HTTP server: address=127.0.0.1:29461 network=tcp writer.go:29: 2021-01-29T19:31:22.852Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: started state syncer === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.866Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.869Z [WARN] TestAgent_RegisterService_InvalidAddress/normal.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.871Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.871Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29465 [Candidate]" term=2 === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.872Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:22.873Z [DEBUG] TestAgent_RegisterService_InvalidAddress/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-dfea3f3f-e16c-8728-8bbb-0ef73dbdfda1 writer.go:29: 2021-01-29T19:31:22.874Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server: member joined, marking health alive: member=Node-dfea3f3f-e16c-8728-8bbb-0ef73dbdfda1 === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.873Z [DEBUG] TestAgent_RegisterService_InvalidAddress/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:22.877Z [DEBUG] TestAgent_RegisterService_InvalidAddress/normal.server.raft: vote granted: from=ff4a4651-4692-a6c6-fe60-b7cc598077d2 term=2 tally=1 writer.go:29: 2021-01-29T19:31:22.878Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:22.892Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29465 [Leader]" writer.go:29: 2021-01-29T19:31:22.901Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:22.903Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server: New leader elected: payload=Node-ff4a4651-4692-a6c6-fe60-b7cc598077d2 writer.go:29: 2021-01-29T19:31:22.917Z [DEBUG] TestAgent_RegisterService_InvalidAddress/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29465 === RUN TestAgent_RegisterService_InvalidAddress/service_manager/addr_0.0.0.0 === RUN TestAgent_RegisterService_InvalidAddress/service_manager/addr_:: === RUN TestAgent_RegisterService_InvalidAddress/service_manager/addr_[::] === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.936Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.940Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:22.942Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:22.943Z [DEBUG] TestAgent_RegisterService_InvalidAddress/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:22.942Z [DEBUG] TestAgent_RegisterService_InvalidAddress/service_manager: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:22.945Z [WARN] TestAgent_RegisterService_InvalidAddress/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:22.946Z [DEBUG] TestAgent_RegisterService_InvalidAddress/service_manager.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.949Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.949Z [WARN] TestAgent_RegisterService_InvalidAddress/service_manager.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.951Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.949Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:22.959Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:22.959Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: consul server down writer.go:29: 2021-01-29T19:31:22.963Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: shutdown complete === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.953Z [DEBUG] TestAgent_RegisterService_InvalidAddress/normal.server: Skipping self join check for node since the cluster is too small: node=Node-ff4a4651-4692-a6c6-fe60-b7cc598077d2 === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.964Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29466 network=tcp writer.go:29: 2021-01-29T19:31:22.966Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29466 network=udp === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:22.965Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server: member joined, marking health alive: member=Node-ff4a4651-4692-a6c6-fe60-b7cc598077d2 === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:22.968Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29467 network=tcp === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:23.042Z [DEBUG] TestAgent_RegisterService_InvalidAddress/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:23.046Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Synced node info writer.go:29: 2021-01-29T19:31:23.295Z [DEBUG] TestAgent_RegisterService_InvalidAddress/normal: Node info in sync writer.go:29: 2021-01-29T19:31:23.296Z [ERROR] TestAgent_RegisterService_InvalidAddress/normal.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:23.299Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Synced service: service=connect-proxy writer.go:29: 2021-01-29T19:31:23.310Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:23.312Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:23.314Z [DEBUG] TestAgent_RegisterService_InvalidAddress/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:23.315Z [WARN] TestAgent_RegisterService_InvalidAddress/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:23.315Z [DEBUG] TestAgent_RegisterService_InvalidAddress/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:23.318Z [WARN] TestAgent_RegisterService_InvalidAddress/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:23.323Z [INFO] TestAgent_RegisterService_InvalidAddress/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:23.323Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: consul server down writer.go:29: 2021-01-29T19:31:23.327Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: shutdown complete writer.go:29: 2021-01-29T19:31:23.328Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Stopping server: protocol=DNS address=127.0.0.1:29460 network=tcp writer.go:29: 2021-01-29T19:31:23.329Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Stopping server: protocol=DNS address=127.0.0.1:29460 network=udp writer.go:29: 2021-01-29T19:31:23.331Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Stopping server: protocol=HTTP address=127.0.0.1:29461 network=tcp === CONT TestAgent_RegisterService_InvalidAddress/service_manager writer.go:29: 2021-01-29T19:31:23.469Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:23.470Z [INFO] TestAgent_RegisterService_InvalidAddress/service_manager: Endpoints down === CONT TestAgent_RegisterService_InvalidAddress/normal writer.go:29: 2021-01-29T19:31:23.832Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:23.835Z [INFO] TestAgent_RegisterService_InvalidAddress/normal: Endpoints down --- PASS: TestAgent_RegisterService_InvalidAddress (0.00s) --- PASS: TestAgent_RegisterService_InvalidAddress/service_manager (0.78s) --- PASS: TestAgent_RegisterService_InvalidAddress/service_manager/addr_0.0.0.0 (0.01s) --- PASS: TestAgent_RegisterService_InvalidAddress/service_manager/addr_:: (0.00s) --- PASS: TestAgent_RegisterService_InvalidAddress/service_manager/addr_[::] (0.00s) --- PASS: TestAgent_RegisterService_InvalidAddress/normal (1.15s) === RUN TestAgent_RegisterService_UnmanagedConnectProxy === RUN TestAgent_RegisterService_UnmanagedConnectProxy/normal === PAUSE TestAgent_RegisterService_UnmanagedConnectProxy/normal === RUN TestAgent_RegisterService_UnmanagedConnectProxy/service_manager === PAUSE TestAgent_RegisterService_UnmanagedConnectProxy/service_manager === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:23.865Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxy/normal: bootstrap = true: do not enable unless necessary === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:23.866Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:23.874Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:23.878Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:23.882Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:23.886Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/normal.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:23.910Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c5520803-89af-a68e-88df-503d55fa40e9 Address:127.0.0.1:29477}]" writer.go:29: 2021-01-29T19:31:23.916Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29477 [Follower]" leader= === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:23.916Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:be5bbd5c-c790-8148-1114-448856ed63a5 Address:127.0.0.1:29483}]" === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:23.916Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.serf.wan: serf: EventMemberJoin: Node-c5520803-89af-a68e-88df-503d55fa40e9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:23.921Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.serf.lan: serf: EventMemberJoin: Node-c5520803-89af-a68e-88df-503d55fa40e9 127.0.0.1 === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:23.919Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29483 [Follower]" leader= === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:23.925Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Started DNS server: address=127.0.0.1:29472 network=udp === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:23.919Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.serf.wan: serf: EventMemberJoin: Node-be5bbd5c-c790-8148-1114-448856ed63a5.dc1 127.0.0.1 === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:23.925Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server: Adding LAN server: server="Node-c5520803-89af-a68e-88df-503d55fa40e9 (Addr: tcp/127.0.0.1:29477) (DC: dc1)" writer.go:29: 2021-01-29T19:31:23.926Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server: Handled event for server in area: event=member-join server=Node-c5520803-89af-a68e-88df-503d55fa40e9.dc1 area=wan writer.go:29: 2021-01-29T19:31:23.927Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Started DNS server: address=127.0.0.1:29472 network=tcp writer.go:29: 2021-01-29T19:31:23.936Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Started HTTP server: address=127.0.0.1:29473 network=tcp writer.go:29: 2021-01-29T19:31:23.938Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: started state syncer === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:23.941Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.serf.lan: serf: EventMemberJoin: Node-be5bbd5c-c790-8148-1114-448856ed63a5 127.0.0.1 writer.go:29: 2021-01-29T19:31:23.945Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Started DNS server: address=127.0.0.1:29478 network=udp writer.go:29: 2021-01-29T19:31:23.946Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server: Adding LAN server: server="Node-be5bbd5c-c790-8148-1114-448856ed63a5 (Addr: tcp/127.0.0.1:29483) (DC: dc1)" writer.go:29: 2021-01-29T19:31:23.946Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server: Handled event for server in area: event=member-join server=Node-be5bbd5c-c790-8148-1114-448856ed63a5.dc1 area=wan writer.go:29: 2021-01-29T19:31:23.957Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Started DNS server: address=127.0.0.1:29478 network=tcp writer.go:29: 2021-01-29T19:31:23.959Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Started HTTP server: address=127.0.0.1:29479 network=tcp writer.go:29: 2021-01-29T19:31:23.961Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: started state syncer === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:23.968Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:23.969Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29477 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:23.971Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:23.972Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.raft: vote granted: from=c5520803-89af-a68e-88df-503d55fa40e9 term=2 tally=1 writer.go:29: 2021-01-29T19:31:23.972Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:23.973Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29477 [Leader]" writer.go:29: 2021-01-29T19:31:23.974Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:23.975Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server: New leader elected: payload=Node-c5520803-89af-a68e-88df-503d55fa40e9 writer.go:29: 2021-01-29T19:31:23.976Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29477 === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:23.981Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:23.982Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29483 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:23.985Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:23.986Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.raft: vote granted: from=be5bbd5c-c790-8148-1114-448856ed63a5 term=2 tally=1 writer.go:29: 2021-01-29T19:31:23.988Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:23.989Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29483 [Leader]" writer.go:29: 2021-01-29T19:31:23.991Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:23.992Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server: New leader elected: payload=Node-be5bbd5c-c790-8148-1114-448856ed63a5 === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:23.993Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:23.993Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29483 === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:23.999Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:24.004Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:24.003Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:24.006Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:24.007Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Node info in sync writer.go:29: 2021-01-29T19:31:24.009Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:24.011Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:24.007Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server: Skipping self join check for node since the cluster is too small: node=Node-c5520803-89af-a68e-88df-503d55fa40e9 === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:24.012Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-be5bbd5c-c790-8148-1114-448856ed63a5 writer.go:29: 2021-01-29T19:31:24.014Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server: member joined, marking health alive: member=Node-be5bbd5c-c790-8148-1114-448856ed63a5 === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:24.012Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server: member joined, marking health alive: member=Node-c5520803-89af-a68e-88df-503d55fa40e9 writer.go:29: 2021-01-29T19:31:24.260Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:24.263Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Synced node info writer.go:29: 2021-01-29T19:31:24.303Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Node info in sync writer.go:29: 2021-01-29T19:31:24.305Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Synced service: service=connect-proxy writer.go:29: 2021-01-29T19:31:24.309Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:24.310Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:24.311Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:24.322Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:24.322Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:24.325Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:24.327Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:24.327Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: consul server down writer.go:29: 2021-01-29T19:31:24.330Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: shutdown complete writer.go:29: 2021-01-29T19:31:24.331Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Stopping server: protocol=DNS address=127.0.0.1:29472 network=tcp writer.go:29: 2021-01-29T19:31:24.332Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Stopping server: protocol=DNS address=127.0.0.1:29472 network=udp writer.go:29: 2021-01-29T19:31:24.333Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Stopping server: protocol=HTTP address=127.0.0.1:29473 network=tcp === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:24.373Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Node info in sync writer.go:29: 2021-01-29T19:31:24.389Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Synced service: service=connect-proxy writer.go:29: 2021-01-29T19:31:24.394Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:24.395Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:24.396Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:24.394Z [ERROR] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:24.397Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:24.397Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:24.412Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:24.416Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:24.417Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: consul server down writer.go:29: 2021-01-29T19:31:24.419Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:24.420Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29478 network=tcp writer.go:29: 2021-01-29T19:31:24.421Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29478 network=udp writer.go:29: 2021-01-29T19:31:24.423Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29479 network=tcp === CONT TestAgent_RegisterService_UnmanagedConnectProxy/normal writer.go:29: 2021-01-29T19:31:24.835Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:24.837Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/normal: Endpoints down === CONT TestAgent_RegisterService_UnmanagedConnectProxy/service_manager writer.go:29: 2021-01-29T19:31:24.925Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:24.926Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxy/service_manager: Endpoints down --- PASS: TestAgent_RegisterService_UnmanagedConnectProxy (0.00s) --- PASS: TestAgent_RegisterService_UnmanagedConnectProxy/normal (1.00s) --- PASS: TestAgent_RegisterService_UnmanagedConnectProxy/service_manager (1.09s) === RUN TestAgent_RegisterServiceDeregisterService_Sidecar === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal === PAUSE TestAgent_RegisterServiceDeregisterService_Sidecar/normal === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager === PAUSE TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:24.966Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: bootstrap = true: do not enable unless necessary === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:24.967Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: bootstrap = true: do not enable unless necessary === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:24.967Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.tlsutil: Update: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:24.967Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:24.969Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:24.968Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:24.976Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e04634bd-3675-3cd6-956a-6ed670e9362a Address:127.0.0.1:29489}]" writer.go:29: 2021-01-29T19:31:24.978Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.raft: entering follower state: follower="Node at 127.0.0.1:29489 [Follower]" leader= === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:24.975Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:05f3a9db-4be3-8b8c-114f-ca9cf08d96a8 Address:127.0.0.1:29495}]" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:24.986Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.serf.wan: serf: EventMemberJoin: Node-e04634bd-3675-3cd6-956a-6ed670e9362a.dc1 127.0.0.1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:24.987Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.raft: entering follower state: follower="Node at 127.0.0.1:29495 [Follower]" leader= writer.go:29: 2021-01-29T19:31:24.990Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.serf.wan: serf: EventMemberJoin: Node-05f3a9db-4be3-8b8c-114f-ca9cf08d96a8.dc1 127.0.0.1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:24.996Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.serf.lan: serf: EventMemberJoin: Node-e04634bd-3675-3cd6-956a-6ed670e9362a 127.0.0.1 writer.go:29: 2021-01-29T19:31:25.003Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server: Adding LAN server: server="Node-e04634bd-3675-3cd6-956a-6ed670e9362a (Addr: tcp/127.0.0.1:29489) (DC: dc1)" writer.go:29: 2021-01-29T19:31:25.004Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server: Handled event for server in area: event=member-join server=Node-e04634bd-3675-3cd6-956a-6ed670e9362a.dc1 area=wan === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.003Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.serf.lan: serf: EventMemberJoin: Node-05f3a9db-4be3-8b8c-114f-ca9cf08d96a8 127.0.0.1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.008Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Started DNS server: address=127.0.0.1:29484 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.012Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server: Adding LAN server: server="Node-05f3a9db-4be3-8b8c-114f-ca9cf08d96a8 (Addr: tcp/127.0.0.1:29495) (DC: dc1)" writer.go:29: 2021-01-29T19:31:25.013Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server: Handled event for server in area: event=member-join server=Node-05f3a9db-4be3-8b8c-114f-ca9cf08d96a8.dc1 area=wan === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.011Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Started DNS server: address=127.0.0.1:29484 network=udp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.018Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Started DNS server: address=127.0.0.1:29490 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.022Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Started HTTP server: address=127.0.0.1:29485 network=tcp writer.go:29: 2021-01-29T19:31:25.023Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: started state syncer === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.020Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Started DNS server: address=127.0.0.1:29490 network=udp writer.go:29: 2021-01-29T19:31:25.026Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Started HTTP server: address=127.0.0.1:29491 network=tcp writer.go:29: 2021-01-29T19:31:25.028Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: started state syncer === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.051Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:25.053Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.raft: entering candidate state: node="Node at 127.0.0.1:29489 [Candidate]" term=2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.055Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:25.057Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.raft: entering candidate state: node="Node at 127.0.0.1:29495 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:25.064Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.raft: votes: needed=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.065Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.raft: votes: needed=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.066Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.raft: vote granted: from=05f3a9db-4be3-8b8c-114f-ca9cf08d96a8 term=2 tally=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.068Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.raft: vote granted: from=e04634bd-3675-3cd6-956a-6ed670e9362a term=2 tally=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.069Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.raft: election won: tally=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.071Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.raft: election won: tally=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.072Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.raft: entering leader state: leader="Node at 127.0.0.1:29495 [Leader]" writer.go:29: 2021-01-29T19:31:25.076Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:25.078Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server: New leader elected: payload=Node-05f3a9db-4be3-8b8c-114f-ca9cf08d96a8 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.075Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.raft: entering leader state: leader="Node at 127.0.0.1:29489 [Leader]" writer.go:29: 2021-01-29T19:31:25.080Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:25.083Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server: New leader elected: payload=Node-e04634bd-3675-3cd6-956a-6ed670e9362a === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.085Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29495 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.087Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29489 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.097Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.098Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.106Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:25.108Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:25.109Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server: Skipping self join check for node since the cluster is too small: node=Node-05f3a9db-4be3-8b8c-114f-ca9cf08d96a8 writer.go:29: 2021-01-29T19:31:25.110Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server: member joined, marking health alive: member=Node-05f3a9db-4be3-8b8c-114f-ca9cf08d96a8 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.123Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:25.127Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:25.132Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server: Skipping self join check for node since the cluster is too small: node=Node-e04634bd-3675-3cd6-956a-6ed670e9362a writer.go:29: 2021-01-29T19:31:25.135Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server: member joined, marking health alive: member=Node-e04634bd-3675-3cd6-956a-6ed670e9362a === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.208Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:25.216Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Synced node info === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.267Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Synced node info writer.go:29: 2021-01-29T19:31:25.270Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Synced service: service=web writer.go:29: 2021-01-29T19:31:25.279Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Requesting shutdown writer.go:29: 2021-01-29T19:31:25.280Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server: shutting down server writer.go:29: 2021-01-29T19:31:25.281Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:25.282Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:25.282Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:25.282Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:25.288Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:25.296Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:25.300Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: consul server down writer.go:29: 2021-01-29T19:31:25.304Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: shutdown complete writer.go:29: 2021-01-29T19:31:25.307Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Stopping server: protocol=DNS address=127.0.0.1:29484 network=tcp writer.go:29: 2021-01-29T19:31:25.311Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Stopping server: protocol=DNS address=127.0.0.1:29484 network=udp writer.go:29: 2021-01-29T19:31:25.314Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Stopping server: protocol=HTTP address=127.0.0.1:29485 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.349Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Node info in sync writer.go:29: 2021-01-29T19:31:25.353Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Synced service: service=web writer.go:29: 2021-01-29T19:31:25.380Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Requesting shutdown writer.go:29: 2021-01-29T19:31:25.383Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server: shutting down server writer.go:29: 2021-01-29T19:31:25.386Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:25.390Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:25.391Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:25.399Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:25.410Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:25.414Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: consul server down writer.go:29: 2021-01-29T19:31:25.417Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: shutdown complete writer.go:29: 2021-01-29T19:31:25.420Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Stopping server: protocol=DNS address=127.0.0.1:29490 network=tcp writer.go:29: 2021-01-29T19:31:25.422Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Stopping server: protocol=DNS address=127.0.0.1:29490 network=udp writer.go:29: 2021-01-29T19:31:25.424Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Stopping server: protocol=HTTP address=127.0.0.1:29491 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.821Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:25.823Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:25.897Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:25.907Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:25.920Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case writer.go:29: 2021-01-29T19:31:25.926Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:25.927Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:25.927Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8d7f4ed4-4c58-7185-33e2-8d04f838ab0f Address:127.0.0.1:29501}]" writer.go:29: 2021-01-29T19:31:25.934Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29501 [Follower]" leader= writer.go:29: 2021-01-29T19:31:25.936Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.serf.wan: serf: EventMemberJoin: Node-8d7f4ed4-4c58-7185-33e2-8d04f838ab0f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:25.947Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.serf.lan: serf: EventMemberJoin: Node-8d7f4ed4-4c58-7185-33e2-8d04f838ab0f 127.0.0.1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:25.955Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: bootstrap = true: do not enable unless necessary === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:25.958Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server: Handled event for server in area: event=member-join server=Node-8d7f4ed4-4c58-7185-33e2-8d04f838ab0f.dc1 area=wan === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:25.957Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.tlsutil: Update: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:25.958Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server: Adding LAN server: server="Node-8d7f4ed4-4c58-7185-33e2-8d04f838ab0f (Addr: tcp/127.0.0.1:29501) (DC: dc1)" writer.go:29: 2021-01-29T19:31:25.958Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Started DNS server: address=127.0.0.1:29496 network=udp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:25.959Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:25.961Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Started DNS server: address=127.0.0.1:29496 network=tcp writer.go:29: 2021-01-29T19:31:25.962Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Started HTTP server: address=127.0.0.1:29497 network=tcp writer.go:29: 2021-01-29T19:31:25.964Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: started state syncer === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:25.979Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b Address:127.0.0.1:29507}]" writer.go:29: 2021-01-29T19:31:25.982Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29507 [Follower]" leader= === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:26.000Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:26.002Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29501 [Candidate]" term=2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:26.000Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.serf.wan: serf: EventMemberJoin: Node-38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b.dc1 127.0.0.1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:26.009Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:26.010Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.raft: vote granted: from=8d7f4ed4-4c58-7185-33e2-8d04f838ab0f term=2 tally=1 writer.go:29: 2021-01-29T19:31:26.011Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:26.012Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29501 [Leader]" writer.go:29: 2021-01-29T19:31:26.013Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:26.013Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server: New leader elected: payload=Node-8d7f4ed4-4c58-7185-33e2-8d04f838ab0f writer.go:29: 2021-01-29T19:31:26.014Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29501 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:26.029Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:26.031Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29507 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:26.032Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.raft: unable to get address for sever, using fallback address: id=38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b fallback=127.0.0.1:29507 error="Could not find address for server id 38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b" writer.go:29: 2021-01-29T19:31:26.033Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.serf.lan: serf: EventMemberJoin: Node-38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b 127.0.0.1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:26.035Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:26.035Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:26.038Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.raft: vote granted: from=38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b term=2 tally=1 writer.go:29: 2021-01-29T19:31:26.039Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:26.038Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Started DNS server: address=127.0.0.1:29502 network=udp writer.go:29: 2021-01-29T19:31:26.038Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server: Adding LAN server: server="Node-38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b (Addr: tcp/127.0.0.1:29507) (DC: dc1)" writer.go:29: 2021-01-29T19:31:26.038Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server: Handled event for server in area: event=member-join server=Node-38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b.dc1 area=wan writer.go:29: 2021-01-29T19:31:26.042Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29507 [Leader]" writer.go:29: 2021-01-29T19:31:26.042Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Started DNS server: address=127.0.0.1:29502 network=tcp writer.go:29: 2021-01-29T19:31:26.044Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:26.045Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server: New leader elected: payload=Node-38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b writer.go:29: 2021-01-29T19:31:26.047Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29507 writer.go:29: 2021-01-29T19:31:26.048Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Started HTTP server: address=127.0.0.1:29503 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:26.049Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:26.051Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:26.054Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Synced node info writer.go:29: 2021-01-29T19:31:26.052Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:26.052Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:26.059Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:26.057Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-8d7f4ed4-4c58-7185-33e2-8d04f838ab0f === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:26.060Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.061Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b writer.go:29: 2021-01-29T19:31:26.062Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server: member joined, marking health alive: member=Node-38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b writer.go:29: 2021-01-29T19:31:26.064Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-38a71d4c-6335-4a3f-07c2-c7ce5c8bb48b === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:26.060Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server: member joined, marking health alive: member=Node-8d7f4ed4-4c58-7185-33e2-8d04f838ab0f === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:26.094Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: added local registration for service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:26.097Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Node info in sync writer.go:29: 2021-01-29T19:31:26.102Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:26.105Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Synced service: service=web writer.go:29: 2021-01-29T19:31:26.109Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:26.113Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:26.142Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: removed service: service=web writer.go:29: 2021-01-29T19:31:26.145Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:26.149Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: removed check: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:26.151Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: removed check: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:26.153Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: removed service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:26.155Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Node info in sync writer.go:29: 2021-01-29T19:31:26.158Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Deregistered service: service=web writer.go:29: 2021-01-29T19:31:26.161Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Deregistered service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:26.168Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:26.170Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:26.171Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.172Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:26.174Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.182Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:26.189Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:26.189Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: consul server down writer.go:29: 2021-01-29T19:31:26.191Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:26.191Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29502 network=tcp writer.go:29: 2021-01-29T19:31:26.192Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29502 network=udp writer.go:29: 2021-01-29T19:31:26.193Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29503 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:26.347Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Synced node info writer.go:29: 2021-01-29T19:31:26.352Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Synced service: service=web writer.go:29: 2021-01-29T19:31:26.382Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:26.393Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:26.399Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:26.407Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: removed service: service=web writer.go:29: 2021-01-29T19:31:26.419Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:26.420Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Node info in sync writer.go:29: 2021-01-29T19:31:26.420Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:26.425Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Deregistered service: service=web writer.go:29: 2021-01-29T19:31:26.441Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:26.442Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:26.444Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:26.446Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: removed check: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:26.448Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: removed check: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:26.450Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: removed service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:26.452Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Node info in sync writer.go:29: 2021-01-29T19:31:26.458Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Deregistered service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:26.461Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:26.472Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:26.473Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.474Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:26.474Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.476Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:26.479Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:26.479Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: consul server down writer.go:29: 2021-01-29T19:31:26.483Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:26.484Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29496 network=tcp writer.go:29: 2021-01-29T19:31:26.486Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29496 network=udp writer.go:29: 2021-01-29T19:31:26.487Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29497 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar writer.go:29: 2021-01-29T19:31:26.694Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:26.695Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:26.705Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:26.706Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:26.707Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:26.708Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:26.728Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9417f46f-20ad-de13-2467-0e1d26998a7e Address:127.0.0.1:29513}]" writer.go:29: 2021-01-29T19:31:26.731Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.raft: entering follower state: follower="Node at 127.0.0.1:29513 [Follower]" leader= writer.go:29: 2021-01-29T19:31:26.736Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.serf.wan: serf: EventMemberJoin: Node-9417f46f-20ad-de13-2467-0e1d26998a7e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:26.739Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.serf.lan: serf: EventMemberJoin: Node-9417f46f-20ad-de13-2467-0e1d26998a7e 127.0.0.1 writer.go:29: 2021-01-29T19:31:26.742Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Handled event for server in area: event=member-join server=Node-9417f46f-20ad-de13-2467-0e1d26998a7e.dc1 area=wan writer.go:29: 2021-01-29T19:31:26.743Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Adding LAN server: server="Node-9417f46f-20ad-de13-2467-0e1d26998a7e (Addr: tcp/127.0.0.1:29513) (DC: dc1)" writer.go:29: 2021-01-29T19:31:26.744Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Started DNS server: address=127.0.0.1:29508 network=tcp writer.go:29: 2021-01-29T19:31:26.747Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Started DNS server: address=127.0.0.1:29508 network=udp writer.go:29: 2021-01-29T19:31:26.749Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Started HTTP server: address=127.0.0.1:29509 network=tcp writer.go:29: 2021-01-29T19:31:26.751Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: started state syncer writer.go:29: 2021-01-29T19:31:26.773Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:26.774Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.raft: entering candidate state: node="Node at 127.0.0.1:29513 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:26.776Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:26.777Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.raft: vote granted: from=9417f46f-20ad-de13-2467-0e1d26998a7e term=2 tally=1 writer.go:29: 2021-01-29T19:31:26.777Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:26.779Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.raft: entering leader state: leader="Node at 127.0.0.1:29513 [Leader]" writer.go:29: 2021-01-29T19:31:26.780Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:26.781Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: New leader elected: payload=Node-9417f46f-20ad-de13-2467-0e1d26998a7e writer.go:29: 2021-01-29T19:31:26.783Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: initializing acls writer.go:29: 2021-01-29T19:31:26.788Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:26.790Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:26.792Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: initializing acls writer.go:29: 2021-01-29T19:31:26.795Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:26.793Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:26.799Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:26.800Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:26.798Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:26.803Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:26.802Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.serf.lan: serf: EventMemberUpdate: Node-9417f46f-20ad-de13-2467-0e1d26998a7e writer.go:29: 2021-01-29T19:31:26.807Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.serf.wan: serf: EventMemberUpdate: Node-9417f46f-20ad-de13-2467-0e1d26998a7e.dc1 writer.go:29: 2021-01-29T19:31:26.807Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Updating LAN server: server="Node-9417f46f-20ad-de13-2467-0e1d26998a7e (Addr: tcp/127.0.0.1:29513) (DC: dc1)" writer.go:29: 2021-01-29T19:31:26.807Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.serf.lan: serf: EventMemberUpdate: Node-9417f46f-20ad-de13-2467-0e1d26998a7e writer.go:29: 2021-01-29T19:31:26.809Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Handled event for server in area: event=member-update server=Node-9417f46f-20ad-de13-2467-0e1d26998a7e.dc1 area=wan writer.go:29: 2021-01-29T19:31:26.810Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:26.813Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Updating LAN server: server="Node-9417f46f-20ad-de13-2467-0e1d26998a7e (Addr: tcp/127.0.0.1:29513) (DC: dc1)" writer.go:29: 2021-01-29T19:31:26.813Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.serf.wan: serf: EventMemberUpdate: Node-9417f46f-20ad-de13-2467-0e1d26998a7e.dc1 writer.go:29: 2021-01-29T19:31:26.823Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Handled event for server in area: event=member-update server=Node-9417f46f-20ad-de13-2467-0e1d26998a7e.dc1 area=wan writer.go:29: 2021-01-29T19:31:26.821Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:26.842Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.844Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Skipping self join check for node since the cluster is too small: node=Node-9417f46f-20ad-de13-2467-0e1d26998a7e writer.go:29: 2021-01-29T19:31:26.845Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: member joined, marking health alive: member=Node-9417f46f-20ad-de13-2467-0e1d26998a7e writer.go:29: 2021-01-29T19:31:26.848Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Skipping self join check for node since the cluster is too small: node=Node-9417f46f-20ad-de13-2467-0e1d26998a7e writer.go:29: 2021-01-29T19:31:26.850Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: Skipping self join check for node since the cluster is too small: node=Node-9417f46f-20ad-de13-2467-0e1d26998a7e writer.go:29: 2021-01-29T19:31:26.959Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:26.962Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Synced node info === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar writer.go:29: 2021-01-29T19:31:26.988Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:26.990Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.014Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:27.015Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:27.017Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:27.042Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:27.063Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d02aebad-00d3-cf93-b87c-5670f3d95591 Address:127.0.0.1:29519}]" writer.go:29: 2021-01-29T19:31:27.066Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.serf.wan: serf: EventMemberJoin: Node-d02aebad-00d3-cf93-b87c-5670f3d95591.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:27.067Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.raft: entering follower state: follower="Node at 127.0.0.1:29519 [Follower]" leader= writer.go:29: 2021-01-29T19:31:27.069Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.serf.lan: serf: EventMemberJoin: Node-d02aebad-00d3-cf93-b87c-5670f3d95591 127.0.0.1 writer.go:29: 2021-01-29T19:31:27.099Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Started DNS server: address=127.0.0.1:29514 network=udp writer.go:29: 2021-01-29T19:31:27.100Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Adding LAN server: server="Node-d02aebad-00d3-cf93-b87c-5670f3d95591 (Addr: tcp/127.0.0.1:29519) (DC: dc1)" writer.go:29: 2021-01-29T19:31:27.102Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Handled event for server in area: event=member-join server=Node-d02aebad-00d3-cf93-b87c-5670f3d95591.dc1 area=wan === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.103Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.acl: dropping node from result due to ACLs: node=Node-9417f46f-20ad-de13-2467-0e1d26998a7e === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.102Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Started DNS server: address=127.0.0.1:29514 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.108Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.acl: dropping node from result due to ACLs: node=Node-9417f46f-20ad-de13-2467-0e1d26998a7e === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.116Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Started HTTP server: address=127.0.0.1:29515 network=tcp writer.go:29: 2021-01-29T19:31:27.117Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: started state syncer writer.go:29: 2021-01-29T19:31:27.118Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:27.119Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.raft: entering candidate state: node="Node at 127.0.0.1:29519 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:27.124Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:27.125Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.raft: vote granted: from=d02aebad-00d3-cf93-b87c-5670f3d95591 term=2 tally=1 writer.go:29: 2021-01-29T19:31:27.126Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:27.128Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.raft: entering leader state: leader="Node at 127.0.0.1:29519 [Leader]" writer.go:29: 2021-01-29T19:31:27.130Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:27.151Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: New leader elected: payload=Node-d02aebad-00d3-cf93-b87c-5670f3d95591 writer.go:29: 2021-01-29T19:31:27.152Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: initializing acls writer.go:29: 2021-01-29T19:31:27.153Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: initializing acls writer.go:29: 2021-01-29T19:31:27.157Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:27.158Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:27.157Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:27.161Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:27.161Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Bootstrapped ACL master token from configuration === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.165Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: added local registration for service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:27.166Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Node info in sync writer.go:29: 2021-01-29T19:31:27.167Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Synced service: service=web writer.go:29: 2021-01-29T19:31:27.169Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:27.170Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:27.170Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Check in sync: check=service:web-sidecar-proxy:2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.175Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:27.176Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:27.177Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:27.176Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:27.178Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:27.179Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.serf.lan: serf: EventMemberUpdate: Node-d02aebad-00d3-cf93-b87c-5670f3d95591 writer.go:29: 2021-01-29T19:31:27.181Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Updating LAN server: server="Node-d02aebad-00d3-cf93-b87c-5670f3d95591 (Addr: tcp/127.0.0.1:29519) (DC: dc1)" writer.go:29: 2021-01-29T19:31:27.181Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.serf.wan: serf: EventMemberUpdate: Node-d02aebad-00d3-cf93-b87c-5670f3d95591.dc1 writer.go:29: 2021-01-29T19:31:27.181Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.serf.lan: serf: EventMemberUpdate: Node-d02aebad-00d3-cf93-b87c-5670f3d95591 writer.go:29: 2021-01-29T19:31:27.187Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.serf.wan: serf: EventMemberUpdate: Node-d02aebad-00d3-cf93-b87c-5670f3d95591.dc1 writer.go:29: 2021-01-29T19:31:27.188Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Handled event for server in area: event=member-update server=Node-d02aebad-00d3-cf93-b87c-5670f3d95591.dc1 area=wan writer.go:29: 2021-01-29T19:31:27.190Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Handled event for server in area: event=member-update server=Node-d02aebad-00d3-cf93-b87c-5670f3d95591.dc1 area=wan writer.go:29: 2021-01-29T19:31:27.189Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:27.189Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Updating LAN server: server="Node-d02aebad-00d3-cf93-b87c-5670f3d95591 (Addr: tcp/127.0.0.1:29519) (DC: dc1)" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.196Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: removed service: service=web === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.198Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.198Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: removed check: check=service:web-sidecar-proxy:1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.199Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:27.200Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Skipping self join check for node since the cluster is too small: node=Node-d02aebad-00d3-cf93-b87c-5670f3d95591 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.199Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: removed check: check=service:web-sidecar-proxy:2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.201Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: member joined, marking health alive: member=Node-d02aebad-00d3-cf93-b87c-5670f3d95591 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.201Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: removed service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:27.203Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Node info in sync === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.204Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Skipping self join check for node since the cluster is too small: node=Node-d02aebad-00d3-cf93-b87c-5670f3d95591 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.204Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.206Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: Skipping self join check for node since the cluster is too small: node=Node-d02aebad-00d3-cf93-b87c-5670f3d95591 writer.go:29: 2021-01-29T19:31:27.214Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.acl: dropping node from result due to ACLs: node=Node-d02aebad-00d3-cf93-b87c-5670f3d95591 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.214Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Deregistered service: service=web writer.go:29: 2021-01-29T19:31:27.225Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Deregistered service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:27.230Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Requesting shutdown writer.go:29: 2021-01-29T19:31:27.232Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server: shutting down server writer.go:29: 2021-01-29T19:31:27.233Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:27.234Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:27.234Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:27.235Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:27.234Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:27.235Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:27.235Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:27.240Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.253Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Synced node info === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.253Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:27.253Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: consul server down writer.go:29: 2021-01-29T19:31:27.256Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: shutdown complete === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.258Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Synced service: service=web === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.258Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Stopping server: protocol=DNS address=127.0.0.1:29508 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.260Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Synced service: service=web-sidecar-proxy === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.259Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Stopping server: protocol=DNS address=127.0.0.1:29508 network=udp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.261Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:27.262Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Check in sync: check=service:web-sidecar-proxy:2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.261Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Stopping server: protocol=HTTP address=127.0.0.1:29509 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.269Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: removed service: service=web writer.go:29: 2021-01-29T19:31:27.270Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: removed check: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:27.271Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: removed check: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:27.279Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: removed service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:27.281Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Node info in sync writer.go:29: 2021-01-29T19:31:27.284Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Deregistered service: service=web writer.go:29: 2021-01-29T19:31:27.286Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Deregistered service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:27.294Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Requesting shutdown writer.go:29: 2021-01-29T19:31:27.300Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server: shutting down server writer.go:29: 2021-01-29T19:31:27.301Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:27.303Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:27.301Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:27.303Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:27.305Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:27.305Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:27.331Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:27.331Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:27.337Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:27.342Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:27.343Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: consul server down writer.go:29: 2021-01-29T19:31:27.349Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: shutdown complete writer.go:29: 2021-01-29T19:31:27.350Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Stopping server: protocol=DNS address=127.0.0.1:29514 network=tcp writer.go:29: 2021-01-29T19:31:27.352Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Stopping server: protocol=DNS address=127.0.0.1:29514 network=udp writer.go:29: 2021-01-29T19:31:27.354Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Stopping server: protocol=HTTP address=127.0.0.1:29515 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.763Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:27.765Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:27.785Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:27.786Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:27.788Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:27.790Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:27.797Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 Address:127.0.0.1:29525}]" writer.go:29: 2021-01-29T19:31:27.800Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.raft: entering follower state: follower="Node at 127.0.0.1:29525 [Follower]" leader= writer.go:29: 2021-01-29T19:31:27.800Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.serf.wan: serf: EventMemberJoin: Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:27.807Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.serf.lan: serf: EventMemberJoin: Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 127.0.0.1 writer.go:29: 2021-01-29T19:31:27.810Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Adding LAN server: server="Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 (Addr: tcp/127.0.0.1:29525) (DC: dc1)" writer.go:29: 2021-01-29T19:31:27.810Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Handled event for server in area: event=member-join server=Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53.dc1 area=wan writer.go:29: 2021-01-29T19:31:27.812Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: Started DNS server: address=127.0.0.1:29520 network=tcp writer.go:29: 2021-01-29T19:31:27.822Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: Started DNS server: address=127.0.0.1:29520 network=udp writer.go:29: 2021-01-29T19:31:27.827Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: Started HTTP server: address=127.0.0.1:29521 network=tcp writer.go:29: 2021-01-29T19:31:27.829Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: started state syncer === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults writer.go:29: 2021-01-29T19:31:27.856Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:27.858Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:27.860Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:27.862Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.raft: entering candidate state: node="Node at 127.0.0.1:29525 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:27.862Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:27.865Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:27.867Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.raft: vote granted: from=5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 term=2 tally=1 writer.go:29: 2021-01-29T19:31:27.870Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:27.871Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.raft: entering leader state: leader="Node at 127.0.0.1:29525 [Leader]" writer.go:29: 2021-01-29T19:31:27.872Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:27.874Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: New leader elected: payload=Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:27.897Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:27.898Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:27.908Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:27.912Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:27.889Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: initializing acls writer.go:29: 2021-01-29T19:31:27.933Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:27.934Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:27.936Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:27.941Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:27.953Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:27.957Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:27.960Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.serf.lan: serf: EventMemberUpdate: Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:27.962Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e2c81beb-8c54-c49e-7959-f9dcac03acc0 Address:127.0.0.1:29531}]" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:27.963Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Updating LAN server: server="Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 (Addr: tcp/127.0.0.1:29525) (DC: dc1)" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:27.964Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.serf.wan: serf: EventMemberJoin: Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:27.965Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.raft: entering follower state: follower="Node at 127.0.0.1:29531 [Follower]" leader= writer.go:29: 2021-01-29T19:31:27.968Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.serf.lan: serf: EventMemberJoin: Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0 127.0.0.1 writer.go:29: 2021-01-29T19:31:27.973Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Adding LAN server: server="Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0 (Addr: tcp/127.0.0.1:29531) (DC: dc1)" writer.go:29: 2021-01-29T19:31:27.973Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Handled event for server in area: event=member-join server=Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0.dc1 area=wan === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:27.973Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.serf.wan: serf: EventMemberUpdate: Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53.dc1 writer.go:29: 2021-01-29T19:31:27.977Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Handled event for server in area: event=member-update server=Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53.dc1 area=wan writer.go:29: 2021-01-29T19:31:27.980Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:27.981Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: Started DNS server: address=127.0.0.1:29526 network=tcp writer.go:29: 2021-01-29T19:31:27.984Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: Started DNS server: address=127.0.0.1:29526 network=udp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:27.986Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:27.987Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: Started HTTP server: address=127.0.0.1:29527 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:27.988Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:27.989Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: started state syncer === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:27.990Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Skipping self join check for node since the cluster is too small: node=Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 writer.go:29: 2021-01-29T19:31:27.992Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: member joined, marking health alive: member=Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 writer.go:29: 2021-01-29T19:31:28.015Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: Skipping self join check for node since the cluster is too small: node=Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:28.019Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:28.020Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.raft: entering candidate state: node="Node at 127.0.0.1:29531 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:28.022Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:28.023Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.raft: vote granted: from=e2c81beb-8c54-c49e-7959-f9dcac03acc0 term=2 tally=1 writer.go:29: 2021-01-29T19:31:28.024Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:28.026Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.raft: entering leader state: leader="Node at 127.0.0.1:29531 [Leader]" writer.go:29: 2021-01-29T19:31:28.024Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:28.028Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:28.028Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: New leader elected: payload=Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0 writer.go:29: 2021-01-29T19:31:28.029Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: initializing acls writer.go:29: 2021-01-29T19:31:28.044Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:28.045Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:28.062Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:28.065Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:28.067Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:28.068Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:28.070Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.serf.lan: serf: EventMemberUpdate: Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0 writer.go:29: 2021-01-29T19:31:28.072Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.serf.wan: serf: EventMemberUpdate: Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0.dc1 writer.go:29: 2021-01-29T19:31:28.072Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Updating LAN server: server="Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0 (Addr: tcp/127.0.0.1:29531) (DC: dc1)" writer.go:29: 2021-01-29T19:31:28.075Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Handled event for server in area: event=member-update server=Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0.dc1 area=wan writer.go:29: 2021-01-29T19:31:28.077Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:28.087Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:28.088Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:28.089Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Skipping self join check for node since the cluster is too small: node=Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0 writer.go:29: 2021-01-29T19:31:28.090Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: member joined, marking health alive: member=Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0 writer.go:29: 2021-01-29T19:31:28.092Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: Skipping self join check for node since the cluster is too small: node=Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0 writer.go:29: 2021-01-29T19:31:28.097Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.acl: dropping node from result due to ACLs: node=Node-e2c81beb-8c54-c49e-7959-f9dcac03acc0 writer.go:29: 2021-01-29T19:31:28.104Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: Requesting shutdown writer.go:29: 2021-01-29T19:31:28.104Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server: shutting down server writer.go:29: 2021-01-29T19:31:28.105Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:28.106Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:28.107Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:28.105Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:28.106Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:28.107Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:28.108Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:28.108Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.leader: stopped routine: routine="acl token reaping" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:28.116Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.acl: dropping node from result due to ACLs: node=Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:28.117Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:28.119Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.acl: dropping node from result due to ACLs: node=Node-5d4ae29f-e29b-f676-f92e-58f7c5e7ef53 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:28.122Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:28.122Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: consul server down writer.go:29: 2021-01-29T19:31:28.125Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: shutdown complete === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:28.127Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: Requesting shutdown === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:28.127Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: Stopping server: protocol=DNS address=127.0.0.1:29526 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:28.129Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server: shutting down server === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:28.130Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: Stopping server: protocol=DNS address=127.0.0.1:29526 network=udp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:28.131Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:28.132Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: Stopping server: protocol=HTTP address=127.0.0.1:29527 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:28.133Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:28.131Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:28.133Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:28.134Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:28.134Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:28.137Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:28.137Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:28.143Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:28.145Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:28.145Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: consul server down writer.go:29: 2021-01-29T19:31:28.147Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: shutdown complete writer.go:29: 2021-01-29T19:31:28.148Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: Stopping server: protocol=DNS address=127.0.0.1:29520 network=tcp writer.go:29: 2021-01-29T19:31:28.149Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: Stopping server: protocol=DNS address=127.0.0.1:29520 network=udp writer.go:29: 2021-01-29T19:31:28.151Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: Stopping server: protocol=HTTP address=127.0.0.1:29521 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied writer.go:29: 2021-01-29T19:31:28.633Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:28.635Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.647Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:28.648Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:28.648Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:28.650Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied writer.go:29: 2021-01-29T19:31:28.652Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:28.653Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.656Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 Address:127.0.0.1:29537}]" writer.go:29: 2021-01-29T19:31:28.657Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29537 [Follower]" leader= writer.go:29: 2021-01-29T19:31:28.660Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.serf.wan: serf: EventMemberJoin: Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:28.664Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.serf.lan: serf: EventMemberJoin: Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 127.0.0.1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.664Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.667Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Adding LAN server: server="Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 (Addr: tcp/127.0.0.1:29537) (DC: dc1)" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.666Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: bootstrap = true: do not enable unless necessary === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.668Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Handled event for server in area: event=member-join server=Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3.dc1 area=wan === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.668Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.tlsutil: Update: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.670Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: Started DNS server: address=127.0.0.1:29532 network=tcp writer.go:29: 2021-01-29T19:31:28.672Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: Started DNS server: address=127.0.0.1:29532 network=udp writer.go:29: 2021-01-29T19:31:28.674Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: Started HTTP server: address=127.0.0.1:29533 network=tcp writer.go:29: 2021-01-29T19:31:28.675Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: started state syncer === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.672Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:28.681Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:380692bf-f02a-1c00-07d0-138111788445 Address:127.0.0.1:29543}]" writer.go:29: 2021-01-29T19:31:28.683Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.serf.wan: serf: EventMemberJoin: Node-380692bf-f02a-1c00-07d0-138111788445.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:28.684Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29543 [Follower]" leader= writer.go:29: 2021-01-29T19:31:28.712Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.serf.lan: serf: EventMemberJoin: Node-380692bf-f02a-1c00-07d0-138111788445 127.0.0.1 writer.go:29: 2021-01-29T19:31:28.740Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Started DNS server: address=127.0.0.1:29538 network=udp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.723Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:28.732Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:28.744Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29537 [Candidate]" term=2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.746Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Handled event for server in area: event=member-join server=Node-380692bf-f02a-1c00-07d0-138111788445.dc1 area=wan writer.go:29: 2021-01-29T19:31:28.747Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:28.748Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29543 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:28.746Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Adding LAN server: server="Node-380692bf-f02a-1c00-07d0-138111788445 (Addr: tcp/127.0.0.1:29543) (DC: dc1)" writer.go:29: 2021-01-29T19:31:28.750Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Started DNS server: address=127.0.0.1:29538 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.760Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.raft: votes: needed=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.760Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Started HTTP server: address=127.0.0.1:29539 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.762Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.raft: vote granted: from=a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 term=2 tally=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.763Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: started state syncer === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.764Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:28.769Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29537 [Leader]" writer.go:29: 2021-01-29T19:31:28.774Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:28.777Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: New leader elected: payload=Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 writer.go:29: 2021-01-29T19:31:28.778Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.787Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.raft: votes: needed=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.787Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: initializing acls === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.788Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.raft: vote granted: from=380692bf-f02a-1c00-07d0-138111788445 term=2 tally=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.790Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:28.790Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.790Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:28.792Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29543 [Leader]" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.793Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Bootstrapped ACL master token from configuration === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.793Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: cluster leadership acquired === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.795Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:28.795Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.795Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: New leader elected: payload=Node-380692bf-f02a-1c00-07d0-138111788445 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.796Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.leader: started routine: routine="acl token reaping" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.797Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: initializing acls === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.797Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.serf.lan: serf: EventMemberUpdate: Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 writer.go:29: 2021-01-29T19:31:28.799Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.serf.wan: serf: EventMemberUpdate: Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3.dc1 writer.go:29: 2021-01-29T19:31:28.799Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Updating LAN server: server="Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 (Addr: tcp/127.0.0.1:29537) (DC: dc1)" writer.go:29: 2021-01-29T19:31:28.800Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Handled event for server in area: event=member-update server=Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3.dc1 area=wan writer.go:29: 2021-01-29T19:31:28.802Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:28.812Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:28.813Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.813Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: initializing acls === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.814Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.814Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Created ACL 'global-management' policy === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.815Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: member joined, marking health alive: member=Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.816Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.817Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.814Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:28.819Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:28.818Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:28.832Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:28.833Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:28.834Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:28.832Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:28.835Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:28.834Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.serf.lan: serf: EventMemberUpdate: Node-380692bf-f02a-1c00-07d0-138111788445 writer.go:29: 2021-01-29T19:31:28.837Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Updating LAN server: server="Node-380692bf-f02a-1c00-07d0-138111788445 (Addr: tcp/127.0.0.1:29543) (DC: dc1)" writer.go:29: 2021-01-29T19:31:28.837Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.serf.wan: serf: EventMemberUpdate: Node-380692bf-f02a-1c00-07d0-138111788445.dc1 writer.go:29: 2021-01-29T19:31:28.837Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.serf.lan: serf: EventMemberUpdate: Node-380692bf-f02a-1c00-07d0-138111788445 writer.go:29: 2021-01-29T19:31:28.840Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.serf.wan: serf: EventMemberUpdate: Node-380692bf-f02a-1c00-07d0-138111788445.dc1 writer.go:29: 2021-01-29T19:31:28.839Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Handled event for server in area: event=member-update server=Node-380692bf-f02a-1c00-07d0-138111788445.dc1 area=wan writer.go:29: 2021-01-29T19:31:28.842Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Handled event for server in area: event=member-update server=Node-380692bf-f02a-1c00-07d0-138111788445.dc1 area=wan writer.go:29: 2021-01-29T19:31:28.840Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:28.843Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Updating LAN server: server="Node-380692bf-f02a-1c00-07d0-138111788445 (Addr: tcp/127.0.0.1:29543) (DC: dc1)" writer.go:29: 2021-01-29T19:31:28.847Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:28.848Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:28.849Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-380692bf-f02a-1c00-07d0-138111788445 writer.go:29: 2021-01-29T19:31:28.850Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: member joined, marking health alive: member=Node-380692bf-f02a-1c00-07d0-138111788445 writer.go:29: 2021-01-29T19:31:28.858Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-380692bf-f02a-1c00-07d0-138111788445 writer.go:29: 2021-01-29T19:31:28.859Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-380692bf-f02a-1c00-07d0-138111788445 writer.go:29: 2021-01-29T19:31:28.875Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:28.876Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Synced node info === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:28.922Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.acl: dropping node from result due to ACLs: node=Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 writer.go:29: 2021-01-29T19:31:28.927Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.acl: dropping node from result due to ACLs: node=Node-a9889a8a-23a5-c5a9-7bfe-f5cd0fbe03a3 writer.go:29: 2021-01-29T19:31:28.951Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:28.952Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:28.953Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:28.954Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:28.955Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:28.955Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:28.957Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:28.958Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:28.959Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:28.961Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:28.973Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:28.973Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: consul server down writer.go:29: 2021-01-29T19:31:28.975Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:28.977Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29532 network=tcp writer.go:29: 2021-01-29T19:31:28.978Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29532 network=udp writer.go:29: 2021-01-29T19:31:28.979Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29533 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:29.125Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.acl: dropping node from result due to ACLs: node=Node-380692bf-f02a-1c00-07d0-138111788445 writer.go:29: 2021-01-29T19:31:29.134Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.acl: dropping node from result due to ACLs: node=Node-380692bf-f02a-1c00-07d0-138111788445 writer.go:29: 2021-01-29T19:31:29.177Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:29.184Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:29.186Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:29.187Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:29.188Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:29.188Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:29.189Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:29.191Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:29.191Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:29.197Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:29.200Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:29.202Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: consul server down writer.go:29: 2021-01-29T19:31:29.202Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:29.203Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29538 network=tcp writer.go:29: 2021-01-29T19:31:29.204Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29538 network=udp writer.go:29: 2021-01-29T19:31:29.205Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29539 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:29.480Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:29.482Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.500Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:29.531Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:29.534Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:29.536Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:29.541Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f Address:127.0.0.1:29549}]" writer.go:29: 2021-01-29T19:31:29.544Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: entering follower state: follower="Node at 127.0.0.1:29549 [Follower]" leader= writer.go:29: 2021-01-29T19:31:29.546Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.wan: serf: EventMemberJoin: Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:29.551Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.lan: serf: EventMemberJoin: Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f 127.0.0.1 writer.go:29: 2021-01-29T19:31:29.558Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Adding LAN server: server="Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f (Addr: tcp/127.0.0.1:29549) (DC: dc1)" writer.go:29: 2021-01-29T19:31:29.559Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Handled event for server in area: event=member-join server=Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f.dc1 area=wan writer.go:29: 2021-01-29T19:31:29.560Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Started DNS server: address=127.0.0.1:29544 network=udp writer.go:29: 2021-01-29T19:31:29.564Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Started DNS server: address=127.0.0.1:29544 network=tcp writer.go:29: 2021-01-29T19:31:29.566Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Started HTTP server: address=127.0.0.1:29545 network=tcp writer.go:29: 2021-01-29T19:31:29.568Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: started state syncer writer.go:29: 2021-01-29T19:31:29.585Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:29.587Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: entering candidate state: node="Node at 127.0.0.1:29549 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:29.590Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:29.592Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: vote granted: from=8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f term=2 tally=1 writer.go:29: 2021-01-29T19:31:29.593Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:29.595Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: entering leader state: leader="Node at 127.0.0.1:29549 [Leader]" writer.go:29: 2021-01-29T19:31:29.598Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:29.599Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: initializing acls writer.go:29: 2021-01-29T19:31:29.599Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: New leader elected: payload=Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f writer.go:29: 2021-01-29T19:31:29.605Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:29.607Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:29.610Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: initializing acls writer.go:29: 2021-01-29T19:31:29.613Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:29.619Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:29.620Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:29.622Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:29.623Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:29.624Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:29.623Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:29.626Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:29.626Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.lan: serf: EventMemberUpdate: Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f writer.go:29: 2021-01-29T19:31:29.628Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Updating LAN server: server="Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f (Addr: tcp/127.0.0.1:29549) (DC: dc1)" writer.go:29: 2021-01-29T19:31:29.628Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.wan: serf: EventMemberUpdate: Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f.dc1 writer.go:29: 2021-01-29T19:31:29.628Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.lan: serf: EventMemberUpdate: Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f writer.go:29: 2021-01-29T19:31:29.635Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Handled event for server in area: event=member-update server=Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f.dc1 area=wan writer.go:29: 2021-01-29T19:31:29.639Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Updating LAN server: server="Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f (Addr: tcp/127.0.0.1:29549) (DC: dc1)" writer.go:29: 2021-01-29T19:31:29.642Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.wan: serf: EventMemberUpdate: Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f.dc1 writer.go:29: 2021-01-29T19:31:29.644Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Handled event for server in area: event=member-update server=Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f.dc1 area=wan writer.go:29: 2021-01-29T19:31:29.663Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:29.670Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:29.672Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:29.674Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Skipping self join check for node since the cluster is too small: node=Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f writer.go:29: 2021-01-29T19:31:29.681Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: member joined, marking health alive: member=Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f writer.go:29: 2021-01-29T19:31:29.684Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Skipping self join check for node since the cluster is too small: node=Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f writer.go:29: 2021-01-29T19:31:29.687Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Skipping self join check for node since the cluster is too small: node=Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar writer.go:29: 2021-01-29T19:31:29.707Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:29.708Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.707Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:29.715Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Synced node info writer.go:29: 2021-01-29T19:31:29.718Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Node info in sync === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.733Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:29.734Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:29.735Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:29.737Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:29.744Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 Address:127.0.0.1:29555}]" writer.go:29: 2021-01-29T19:31:29.747Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: entering follower state: follower="Node at 127.0.0.1:29555 [Follower]" leader= writer.go:29: 2021-01-29T19:31:29.753Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.wan: serf: EventMemberJoin: Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:29.762Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.lan: serf: EventMemberJoin: Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 127.0.0.1 writer.go:29: 2021-01-29T19:31:29.767Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Handled event for server in area: event=member-join server=Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6.dc1 area=wan writer.go:29: 2021-01-29T19:31:29.767Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Adding LAN server: server="Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 (Addr: tcp/127.0.0.1:29555) (DC: dc1)" writer.go:29: 2021-01-29T19:31:29.773Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Started DNS server: address=127.0.0.1:29550 network=tcp writer.go:29: 2021-01-29T19:31:29.777Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Started DNS server: address=127.0.0.1:29550 network=udp writer.go:29: 2021-01-29T19:31:29.781Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Started HTTP server: address=127.0.0.1:29551 network=tcp writer.go:29: 2021-01-29T19:31:29.782Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: started state syncer === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.785Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.acl: dropping node from result due to ACLs: node=Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f writer.go:29: 2021-01-29T19:31:29.789Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.acl: dropping node from result due to ACLs: node=Node-8dd0fff7-3737-f3a5-9b0f-4c07e2bfd66f writer.go:29: 2021-01-29T19:31:29.790Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:29.792Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Node info in sync writer.go:29: 2021-01-29T19:31:29.800Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Requesting shutdown writer.go:29: 2021-01-29T19:31:29.801Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: shutting down server writer.go:29: 2021-01-29T19:31:29.802Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:29.803Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:29.804Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:29.803Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:29.804Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:29.805Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:29.806Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.822Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:29.823Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:29.824Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: entering candidate state: node="Node at 127.0.0.1:29555 [Candidate]" term=2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.811Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.834Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: votes: needed=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.835Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.router.manager: shutting down === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.836Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: vote granted: from=d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 term=2 tally=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.835Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: consul server down === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.838Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: election won: tally=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.838Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: shutdown complete writer.go:29: 2021-01-29T19:31:29.852Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Stopping server: protocol=DNS address=127.0.0.1:29544 network=tcp writer.go:29: 2021-01-29T19:31:29.855Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Stopping server: protocol=DNS address=127.0.0.1:29544 network=udp writer.go:29: 2021-01-29T19:31:29.857Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Stopping server: protocol=HTTP address=127.0.0.1:29545 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:29.839Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.raft: entering leader state: leader="Node at 127.0.0.1:29555 [Leader]" writer.go:29: 2021-01-29T19:31:29.859Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:29.862Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: New leader elected: payload=Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 writer.go:29: 2021-01-29T19:31:29.875Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: initializing acls writer.go:29: 2021-01-29T19:31:29.893Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:29.894Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:29.897Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:29.899Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:29.901Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:29.922Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:29.922Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: initializing acls writer.go:29: 2021-01-29T19:31:29.925Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:29.923Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.lan: serf: EventMemberUpdate: Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 writer.go:29: 2021-01-29T19:31:29.927Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:29.929Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Updating LAN server: server="Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 (Addr: tcp/127.0.0.1:29555) (DC: dc1)" writer.go:29: 2021-01-29T19:31:29.929Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.wan: serf: EventMemberUpdate: Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6.dc1 writer.go:29: 2021-01-29T19:31:29.931Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.lan: serf: EventMemberUpdate: Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 writer.go:29: 2021-01-29T19:31:29.934Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Handled event for server in area: event=member-update server=Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6.dc1 area=wan writer.go:29: 2021-01-29T19:31:29.935Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:29.936Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.wan: serf: EventMemberUpdate: Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6.dc1 writer.go:29: 2021-01-29T19:31:29.939Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Handled event for server in area: event=member-update server=Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6.dc1 area=wan writer.go:29: 2021-01-29T19:31:29.936Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Updating LAN server: server="Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 (Addr: tcp/127.0.0.1:29555) (DC: dc1)" writer.go:29: 2021-01-29T19:31:29.941Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:29.943Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:29.944Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Skipping self join check for node since the cluster is too small: node=Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 writer.go:29: 2021-01-29T19:31:29.946Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: member joined, marking health alive: member=Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 writer.go:29: 2021-01-29T19:31:29.948Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Skipping self join check for node since the cluster is too small: node=Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 writer.go:29: 2021-01-29T19:31:29.949Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: Skipping self join check for node since the cluster is too small: node=Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 writer.go:29: 2021-01-29T19:31:30.055Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:30.058Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Synced node info writer.go:29: 2021-01-29T19:31:30.059Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Node info in sync writer.go:29: 2021-01-29T19:31:30.138Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.acl: dropping node from result due to ACLs: node=Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 writer.go:29: 2021-01-29T19:31:30.143Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.acl: dropping node from result due to ACLs: node=Node-d31604e9-be9e-74b2-1a4f-ab7a0ca15af6 writer.go:29: 2021-01-29T19:31:30.168Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Requesting shutdown writer.go:29: 2021-01-29T19:31:30.169Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server: shutting down server writer.go:29: 2021-01-29T19:31:30.171Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:30.172Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:30.173Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:30.172Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:30.173Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:30.174Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:30.174Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:30.179Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:30.182Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:30.183Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: consul server down writer.go:29: 2021-01-29T19:31:30.184Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: shutdown complete writer.go:29: 2021-01-29T19:31:30.185Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Stopping server: protocol=DNS address=127.0.0.1:29550 network=tcp writer.go:29: 2021-01-29T19:31:30.186Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Stopping server: protocol=DNS address=127.0.0.1:29550 network=udp writer.go:29: 2021-01-29T19:31:30.187Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Stopping server: protocol=HTTP address=127.0.0.1:29551 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:30.363Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:30.365Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar writer.go:29: 2021-01-29T19:31:30.384Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:30.386Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:30.389Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:30.391Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:30.445Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:dad4d492-851e-9505-3ae1-eafad12d5615 Address:127.0.0.1:29561}]" writer.go:29: 2021-01-29T19:31:30.447Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.wan: serf: EventMemberJoin: Node-dad4d492-851e-9505-3ae1-eafad12d5615.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:30.449Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.lan: serf: EventMemberJoin: Node-dad4d492-851e-9505-3ae1-eafad12d5615 127.0.0.1 writer.go:29: 2021-01-29T19:31:30.447Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29561 [Follower]" leader= writer.go:29: 2021-01-29T19:31:30.454Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: Started DNS server: address=127.0.0.1:29556 network=udp writer.go:29: 2021-01-29T19:31:30.456Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Adding LAN server: server="Node-dad4d492-851e-9505-3ae1-eafad12d5615 (Addr: tcp/127.0.0.1:29561) (DC: dc1)" writer.go:29: 2021-01-29T19:31:30.458Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Handled event for server in area: event=member-join server=Node-dad4d492-851e-9505-3ae1-eafad12d5615.dc1 area=wan writer.go:29: 2021-01-29T19:31:30.461Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: Started DNS server: address=127.0.0.1:29556 network=tcp writer.go:29: 2021-01-29T19:31:30.465Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: Started HTTP server: address=127.0.0.1:29557 network=tcp writer.go:29: 2021-01-29T19:31:30.467Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:30.491Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:30.494Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29561 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:30.510Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:30.517Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:30.520Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: vote granted: from=dad4d492-851e-9505-3ae1-eafad12d5615 term=2 tally=1 writer.go:29: 2021-01-29T19:31:30.523Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:30.530Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29561 [Leader]" writer.go:29: 2021-01-29T19:31:30.534Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:30.538Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: New leader elected: payload=Node-dad4d492-851e-9505-3ae1-eafad12d5615 writer.go:29: 2021-01-29T19:31:30.542Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: initializing acls writer.go:29: 2021-01-29T19:31:30.560Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:30.562Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:30.568Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:30.574Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:30.581Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:30.586Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:30.590Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.lan: serf: EventMemberUpdate: Node-dad4d492-851e-9505-3ae1-eafad12d5615 writer.go:29: 2021-01-29T19:31:30.596Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.wan: serf: EventMemberUpdate: Node-dad4d492-851e-9505-3ae1-eafad12d5615.dc1 writer.go:29: 2021-01-29T19:31:30.597Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Updating LAN server: server="Node-dad4d492-851e-9505-3ae1-eafad12d5615 (Addr: tcp/127.0.0.1:29561) (DC: dc1)" writer.go:29: 2021-01-29T19:31:30.599Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Handled event for server in area: event=member-update server=Node-dad4d492-851e-9505-3ae1-eafad12d5615.dc1 area=wan writer.go:29: 2021-01-29T19:31:30.605Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:30.621Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:30.630Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:30.631Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-dad4d492-851e-9505-3ae1-eafad12d5615 writer.go:29: 2021-01-29T19:31:30.632Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: member joined, marking health alive: member=Node-dad4d492-851e-9505-3ae1-eafad12d5615 writer.go:29: 2021-01-29T19:31:30.643Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-dad4d492-851e-9505-3ae1-eafad12d5615 writer.go:29: 2021-01-29T19:31:30.662Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.acl: dropping node from result due to ACLs: node=Node-dad4d492-851e-9505-3ae1-eafad12d5615 writer.go:29: 2021-01-29T19:31:30.674Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:30.675Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:30.676Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:30.677Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:30.678Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:30.679Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:30.678Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:30.678Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:30.678Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:30.679Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:30.681Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination writer.go:29: 2021-01-29T19:31:30.688Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:30.689Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar writer.go:29: 2021-01-29T19:31:30.689Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:30.689Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: consul server down writer.go:29: 2021-01-29T19:31:30.692Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:30.693Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29556 network=tcp writer.go:29: 2021-01-29T19:31:30.694Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29556 network=udp writer.go:29: 2021-01-29T19:31:30.695Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29557 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar writer.go:29: 2021-01-29T19:31:30.710Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:30.711Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:30.712Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:30.714Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:30.721Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:220982b1-1faf-f9c0-d1b9-bda41085c477 Address:127.0.0.1:29567}]" writer.go:29: 2021-01-29T19:31:30.723Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29567 [Follower]" leader= writer.go:29: 2021-01-29T19:31:30.723Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.wan: serf: EventMemberJoin: Node-220982b1-1faf-f9c0-d1b9-bda41085c477.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:30.726Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.lan: serf: EventMemberJoin: Node-220982b1-1faf-f9c0-d1b9-bda41085c477 127.0.0.1 writer.go:29: 2021-01-29T19:31:30.728Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Adding LAN server: server="Node-220982b1-1faf-f9c0-d1b9-bda41085c477 (Addr: tcp/127.0.0.1:29567) (DC: dc1)" writer.go:29: 2021-01-29T19:31:30.729Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Handled event for server in area: event=member-join server=Node-220982b1-1faf-f9c0-d1b9-bda41085c477.dc1 area=wan writer.go:29: 2021-01-29T19:31:30.730Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Started DNS server: address=127.0.0.1:29562 network=udp writer.go:29: 2021-01-29T19:31:30.732Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Started DNS server: address=127.0.0.1:29562 network=tcp writer.go:29: 2021-01-29T19:31:30.734Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Started HTTP server: address=127.0.0.1:29563 network=tcp writer.go:29: 2021-01-29T19:31:30.735Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:30.780Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:30.786Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:30.787Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29567 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:30.789Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:30.791Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: vote granted: from=220982b1-1faf-f9c0-d1b9-bda41085c477 term=2 tally=1 writer.go:29: 2021-01-29T19:31:30.793Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:30.794Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29567 [Leader]" writer.go:29: 2021-01-29T19:31:30.796Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:30.797Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: New leader elected: payload=Node-220982b1-1faf-f9c0-d1b9-bda41085c477 writer.go:29: 2021-01-29T19:31:30.799Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: initializing acls writer.go:29: 2021-01-29T19:31:30.805Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:30.806Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:30.808Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:30.810Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:30.811Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:30.812Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:30.813Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.lan: serf: EventMemberUpdate: Node-220982b1-1faf-f9c0-d1b9-bda41085c477 writer.go:29: 2021-01-29T19:31:30.814Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.wan: serf: EventMemberUpdate: Node-220982b1-1faf-f9c0-d1b9-bda41085c477.dc1 writer.go:29: 2021-01-29T19:31:30.815Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Updating LAN server: server="Node-220982b1-1faf-f9c0-d1b9-bda41085c477 (Addr: tcp/127.0.0.1:29567) (DC: dc1)" writer.go:29: 2021-01-29T19:31:30.816Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Handled event for server in area: event=member-update server=Node-220982b1-1faf-f9c0-d1b9-bda41085c477.dc1 area=wan writer.go:29: 2021-01-29T19:31:30.819Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:30.824Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:30.826Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:30.828Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-220982b1-1faf-f9c0-d1b9-bda41085c477 writer.go:29: 2021-01-29T19:31:30.829Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: member joined, marking health alive: member=Node-220982b1-1faf-f9c0-d1b9-bda41085c477 writer.go:29: 2021-01-29T19:31:30.832Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-220982b1-1faf-f9c0-d1b9-bda41085c477 writer.go:29: 2021-01-29T19:31:30.977Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:30.981Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Synced node info writer.go:29: 2021-01-29T19:31:31.041Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.acl: dropping node from result due to ACLs: node=Node-220982b1-1faf-f9c0-d1b9-bda41085c477 writer.go:29: 2021-01-29T19:31:31.048Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.acl: dropping node from result due to ACLs: node=Node-220982b1-1faf-f9c0-d1b9-bda41085c477 writer.go:29: 2021-01-29T19:31:31.058Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:31.060Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:31.061Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:31.062Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:31.063Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:31.064Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:31.064Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:31.064Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:31.064Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:31.069Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:31.072Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:31.072Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: consul server down writer.go:29: 2021-01-29T19:31:31.075Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:31.077Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29562 network=tcp writer.go:29: 2021-01-29T19:31:31.079Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29562 network=udp writer.go:29: 2021-01-29T19:31:31.080Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29563 network=tcp writer.go:29: 2021-01-29T19:31:31.084Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:31.086Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.147Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:31.148Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:31.153Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:31.156Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:31.172Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a508369c-ae95-8590-facf-e65caf658965 Address:127.0.0.1:29573}]" writer.go:29: 2021-01-29T19:31:31.175Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29573 [Follower]" leader= writer.go:29: 2021-01-29T19:31:31.176Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.wan: serf: EventMemberJoin: Node-a508369c-ae95-8590-facf-e65caf658965.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:31.182Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.lan: serf: EventMemberJoin: Node-a508369c-ae95-8590-facf-e65caf658965 127.0.0.1 writer.go:29: 2021-01-29T19:31:31.186Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Started DNS server: address=127.0.0.1:29568 network=udp writer.go:29: 2021-01-29T19:31:31.188Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Adding LAN server: server="Node-a508369c-ae95-8590-facf-e65caf658965 (Addr: tcp/127.0.0.1:29573) (DC: dc1)" writer.go:29: 2021-01-29T19:31:31.189Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Handled event for server in area: event=member-join server=Node-a508369c-ae95-8590-facf-e65caf658965.dc1 area=wan writer.go:29: 2021-01-29T19:31:31.193Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Started DNS server: address=127.0.0.1:29568 network=tcp writer.go:29: 2021-01-29T19:31:31.195Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Started HTTP server: address=127.0.0.1:29569 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar writer.go:29: 2021-01-29T19:31:31.196Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:31.198Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.199Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: started state syncer === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.211Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.214Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:31.218Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29573 [Candidate]" term=2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.213Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:31.221Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:31.222Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.225Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:31.226Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: vote granted: from=a508369c-ae95-8590-facf-e65caf658965 term=2 tally=1 writer.go:29: 2021-01-29T19:31:31.227Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:31.229Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29573 [Leader]" writer.go:29: 2021-01-29T19:31:31.230Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:31.232Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: New leader elected: payload=Node-a508369c-ae95-8590-facf-e65caf658965 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.233Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2c7dc735-2b30-f489-fb2f-1c7a77aedd97 Address:127.0.0.1:29579}]" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.233Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: initializing acls === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.235Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29579 [Follower]" leader= === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.237Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Created ACL 'global-management' policy === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.236Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.wan: serf: EventMemberJoin: Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97.dc1 127.0.0.1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.238Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.239Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.lan: serf: EventMemberJoin: Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97 127.0.0.1 writer.go:29: 2021-01-29T19:31:31.241Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Handled event for server in area: event=member-join server=Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97.dc1 area=wan === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.241Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: initializing acls === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.242Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Started DNS server: address=127.0.0.1:29574 network=udp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.243Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.244Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Started DNS server: address=127.0.0.1:29574 network=tcp writer.go:29: 2021-01-29T19:31:31.246Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Started HTTP server: address=127.0.0.1:29575 network=tcp writer.go:29: 2021-01-29T19:31:31.248Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:31.249Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Adding LAN server: server="Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97 (Addr: tcp/127.0.0.1:29579) (DC: dc1)" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.258Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:31.258Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:31.260Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:31.260Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:31.259Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:31.262Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:31.261Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.lan: serf: EventMemberUpdate: Node-a508369c-ae95-8590-facf-e65caf658965 writer.go:29: 2021-01-29T19:31:31.264Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.wan: serf: EventMemberUpdate: Node-a508369c-ae95-8590-facf-e65caf658965.dc1 writer.go:29: 2021-01-29T19:31:31.264Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Updating LAN server: server="Node-a508369c-ae95-8590-facf-e65caf658965 (Addr: tcp/127.0.0.1:29573) (DC: dc1)" writer.go:29: 2021-01-29T19:31:31.264Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.lan: serf: EventMemberUpdate: Node-a508369c-ae95-8590-facf-e65caf658965 writer.go:29: 2021-01-29T19:31:31.273Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.wan: serf: EventMemberUpdate: Node-a508369c-ae95-8590-facf-e65caf658965.dc1 writer.go:29: 2021-01-29T19:31:31.271Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:31.271Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Handled event for server in area: event=member-update server=Node-a508369c-ae95-8590-facf-e65caf658965.dc1 area=wan writer.go:29: 2021-01-29T19:31:31.276Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Handled event for server in area: event=member-update server=Node-a508369c-ae95-8590-facf-e65caf658965.dc1 area=wan writer.go:29: 2021-01-29T19:31:31.274Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Updating LAN server: server="Node-a508369c-ae95-8590-facf-e65caf658965 (Addr: tcp/127.0.0.1:29573) (DC: dc1)" writer.go:29: 2021-01-29T19:31:31.278Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:31.280Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:31.281Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-a508369c-ae95-8590-facf-e65caf658965 writer.go:29: 2021-01-29T19:31:31.283Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: member joined, marking health alive: member=Node-a508369c-ae95-8590-facf-e65caf658965 writer.go:29: 2021-01-29T19:31:31.285Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-a508369c-ae95-8590-facf-e65caf658965 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.286Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.287Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-a508369c-ae95-8590-facf-e65caf658965 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.287Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29579 [Candidate]" term=2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.287Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.acl: dropping node from result due to ACLs: node=Node-a508369c-ae95-8590-facf-e65caf658965 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.289Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:31.290Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: vote granted: from=2c7dc735-2b30-f489-fb2f-1c7a77aedd97 term=2 tally=1 writer.go:29: 2021-01-29T19:31:31.291Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:31.292Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29579 [Leader]" writer.go:29: 2021-01-29T19:31:31.292Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:31.294Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:31.295Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: New leader elected: payload=Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97 writer.go:29: 2021-01-29T19:31:31.296Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: initializing acls writer.go:29: 2021-01-29T19:31:31.305Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:31.306Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:31.311Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:31.316Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Created ACL anonymous token from configuration === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.319Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: added local registration for service: service=web-sidecar-proxy === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.318Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:31.321Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: started routine: routine="acl token reaping" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.324Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Synced node info writer.go:29: 2021-01-29T19:31:31.326Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Synced service: service=web writer.go:29: 2021-01-29T19:31:31.328Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Synced service: service=web-sidecar-proxy === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.324Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.lan: serf: EventMemberUpdate: Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.329Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Check in sync: check=service:web-sidecar-proxy:1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.329Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Updating LAN server: server="Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97 (Addr: tcp/127.0.0.1:29579) (DC: dc1)" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.330Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Check in sync: check=service:web-sidecar-proxy:2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.330Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.wan: serf: EventMemberUpdate: Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97.dc1 writer.go:29: 2021-01-29T19:31:31.332Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Handled event for server in area: event=member-update server=Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97.dc1 area=wan writer.go:29: 2021-01-29T19:31:31.334Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:31.340Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:31.341Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:31.342Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.333Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:31.350Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Synced node info writer.go:29: 2021-01-29T19:31:31.356Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Synced service: service=web === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.343Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: member joined, marking health alive: member=Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.358Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:31.359Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:31.360Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:31.365Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: removed service: service=web writer.go:29: 2021-01-29T19:31:31.366Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: removed check: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:31.367Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: removed check: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:31.368Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: removed service: service=web-sidecar-proxy === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.368Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.369Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Node info in sync writer.go:29: 2021-01-29T19:31:31.370Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:31.370Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Deregistered service: service=web writer.go:29: 2021-01-29T19:31:31.377Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Deregistered service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:31.380Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:31.380Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:31.382Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:31.383Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:31.384Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:31.383Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:31.385Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:31.393Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:31.393Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:31.399Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:31.401Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:31.402Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: consul server down writer.go:29: 2021-01-29T19:31:31.403Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:31.404Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29568 network=tcp writer.go:29: 2021-01-29T19:31:31.404Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29568 network=udp writer.go:29: 2021-01-29T19:31:31.405Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29569 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.523Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.acl: dropping node from result due to ACLs: node=Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97 writer.go:29: 2021-01-29T19:31:31.526Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.acl: dropping node from result due to ACLs: node=Node-2c7dc735-2b30-f489-fb2f-1c7a77aedd97 writer.go:29: 2021-01-29T19:31:31.560Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Synced node info writer.go:29: 2021-01-29T19:31:31.562Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Synced service: service=web writer.go:29: 2021-01-29T19:31:31.567Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:31.568Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:31.570Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:31.588Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:31.589Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: removed service: service=web writer.go:29: 2021-01-29T19:31:31.596Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: removed check: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:31.598Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: removed check: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:31.600Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: removed service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:31.608Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Node info in sync writer.go:29: 2021-01-29T19:31:31.610Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Deregistered service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:31.613Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Deregistered service: service=web writer.go:29: 2021-01-29T19:31:31.619Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:31.628Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:31.631Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:31.633Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:31.631Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:31.633Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:31.639Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:31.639Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:31.653Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:31.653Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:31.659Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:31.663Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:31.663Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: consul server down writer.go:29: 2021-01-29T19:31:31.665Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:31.666Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29574 network=tcp writer.go:29: 2021-01-29T19:31:31.667Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29574 network=udp writer.go:29: 2021-01-29T19:31:31.667Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29575 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:31.906Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:31.909Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:31.920Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:31.921Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:31.932Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:31.936Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0ad6f808-40f6-14e2-7832-3935e636bef5 Address:127.0.0.1:29585}]" writer.go:29: 2021-01-29T19:31:31.938Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.serf.wan: serf: EventMemberJoin: Node-0ad6f808-40f6-14e2-7832-3935e636bef5.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:31.938Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29585 [Follower]" leader= writer.go:29: 2021-01-29T19:31:31.940Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.serf.lan: serf: EventMemberJoin: Node-0ad6f808-40f6-14e2-7832-3935e636bef5 127.0.0.1 writer.go:29: 2021-01-29T19:31:31.944Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server: Handled event for server in area: event=member-join server=Node-0ad6f808-40f6-14e2-7832-3935e636bef5.dc1 area=wan writer.go:29: 2021-01-29T19:31:31.944Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server: Adding LAN server: server="Node-0ad6f808-40f6-14e2-7832-3935e636bef5 (Addr: tcp/127.0.0.1:29585) (DC: dc1)" writer.go:29: 2021-01-29T19:31:31.946Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Started DNS server: address=127.0.0.1:29580 network=tcp writer.go:29: 2021-01-29T19:31:31.948Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Started DNS server: address=127.0.0.1:29580 network=udp writer.go:29: 2021-01-29T19:31:31.952Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Started HTTP server: address=127.0.0.1:29581 network=tcp writer.go:29: 2021-01-29T19:31:31.953Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:32.014Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:32.017Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29585 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:32.021Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:32.023Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.raft: vote granted: from=0ad6f808-40f6-14e2-7832-3935e636bef5 term=2 tally=1 writer.go:29: 2021-01-29T19:31:32.025Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:32.025Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29585 [Leader]" writer.go:29: 2021-01-29T19:31:32.026Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:32.027Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server: New leader elected: payload=Node-0ad6f808-40f6-14e2-7832-3935e636bef5 writer.go:29: 2021-01-29T19:31:32.028Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29585 writer.go:29: 2021-01-29T19:31:32.033Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:32.040Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:32.041Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.042Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-0ad6f808-40f6-14e2-7832-3935e636bef5 writer.go:29: 2021-01-29T19:31:32.044Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server: member joined, marking health alive: member=Node-0ad6f808-40f6-14e2-7832-3935e636bef5 writer.go:29: 2021-01-29T19:31:32.061Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:32.063Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Synced node info writer.go:29: 2021-01-29T19:31:32.064Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Node info in sync === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar writer.go:29: 2021-01-29T19:31:32.168Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:32.170Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.189Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:32.190Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:32.191Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.205Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:32.206Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:32.207Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.208Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:32.209Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.210Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.212Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:062e610c-d73d-48db-f7c5-a3c896e7701a Address:127.0.0.1:29591}]" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.212Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:32.214Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: consul server down === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.214Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.serf.wan: serf: EventMemberJoin: Node-062e610c-d73d-48db-f7c5-a3c896e7701a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:32.216Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29591 [Follower]" leader= === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.215Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: shutdown complete === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.217Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.serf.lan: serf: EventMemberJoin: Node-062e610c-d73d-48db-f7c5-a3c896e7701a 127.0.0.1 writer.go:29: 2021-01-29T19:31:32.226Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server: Adding LAN server: server="Node-062e610c-d73d-48db-f7c5-a3c896e7701a (Addr: tcp/127.0.0.1:29591) (DC: dc1)" writer.go:29: 2021-01-29T19:31:32.226Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server: Handled event for server in area: event=member-join server=Node-062e610c-d73d-48db-f7c5-a3c896e7701a.dc1 area=wan === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.224Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29580 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.229Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Started DNS server: address=127.0.0.1:29586 network=tcp writer.go:29: 2021-01-29T19:31:32.230Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Started DNS server: address=127.0.0.1:29586 network=udp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.228Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29580 network=udp writer.go:29: 2021-01-29T19:31:32.233Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29581 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.232Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Started HTTP server: address=127.0.0.1:29587 network=tcp writer.go:29: 2021-01-29T19:31:32.236Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:32.277Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:32.279Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29591 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:32.282Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:32.284Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.raft: vote granted: from=062e610c-d73d-48db-f7c5-a3c896e7701a term=2 tally=1 writer.go:29: 2021-01-29T19:31:32.286Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:32.287Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29591 [Leader]" writer.go:29: 2021-01-29T19:31:32.290Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:32.291Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server: New leader elected: payload=Node-062e610c-d73d-48db-f7c5-a3c896e7701a writer.go:29: 2021-01-29T19:31:32.295Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29591 writer.go:29: 2021-01-29T19:31:32.310Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:32.326Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:32.329Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.331Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-062e610c-d73d-48db-f7c5-a3c896e7701a writer.go:29: 2021-01-29T19:31:32.333Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server: member joined, marking health alive: member=Node-062e610c-d73d-48db-f7c5-a3c896e7701a writer.go:29: 2021-01-29T19:31:32.493Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:32.509Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Synced node info writer.go:29: 2021-01-29T19:31:32.510Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Node info in sync writer.go:29: 2021-01-29T19:31:32.603Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:32.604Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:32.605Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.607Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:32.609Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.629Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:32.632Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:32.634Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: consul server down writer.go:29: 2021-01-29T19:31:32.635Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:32.637Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29586 network=tcp writer.go:29: 2021-01-29T19:31:32.638Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29586 network=udp writer.go:29: 2021-01-29T19:31:32.639Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29587 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:32.734Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:32.736Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar writer.go:29: 2021-01-29T19:31:32.764Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:32.768Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:32.774Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:32.789Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:65b1fabd-faff-78f7-9ae8-10f856d5a8e8 Address:127.0.0.1:29597}]" writer.go:29: 2021-01-29T19:31:32.794Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29597 [Follower]" leader= writer.go:29: 2021-01-29T19:31:32.796Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.serf.wan: serf: EventMemberJoin: Node-65b1fabd-faff-78f7-9ae8-10f856d5a8e8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:32.799Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.serf.lan: serf: EventMemberJoin: Node-65b1fabd-faff-78f7-9ae8-10f856d5a8e8 127.0.0.1 writer.go:29: 2021-01-29T19:31:32.802Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server: Adding LAN server: server="Node-65b1fabd-faff-78f7-9ae8-10f856d5a8e8 (Addr: tcp/127.0.0.1:29597) (DC: dc1)" writer.go:29: 2021-01-29T19:31:32.807Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server: Handled event for server in area: event=member-join server=Node-65b1fabd-faff-78f7-9ae8-10f856d5a8e8.dc1 area=wan writer.go:29: 2021-01-29T19:31:32.808Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: Started DNS server: address=127.0.0.1:29592 network=udp writer.go:29: 2021-01-29T19:31:32.810Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: Started DNS server: address=127.0.0.1:29592 network=tcp writer.go:29: 2021-01-29T19:31:32.814Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: Started HTTP server: address=127.0.0.1:29593 network=tcp writer.go:29: 2021-01-29T19:31:32.816Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:32.865Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:32.867Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29597 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:32.879Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:32.881Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.raft: vote granted: from=65b1fabd-faff-78f7-9ae8-10f856d5a8e8 term=2 tally=1 writer.go:29: 2021-01-29T19:31:32.884Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:32.886Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29597 [Leader]" writer.go:29: 2021-01-29T19:31:32.887Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:32.889Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server: New leader elected: payload=Node-65b1fabd-faff-78f7-9ae8-10f856d5a8e8 writer.go:29: 2021-01-29T19:31:32.890Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29597 writer.go:29: 2021-01-29T19:31:32.896Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:32.903Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:32.904Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.906Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-65b1fabd-faff-78f7-9ae8-10f856d5a8e8 writer.go:29: 2021-01-29T19:31:32.907Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server: member joined, marking health alive: member=Node-65b1fabd-faff-78f7-9ae8-10f856d5a8e8 writer.go:29: 2021-01-29T19:31:32.979Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:32.981Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:32.982Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.984Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:32.985Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.983Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:32.986Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:32.999Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:33.000Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: consul server down writer.go:29: 2021-01-29T19:31:33.002Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:33.004Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29592 network=tcp writer.go:29: 2021-01-29T19:31:33.005Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29592 network=udp writer.go:29: 2021-01-29T19:31:33.007Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29593 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar writer.go:29: 2021-01-29T19:31:33.141Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:33.142Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar writer.go:29: 2021-01-29T19:31:33.172Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:33.205Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:33.210Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:33.225Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b70168a6-d2fe-c6ee-9991-44ffaf6aaf94 Address:127.0.0.1:29603}]" writer.go:29: 2021-01-29T19:31:33.229Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29603 [Follower]" leader= writer.go:29: 2021-01-29T19:31:33.230Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.serf.wan: serf: EventMemberJoin: Node-b70168a6-d2fe-c6ee-9991-44ffaf6aaf94.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:33.237Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.serf.lan: serf: EventMemberJoin: Node-b70168a6-d2fe-c6ee-9991-44ffaf6aaf94 127.0.0.1 writer.go:29: 2021-01-29T19:31:33.241Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server: Adding LAN server: server="Node-b70168a6-d2fe-c6ee-9991-44ffaf6aaf94 (Addr: tcp/127.0.0.1:29603) (DC: dc1)" writer.go:29: 2021-01-29T19:31:33.243Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server: Handled event for server in area: event=member-join server=Node-b70168a6-d2fe-c6ee-9991-44ffaf6aaf94.dc1 area=wan writer.go:29: 2021-01-29T19:31:33.247Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Started DNS server: address=127.0.0.1:29598 network=tcp writer.go:29: 2021-01-29T19:31:33.250Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Started DNS server: address=127.0.0.1:29598 network=udp writer.go:29: 2021-01-29T19:31:33.252Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Started HTTP server: address=127.0.0.1:29599 network=tcp writer.go:29: 2021-01-29T19:31:33.254Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:33.282Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:33.283Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29603 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:33.286Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:33.288Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.raft: vote granted: from=b70168a6-d2fe-c6ee-9991-44ffaf6aaf94 term=2 tally=1 writer.go:29: 2021-01-29T19:31:33.290Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:33.292Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29603 [Leader]" writer.go:29: 2021-01-29T19:31:33.293Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:33.295Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server: New leader elected: payload=Node-b70168a6-d2fe-c6ee-9991-44ffaf6aaf94 writer.go:29: 2021-01-29T19:31:33.296Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29603 writer.go:29: 2021-01-29T19:31:33.309Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:33.315Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:33.316Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:33.318Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-b70168a6-d2fe-c6ee-9991-44ffaf6aaf94 writer.go:29: 2021-01-29T19:31:33.320Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server: member joined, marking health alive: member=Node-b70168a6-d2fe-c6ee-9991-44ffaf6aaf94 writer.go:29: 2021-01-29T19:31:33.341Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:33.345Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Synced node info writer.go:29: 2021-01-29T19:31:33.347Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Node info in sync writer.go:29: 2021-01-29T19:31:33.451Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:33.452Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Node info in sync === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar writer.go:29: 2021-01-29T19:31:33.509Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:33.511Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar writer.go:29: 2021-01-29T19:31:33.525Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:33.526Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:33.528Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:33.529Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:33.529Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:33.531Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:33.535Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:33.535Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: consul server down writer.go:29: 2021-01-29T19:31:33.538Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: shutdown complete === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:33.537Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:33.540Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.tlsutil: Update: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar writer.go:29: 2021-01-29T19:31:33.539Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29598 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:33.543Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar writer.go:29: 2021-01-29T19:31:33.544Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29598 network=udp writer.go:29: 2021-01-29T19:31:33.546Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29599 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:33.549Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f234b736-2501-1dfc-e201-40181356a7db Address:127.0.0.1:29609}]" writer.go:29: 2021-01-29T19:31:33.551Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29609 [Follower]" leader= writer.go:29: 2021-01-29T19:31:33.551Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.serf.wan: serf: EventMemberJoin: Node-f234b736-2501-1dfc-e201-40181356a7db.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:33.555Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.serf.lan: serf: EventMemberJoin: Node-f234b736-2501-1dfc-e201-40181356a7db 127.0.0.1 writer.go:29: 2021-01-29T19:31:33.556Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server: Handled event for server in area: event=member-join server=Node-f234b736-2501-1dfc-e201-40181356a7db.dc1 area=wan writer.go:29: 2021-01-29T19:31:33.557Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Started DNS server: address=127.0.0.1:29604 network=udp writer.go:29: 2021-01-29T19:31:33.558Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server: Adding LAN server: server="Node-f234b736-2501-1dfc-e201-40181356a7db (Addr: tcp/127.0.0.1:29609) (DC: dc1)" writer.go:29: 2021-01-29T19:31:33.558Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Started DNS server: address=127.0.0.1:29604 network=tcp writer.go:29: 2021-01-29T19:31:33.561Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Started HTTP server: address=127.0.0.1:29605 network=tcp writer.go:29: 2021-01-29T19:31:33.562Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:33.604Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:33.605Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29609 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:33.608Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:33.609Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.raft: vote granted: from=f234b736-2501-1dfc-e201-40181356a7db term=2 tally=1 writer.go:29: 2021-01-29T19:31:33.610Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:33.612Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29609 [Leader]" writer.go:29: 2021-01-29T19:31:33.613Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:33.614Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server: New leader elected: payload=Node-f234b736-2501-1dfc-e201-40181356a7db writer.go:29: 2021-01-29T19:31:33.618Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29609 writer.go:29: 2021-01-29T19:31:33.643Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:33.656Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:33.659Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:33.662Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-f234b736-2501-1dfc-e201-40181356a7db writer.go:29: 2021-01-29T19:31:33.665Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server: member joined, marking health alive: member=Node-f234b736-2501-1dfc-e201-40181356a7db writer.go:29: 2021-01-29T19:31:33.726Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:33.739Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Synced node info writer.go:29: 2021-01-29T19:31:33.741Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Node info in sync writer.go:29: 2021-01-29T19:31:33.901Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:33.903Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:33.904Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:33.905Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:33.905Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:33.908Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:33.912Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:33.912Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: consul server down writer.go:29: 2021-01-29T19:31:33.916Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:33.918Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29604 network=tcp writer.go:29: 2021-01-29T19:31:33.920Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29604 network=udp writer.go:29: 2021-01-29T19:31:33.922Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29605 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar writer.go:29: 2021-01-29T19:31:34.047Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:34.049Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.075Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:34.085Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:34.091Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:34.098Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:78cdefa7-2d1c-26ed-4da9-80b20f2db984 Address:127.0.0.1:29615}]" writer.go:29: 2021-01-29T19:31:34.100Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29615 [Follower]" leader= writer.go:29: 2021-01-29T19:31:34.103Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.serf.wan: serf: EventMemberJoin: Node-78cdefa7-2d1c-26ed-4da9-80b20f2db984.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:34.107Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.serf.lan: serf: EventMemberJoin: Node-78cdefa7-2d1c-26ed-4da9-80b20f2db984 127.0.0.1 writer.go:29: 2021-01-29T19:31:34.112Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server: Adding LAN server: server="Node-78cdefa7-2d1c-26ed-4da9-80b20f2db984 (Addr: tcp/127.0.0.1:29615) (DC: dc1)" writer.go:29: 2021-01-29T19:31:34.112Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server: Handled event for server in area: event=member-join server=Node-78cdefa7-2d1c-26ed-4da9-80b20f2db984.dc1 area=wan writer.go:29: 2021-01-29T19:31:34.115Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Started DNS server: address=127.0.0.1:29610 network=tcp writer.go:29: 2021-01-29T19:31:34.118Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Started DNS server: address=127.0.0.1:29610 network=udp writer.go:29: 2021-01-29T19:31:34.121Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Started HTTP server: address=127.0.0.1:29611 network=tcp writer.go:29: 2021-01-29T19:31:34.122Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:34.179Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:34.181Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29615 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:34.199Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:34.201Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.raft: vote granted: from=78cdefa7-2d1c-26ed-4da9-80b20f2db984 term=2 tally=1 writer.go:29: 2021-01-29T19:31:34.203Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:34.204Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29615 [Leader]" writer.go:29: 2021-01-29T19:31:34.206Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:34.207Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server: New leader elected: payload=Node-78cdefa7-2d1c-26ed-4da9-80b20f2db984 writer.go:29: 2021-01-29T19:31:34.208Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29615 writer.go:29: 2021-01-29T19:31:34.218Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:34.253Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:34.254Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:34.255Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-78cdefa7-2d1c-26ed-4da9-80b20f2db984 writer.go:29: 2021-01-29T19:31:34.257Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server: member joined, marking health alive: member=Node-78cdefa7-2d1c-26ed-4da9-80b20f2db984 writer.go:29: 2021-01-29T19:31:34.338Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:34.341Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Synced node info === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.424Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:34.425Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.469Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:34.488Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:34.493Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:34.554Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c6c631cf-ffd8-9e15-46b2-64f3896cbf48 Address:127.0.0.1:29621}]" writer.go:29: 2021-01-29T19:31:34.557Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.serf.wan: serf: EventMemberJoin: Node-c6c631cf-ffd8-9e15-46b2-64f3896cbf48.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:34.562Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29621 [Follower]" leader= writer.go:29: 2021-01-29T19:31:34.562Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.serf.lan: serf: EventMemberJoin: Node-c6c631cf-ffd8-9e15-46b2-64f3896cbf48 127.0.0.1 writer.go:29: 2021-01-29T19:31:34.569Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server: Handled event for server in area: event=member-join server=Node-c6c631cf-ffd8-9e15-46b2-64f3896cbf48.dc1 area=wan === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.560Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Requesting shutdown === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.569Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server: Adding LAN server: server="Node-c6c631cf-ffd8-9e15-46b2-64f3896cbf48 (Addr: tcp/127.0.0.1:29621) (DC: dc1)" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.572Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server: shutting down server === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.571Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: Started DNS server: address=127.0.0.1:29616 network=udp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.573Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:34.574Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:34.574Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.577Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: Started DNS server: address=127.0.0.1:29616 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.578Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.581Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: Started HTTP server: address=127.0.0.1:29617 network=tcp writer.go:29: 2021-01-29T19:31:34.583Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: started state syncer === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.588Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:34.588Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: consul server down writer.go:29: 2021-01-29T19:31:34.591Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:34.593Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29610 network=tcp writer.go:29: 2021-01-29T19:31:34.595Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29610 network=udp writer.go:29: 2021-01-29T19:31:34.597Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29611 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar writer.go:29: 2021-01-29T19:31:34.625Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:34.627Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29621 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:34.630Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:34.631Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.raft: vote granted: from=c6c631cf-ffd8-9e15-46b2-64f3896cbf48 term=2 tally=1 writer.go:29: 2021-01-29T19:31:34.633Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:34.634Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29621 [Leader]" writer.go:29: 2021-01-29T19:31:34.653Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:34.655Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server: New leader elected: payload=Node-c6c631cf-ffd8-9e15-46b2-64f3896cbf48 writer.go:29: 2021-01-29T19:31:34.658Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29621 writer.go:29: 2021-01-29T19:31:34.667Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:34.677Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:34.679Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:34.681Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-c6c631cf-ffd8-9e15-46b2-64f3896cbf48 writer.go:29: 2021-01-29T19:31:34.687Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server: member joined, marking health alive: member=Node-c6c631cf-ffd8-9e15-46b2-64f3896cbf48 writer.go:29: 2021-01-29T19:31:34.911Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:34.912Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:34.913Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:34.915Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:34.913Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:34.915Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:34.918Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:34.921Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:34.921Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: consul server down writer.go:29: 2021-01-29T19:31:34.924Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:34.925Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29616 network=tcp writer.go:29: 2021-01-29T19:31:34.927Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29616 network=udp writer.go:29: 2021-01-29T19:31:34.928Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29617 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar writer.go:29: 2021-01-29T19:31:35.099Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:35.101Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar writer.go:29: 2021-01-29T19:31:35.110Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:35.111Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:35.114Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:35.121Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:55a99e7d-420c-5a21-f0b8-a2e8818bb9e4 Address:127.0.0.1:29627}]" writer.go:29: 2021-01-29T19:31:35.123Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29627 [Follower]" leader= writer.go:29: 2021-01-29T19:31:35.125Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.serf.wan: serf: EventMemberJoin: Node-55a99e7d-420c-5a21-f0b8-a2e8818bb9e4.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:35.156Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.serf.lan: serf: EventMemberJoin: Node-55a99e7d-420c-5a21-f0b8-a2e8818bb9e4 127.0.0.1 writer.go:29: 2021-01-29T19:31:35.169Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server: Handled event for server in area: event=member-join server=Node-55a99e7d-420c-5a21-f0b8-a2e8818bb9e4.dc1 area=wan writer.go:29: 2021-01-29T19:31:35.169Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server: Adding LAN server: server="Node-55a99e7d-420c-5a21-f0b8-a2e8818bb9e4 (Addr: tcp/127.0.0.1:29627) (DC: dc1)" writer.go:29: 2021-01-29T19:31:35.170Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: Started DNS server: address=127.0.0.1:29622 network=udp writer.go:29: 2021-01-29T19:31:35.173Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: Started DNS server: address=127.0.0.1:29622 network=tcp writer.go:29: 2021-01-29T19:31:35.173Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:35.176Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29627 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:35.176Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: Started HTTP server: address=127.0.0.1:29623 network=tcp writer.go:29: 2021-01-29T19:31:35.181Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: started state syncer writer.go:29: 2021-01-29T19:31:35.183Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:35.185Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.raft: vote granted: from=55a99e7d-420c-5a21-f0b8-a2e8818bb9e4 term=2 tally=1 writer.go:29: 2021-01-29T19:31:35.186Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:35.188Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29627 [Leader]" writer.go:29: 2021-01-29T19:31:35.201Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:35.201Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server: New leader elected: payload=Node-55a99e7d-420c-5a21-f0b8-a2e8818bb9e4 writer.go:29: 2021-01-29T19:31:35.203Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29627 writer.go:29: 2021-01-29T19:31:35.213Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:35.219Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:35.221Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:35.223Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-55a99e7d-420c-5a21-f0b8-a2e8818bb9e4 writer.go:29: 2021-01-29T19:31:35.225Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server: member joined, marking health alive: member=Node-55a99e7d-420c-5a21-f0b8-a2e8818bb9e4 writer.go:29: 2021-01-29T19:31:35.336Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:31:35.337Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:31:35.339Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:35.340Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:35.339Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:35.340Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:35.344Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:35.350Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:35.350Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: consul server down writer.go:29: 2021-01-29T19:31:35.353Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: shutdown complete writer.go:29: 2021-01-29T19:31:35.354Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29622 network=tcp writer.go:29: 2021-01-29T19:31:35.356Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: Stopping server: protocol=DNS address=127.0.0.1:29622 network=udp writer.go:29: 2021-01-29T19:31:35.357Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29623 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar writer.go:29: 2021-01-29T19:31:35.429Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:35.431Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:35.447Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:35.461Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:35.463Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:35.485Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a12007c4-a07f-952c-b372-eb5ba692d6d1 Address:127.0.0.1:29633}]" writer.go:29: 2021-01-29T19:31:35.488Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: entering follower state: follower="Node at 127.0.0.1:29633 [Follower]" leader= writer.go:29: 2021-01-29T19:31:35.489Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.serf.wan: serf: EventMemberJoin: Node-a12007c4-a07f-952c-b372-eb5ba692d6d1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:35.492Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.serf.lan: serf: EventMemberJoin: Node-a12007c4-a07f-952c-b372-eb5ba692d6d1 127.0.0.1 writer.go:29: 2021-01-29T19:31:35.494Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: Handled event for server in area: event=member-join server=Node-a12007c4-a07f-952c-b372-eb5ba692d6d1.dc1 area=wan writer.go:29: 2021-01-29T19:31:35.494Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: Adding LAN server: server="Node-a12007c4-a07f-952c-b372-eb5ba692d6d1 (Addr: tcp/127.0.0.1:29633) (DC: dc1)" writer.go:29: 2021-01-29T19:31:35.495Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Started DNS server: address=127.0.0.1:29628 network=udp writer.go:29: 2021-01-29T19:31:35.501Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Started DNS server: address=127.0.0.1:29628 network=tcp writer.go:29: 2021-01-29T19:31:35.504Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Started HTTP server: address=127.0.0.1:29629 network=tcp writer.go:29: 2021-01-29T19:31:35.506Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: started state syncer writer.go:29: 2021-01-29T19:31:35.530Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:35.532Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: entering candidate state: node="Node at 127.0.0.1:29633 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:35.535Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:35.540Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: vote granted: from=a12007c4-a07f-952c-b372-eb5ba692d6d1 term=2 tally=1 writer.go:29: 2021-01-29T19:31:35.542Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:35.543Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: entering leader state: leader="Node at 127.0.0.1:29633 [Leader]" writer.go:29: 2021-01-29T19:31:35.562Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:35.563Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: New leader elected: payload=Node-a12007c4-a07f-952c-b372-eb5ba692d6d1 writer.go:29: 2021-01-29T19:31:35.564Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29633 writer.go:29: 2021-01-29T19:31:35.571Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:35.610Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:35.612Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:35.613Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: Skipping self join check for node since the cluster is too small: node=Node-a12007c4-a07f-952c-b372-eb5ba692d6d1 writer.go:29: 2021-01-29T19:31:35.614Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: member joined, marking health alive: member=Node-a12007c4-a07f-952c-b372-eb5ba692d6d1 writer.go:29: 2021-01-29T19:31:35.640Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Synced node info writer.go:29: 2021-01-29T19:31:35.642Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:35.644Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Synced service: service=web writer.go:29: 2021-01-29T19:31:35.653Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: removed service: service=web writer.go:29: 2021-01-29T19:31:35.663Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Node info in sync writer.go:29: 2021-01-29T19:31:35.664Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Service in sync: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:35.672Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Deregistered service: service=web writer.go:29: 2021-01-29T19:31:35.678Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Requesting shutdown writer.go:29: 2021-01-29T19:31:35.680Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: shutting down server writer.go:29: 2021-01-29T19:31:35.681Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:35.683Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:35.685Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:35.686Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:35.689Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:35.693Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:35.693Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: consul server down writer.go:29: 2021-01-29T19:31:35.696Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: shutdown complete writer.go:29: 2021-01-29T19:31:35.696Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Stopping server: protocol=DNS address=127.0.0.1:29628 network=tcp writer.go:29: 2021-01-29T19:31:35.697Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Stopping server: protocol=DNS address=127.0.0.1:29628 network=udp writer.go:29: 2021-01-29T19:31:35.698Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Stopping server: protocol=HTTP address=127.0.0.1:29629 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar writer.go:29: 2021-01-29T19:31:35.858Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:35.860Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:35.903Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:35.907Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:35.928Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:35.975Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:95ea3ebf-d10c-f08e-24d9-9d97de28f6c9 Address:127.0.0.1:29639}]" writer.go:29: 2021-01-29T19:31:35.977Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: entering follower state: follower="Node at 127.0.0.1:29639 [Follower]" leader= writer.go:29: 2021-01-29T19:31:35.978Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.serf.wan: serf: EventMemberJoin: Node-95ea3ebf-d10c-f08e-24d9-9d97de28f6c9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:35.983Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.serf.lan: serf: EventMemberJoin: Node-95ea3ebf-d10c-f08e-24d9-9d97de28f6c9 127.0.0.1 writer.go:29: 2021-01-29T19:31:35.987Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Started DNS server: address=127.0.0.1:29634 network=udp writer.go:29: 2021-01-29T19:31:35.987Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: Adding LAN server: server="Node-95ea3ebf-d10c-f08e-24d9-9d97de28f6c9 (Addr: tcp/127.0.0.1:29639) (DC: dc1)" writer.go:29: 2021-01-29T19:31:35.988Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: Handled event for server in area: event=member-join server=Node-95ea3ebf-d10c-f08e-24d9-9d97de28f6c9.dc1 area=wan writer.go:29: 2021-01-29T19:31:35.989Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Started DNS server: address=127.0.0.1:29634 network=tcp writer.go:29: 2021-01-29T19:31:35.992Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Started HTTP server: address=127.0.0.1:29635 network=tcp writer.go:29: 2021-01-29T19:31:35.993Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: started state syncer writer.go:29: 2021-01-29T19:31:36.031Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:36.033Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: entering candidate state: node="Node at 127.0.0.1:29639 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:36.037Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:36.050Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: vote granted: from=95ea3ebf-d10c-f08e-24d9-9d97de28f6c9 term=2 tally=1 writer.go:29: 2021-01-29T19:31:36.051Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:36.052Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.raft: entering leader state: leader="Node at 127.0.0.1:29639 [Leader]" writer.go:29: 2021-01-29T19:31:36.053Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:36.054Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: New leader elected: payload=Node-95ea3ebf-d10c-f08e-24d9-9d97de28f6c9 writer.go:29: 2021-01-29T19:31:36.056Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29639 writer.go:29: 2021-01-29T19:31:36.062Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:36.067Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:36.068Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:36.069Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: Skipping self join check for node since the cluster is too small: node=Node-95ea3ebf-d10c-f08e-24d9-9d97de28f6c9 writer.go:29: 2021-01-29T19:31:36.070Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: member joined, marking health alive: member=Node-95ea3ebf-d10c-f08e-24d9-9d97de28f6c9 writer.go:29: 2021-01-29T19:31:36.081Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Synced node info writer.go:29: 2021-01-29T19:31:36.092Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Node info in sync === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:36.200Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:36.202Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.217Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:36.218Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:36.220Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:36.244Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8f8b7cbd-c0e1-cc80-8cfd-bfedb03f0f04 Address:127.0.0.1:29645}]" writer.go:29: 2021-01-29T19:31:36.248Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.raft: entering follower state: follower="Node at 127.0.0.1:29645 [Follower]" leader= writer.go:29: 2021-01-29T19:31:36.249Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.serf.wan: serf: EventMemberJoin: Node-8f8b7cbd-c0e1-cc80-8cfd-bfedb03f0f04.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:36.256Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.serf.lan: serf: EventMemberJoin: Node-8f8b7cbd-c0e1-cc80-8cfd-bfedb03f0f04 127.0.0.1 writer.go:29: 2021-01-29T19:31:36.260Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server: Adding LAN server: server="Node-8f8b7cbd-c0e1-cc80-8cfd-bfedb03f0f04 (Addr: tcp/127.0.0.1:29645) (DC: dc1)" writer.go:29: 2021-01-29T19:31:36.261Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Started DNS server: address=127.0.0.1:29640 network=udp writer.go:29: 2021-01-29T19:31:36.263Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Started DNS server: address=127.0.0.1:29640 network=tcp writer.go:29: 2021-01-29T19:31:36.262Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server: Handled event for server in area: event=member-join server=Node-8f8b7cbd-c0e1-cc80-8cfd-bfedb03f0f04.dc1 area=wan writer.go:29: 2021-01-29T19:31:36.266Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Started HTTP server: address=127.0.0.1:29641 network=tcp writer.go:29: 2021-01-29T19:31:36.269Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: started state syncer writer.go:29: 2021-01-29T19:31:36.306Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:36.308Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.raft: entering candidate state: node="Node at 127.0.0.1:29645 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:36.311Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:36.313Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.raft: vote granted: from=8f8b7cbd-c0e1-cc80-8cfd-bfedb03f0f04 term=2 tally=1 writer.go:29: 2021-01-29T19:31:36.315Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:36.317Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.raft: entering leader state: leader="Node at 127.0.0.1:29645 [Leader]" writer.go:29: 2021-01-29T19:31:36.318Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:36.320Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server: New leader elected: payload=Node-8f8b7cbd-c0e1-cc80-8cfd-bfedb03f0f04 writer.go:29: 2021-01-29T19:31:36.333Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29645 writer.go:29: 2021-01-29T19:31:36.340Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:36.350Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:36.351Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:36.353Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server: Skipping self join check for node since the cluster is too small: node=Node-8f8b7cbd-c0e1-cc80-8cfd-bfedb03f0f04 writer.go:29: 2021-01-29T19:31:36.354Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server: member joined, marking health alive: member=Node-8f8b7cbd-c0e1-cc80-8cfd-bfedb03f0f04 writer.go:29: 2021-01-29T19:31:36.414Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:36.415Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Node info in sync === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.417Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Synced node info writer.go:29: 2021-01-29T19:31:36.419Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Node info in sync === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:36.418Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:36.422Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Synced service: service=web writer.go:29: 2021-01-29T19:31:36.447Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:36.451Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Node info in sync writer.go:29: 2021-01-29T19:31:36.454Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Service in sync: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:36.457Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Service in sync: service=web writer.go:29: 2021-01-29T19:31:36.458Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Node info in sync writer.go:29: 2021-01-29T19:31:36.459Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Service in sync: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:36.460Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Service in sync: service=web writer.go:29: 2021-01-29T19:31:36.463Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: removed service: service=web writer.go:29: 2021-01-29T19:31:36.465Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Node info in sync writer.go:29: 2021-01-29T19:31:36.466Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Service in sync: service=web-sidecar-proxy === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.482Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: added local registration for service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:36.484Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Node info in sync === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:36.485Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Deregistered service: service=web === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.487Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:36.491Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Synced service: service=web === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:36.491Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Node info in sync === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.493Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Check in sync: check=service:web-sidecar-proxy:1 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:36.494Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Service in sync: service=web-sidecar-proxy === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.495Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Check in sync: check=service:web-sidecar-proxy:2 === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:36.511Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Requesting shutdown writer.go:29: 2021-01-29T19:31:36.515Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server: shutting down server writer.go:29: 2021-01-29T19:31:36.517Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:36.522Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:36.522Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:36.528Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:36.532Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:36.534Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: consul server down writer.go:29: 2021-01-29T19:31:36.536Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: shutdown complete writer.go:29: 2021-01-29T19:31:36.537Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Stopping server: protocol=DNS address=127.0.0.1:29634 network=tcp writer.go:29: 2021-01-29T19:31:36.539Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Stopping server: protocol=DNS address=127.0.0.1:29634 network=udp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.540Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: removed service: service=web writer.go:29: 2021-01-29T19:31:36.543Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: removed check: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:36.545Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: removed check: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:36.547Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: removed service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:36.549Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Node info in sync === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:36.541Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Stopping server: protocol=HTTP address=127.0.0.1:29635 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.551Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:36.556Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Waiting for endpoints to shut down === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.552Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Deregistered service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:36.568Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Deregistered service: service=web === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered writer.go:29: 2021-01-29T19:31:36.565Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.579Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Requesting shutdown writer.go:29: 2021-01-29T19:31:36.587Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server: shutting down server writer.go:29: 2021-01-29T19:31:36.590Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:36.592Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:36.592Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:36.603Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:36.605Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:36.605Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: consul server down writer.go:29: 2021-01-29T19:31:36.610Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: shutdown complete writer.go:29: 2021-01-29T19:31:36.612Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Stopping server: protocol=DNS address=127.0.0.1:29640 network=tcp writer.go:29: 2021-01-29T19:31:36.615Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Stopping server: protocol=DNS address=127.0.0.1:29640 network=udp writer.go:29: 2021-01-29T19:31:36.616Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Stopping server: protocol=HTTP address=127.0.0.1:29641 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:36.622Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:36.624Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:36.627Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:36.633Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c2e82a77-9e2a-08c9-078e-cbe9072298a2 Address:127.0.0.1:29651}]" writer.go:29: 2021-01-29T19:31:36.635Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.raft: entering follower state: follower="Node at 127.0.0.1:29651 [Follower]" leader= writer.go:29: 2021-01-29T19:31:36.636Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.serf.wan: serf: EventMemberJoin: Node-c2e82a77-9e2a-08c9-078e-cbe9072298a2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:36.643Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.serf.lan: serf: EventMemberJoin: Node-c2e82a77-9e2a-08c9-078e-cbe9072298a2 127.0.0.1 writer.go:29: 2021-01-29T19:31:36.649Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server: Adding LAN server: server="Node-c2e82a77-9e2a-08c9-078e-cbe9072298a2 (Addr: tcp/127.0.0.1:29651) (DC: dc1)" writer.go:29: 2021-01-29T19:31:36.651Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server: Handled event for server in area: event=member-join server=Node-c2e82a77-9e2a-08c9-078e-cbe9072298a2.dc1 area=wan writer.go:29: 2021-01-29T19:31:36.653Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Started DNS server: address=127.0.0.1:29646 network=tcp writer.go:29: 2021-01-29T19:31:36.656Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Started DNS server: address=127.0.0.1:29646 network=udp writer.go:29: 2021-01-29T19:31:36.659Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Started HTTP server: address=127.0.0.1:29647 network=tcp writer.go:29: 2021-01-29T19:31:36.660Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: started state syncer writer.go:29: 2021-01-29T19:31:36.678Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:36.679Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.raft: entering candidate state: node="Node at 127.0.0.1:29651 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:36.683Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:36.685Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.raft: vote granted: from=c2e82a77-9e2a-08c9-078e-cbe9072298a2 term=2 tally=1 writer.go:29: 2021-01-29T19:31:36.687Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:36.688Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.raft: entering leader state: leader="Node at 127.0.0.1:29651 [Leader]" writer.go:29: 2021-01-29T19:31:36.692Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:36.693Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server: New leader elected: payload=Node-c2e82a77-9e2a-08c9-078e-cbe9072298a2 writer.go:29: 2021-01-29T19:31:36.693Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29651 writer.go:29: 2021-01-29T19:31:36.710Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:36.739Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:36.741Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:36.742Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server: Skipping self join check for node since the cluster is too small: node=Node-c2e82a77-9e2a-08c9-078e-cbe9072298a2 writer.go:29: 2021-01-29T19:31:36.744Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server: member joined, marking health alive: member=Node-c2e82a77-9e2a-08c9-078e-cbe9072298a2 writer.go:29: 2021-01-29T19:31:37.040Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Synced node info writer.go:29: 2021-01-29T19:31:37.043Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.045Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Synced service: service=web writer.go:29: 2021-01-29T19:31:37.047Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:37.049Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:37.051Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:37.055Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Node info in sync writer.go:29: 2021-01-29T19:31:37.057Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Service in sync: service=web writer.go:29: 2021-01-29T19:31:37.059Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Service in sync: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.061Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:37.063Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:37.065Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Node info in sync writer.go:29: 2021-01-29T19:31:37.067Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Service in sync: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.068Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Service in sync: service=web writer.go:29: 2021-01-29T19:31:37.068Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:37.070Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:37.078Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: removed service: service=web writer.go:29: 2021-01-29T19:31:37.078Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:37.084Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: removed check: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:31:37.086Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: removed check: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:31:37.088Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: removed service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.090Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Node info in sync writer.go:29: 2021-01-29T19:31:37.092Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Deregistered service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.095Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Deregistered service: service=web writer.go:29: 2021-01-29T19:31:37.104Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Requesting shutdown writer.go:29: 2021-01-29T19:31:37.106Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server: shutting down server writer.go:29: 2021-01-29T19:31:37.108Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:37.110Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:37.112Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:37.116Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:37.120Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work.server.router.manager: shutting down === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:37.119Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Waiting for endpoints to shut down === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:37.121Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: consul server down writer.go:29: 2021-01-29T19:31:37.122Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: shutdown complete === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:37.121Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:37.123Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Stopping server: protocol=DNS address=127.0.0.1:29646 network=tcp writer.go:29: 2021-01-29T19:31:37.125Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Stopping server: protocol=DNS address=127.0.0.1:29646 network=udp writer.go:29: 2021-01-29T19:31:37.126Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Stopping server: protocol=HTTP address=127.0.0.1:29647 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it writer.go:29: 2021-01-29T19:31:37.168Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:37.170Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:37.194Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:37.252Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d62574a6-25fd-7e70-a7f6-02d7594bb09c Address:127.0.0.1:29657}]" writer.go:29: 2021-01-29T19:31:37.256Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: entering follower state: follower="Node at 127.0.0.1:29657 [Follower]" leader= writer.go:29: 2021-01-29T19:31:37.288Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.serf.wan: serf: EventMemberJoin: Node-d62574a6-25fd-7e70-a7f6-02d7594bb09c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:37.291Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.serf.lan: serf: EventMemberJoin: Node-d62574a6-25fd-7e70-a7f6-02d7594bb09c 127.0.0.1 writer.go:29: 2021-01-29T19:31:37.303Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server: Handled event for server in area: event=member-join server=Node-d62574a6-25fd-7e70-a7f6-02d7594bb09c.dc1 area=wan writer.go:29: 2021-01-29T19:31:37.303Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server: Adding LAN server: server="Node-d62574a6-25fd-7e70-a7f6-02d7594bb09c (Addr: tcp/127.0.0.1:29657) (DC: dc1)" writer.go:29: 2021-01-29T19:31:37.311Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Started DNS server: address=127.0.0.1:29652 network=udp writer.go:29: 2021-01-29T19:31:37.316Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Started DNS server: address=127.0.0.1:29652 network=tcp writer.go:29: 2021-01-29T19:31:37.321Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Started HTTP server: address=127.0.0.1:29653 network=tcp writer.go:29: 2021-01-29T19:31:37.318Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:37.326Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: entering candidate state: node="Node at 127.0.0.1:29657 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:37.323Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: started state syncer writer.go:29: 2021-01-29T19:31:37.329Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:37.332Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: vote granted: from=d62574a6-25fd-7e70-a7f6-02d7594bb09c term=2 tally=1 writer.go:29: 2021-01-29T19:31:37.333Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:37.335Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: entering leader state: leader="Node at 127.0.0.1:29657 [Leader]" writer.go:29: 2021-01-29T19:31:37.337Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:37.339Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server: New leader elected: payload=Node-d62574a6-25fd-7e70-a7f6-02d7594bb09c writer.go:29: 2021-01-29T19:31:37.342Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29657 writer.go:29: 2021-01-29T19:31:37.348Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:37.359Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:37.361Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:37.363Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server: Skipping self join check for node since the cluster is too small: node=Node-d62574a6-25fd-7e70-a7f6-02d7594bb09c writer.go:29: 2021-01-29T19:31:37.365Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server: member joined, marking health alive: member=Node-d62574a6-25fd-7e70-a7f6-02d7594bb09c writer.go:29: 2021-01-29T19:31:37.404Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: added local registration for service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.411Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Synced node info writer.go:29: 2021-01-29T19:31:37.413Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Synced service: service=web writer.go:29: 2021-01-29T19:31:37.416Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.436Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:37.438Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:37.439Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: removed service: service=web writer.go:29: 2021-01-29T19:31:37.442Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Node info in sync writer.go:29: 2021-01-29T19:31:37.449Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Deregistered service: service=web writer.go:29: 2021-01-29T19:31:37.451Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Service in sync: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.453Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: removed service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.456Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Node info in sync writer.go:29: 2021-01-29T19:31:37.461Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Deregistered service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.465Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Requesting shutdown writer.go:29: 2021-01-29T19:31:37.467Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server: shutting down server writer.go:29: 2021-01-29T19:31:37.469Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:37.471Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:37.472Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:37.474Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:37.478Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:37.478Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: consul server down writer.go:29: 2021-01-29T19:31:37.481Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: shutdown complete writer.go:29: 2021-01-29T19:31:37.482Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Stopping server: protocol=DNS address=127.0.0.1:29652 network=tcp writer.go:29: 2021-01-29T19:31:37.484Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Stopping server: protocol=DNS address=127.0.0.1:29652 network=udp writer.go:29: 2021-01-29T19:31:37.489Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Stopping server: protocol=HTTP address=127.0.0.1:29653 network=tcp === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work writer.go:29: 2021-01-29T19:31:37.628Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:37.630Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work: Endpoints down === RUN TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it writer.go:29: 2021-01-29T19:31:37.641Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:37.652Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:37.654Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:37.659Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:38e283d5-cde5-72d1-9a9b-b9e5d462dd8c Address:127.0.0.1:29663}]" writer.go:29: 2021-01-29T19:31:37.660Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.serf.wan: serf: EventMemberJoin: Node-38e283d5-cde5-72d1-9a9b-b9e5d462dd8c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:37.661Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: entering follower state: follower="Node at 127.0.0.1:29663 [Follower]" leader= writer.go:29: 2021-01-29T19:31:37.663Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.serf.lan: serf: EventMemberJoin: Node-38e283d5-cde5-72d1-9a9b-b9e5d462dd8c 127.0.0.1 writer.go:29: 2021-01-29T19:31:37.671Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server: Adding LAN server: server="Node-38e283d5-cde5-72d1-9a9b-b9e5d462dd8c (Addr: tcp/127.0.0.1:29663) (DC: dc1)" writer.go:29: 2021-01-29T19:31:37.673Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server: Handled event for server in area: event=member-join server=Node-38e283d5-cde5-72d1-9a9b-b9e5d462dd8c.dc1 area=wan writer.go:29: 2021-01-29T19:31:37.679Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Started DNS server: address=127.0.0.1:29658 network=tcp writer.go:29: 2021-01-29T19:31:37.682Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Started DNS server: address=127.0.0.1:29658 network=udp writer.go:29: 2021-01-29T19:31:37.685Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Started HTTP server: address=127.0.0.1:29659 network=tcp writer.go:29: 2021-01-29T19:31:37.687Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: started state syncer writer.go:29: 2021-01-29T19:31:37.723Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:37.724Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: entering candidate state: node="Node at 127.0.0.1:29663 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:37.739Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:37.740Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: vote granted: from=38e283d5-cde5-72d1-9a9b-b9e5d462dd8c term=2 tally=1 writer.go:29: 2021-01-29T19:31:37.742Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:37.744Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.raft: entering leader state: leader="Node at 127.0.0.1:29663 [Leader]" writer.go:29: 2021-01-29T19:31:37.746Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:37.749Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server: New leader elected: payload=Node-38e283d5-cde5-72d1-9a9b-b9e5d462dd8c writer.go:29: 2021-01-29T19:31:37.752Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29663 writer.go:29: 2021-01-29T19:31:37.762Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:37.771Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:37.774Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:37.776Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server: Skipping self join check for node since the cluster is too small: node=Node-38e283d5-cde5-72d1-9a9b-b9e5d462dd8c writer.go:29: 2021-01-29T19:31:37.782Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server: member joined, marking health alive: member=Node-38e283d5-cde5-72d1-9a9b-b9e5d462dd8c writer.go:29: 2021-01-29T19:31:37.867Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:37.869Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Synced node info writer.go:29: 2021-01-29T19:31:37.968Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Node info in sync writer.go:29: 2021-01-29T19:31:37.971Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Synced service: service=web writer.go:29: 2021-01-29T19:31:37.973Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:37.991Z [ERROR] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it writer.go:29: 2021-01-29T19:31:37.992Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:37.994Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it: Endpoints down === CONT TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it writer.go:29: 2021-01-29T19:31:37.996Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: removed service: service=web writer.go:29: 2021-01-29T19:31:38.003Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: removed service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:38.006Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Node info in sync writer.go:29: 2021-01-29T19:31:38.009Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Deregistered service: service=web writer.go:29: 2021-01-29T19:31:38.012Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Deregistered service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:31:38.019Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Requesting shutdown writer.go:29: 2021-01-29T19:31:38.021Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server: shutting down server writer.go:29: 2021-01-29T19:31:38.023Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:38.025Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:38.025Z [DEBUG] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:38.030Z [WARN] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:38.035Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:38.035Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: consul server down writer.go:29: 2021-01-29T19:31:38.037Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: shutdown complete writer.go:29: 2021-01-29T19:31:38.038Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Stopping server: protocol=DNS address=127.0.0.1:29658 network=tcp writer.go:29: 2021-01-29T19:31:38.039Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Stopping server: protocol=DNS address=127.0.0.1:29658 network=udp writer.go:29: 2021-01-29T19:31:38.040Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Stopping server: protocol=HTTP address=127.0.0.1:29659 network=tcp writer.go:29: 2021-01-29T19:31:38.541Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:38.543Z [INFO] TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it: Endpoints down --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar (0.00s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager (13.06s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/sanity_check_no_sidecar_case (1.00s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/default_sidecar (0.77s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_defaults (1.07s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_denied (0.88s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_sidecar (1.06s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination (0.98s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_not_for_overridden_sidecar (0.40s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/ACL_OK_for_service_but_and_overridden_for_sidecar (0.82s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_definition_in_sidecar (0.83s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_definitions_in_sidecar (0.77s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_check_status_in_sidecar (0.91s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/invalid_checks_status_in_sidecar (1.01s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered (0.77s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/updates_to_sidecar_should_work (0.92s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/service_manager/update_that_removes_sidecar_should_NOT_deregister_it (0.87s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal (13.61s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/sanity_check_no_sidecar_case (0.89s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/default_sidecar (1.17s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_defaults (0.87s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_denied (0.77s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_sidecar (0.85s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_and_sidecar_but_not_sidecar's_overridden_destination (0.88s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_not_for_overridden_sidecar (0.83s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/ACL_OK_for_service_but_and_overridden_for_sidecar (0.97s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_definition_in_sidecar (0.97s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_definitions_in_sidecar (0.91s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_check_status_in_sidecar (1.05s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/invalid_checks_status_in_sidecar (0.76s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/another_service_registered_with_same_ID_as_a_sidecar_should_not_be_deregistered (0.71s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/updates_to_sidecar_should_work (1.06s) --- PASS: TestAgent_RegisterServiceDeregisterService_Sidecar/normal/update_that_removes_sidecar_should_NOT_deregister_it (0.91s) === RUN TestAgent_RegisterService_UnmanagedConnectProxyInvalid === RUN TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal === PAUSE TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal === RUN TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager === PAUSE TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.564Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:38.566Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.tlsutil: Update: version=1 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.568Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: bootstrap = true: do not enable unless necessary === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.569Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.570Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:38.574Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:38.592Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c4c16ce9-bfff-ec10-fba6-3782de9dc673 Address:127.0.0.1:29675}]" === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.592Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3cbf6f5e-d631-d8a5-9deb-33440acba073 Address:127.0.0.1:29669}]" === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.594Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29675 [Follower]" leader= === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.595Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29669 [Follower]" leader= === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.594Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.serf.wan: serf: EventMemberJoin: Node-c4c16ce9-bfff-ec10-fba6-3782de9dc673.dc1 127.0.0.1 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.596Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.serf.wan: serf: EventMemberJoin: Node-3cbf6f5e-d631-d8a5-9deb-33440acba073.dc1 127.0.0.1 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.599Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.serf.lan: serf: EventMemberJoin: Node-c4c16ce9-bfff-ec10-fba6-3782de9dc673 127.0.0.1 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.612Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.serf.lan: serf: EventMemberJoin: Node-3cbf6f5e-d631-d8a5-9deb-33440acba073 127.0.0.1 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.616Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Started DNS server: address=127.0.0.1:29670 network=udp === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.619Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Started DNS server: address=127.0.0.1:29664 network=udp === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.622Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server: Handled event for server in area: event=member-join server=Node-c4c16ce9-bfff-ec10-fba6-3782de9dc673.dc1 area=wan writer.go:29: 2021-01-29T19:31:38.623Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server: Adding LAN server: server="Node-c4c16ce9-bfff-ec10-fba6-3782de9dc673 (Addr: tcp/127.0.0.1:29675) (DC: dc1)" === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.622Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server: Adding LAN server: server="Node-3cbf6f5e-d631-d8a5-9deb-33440acba073 (Addr: tcp/127.0.0.1:29669) (DC: dc1)" === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.632Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Started DNS server: address=127.0.0.1:29670 network=tcp writer.go:29: 2021-01-29T19:31:38.666Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Started HTTP server: address=127.0.0.1:29671 network=tcp writer.go:29: 2021-01-29T19:31:38.667Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: started state syncer writer.go:29: 2021-01-29T19:31:38.656Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.630Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server: Handled event for server in area: event=member-join server=Node-3cbf6f5e-d631-d8a5-9deb-33440acba073.dc1 area=wan === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.670Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29675 [Candidate]" term=2 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.631Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Started DNS server: address=127.0.0.1:29664 network=tcp writer.go:29: 2021-01-29T19:31:38.661Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:38.680Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29669 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:38.682Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Started HTTP server: address=127.0.0.1:29665 network=tcp writer.go:29: 2021-01-29T19:31:38.702Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: started state syncer writer.go:29: 2021-01-29T19:31:38.685Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:38.714Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.raft: vote granted: from=3cbf6f5e-d631-d8a5-9deb-33440acba073 term=2 tally=1 writer.go:29: 2021-01-29T19:31:38.715Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:38.716Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29669 [Leader]" writer.go:29: 2021-01-29T19:31:38.718Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:38.719Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server: New leader elected: payload=Node-3cbf6f5e-d631-d8a5-9deb-33440acba073 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.722Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:38.723Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.raft: vote granted: from=c4c16ce9-bfff-ec10-fba6-3782de9dc673 term=2 tally=1 writer.go:29: 2021-01-29T19:31:38.724Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:38.725Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29675 [Leader]" writer.go:29: 2021-01-29T19:31:38.726Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:38.728Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server: New leader elected: payload=Node-c4c16ce9-bfff-ec10-fba6-3782de9dc673 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.728Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29669 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.731Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29675 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.735Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.739Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.741Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:38.744Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.747Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.746Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server: Skipping self join check for node since the cluster is too small: node=Node-3cbf6f5e-d631-d8a5-9deb-33440acba073 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.748Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.748Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server: member joined, marking health alive: member=Node-3cbf6f5e-d631-d8a5-9deb-33440acba073 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.749Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-c4c16ce9-bfff-ec10-fba6-3782de9dc673 writer.go:29: 2021-01-29T19:31:38.750Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server: member joined, marking health alive: member=Node-c4c16ce9-bfff-ec10-fba6-3782de9dc673 === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:38.896Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:38.897Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:38.897Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:38.900Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:38.901Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:38.900Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Synced node info writer.go:29: 2021-01-29T19:31:38.903Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:38.905Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:38.907Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:38.908Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: consul server down writer.go:29: 2021-01-29T19:31:38.909Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: shutdown complete writer.go:29: 2021-01-29T19:31:38.911Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Stopping server: protocol=DNS address=127.0.0.1:29664 network=tcp writer.go:29: 2021-01-29T19:31:38.912Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Stopping server: protocol=DNS address=127.0.0.1:29664 network=udp writer.go:29: 2021-01-29T19:31:38.914Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Stopping server: protocol=HTTP address=127.0.0.1:29665 network=tcp === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:38.922Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:38.924Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:38.988Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:38.989Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:38.990Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:38.991Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:38.991Z [DEBUG] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:38.993Z [WARN] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:38.996Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:38.996Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: consul server down writer.go:29: 2021-01-29T19:31:38.998Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:38.999Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29670 network=tcp writer.go:29: 2021-01-29T19:31:39.000Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29670 network=udp writer.go:29: 2021-01-29T19:31:39.002Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29671 network=tcp === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal writer.go:29: 2021-01-29T19:31:39.415Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:39.417Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal: Endpoints down === CONT TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager writer.go:29: 2021-01-29T19:31:39.504Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:39.505Z [INFO] TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager: Endpoints down --- PASS: TestAgent_RegisterService_UnmanagedConnectProxyInvalid (0.00s) --- PASS: TestAgent_RegisterService_UnmanagedConnectProxyInvalid/normal (0.87s) --- PASS: TestAgent_RegisterService_UnmanagedConnectProxyInvalid/service_manager (0.96s) === RUN TestAgent_RegisterService_ConnectNative === RUN TestAgent_RegisterService_ConnectNative/normal === PAUSE TestAgent_RegisterService_ConnectNative/normal === RUN TestAgent_RegisterService_ConnectNative/service_manager === PAUSE TestAgent_RegisterService_ConnectNative/service_manager === CONT TestAgent_RegisterService_ConnectNative/normal === CONT TestAgent_RegisterService_ConnectNative/service_manager === CONT TestAgent_RegisterService_ConnectNative/normal writer.go:29: 2021-01-29T19:31:39.522Z [WARN] TestAgent_RegisterService_ConnectNative/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:39.524Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:39.526Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:39.535Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5a90eff1-4ed2-3767-943d-6b5a4dce66a2 Address:127.0.0.1:29681}]" writer.go:29: 2021-01-29T19:31:39.538Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29681 [Follower]" leader= writer.go:29: 2021-01-29T19:31:39.538Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server.serf.wan: serf: EventMemberJoin: Node-5a90eff1-4ed2-3767-943d-6b5a4dce66a2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:39.551Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server.serf.lan: serf: EventMemberJoin: Node-5a90eff1-4ed2-3767-943d-6b5a4dce66a2 127.0.0.1 writer.go:29: 2021-01-29T19:31:39.573Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server: Adding LAN server: server="Node-5a90eff1-4ed2-3767-943d-6b5a4dce66a2 (Addr: tcp/127.0.0.1:29681) (DC: dc1)" writer.go:29: 2021-01-29T19:31:39.575Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server: Handled event for server in area: event=member-join server=Node-5a90eff1-4ed2-3767-943d-6b5a4dce66a2.dc1 area=wan === CONT TestAgent_RegisterService_ConnectNative/service_manager writer.go:29: 2021-01-29T19:31:39.576Z [WARN] TestAgent_RegisterService_ConnectNative/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:39.577Z [DEBUG] TestAgent_RegisterService_ConnectNative/service_manager.tlsutil: Update: version=1 === CONT TestAgent_RegisterService_ConnectNative/normal writer.go:29: 2021-01-29T19:31:39.578Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Started DNS server: address=127.0.0.1:29676 network=udp writer.go:29: 2021-01-29T19:31:39.579Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Started DNS server: address=127.0.0.1:29676 network=tcp === CONT TestAgent_RegisterService_ConnectNative/service_manager writer.go:29: 2021-01-29T19:31:39.579Z [DEBUG] TestAgent_RegisterService_ConnectNative/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ConnectNative/normal writer.go:29: 2021-01-29T19:31:39.581Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Started HTTP server: address=127.0.0.1:29677 network=tcp writer.go:29: 2021-01-29T19:31:39.579Z [WARN] TestAgent_RegisterService_ConnectNative/normal.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_RegisterService_ConnectNative/service_manager writer.go:29: 2021-01-29T19:31:39.585Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f853c97f-125a-a0e1-3f83-68a4698cb48c Address:127.0.0.1:29687}]" writer.go:29: 2021-01-29T19:31:39.586Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29687 [Follower]" leader= === CONT TestAgent_RegisterService_ConnectNative/normal writer.go:29: 2021-01-29T19:31:39.584Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29681 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:39.582Z [INFO] TestAgent_RegisterService_ConnectNative/normal: started state syncer === CONT TestAgent_RegisterService_ConnectNative/service_manager writer.go:29: 2021-01-29T19:31:39.587Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server.serf.wan: serf: EventMemberJoin: Node-f853c97f-125a-a0e1-3f83-68a4698cb48c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:39.591Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server.serf.lan: serf: EventMemberJoin: Node-f853c97f-125a-a0e1-3f83-68a4698cb48c 127.0.0.1 writer.go:29: 2021-01-29T19:31:39.595Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server: Adding LAN server: server="Node-f853c97f-125a-a0e1-3f83-68a4698cb48c (Addr: tcp/127.0.0.1:29687) (DC: dc1)" writer.go:29: 2021-01-29T19:31:39.596Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server: Handled event for server in area: event=member-join server=Node-f853c97f-125a-a0e1-3f83-68a4698cb48c.dc1 area=wan writer.go:29: 2021-01-29T19:31:39.600Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Started DNS server: address=127.0.0.1:29682 network=tcp writer.go:29: 2021-01-29T19:31:39.602Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Started DNS server: address=127.0.0.1:29682 network=udp writer.go:29: 2021-01-29T19:31:39.604Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Started HTTP server: address=127.0.0.1:29683 network=tcp writer.go:29: 2021-01-29T19:31:39.605Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: started state syncer === CONT TestAgent_RegisterService_ConnectNative/normal writer.go:29: 2021-01-29T19:31:39.590Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:39.607Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal.server.raft: vote granted: from=5a90eff1-4ed2-3767-943d-6b5a4dce66a2 term=2 tally=1 writer.go:29: 2021-01-29T19:31:39.608Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:39.608Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29681 [Leader]" writer.go:29: 2021-01-29T19:31:39.609Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:39.611Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server: New leader elected: payload=Node-5a90eff1-4ed2-3767-943d-6b5a4dce66a2 writer.go:29: 2021-01-29T19:31:39.618Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29681 writer.go:29: 2021-01-29T19:31:39.633Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:39.647Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:39.648Z [INFO] TestAgent_RegisterService_ConnectNative/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.649Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal.server: Skipping self join check for node since the cluster is too small: node=Node-5a90eff1-4ed2-3767-943d-6b5a4dce66a2 writer.go:29: 2021-01-29T19:31:39.651Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server: member joined, marking health alive: member=Node-5a90eff1-4ed2-3767-943d-6b5a4dce66a2 === CONT TestAgent_RegisterService_ConnectNative/service_manager writer.go:29: 2021-01-29T19:31:39.656Z [WARN] TestAgent_RegisterService_ConnectNative/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:39.659Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29687 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:39.665Z [DEBUG] TestAgent_RegisterService_ConnectNative/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:39.668Z [DEBUG] TestAgent_RegisterService_ConnectNative/service_manager.server.raft: vote granted: from=f853c97f-125a-a0e1-3f83-68a4698cb48c term=2 tally=1 writer.go:29: 2021-01-29T19:31:39.670Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:39.672Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29687 [Leader]" writer.go:29: 2021-01-29T19:31:39.673Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:39.675Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server: New leader elected: payload=Node-f853c97f-125a-a0e1-3f83-68a4698cb48c writer.go:29: 2021-01-29T19:31:39.676Z [DEBUG] TestAgent_RegisterService_ConnectNative/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29687 writer.go:29: 2021-01-29T19:31:39.700Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:39.709Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:39.710Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.712Z [DEBUG] TestAgent_RegisterService_ConnectNative/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-f853c97f-125a-a0e1-3f83-68a4698cb48c writer.go:29: 2021-01-29T19:31:39.713Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server: member joined, marking health alive: member=Node-f853c97f-125a-a0e1-3f83-68a4698cb48c === CONT TestAgent_RegisterService_ConnectNative/normal writer.go:29: 2021-01-29T19:31:39.719Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:39.720Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Synced node info writer.go:29: 2021-01-29T19:31:39.722Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal: Node info in sync writer.go:29: 2021-01-29T19:31:39.795Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal: Node info in sync writer.go:29: 2021-01-29T19:31:39.803Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Synced service: service=web writer.go:29: 2021-01-29T19:31:39.805Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal: Check in sync: check=service:web writer.go:29: 2021-01-29T19:31:39.810Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:39.812Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:39.814Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.815Z [WARN] TestAgent_RegisterService_ConnectNative/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.817Z [DEBUG] TestAgent_RegisterService_ConnectNative/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.821Z [WARN] TestAgent_RegisterService_ConnectNative/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.824Z [INFO] TestAgent_RegisterService_ConnectNative/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:39.824Z [INFO] TestAgent_RegisterService_ConnectNative/normal: consul server down writer.go:29: 2021-01-29T19:31:39.828Z [INFO] TestAgent_RegisterService_ConnectNative/normal: shutdown complete writer.go:29: 2021-01-29T19:31:39.829Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Stopping server: protocol=DNS address=127.0.0.1:29676 network=tcp writer.go:29: 2021-01-29T19:31:39.831Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Stopping server: protocol=DNS address=127.0.0.1:29676 network=udp writer.go:29: 2021-01-29T19:31:39.832Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Stopping server: protocol=HTTP address=127.0.0.1:29677 network=tcp === CONT TestAgent_RegisterService_ConnectNative/service_manager writer.go:29: 2021-01-29T19:31:39.880Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:39.882Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Synced service: service=web writer.go:29: 2021-01-29T19:31:39.883Z [DEBUG] TestAgent_RegisterService_ConnectNative/service_manager: Check in sync: check=service:web writer.go:29: 2021-01-29T19:31:39.887Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:39.888Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:39.890Z [DEBUG] TestAgent_RegisterService_ConnectNative/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.892Z [WARN] TestAgent_RegisterService_ConnectNative/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.892Z [ERROR] TestAgent_RegisterService_ConnectNative/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:39.896Z [DEBUG] TestAgent_RegisterService_ConnectNative/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.899Z [WARN] TestAgent_RegisterService_ConnectNative/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.901Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:39.902Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: consul server down writer.go:29: 2021-01-29T19:31:39.905Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:39.906Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29682 network=tcp writer.go:29: 2021-01-29T19:31:39.910Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29682 network=udp writer.go:29: 2021-01-29T19:31:39.913Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29683 network=tcp === CONT TestAgent_RegisterService_ConnectNative/normal writer.go:29: 2021-01-29T19:31:40.334Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:40.335Z [INFO] TestAgent_RegisterService_ConnectNative/normal: Endpoints down === CONT TestAgent_RegisterService_ConnectNative/service_manager writer.go:29: 2021-01-29T19:31:40.416Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:40.417Z [INFO] TestAgent_RegisterService_ConnectNative/service_manager: Endpoints down --- PASS: TestAgent_RegisterService_ConnectNative (0.00s) --- PASS: TestAgent_RegisterService_ConnectNative/normal (0.83s) --- PASS: TestAgent_RegisterService_ConnectNative/service_manager (0.91s) === RUN TestAgent_RegisterService_ScriptCheck_ExecDisable === RUN TestAgent_RegisterService_ScriptCheck_ExecDisable/normal === PAUSE TestAgent_RegisterService_ScriptCheck_ExecDisable/normal === RUN TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager === PAUSE TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal writer.go:29: 2021-01-29T19:31:40.430Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:40.431Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:40.432Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager writer.go:29: 2021-01-29T19:31:40.440Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:40.443Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.tlsutil: Update: version=1 === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal writer.go:29: 2021-01-29T19:31:40.445Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2fc6ce35-4e9b-1a31-342b-6542f6ee5978 Address:127.0.0.1:29693}]" === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager writer.go:29: 2021-01-29T19:31:40.446Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal writer.go:29: 2021-01-29T19:31:40.450Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.serf.wan: serf: EventMemberJoin: Node-2fc6ce35-4e9b-1a31-342b-6542f6ee5978.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:40.451Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29693 [Follower]" leader= writer.go:29: 2021-01-29T19:31:40.471Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.serf.lan: serf: EventMemberJoin: Node-2fc6ce35-4e9b-1a31-342b-6542f6ee5978 127.0.0.1 writer.go:29: 2021-01-29T19:31:40.477Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server: Handled event for server in area: event=member-join server=Node-2fc6ce35-4e9b-1a31-342b-6542f6ee5978.dc1 area=wan writer.go:29: 2021-01-29T19:31:40.478Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: Started DNS server: address=127.0.0.1:29688 network=udp writer.go:29: 2021-01-29T19:31:40.479Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server: Adding LAN server: server="Node-2fc6ce35-4e9b-1a31-342b-6542f6ee5978 (Addr: tcp/127.0.0.1:29693) (DC: dc1)" writer.go:29: 2021-01-29T19:31:40.502Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: Started DNS server: address=127.0.0.1:29688 network=tcp writer.go:29: 2021-01-29T19:31:40.519Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: Started HTTP server: address=127.0.0.1:29689 network=tcp writer.go:29: 2021-01-29T19:31:40.522Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: started state syncer === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager writer.go:29: 2021-01-29T19:31:40.522Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4225ad76-002f-9d2b-6952-a17f5e2f1007 Address:127.0.0.1:29699}]" writer.go:29: 2021-01-29T19:31:40.524Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29699 [Follower]" leader= writer.go:29: 2021-01-29T19:31:40.525Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.serf.wan: serf: EventMemberJoin: Node-4225ad76-002f-9d2b-6952-a17f5e2f1007.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:40.528Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.serf.lan: serf: EventMemberJoin: Node-4225ad76-002f-9d2b-6952-a17f5e2f1007 127.0.0.1 writer.go:29: 2021-01-29T19:31:40.531Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: Started DNS server: address=127.0.0.1:29694 network=udp writer.go:29: 2021-01-29T19:31:40.531Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server: Adding LAN server: server="Node-4225ad76-002f-9d2b-6952-a17f5e2f1007 (Addr: tcp/127.0.0.1:29699) (DC: dc1)" writer.go:29: 2021-01-29T19:31:40.532Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server: Handled event for server in area: event=member-join server=Node-4225ad76-002f-9d2b-6952-a17f5e2f1007.dc1 area=wan writer.go:29: 2021-01-29T19:31:40.533Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: Started DNS server: address=127.0.0.1:29694 network=tcp writer.go:29: 2021-01-29T19:31:40.537Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: Started HTTP server: address=127.0.0.1:29695 network=tcp writer.go:29: 2021-01-29T19:31:40.538Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: started state syncer === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal writer.go:29: 2021-01-29T19:31:40.539Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:40.543Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29693 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:40.546Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:40.548Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.raft: vote granted: from=2fc6ce35-4e9b-1a31-342b-6542f6ee5978 term=2 tally=1 writer.go:29: 2021-01-29T19:31:40.550Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:40.553Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29693 [Leader]" writer.go:29: 2021-01-29T19:31:40.554Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:40.556Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server: New leader elected: payload=Node-2fc6ce35-4e9b-1a31-342b-6542f6ee5978 writer.go:29: 2021-01-29T19:31:40.561Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29693 writer.go:29: 2021-01-29T19:31:40.573Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager writer.go:29: 2021-01-29T19:31:40.580Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal writer.go:29: 2021-01-29T19:31:40.580Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager writer.go:29: 2021-01-29T19:31:40.581Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29699 [Candidate]" term=2 === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal writer.go:29: 2021-01-29T19:31:40.582Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager writer.go:29: 2021-01-29T19:31:40.583Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:40.584Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.raft: vote granted: from=4225ad76-002f-9d2b-6952-a17f5e2f1007 term=2 tally=1 === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal writer.go:29: 2021-01-29T19:31:40.583Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server: Skipping self join check for node since the cluster is too small: node=Node-2fc6ce35-4e9b-1a31-342b-6542f6ee5978 === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager writer.go:29: 2021-01-29T19:31:40.584Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.raft: election won: tally=1 === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal writer.go:29: 2021-01-29T19:31:40.585Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server: member joined, marking health alive: member=Node-2fc6ce35-4e9b-1a31-342b-6542f6ee5978 === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager writer.go:29: 2021-01-29T19:31:40.585Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29699 [Leader]" writer.go:29: 2021-01-29T19:31:40.586Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:40.587Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server: New leader elected: payload=Node-4225ad76-002f-9d2b-6952-a17f5e2f1007 writer.go:29: 2021-01-29T19:31:40.588Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29699 writer.go:29: 2021-01-29T19:31:40.602Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:40.607Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:40.609Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:40.610Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-4225ad76-002f-9d2b-6952-a17f5e2f1007 writer.go:29: 2021-01-29T19:31:40.612Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server: member joined, marking health alive: member=Node-4225ad76-002f-9d2b-6952-a17f5e2f1007 === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal writer.go:29: 2021-01-29T19:31:40.698Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:40.699Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:40.701Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:40.702Z [ERROR] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:40.703Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:40.703Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:40.710Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:40.716Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:40.717Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: consul server down writer.go:29: 2021-01-29T19:31:40.718Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: shutdown complete writer.go:29: 2021-01-29T19:31:40.719Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: Stopping server: protocol=DNS address=127.0.0.1:29688 network=tcp writer.go:29: 2021-01-29T19:31:40.720Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: Stopping server: protocol=DNS address=127.0.0.1:29688 network=udp writer.go:29: 2021-01-29T19:31:40.721Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: Stopping server: protocol=HTTP address=127.0.0.1:29689 network=tcp === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager writer.go:29: 2021-01-29T19:31:40.950Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:40.952Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:40.954Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:40.956Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:40.956Z [ERROR] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:40.956Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:40.963Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:40.965Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:40.965Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: consul server down writer.go:29: 2021-01-29T19:31:40.967Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:40.968Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29694 network=tcp writer.go:29: 2021-01-29T19:31:40.968Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29694 network=udp writer.go:29: 2021-01-29T19:31:40.969Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29695 network=tcp === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/normal writer.go:29: 2021-01-29T19:31:41.223Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:41.225Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/normal: Endpoints down === CONT TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager writer.go:29: 2021-01-29T19:31:41.472Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:41.473Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager: Endpoints down --- PASS: TestAgent_RegisterService_ScriptCheck_ExecDisable (0.00s) --- PASS: TestAgent_RegisterService_ScriptCheck_ExecDisable/normal (0.81s) --- PASS: TestAgent_RegisterService_ScriptCheck_ExecDisable/service_manager (1.05s) === RUN TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable === RUN TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal === PAUSE TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal === RUN TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager === PAUSE TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.489Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:41.490Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:41.492Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.508Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: bootstrap = true: do not enable unless necessary === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.510Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3074ad6d-476f-2844-9e3b-63257d5fcce0 Address:127.0.0.1:29711}]" === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.510Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.tlsutil: Update: version=1 === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.512Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29711 [Follower]" leader= === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.512Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.515Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.serf.wan: serf: EventMemberJoin: Node-3074ad6d-476f-2844-9e3b-63257d5fcce0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:41.517Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.serf.lan: serf: EventMemberJoin: Node-3074ad6d-476f-2844-9e3b-63257d5fcce0 127.0.0.1 writer.go:29: 2021-01-29T19:31:41.520Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Started DNS server: address=127.0.0.1:29706 network=udp writer.go:29: 2021-01-29T19:31:41.521Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server: Adding LAN server: server="Node-3074ad6d-476f-2844-9e3b-63257d5fcce0 (Addr: tcp/127.0.0.1:29711) (DC: dc1)" === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.519Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2dd1117e-f381-e345-eaf1-0cf4b2c9c7b9 Address:127.0.0.1:29705}]" === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.521Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server: Handled event for server in area: event=member-join server=Node-3074ad6d-476f-2844-9e3b-63257d5fcce0.dc1 area=wan writer.go:29: 2021-01-29T19:31:41.522Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Started DNS server: address=127.0.0.1:29706 network=tcp writer.go:29: 2021-01-29T19:31:41.527Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Started HTTP server: address=127.0.0.1:29707 network=tcp writer.go:29: 2021-01-29T19:31:41.528Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: started state syncer === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.530Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.serf.wan: serf: EventMemberJoin: Node-2dd1117e-f381-e345-eaf1-0cf4b2c9c7b9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:41.533Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29705 [Follower]" leader= writer.go:29: 2021-01-29T19:31:41.540Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.serf.lan: serf: EventMemberJoin: Node-2dd1117e-f381-e345-eaf1-0cf4b2c9c7b9 127.0.0.1 writer.go:29: 2021-01-29T19:31:41.544Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server: Handled event for server in area: event=member-join server=Node-2dd1117e-f381-e345-eaf1-0cf4b2c9c7b9.dc1 area=wan writer.go:29: 2021-01-29T19:31:41.544Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server: Adding LAN server: server="Node-2dd1117e-f381-e345-eaf1-0cf4b2c9c7b9 (Addr: tcp/127.0.0.1:29705) (DC: dc1)" writer.go:29: 2021-01-29T19:31:41.548Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: Started DNS server: address=127.0.0.1:29700 network=tcp writer.go:29: 2021-01-29T19:31:41.549Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: Started DNS server: address=127.0.0.1:29700 network=udp writer.go:29: 2021-01-29T19:31:41.551Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: Started HTTP server: address=127.0.0.1:29701 network=tcp writer.go:29: 2021-01-29T19:31:41.552Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: started state syncer === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.579Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:41.581Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29711 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:41.584Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:41.586Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.raft: vote granted: from=3074ad6d-476f-2844-9e3b-63257d5fcce0 term=2 tally=1 writer.go:29: 2021-01-29T19:31:41.587Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.raft: election won: tally=1 === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.594Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:41.596Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29705 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:41.598Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:41.599Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.raft: vote granted: from=2dd1117e-f381-e345-eaf1-0cf4b2c9c7b9 term=2 tally=1 writer.go:29: 2021-01-29T19:31:41.601Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.raft: election won: tally=1 === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.589Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29711 [Leader]" writer.go:29: 2021-01-29T19:31:41.603Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:41.605Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server: New leader elected: payload=Node-3074ad6d-476f-2844-9e3b-63257d5fcce0 === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.602Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29705 [Leader]" writer.go:29: 2021-01-29T19:31:41.606Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:41.608Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server: New leader elected: payload=Node-2dd1117e-f381-e345-eaf1-0cf4b2c9c7b9 === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.608Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29711 writer.go:29: 2021-01-29T19:31:41.614Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.613Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29705 === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.619Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.619Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.620Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.622Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-3074ad6d-476f-2844-9e3b-63257d5fcce0 writer.go:29: 2021-01-29T19:31:41.624Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server: member joined, marking health alive: member=Node-3074ad6d-476f-2844-9e3b-63257d5fcce0 === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.630Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:41.632Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.634Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server: Skipping self join check for node since the cluster is too small: node=Node-2dd1117e-f381-e345-eaf1-0cf4b2c9c7b9 writer.go:29: 2021-01-29T19:31:41.635Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server: member joined, marking health alive: member=Node-2dd1117e-f381-e345-eaf1-0cf4b2c9c7b9 === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:41.663Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:41.665Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:41.684Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:41.685Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:41.686Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.687Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:41.687Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.690Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:41.693Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:41.693Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: consul server down writer.go:29: 2021-01-29T19:31:41.696Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:41.698Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29706 network=tcp writer.go:29: 2021-01-29T19:31:41.699Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29706 network=udp writer.go:29: 2021-01-29T19:31:41.701Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29707 network=tcp === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:41.741Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:41.742Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:41.744Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.745Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:41.744Z [ERROR] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:41.745Z [DEBUG] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.748Z [WARN] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:41.755Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:41.755Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: consul server down writer.go:29: 2021-01-29T19:31:41.758Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: shutdown complete writer.go:29: 2021-01-29T19:31:41.760Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: Stopping server: protocol=DNS address=127.0.0.1:29700 network=tcp writer.go:29: 2021-01-29T19:31:41.762Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: Stopping server: protocol=DNS address=127.0.0.1:29700 network=udp writer.go:29: 2021-01-29T19:31:41.764Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: Stopping server: protocol=HTTP address=127.0.0.1:29701 network=tcp === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager writer.go:29: 2021-01-29T19:31:42.203Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:42.204Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager: Endpoints down === CONT TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal writer.go:29: 2021-01-29T19:31:42.266Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:42.267Z [INFO] TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal: Endpoints down --- PASS: TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable (0.00s) --- PASS: TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/service_manager (0.72s) --- PASS: TestAgent_RegisterService_ScriptCheck_ExecRemoteDisable/normal (0.79s) === RUN TestAgent_DeregisterService === PAUSE TestAgent_DeregisterService === RUN TestAgent_DeregisterService_ACLDeny === PAUSE TestAgent_DeregisterService_ACLDeny === RUN TestAgent_ServiceMaintenance_BadRequest === PAUSE TestAgent_ServiceMaintenance_BadRequest === RUN TestAgent_ServiceMaintenance_Enable agent_endpoint_test.go:3953: DM-skipped --- SKIP: TestAgent_ServiceMaintenance_Enable (0.00s) === RUN TestAgent_ServiceMaintenance_Disable === PAUSE TestAgent_ServiceMaintenance_Disable === RUN TestAgent_ServiceMaintenance_ACLDeny agent_endpoint_test.go:4034: DM-skipped --- SKIP: TestAgent_ServiceMaintenance_ACLDeny (0.00s) === RUN TestAgent_NodeMaintenance_BadRequest === PAUSE TestAgent_NodeMaintenance_BadRequest === RUN TestAgent_NodeMaintenance_Enable === PAUSE TestAgent_NodeMaintenance_Enable === RUN TestAgent_NodeMaintenance_Disable === PAUSE TestAgent_NodeMaintenance_Disable === RUN TestAgent_NodeMaintenance_ACLDeny === PAUSE TestAgent_NodeMaintenance_ACLDeny === RUN TestAgent_RegisterCheck_Service === PAUSE TestAgent_RegisterCheck_Service === RUN TestAgent_Monitor agent_endpoint_test.go:4212: DM-skipped --- SKIP: TestAgent_Monitor (0.00s) === RUN TestAgent_Monitor_ACLDeny === PAUSE TestAgent_Monitor_ACLDeny === RUN TestAgent_TokenTriggersFullSync === PAUSE TestAgent_TokenTriggersFullSync === RUN TestAgent_Token === PAUSE TestAgent_Token === RUN TestAgentConnectCARoots_empty === PAUSE TestAgentConnectCARoots_empty === RUN TestAgentConnectCARoots_list === PAUSE TestAgentConnectCARoots_list === RUN TestAgentConnectCALeafCert_aclDefaultDeny === PAUSE TestAgentConnectCALeafCert_aclDefaultDeny === RUN TestAgentConnectCALeafCert_aclServiceWrite === PAUSE TestAgentConnectCALeafCert_aclServiceWrite === RUN TestAgentConnectCALeafCert_aclServiceReadDeny === PAUSE TestAgentConnectCALeafCert_aclServiceReadDeny === RUN TestAgentConnectCALeafCert_good agent_endpoint_test.go:4987: DM-skipped --- SKIP: TestAgentConnectCALeafCert_good (0.00s) === RUN TestAgentConnectCALeafCert_goodNotLocal agent_endpoint_test.go:5091: DM-skipped --- SKIP: TestAgentConnectCALeafCert_goodNotLocal (0.00s) === RUN TestAgentConnectCALeafCert_secondaryDC_good === PAUSE TestAgentConnectCALeafCert_secondaryDC_good === RUN TestAgentConnectAuthorize_badBody === PAUSE TestAgentConnectAuthorize_badBody === RUN TestAgentConnectAuthorize_noTarget === PAUSE TestAgentConnectAuthorize_noTarget === RUN TestAgentConnectAuthorize_idInvalidFormat === PAUSE TestAgentConnectAuthorize_idInvalidFormat === RUN TestAgentConnectAuthorize_idNotService === PAUSE TestAgentConnectAuthorize_idNotService === RUN TestAgentConnectAuthorize_allow === PAUSE TestAgentConnectAuthorize_allow === RUN TestAgentConnectAuthorize_deny === PAUSE TestAgentConnectAuthorize_deny === RUN TestAgentConnectAuthorize_allowTrustDomain === PAUSE TestAgentConnectAuthorize_allowTrustDomain === RUN TestAgentConnectAuthorize_denyWildcard === PAUSE TestAgentConnectAuthorize_denyWildcard === RUN TestAgentConnectAuthorize_serviceWrite === PAUSE TestAgentConnectAuthorize_serviceWrite === RUN TestAgentConnectAuthorize_defaultDeny === PAUSE TestAgentConnectAuthorize_defaultDeny === RUN TestAgentConnectAuthorize_defaultAllow === PAUSE TestAgentConnectAuthorize_defaultAllow === RUN TestAgent_Host === PAUSE TestAgent_Host === RUN TestAgent_HostBadACL === PAUSE TestAgent_HostBadACL === RUN TestAgent_Services_ExposeConfig === PAUSE TestAgent_Services_ExposeConfig === RUN TestAgent_MultiStartStop === RUN TestAgent_MultiStartStop/#00 === PAUSE TestAgent_MultiStartStop/#00 === RUN TestAgent_MultiStartStop/#01 === PAUSE TestAgent_MultiStartStop/#01 === RUN TestAgent_MultiStartStop/#02 === PAUSE TestAgent_MultiStartStop/#02 === RUN TestAgent_MultiStartStop/#03 === PAUSE TestAgent_MultiStartStop/#03 === RUN TestAgent_MultiStartStop/#04 === PAUSE TestAgent_MultiStartStop/#04 === RUN TestAgent_MultiStartStop/#05 === PAUSE TestAgent_MultiStartStop/#05 === RUN TestAgent_MultiStartStop/#06 === PAUSE TestAgent_MultiStartStop/#06 === RUN TestAgent_MultiStartStop/#07 === PAUSE TestAgent_MultiStartStop/#07 === RUN TestAgent_MultiStartStop/#08 === PAUSE TestAgent_MultiStartStop/#08 === RUN TestAgent_MultiStartStop/#09 === PAUSE TestAgent_MultiStartStop/#09 === CONT TestAgent_MultiStartStop/#00 === CONT TestAgent_MultiStartStop/#05 === CONT TestAgent_MultiStartStop/#08 === CONT TestAgent_MultiStartStop/#03 === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.379Z [WARN] TestAgent_MultiStartStop/#08: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:42.393Z [DEBUG] TestAgent_MultiStartStop/#08.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:42.410Z [DEBUG] TestAgent_MultiStartStop/#08.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:42.429Z [INFO] TestAgent_MultiStartStop/#08.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cb057f30-c75d-5ca1-a4d5-51366fa4e348 Address:127.0.0.1:29717}]" writer.go:29: 2021-01-29T19:31:42.435Z [INFO] TestAgent_MultiStartStop/#08.server.raft: entering follower state: follower="Node at 127.0.0.1:29717 [Follower]" leader= writer.go:29: 2021-01-29T19:31:42.502Z [WARN] TestAgent_MultiStartStop/#08.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.506Z [WARN] TestAgent_MultiStartStop/#03: bootstrap = true: do not enable unless necessary === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.505Z [INFO] TestAgent_MultiStartStop/#08.server.raft: entering candidate state: node="Node at 127.0.0.1:29717 [Candidate]" term=2 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.518Z [DEBUG] TestAgent_MultiStartStop/#03.tlsutil: Update: version=1 === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.518Z [WARN] TestAgent_MultiStartStop/#08.server.raft: unable to get address for sever, using fallback address: id=cb057f30-c75d-5ca1-a4d5-51366fa4e348 fallback=127.0.0.1:29717 error="Could not find address for server id cb057f30-c75d-5ca1-a4d5-51366fa4e348" === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.521Z [DEBUG] TestAgent_MultiStartStop/#03.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.520Z [INFO] TestAgent_MultiStartStop/#08.server.serf.wan: serf: EventMemberJoin: Node-cb057f30-c75d-5ca1-a4d5-51366fa4e348.dc1 127.0.0.1 === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.522Z [WARN] TestAgent_MultiStartStop/#05: bootstrap = true: do not enable unless necessary === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.521Z [DEBUG] TestAgent_MultiStartStop/#08.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:42.526Z [DEBUG] TestAgent_MultiStartStop/#08.server.raft: vote granted: from=cb057f30-c75d-5ca1-a4d5-51366fa4e348 term=2 tally=1 writer.go:29: 2021-01-29T19:31:42.527Z [INFO] TestAgent_MultiStartStop/#08.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:42.528Z [INFO] TestAgent_MultiStartStop/#08.server.raft: entering leader state: leader="Node at 127.0.0.1:29717 [Leader]" === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.528Z [INFO] TestAgent_MultiStartStop/#03.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0d3b2cdc-9348-5055-f5e1-a2ccfe4c3161 Address:127.0.0.1:29729}]" === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.526Z [DEBUG] TestAgent_MultiStartStop/#05.tlsutil: Update: version=1 === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.525Z [INFO] TestAgent_MultiStartStop/#08.server.serf.lan: serf: EventMemberJoin: Node-cb057f30-c75d-5ca1-a4d5-51366fa4e348 127.0.0.1 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.530Z [INFO] TestAgent_MultiStartStop/#03.server.serf.wan: serf: EventMemberJoin: Node-0d3b2cdc-9348-5055-f5e1-a2ccfe4c3161.dc1 127.0.0.1 === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.530Z [DEBUG] TestAgent_MultiStartStop/#05.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.532Z [INFO] TestAgent_MultiStartStop/#03.server.serf.lan: serf: EventMemberJoin: Node-0d3b2cdc-9348-5055-f5e1-a2ccfe4c3161 127.0.0.1 === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.531Z [WARN] TestAgent_MultiStartStop/#00: bootstrap = true: do not enable unless necessary === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.532Z [INFO] TestAgent_MultiStartStop/#08: Started DNS server: address=127.0.0.1:29712 network=udp writer.go:29: 2021-01-29T19:31:42.534Z [INFO] TestAgent_MultiStartStop/#08.server: Adding LAN server: server="Node-cb057f30-c75d-5ca1-a4d5-51366fa4e348 (Addr: tcp/127.0.0.1:29717) (DC: dc1)" writer.go:29: 2021-01-29T19:31:42.535Z [INFO] TestAgent_MultiStartStop/#08.server: Handled event for server in area: event=member-join server=Node-cb057f30-c75d-5ca1-a4d5-51366fa4e348.dc1 area=wan === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.533Z [INFO] TestAgent_MultiStartStop/#03.server.raft: entering follower state: follower="Node at 127.0.0.1:29729 [Follower]" leader= === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.536Z [INFO] TestAgent_MultiStartStop/#08.server: cluster leadership acquired === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.534Z [INFO] TestAgent_MultiStartStop/#03: Started DNS server: address=127.0.0.1:29724 network=udp === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.533Z [DEBUG] TestAgent_MultiStartStop/#00.tlsutil: Update: version=1 === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.539Z [INFO] TestAgent_MultiStartStop/#08: Started DNS server: address=127.0.0.1:29712 network=tcp === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.539Z [INFO] TestAgent_MultiStartStop/#03: Started DNS server: address=127.0.0.1:29724 network=tcp === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.540Z [DEBUG] TestAgent_MultiStartStop/#00.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.536Z [INFO] TestAgent_MultiStartStop/#03.server: Adding LAN server: server="Node-0d3b2cdc-9348-5055-f5e1-a2ccfe4c3161 (Addr: tcp/127.0.0.1:29729) (DC: dc1)" writer.go:29: 2021-01-29T19:31:42.540Z [INFO] TestAgent_MultiStartStop/#03.server: Handled event for server in area: event=member-join server=Node-0d3b2cdc-9348-5055-f5e1-a2ccfe4c3161.dc1 area=wan writer.go:29: 2021-01-29T19:31:42.543Z [INFO] TestAgent_MultiStartStop/#03: Started HTTP server: address=127.0.0.1:29725 network=tcp writer.go:29: 2021-01-29T19:31:42.545Z [INFO] TestAgent_MultiStartStop/#03: started state syncer === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.541Z [DEBUG] TestAgent_MultiStartStop/#08.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29717 writer.go:29: 2021-01-29T19:31:42.541Z [INFO] TestAgent_MultiStartStop/#08: Started HTTP server: address=127.0.0.1:29713 network=tcp === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.546Z [INFO] TestAgent_MultiStartStop/#05.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:820c3af5-5718-56b7-d3be-c83983e80385 Address:127.0.0.1:29723}]" === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.541Z [INFO] TestAgent_MultiStartStop/#08.server: New leader elected: payload=Node-cb057f30-c75d-5ca1-a4d5-51366fa4e348 writer.go:29: 2021-01-29T19:31:42.548Z [INFO] TestAgent_MultiStartStop/#08: started state syncer === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.556Z [INFO] TestAgent_MultiStartStop/#00.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:15dc79e9-789d-2e46-1a1e-a65525486ec2 Address:127.0.0.1:29735}]" === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.557Z [INFO] TestAgent_MultiStartStop/#08: Synced node info === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.557Z [INFO] TestAgent_MultiStartStop/#05.server.serf.wan: serf: EventMemberJoin: Node-820c3af5-5718-56b7-d3be-c83983e80385.dc1 127.0.0.1 === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.558Z [DEBUG] TestAgent_MultiStartStop/#08: Node info in sync === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.559Z [INFO] TestAgent_MultiStartStop/#00.server.serf.wan: serf: EventMemberJoin: Node-15dc79e9-789d-2e46-1a1e-a65525486ec2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:42.560Z [INFO] TestAgent_MultiStartStop/#00.server.raft: entering follower state: follower="Node at 127.0.0.1:29735 [Follower]" leader= === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.562Z [INFO] TestAgent_MultiStartStop/#05.server.raft: entering follower state: follower="Node at 127.0.0.1:29723 [Follower]" leader= === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.565Z [INFO] TestAgent_MultiStartStop/#00.server.serf.lan: serf: EventMemberJoin: Node-15dc79e9-789d-2e46-1a1e-a65525486ec2 127.0.0.1 writer.go:29: 2021-01-29T19:31:42.567Z [INFO] TestAgent_MultiStartStop/#00: Started DNS server: address=127.0.0.1:29730 network=udp === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.564Z [INFO] TestAgent_MultiStartStop/#05.server.serf.lan: serf: EventMemberJoin: Node-820c3af5-5718-56b7-d3be-c83983e80385 127.0.0.1 === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.568Z [INFO] TestAgent_MultiStartStop/#00.server: Handled event for server in area: event=member-join server=Node-15dc79e9-789d-2e46-1a1e-a65525486ec2.dc1 area=wan writer.go:29: 2021-01-29T19:31:42.569Z [INFO] TestAgent_MultiStartStop/#00.server: Adding LAN server: server="Node-15dc79e9-789d-2e46-1a1e-a65525486ec2 (Addr: tcp/127.0.0.1:29735) (DC: dc1)" === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.572Z [INFO] TestAgent_MultiStartStop/#05.server: Handled event for server in area: event=member-join server=Node-820c3af5-5718-56b7-d3be-c83983e80385.dc1 area=wan === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.571Z [INFO] TestAgent_MultiStartStop/#00: Started DNS server: address=127.0.0.1:29730 network=tcp === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.572Z [INFO] TestAgent_MultiStartStop/#05.server: Adding LAN server: server="Node-820c3af5-5718-56b7-d3be-c83983e80385 (Addr: tcp/127.0.0.1:29723) (DC: dc1)" writer.go:29: 2021-01-29T19:31:42.573Z [INFO] TestAgent_MultiStartStop/#05: Started DNS server: address=127.0.0.1:29718 network=udp === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.574Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.575Z [INFO] TestAgent_MultiStartStop/#05: Started DNS server: address=127.0.0.1:29718 network=tcp === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.575Z [INFO] TestAgent_MultiStartStop/#00: Started HTTP server: address=127.0.0.1:29731 network=tcp === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.577Z [INFO] TestAgent_MultiStartStop/#05: Started HTTP server: address=127.0.0.1:29719 network=tcp writer.go:29: 2021-01-29T19:31:42.578Z [INFO] TestAgent_MultiStartStop/#05: started state syncer === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.577Z [INFO] TestAgent_MultiStartStop/#00: started state syncer === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.579Z [WARN] TestAgent_MultiStartStop/#03.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:42.579Z [INFO] TestAgent_MultiStartStop/#03.server.raft: entering candidate state: node="Node at 127.0.0.1:29729 [Candidate]" term=2 === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.580Z [INFO] TestAgent_MultiStartStop/#08.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:42.581Z [INFO] TestAgent_MultiStartStop/#08.leader: started routine: routine="CA root pruning" === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.581Z [DEBUG] TestAgent_MultiStartStop/#03.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:42.582Z [DEBUG] TestAgent_MultiStartStop/#03.server.raft: vote granted: from=0d3b2cdc-9348-5055-f5e1-a2ccfe4c3161 term=2 tally=1 === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.582Z [DEBUG] TestAgent_MultiStartStop/#08.server: Skipping self join check for node since the cluster is too small: node=Node-cb057f30-c75d-5ca1-a4d5-51366fa4e348 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.583Z [INFO] TestAgent_MultiStartStop/#03.server.raft: election won: tally=1 === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.583Z [INFO] TestAgent_MultiStartStop/#08.server: member joined, marking health alive: member=Node-cb057f30-c75d-5ca1-a4d5-51366fa4e348 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.584Z [INFO] TestAgent_MultiStartStop/#03.server.raft: entering leader state: leader="Node at 127.0.0.1:29729 [Leader]" writer.go:29: 2021-01-29T19:31:42.585Z [INFO] TestAgent_MultiStartStop/#03.server: cluster leadership acquired === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.586Z [DEBUG] TestAgent_MultiStartStop/#08.server: Skipping self join check for node since the cluster is too small: node=Node-cb057f30-c75d-5ca1-a4d5-51366fa4e348 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.586Z [INFO] TestAgent_MultiStartStop/#03.server: New leader elected: payload=Node-0d3b2cdc-9348-5055-f5e1-a2ccfe4c3161 writer.go:29: 2021-01-29T19:31:42.586Z [DEBUG] TestAgent_MultiStartStop/#03.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29729 writer.go:29: 2021-01-29T19:31:42.594Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:42.600Z [INFO] TestAgent_MultiStartStop/#03.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.601Z [WARN] TestAgent_MultiStartStop/#05.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:42.602Z [INFO] TestAgent_MultiStartStop/#05.server.raft: entering candidate state: node="Node at 127.0.0.1:29723 [Candidate]" term=2 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.602Z [INFO] TestAgent_MultiStartStop/#03.leader: started routine: routine="CA root pruning" === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.604Z [DEBUG] TestAgent_MultiStartStop/#05.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:42.605Z [DEBUG] TestAgent_MultiStartStop/#05.server.raft: vote granted: from=820c3af5-5718-56b7-d3be-c83983e80385 term=2 tally=1 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.604Z [DEBUG] TestAgent_MultiStartStop/#03.server: Skipping self join check for node since the cluster is too small: node=Node-0d3b2cdc-9348-5055-f5e1-a2ccfe4c3161 === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.606Z [INFO] TestAgent_MultiStartStop/#05.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:42.607Z [INFO] TestAgent_MultiStartStop/#05.server.raft: entering leader state: leader="Node at 127.0.0.1:29723 [Leader]" writer.go:29: 2021-01-29T19:31:42.608Z [INFO] TestAgent_MultiStartStop/#05.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:42.609Z [INFO] TestAgent_MultiStartStop/#05.server: New leader elected: payload=Node-820c3af5-5718-56b7-d3be-c83983e80385 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.607Z [INFO] TestAgent_MultiStartStop/#03.server: member joined, marking health alive: member=Node-0d3b2cdc-9348-5055-f5e1-a2ccfe4c3161 === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.620Z [WARN] TestAgent_MultiStartStop/#00.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:42.622Z [INFO] TestAgent_MultiStartStop/#00.server.raft: entering candidate state: node="Node at 127.0.0.1:29735 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:42.625Z [DEBUG] TestAgent_MultiStartStop/#00.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:42.626Z [DEBUG] TestAgent_MultiStartStop/#00.server.raft: vote granted: from=15dc79e9-789d-2e46-1a1e-a65525486ec2 term=2 tally=1 === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.625Z [DEBUG] TestAgent_MultiStartStop/#05.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29723 === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.626Z [INFO] TestAgent_MultiStartStop/#00.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:42.627Z [INFO] TestAgent_MultiStartStop/#00.server.raft: entering leader state: leader="Node at 127.0.0.1:29735 [Leader]" writer.go:29: 2021-01-29T19:31:42.628Z [INFO] TestAgent_MultiStartStop/#00.server: cluster leadership acquired === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.628Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.629Z [INFO] TestAgent_MultiStartStop/#00.server: New leader elected: payload=Node-15dc79e9-789d-2e46-1a1e-a65525486ec2 writer.go:29: 2021-01-29T19:31:42.640Z [DEBUG] TestAgent_MultiStartStop/#00.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29735 === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.643Z [INFO] TestAgent_MultiStartStop/#05.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.644Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.645Z [INFO] TestAgent_MultiStartStop/#05.leader: started routine: routine="CA root pruning" === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.646Z [INFO] TestAgent_MultiStartStop/#00: Synced node info writer.go:29: 2021-01-29T19:31:42.647Z [DEBUG] TestAgent_MultiStartStop/#00: Node info in sync === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.646Z [DEBUG] TestAgent_MultiStartStop/#05.server: Skipping self join check for node since the cluster is too small: node=Node-820c3af5-5718-56b7-d3be-c83983e80385 === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.649Z [INFO] TestAgent_MultiStartStop/#00.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:42.650Z [INFO] TestAgent_MultiStartStop/#00.leader: started routine: routine="CA root pruning" === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.649Z [INFO] TestAgent_MultiStartStop/#05.server: member joined, marking health alive: member=Node-820c3af5-5718-56b7-d3be-c83983e80385 === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:42.651Z [DEBUG] TestAgent_MultiStartStop/#00.server: Skipping self join check for node since the cluster is too small: node=Node-15dc79e9-789d-2e46-1a1e-a65525486ec2 writer.go:29: 2021-01-29T19:31:42.652Z [INFO] TestAgent_MultiStartStop/#00.server: member joined, marking health alive: member=Node-15dc79e9-789d-2e46-1a1e-a65525486ec2 === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.705Z [DEBUG] TestAgent_MultiStartStop/#03: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:42.708Z [INFO] TestAgent_MultiStartStop/#03: Synced node info writer.go:29: 2021-01-29T19:31:42.710Z [DEBUG] TestAgent_MultiStartStop/#03: Node info in sync === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.736Z [DEBUG] TestAgent_MultiStartStop/#05: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:42.738Z [INFO] TestAgent_MultiStartStop/#05: Synced node info === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:42.799Z [INFO] TestAgent_MultiStartStop/#08: Requesting shutdown writer.go:29: 2021-01-29T19:31:42.802Z [INFO] TestAgent_MultiStartStop/#08.server: shutting down server writer.go:29: 2021-01-29T19:31:42.804Z [DEBUG] TestAgent_MultiStartStop/#08.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:42.806Z [WARN] TestAgent_MultiStartStop/#08.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:42.806Z [DEBUG] TestAgent_MultiStartStop/#08.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:42.810Z [WARN] TestAgent_MultiStartStop/#08.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:42.813Z [INFO] TestAgent_MultiStartStop/#08.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:42.813Z [INFO] TestAgent_MultiStartStop/#08: consul server down writer.go:29: 2021-01-29T19:31:42.817Z [INFO] TestAgent_MultiStartStop/#08: shutdown complete writer.go:29: 2021-01-29T19:31:42.819Z [INFO] TestAgent_MultiStartStop/#08: Stopping server: protocol=DNS address=127.0.0.1:29712 network=tcp writer.go:29: 2021-01-29T19:31:42.821Z [INFO] TestAgent_MultiStartStop/#08: Stopping server: protocol=DNS address=127.0.0.1:29712 network=udp writer.go:29: 2021-01-29T19:31:42.822Z [INFO] TestAgent_MultiStartStop/#08: Stopping server: protocol=HTTP address=127.0.0.1:29713 network=tcp === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:42.872Z [INFO] TestAgent_MultiStartStop/#05: Requesting shutdown writer.go:29: 2021-01-29T19:31:42.875Z [INFO] TestAgent_MultiStartStop/#05.server: shutting down server writer.go:29: 2021-01-29T19:31:42.876Z [DEBUG] TestAgent_MultiStartStop/#05.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:42.878Z [WARN] TestAgent_MultiStartStop/#05.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:42.878Z [DEBUG] TestAgent_MultiStartStop/#05.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:42.881Z [WARN] TestAgent_MultiStartStop/#05.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:42.885Z [INFO] TestAgent_MultiStartStop/#05.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:42.885Z [INFO] TestAgent_MultiStartStop/#05: consul server down writer.go:29: 2021-01-29T19:31:42.888Z [INFO] TestAgent_MultiStartStop/#05: shutdown complete writer.go:29: 2021-01-29T19:31:42.889Z [INFO] TestAgent_MultiStartStop/#05: Stopping server: protocol=DNS address=127.0.0.1:29718 network=tcp writer.go:29: 2021-01-29T19:31:42.890Z [INFO] TestAgent_MultiStartStop/#05: Stopping server: protocol=DNS address=127.0.0.1:29718 network=udp writer.go:29: 2021-01-29T19:31:42.891Z [INFO] TestAgent_MultiStartStop/#05: Stopping server: protocol=HTTP address=127.0.0.1:29719 network=tcp === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:42.940Z [INFO] TestAgent_MultiStartStop/#03: Requesting shutdown writer.go:29: 2021-01-29T19:31:42.942Z [INFO] TestAgent_MultiStartStop/#03.server: shutting down server writer.go:29: 2021-01-29T19:31:42.944Z [DEBUG] TestAgent_MultiStartStop/#03.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:42.945Z [WARN] TestAgent_MultiStartStop/#03.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:42.945Z [DEBUG] TestAgent_MultiStartStop/#03.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:42.949Z [WARN] TestAgent_MultiStartStop/#03.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:42.953Z [INFO] TestAgent_MultiStartStop/#03.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:42.953Z [INFO] TestAgent_MultiStartStop/#03: consul server down writer.go:29: 2021-01-29T19:31:42.956Z [INFO] TestAgent_MultiStartStop/#03: shutdown complete writer.go:29: 2021-01-29T19:31:42.957Z [INFO] TestAgent_MultiStartStop/#03: Stopping server: protocol=DNS address=127.0.0.1:29724 network=tcp writer.go:29: 2021-01-29T19:31:42.959Z [INFO] TestAgent_MultiStartStop/#03: Stopping server: protocol=DNS address=127.0.0.1:29724 network=udp writer.go:29: 2021-01-29T19:31:42.961Z [INFO] TestAgent_MultiStartStop/#03: Stopping server: protocol=HTTP address=127.0.0.1:29725 network=tcp === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:43.007Z [INFO] TestAgent_MultiStartStop/#00: Requesting shutdown writer.go:29: 2021-01-29T19:31:43.008Z [INFO] TestAgent_MultiStartStop/#00.server: shutting down server writer.go:29: 2021-01-29T19:31:43.010Z [DEBUG] TestAgent_MultiStartStop/#00.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:43.011Z [DEBUG] TestAgent_MultiStartStop/#00.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:43.011Z [WARN] TestAgent_MultiStartStop/#00.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:43.015Z [WARN] TestAgent_MultiStartStop/#00.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:43.016Z [INFO] TestAgent_MultiStartStop/#00.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:43.017Z [INFO] TestAgent_MultiStartStop/#00: consul server down writer.go:29: 2021-01-29T19:31:43.019Z [INFO] TestAgent_MultiStartStop/#00: shutdown complete writer.go:29: 2021-01-29T19:31:43.021Z [INFO] TestAgent_MultiStartStop/#00: Stopping server: protocol=DNS address=127.0.0.1:29730 network=tcp writer.go:29: 2021-01-29T19:31:43.022Z [INFO] TestAgent_MultiStartStop/#00: Stopping server: protocol=DNS address=127.0.0.1:29730 network=udp writer.go:29: 2021-01-29T19:31:43.023Z [INFO] TestAgent_MultiStartStop/#00: Stopping server: protocol=HTTP address=127.0.0.1:29731 network=tcp === CONT TestAgent_MultiStartStop/#08 writer.go:29: 2021-01-29T19:31:43.324Z [INFO] TestAgent_MultiStartStop/#08: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:43.326Z [INFO] TestAgent_MultiStartStop/#08: Endpoints down === CONT TestAgent_MultiStartStop/#04 writer.go:29: 2021-01-29T19:31:43.342Z [WARN] TestAgent_MultiStartStop/#04: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:43.344Z [DEBUG] TestAgent_MultiStartStop/#04.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:43.345Z [DEBUG] TestAgent_MultiStartStop/#04.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:43.365Z [INFO] TestAgent_MultiStartStop/#04.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d53c33ba-e89a-1440-bae6-785c02ca058a Address:127.0.0.1:29741}]" writer.go:29: 2021-01-29T19:31:43.368Z [INFO] TestAgent_MultiStartStop/#04.server.raft: entering follower state: follower="Node at 127.0.0.1:29741 [Follower]" leader= writer.go:29: 2021-01-29T19:31:43.369Z [INFO] TestAgent_MultiStartStop/#04.server.serf.wan: serf: EventMemberJoin: Node-d53c33ba-e89a-1440-bae6-785c02ca058a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:43.372Z [INFO] TestAgent_MultiStartStop/#04.server.serf.lan: serf: EventMemberJoin: Node-d53c33ba-e89a-1440-bae6-785c02ca058a 127.0.0.1 writer.go:29: 2021-01-29T19:31:43.374Z [INFO] TestAgent_MultiStartStop/#04.server: Adding LAN server: server="Node-d53c33ba-e89a-1440-bae6-785c02ca058a (Addr: tcp/127.0.0.1:29741) (DC: dc1)" writer.go:29: 2021-01-29T19:31:43.375Z [INFO] TestAgent_MultiStartStop/#04.server: Handled event for server in area: event=member-join server=Node-d53c33ba-e89a-1440-bae6-785c02ca058a.dc1 area=wan writer.go:29: 2021-01-29T19:31:43.379Z [INFO] TestAgent_MultiStartStop/#04: Started DNS server: address=127.0.0.1:29736 network=udp writer.go:29: 2021-01-29T19:31:43.381Z [INFO] TestAgent_MultiStartStop/#04: Started DNS server: address=127.0.0.1:29736 network=tcp writer.go:29: 2021-01-29T19:31:43.386Z [INFO] TestAgent_MultiStartStop/#04: Started HTTP server: address=127.0.0.1:29737 network=tcp writer.go:29: 2021-01-29T19:31:43.388Z [INFO] TestAgent_MultiStartStop/#04: started state syncer === CONT TestAgent_MultiStartStop/#05 writer.go:29: 2021-01-29T19:31:43.392Z [INFO] TestAgent_MultiStartStop/#05: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:43.394Z [INFO] TestAgent_MultiStartStop/#05: Endpoints down === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.413Z [WARN] TestAgent_MultiStartStop/#02: bootstrap = true: do not enable unless necessary === CONT TestAgent_MultiStartStop/#04 writer.go:29: 2021-01-29T19:31:43.416Z [WARN] TestAgent_MultiStartStop/#04.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:43.417Z [INFO] TestAgent_MultiStartStop/#04.server.raft: entering candidate state: node="Node at 127.0.0.1:29741 [Candidate]" term=2 === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.416Z [DEBUG] TestAgent_MultiStartStop/#02.tlsutil: Update: version=1 === CONT TestAgent_MultiStartStop/#04 writer.go:29: 2021-01-29T19:31:43.419Z [DEBUG] TestAgent_MultiStartStop/#04.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:43.419Z [DEBUG] TestAgent_MultiStartStop/#04.server.raft: vote granted: from=d53c33ba-e89a-1440-bae6-785c02ca058a term=2 tally=1 === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.419Z [DEBUG] TestAgent_MultiStartStop/#02.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_MultiStartStop/#04 writer.go:29: 2021-01-29T19:31:43.420Z [INFO] TestAgent_MultiStartStop/#04.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:43.422Z [INFO] TestAgent_MultiStartStop/#04.server.raft: entering leader state: leader="Node at 127.0.0.1:29741 [Leader]" writer.go:29: 2021-01-29T19:31:43.423Z [INFO] TestAgent_MultiStartStop/#04.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:43.426Z [DEBUG] TestAgent_MultiStartStop/#04.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29741 writer.go:29: 2021-01-29T19:31:43.432Z [INFO] TestAgent_MultiStartStop/#04.server: New leader elected: payload=Node-d53c33ba-e89a-1440-bae6-785c02ca058a writer.go:29: 2021-01-29T19:31:43.435Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.440Z [INFO] TestAgent_MultiStartStop/#02.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f5bedfe4-a308-1914-6279-c7ec43d853c9 Address:127.0.0.1:29747}]" writer.go:29: 2021-01-29T19:31:43.443Z [INFO] TestAgent_MultiStartStop/#02.server.serf.wan: serf: EventMemberJoin: Node-f5bedfe4-a308-1914-6279-c7ec43d853c9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:43.447Z [INFO] TestAgent_MultiStartStop/#02.server.serf.lan: serf: EventMemberJoin: Node-f5bedfe4-a308-1914-6279-c7ec43d853c9 127.0.0.1 writer.go:29: 2021-01-29T19:31:43.443Z [INFO] TestAgent_MultiStartStop/#02.server.raft: entering follower state: follower="Node at 127.0.0.1:29747 [Follower]" leader= === CONT TestAgent_MultiStartStop/#04 writer.go:29: 2021-01-29T19:31:43.453Z [INFO] TestAgent_MultiStartStop/#04.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:43.454Z [INFO] TestAgent_MultiStartStop/#04.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:43.455Z [DEBUG] TestAgent_MultiStartStop/#04.server: Skipping self join check for node since the cluster is too small: node=Node-d53c33ba-e89a-1440-bae6-785c02ca058a writer.go:29: 2021-01-29T19:31:43.456Z [INFO] TestAgent_MultiStartStop/#04.server: member joined, marking health alive: member=Node-d53c33ba-e89a-1440-bae6-785c02ca058a === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.456Z [INFO] TestAgent_MultiStartStop/#02.server: Adding LAN server: server="Node-f5bedfe4-a308-1914-6279-c7ec43d853c9 (Addr: tcp/127.0.0.1:29747) (DC: dc1)" === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:43.471Z [INFO] TestAgent_MultiStartStop/#03: Waiting for endpoints to shut down === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.456Z [INFO] TestAgent_MultiStartStop/#02: Started DNS server: address=127.0.0.1:29742 network=udp === CONT TestAgent_MultiStartStop/#03 writer.go:29: 2021-01-29T19:31:43.472Z [INFO] TestAgent_MultiStartStop/#03: Endpoints down === CONT TestAgent_MultiStartStop/#07 === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.472Z [INFO] TestAgent_MultiStartStop/#02: Started DNS server: address=127.0.0.1:29742 network=tcp writer.go:29: 2021-01-29T19:31:43.477Z [INFO] TestAgent_MultiStartStop/#02: Started HTTP server: address=127.0.0.1:29743 network=tcp writer.go:29: 2021-01-29T19:31:43.478Z [INFO] TestAgent_MultiStartStop/#02: started state syncer writer.go:29: 2021-01-29T19:31:43.457Z [INFO] TestAgent_MultiStartStop/#02.server: Handled event for server in area: event=member-join server=Node-f5bedfe4-a308-1914-6279-c7ec43d853c9.dc1 area=wan writer.go:29: 2021-01-29T19:31:43.516Z [WARN] TestAgent_MultiStartStop/#02.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:43.518Z [INFO] TestAgent_MultiStartStop/#02.server.raft: entering candidate state: node="Node at 127.0.0.1:29747 [Candidate]" term=2 === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.523Z [WARN] TestAgent_MultiStartStop/#07: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:43.526Z [DEBUG] TestAgent_MultiStartStop/#07.tlsutil: Update: version=1 === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:43.526Z [INFO] TestAgent_MultiStartStop/#00: Waiting for endpoints to shut down === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.534Z [DEBUG] TestAgent_MultiStartStop/#02.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:43.538Z [DEBUG] TestAgent_MultiStartStop/#02.server.raft: vote granted: from=f5bedfe4-a308-1914-6279-c7ec43d853c9 term=2 tally=1 === CONT TestAgent_MultiStartStop/#00 writer.go:29: 2021-01-29T19:31:43.537Z [INFO] TestAgent_MultiStartStop/#00: Endpoints down === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.542Z [INFO] TestAgent_MultiStartStop/#02.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:43.545Z [INFO] TestAgent_MultiStartStop/#02.server.raft: entering leader state: leader="Node at 127.0.0.1:29747 [Leader]" writer.go:29: 2021-01-29T19:31:43.549Z [INFO] TestAgent_MultiStartStop/#02.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:43.552Z [INFO] TestAgent_MultiStartStop/#02.server: New leader elected: payload=Node-f5bedfe4-a308-1914-6279-c7ec43d853c9 === CONT TestAgent_MultiStartStop/#09 === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.562Z [DEBUG] TestAgent_MultiStartStop/#07.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.563Z [DEBUG] TestAgent_MultiStartStop/#02.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29747 === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.571Z [INFO] TestAgent_MultiStartStop/#07.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b3236944-ede9-e58a-9f13-ea834b00b37c Address:127.0.0.1:29753}]" writer.go:29: 2021-01-29T19:31:43.573Z [INFO] TestAgent_MultiStartStop/#07.server.serf.wan: serf: EventMemberJoin: Node-b3236944-ede9-e58a-9f13-ea834b00b37c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:43.573Z [INFO] TestAgent_MultiStartStop/#07.server.raft: entering follower state: follower="Node at 127.0.0.1:29753 [Follower]" leader= === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.568Z [INFO] TestAgent_MultiStartStop/#02: Synced node info === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.576Z [INFO] TestAgent_MultiStartStop/#07.server.serf.lan: serf: EventMemberJoin: Node-b3236944-ede9-e58a-9f13-ea834b00b37c 127.0.0.1 === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.580Z [WARN] TestAgent_MultiStartStop/#09: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:43.581Z [DEBUG] TestAgent_MultiStartStop/#09.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:43.583Z [DEBUG] TestAgent_MultiStartStop/#09.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.578Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.583Z [INFO] TestAgent_MultiStartStop/#07: Started DNS server: address=127.0.0.1:29748 network=udp writer.go:29: 2021-01-29T19:31:43.584Z [INFO] TestAgent_MultiStartStop/#07.server: Adding LAN server: server="Node-b3236944-ede9-e58a-9f13-ea834b00b37c (Addr: tcp/127.0.0.1:29753) (DC: dc1)" writer.go:29: 2021-01-29T19:31:43.584Z [INFO] TestAgent_MultiStartStop/#07.server: Handled event for server in area: event=member-join server=Node-b3236944-ede9-e58a-9f13-ea834b00b37c.dc1 area=wan === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.588Z [INFO] TestAgent_MultiStartStop/#09.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5717364d-ca19-6873-f231-fedcdae8af30 Address:127.0.0.1:29759}]" === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.587Z [INFO] TestAgent_MultiStartStop/#07: Started DNS server: address=127.0.0.1:29748 network=tcp === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.591Z [INFO] TestAgent_MultiStartStop/#09.server.serf.wan: serf: EventMemberJoin: Node-5717364d-ca19-6873-f231-fedcdae8af30.dc1 127.0.0.1 === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.592Z [INFO] TestAgent_MultiStartStop/#07: Started HTTP server: address=127.0.0.1:29749 network=tcp writer.go:29: 2021-01-29T19:31:43.594Z [INFO] TestAgent_MultiStartStop/#07: started state syncer === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.593Z [INFO] TestAgent_MultiStartStop/#09.server.raft: entering follower state: follower="Node at 127.0.0.1:29759 [Follower]" leader= === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.596Z [INFO] TestAgent_MultiStartStop/#02.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:43.598Z [INFO] TestAgent_MultiStartStop/#02.leader: started routine: routine="CA root pruning" === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.594Z [INFO] TestAgent_MultiStartStop/#09.server.serf.lan: serf: EventMemberJoin: Node-5717364d-ca19-6873-f231-fedcdae8af30 127.0.0.1 === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.600Z [DEBUG] TestAgent_MultiStartStop/#02.server: Skipping self join check for node since the cluster is too small: node=Node-f5bedfe4-a308-1914-6279-c7ec43d853c9 === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.601Z [INFO] TestAgent_MultiStartStop/#09.server: Handled event for server in area: event=member-join server=Node-5717364d-ca19-6873-f231-fedcdae8af30.dc1 area=wan === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.602Z [INFO] TestAgent_MultiStartStop/#02.server: member joined, marking health alive: member=Node-f5bedfe4-a308-1914-6279-c7ec43d853c9 === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.601Z [INFO] TestAgent_MultiStartStop/#09.server: Adding LAN server: server="Node-5717364d-ca19-6873-f231-fedcdae8af30 (Addr: tcp/127.0.0.1:29759) (DC: dc1)" writer.go:29: 2021-01-29T19:31:43.603Z [INFO] TestAgent_MultiStartStop/#09: Started DNS server: address=127.0.0.1:29754 network=udp writer.go:29: 2021-01-29T19:31:43.606Z [INFO] TestAgent_MultiStartStop/#09: Started DNS server: address=127.0.0.1:29754 network=tcp writer.go:29: 2021-01-29T19:31:43.608Z [INFO] TestAgent_MultiStartStop/#09: Started HTTP server: address=127.0.0.1:29755 network=tcp writer.go:29: 2021-01-29T19:31:43.610Z [INFO] TestAgent_MultiStartStop/#09: started state syncer === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.630Z [WARN] TestAgent_MultiStartStop/#07.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:43.633Z [INFO] TestAgent_MultiStartStop/#07.server.raft: entering candidate state: node="Node at 127.0.0.1:29753 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:43.638Z [DEBUG] TestAgent_MultiStartStop/#07.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:43.640Z [DEBUG] TestAgent_MultiStartStop/#07.server.raft: vote granted: from=b3236944-ede9-e58a-9f13-ea834b00b37c term=2 tally=1 writer.go:29: 2021-01-29T19:31:43.642Z [INFO] TestAgent_MultiStartStop/#07.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:43.644Z [INFO] TestAgent_MultiStartStop/#07.server.raft: entering leader state: leader="Node at 127.0.0.1:29753 [Leader]" writer.go:29: 2021-01-29T19:31:43.645Z [INFO] TestAgent_MultiStartStop/#07.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:43.646Z [INFO] TestAgent_MultiStartStop/#07.server: New leader elected: payload=Node-b3236944-ede9-e58a-9f13-ea834b00b37c === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.653Z [WARN] TestAgent_MultiStartStop/#09.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:43.653Z [INFO] TestAgent_MultiStartStop/#09.server.raft: entering candidate state: node="Node at 127.0.0.1:29759 [Candidate]" term=2 === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.650Z [DEBUG] TestAgent_MultiStartStop/#07.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29753 === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.658Z [DEBUG] TestAgent_MultiStartStop/#09.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:43.660Z [DEBUG] TestAgent_MultiStartStop/#09.server.raft: vote granted: from=5717364d-ca19-6873-f231-fedcdae8af30 term=2 tally=1 writer.go:29: 2021-01-29T19:31:43.661Z [INFO] TestAgent_MultiStartStop/#09.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:43.663Z [INFO] TestAgent_MultiStartStop/#09.server.raft: entering leader state: leader="Node at 127.0.0.1:29759 [Leader]" writer.go:29: 2021-01-29T19:31:43.672Z [INFO] TestAgent_MultiStartStop/#09.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:43.673Z [INFO] TestAgent_MultiStartStop/#09.server: New leader elected: payload=Node-5717364d-ca19-6873-f231-fedcdae8af30 === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.689Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.691Z [DEBUG] TestAgent_MultiStartStop/#09.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29759 writer.go:29: 2021-01-29T19:31:43.702Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.708Z [INFO] TestAgent_MultiStartStop/#07.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:43.709Z [INFO] TestAgent_MultiStartStop/#07.leader: started routine: routine="CA root pruning" === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.708Z [INFO] TestAgent_MultiStartStop/#09.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.710Z [DEBUG] TestAgent_MultiStartStop/#07.server: Skipping self join check for node since the cluster is too small: node=Node-b3236944-ede9-e58a-9f13-ea834b00b37c writer.go:29: 2021-01-29T19:31:43.711Z [INFO] TestAgent_MultiStartStop/#07.server: member joined, marking health alive: member=Node-b3236944-ede9-e58a-9f13-ea834b00b37c === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.710Z [INFO] TestAgent_MultiStartStop/#09.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:43.712Z [DEBUG] TestAgent_MultiStartStop/#09.server: Skipping self join check for node since the cluster is too small: node=Node-5717364d-ca19-6873-f231-fedcdae8af30 writer.go:29: 2021-01-29T19:31:43.713Z [INFO] TestAgent_MultiStartStop/#09.server: member joined, marking health alive: member=Node-5717364d-ca19-6873-f231-fedcdae8af30 === CONT TestAgent_MultiStartStop/#04 writer.go:29: 2021-01-29T19:31:43.786Z [DEBUG] TestAgent_MultiStartStop/#04: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:43.790Z [INFO] TestAgent_MultiStartStop/#04: Synced node info === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.959Z [INFO] TestAgent_MultiStartStop/#09: Requesting shutdown writer.go:29: 2021-01-29T19:31:43.962Z [INFO] TestAgent_MultiStartStop/#09.server: shutting down server writer.go:29: 2021-01-29T19:31:43.963Z [DEBUG] TestAgent_MultiStartStop/#09.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:43.965Z [WARN] TestAgent_MultiStartStop/#09.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.968Z [INFO] TestAgent_MultiStartStop/#02: Requesting shutdown writer.go:29: 2021-01-29T19:31:43.969Z [INFO] TestAgent_MultiStartStop/#02.server: shutting down server writer.go:29: 2021-01-29T19:31:43.970Z [DEBUG] TestAgent_MultiStartStop/#02.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:43.972Z [WARN] TestAgent_MultiStartStop/#02.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:43.972Z [DEBUG] TestAgent_MultiStartStop/#02.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:43.976Z [WARN] TestAgent_MultiStartStop/#02.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.966Z [ERROR] TestAgent_MultiStartStop/#09.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:43.966Z [DEBUG] TestAgent_MultiStartStop/#09.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.982Z [INFO] TestAgent_MultiStartStop/#02.server.router.manager: shutting down === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.968Z [WARN] TestAgent_MultiStartStop/#09.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.982Z [INFO] TestAgent_MultiStartStop/#02: consul server down === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.989Z [DEBUG] TestAgent_MultiStartStop/#07: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.991Z [INFO] TestAgent_MultiStartStop/#09.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:43.991Z [INFO] TestAgent_MultiStartStop/#09: consul server down === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.991Z [INFO] TestAgent_MultiStartStop/#02: shutdown complete === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.993Z [INFO] TestAgent_MultiStartStop/#09: shutdown complete === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.993Z [INFO] TestAgent_MultiStartStop/#07: Synced node info === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.994Z [INFO] TestAgent_MultiStartStop/#09: Stopping server: protocol=DNS address=127.0.0.1:29754 network=tcp === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.993Z [INFO] TestAgent_MultiStartStop/#02: Stopping server: protocol=DNS address=127.0.0.1:29742 network=tcp === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:43.995Z [INFO] TestAgent_MultiStartStop/#09: Stopping server: protocol=DNS address=127.0.0.1:29754 network=udp writer.go:29: 2021-01-29T19:31:43.996Z [INFO] TestAgent_MultiStartStop/#09: Stopping server: protocol=HTTP address=127.0.0.1:29755 network=tcp === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:43.995Z [DEBUG] TestAgent_MultiStartStop/#07: Node info in sync === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:43.996Z [INFO] TestAgent_MultiStartStop/#02: Stopping server: protocol=DNS address=127.0.0.1:29742 network=udp writer.go:29: 2021-01-29T19:31:43.998Z [INFO] TestAgent_MultiStartStop/#02: Stopping server: protocol=HTTP address=127.0.0.1:29743 network=tcp === CONT TestAgent_MultiStartStop/#04 writer.go:29: 2021-01-29T19:31:44.016Z [INFO] TestAgent_MultiStartStop/#04: Requesting shutdown writer.go:29: 2021-01-29T19:31:44.019Z [INFO] TestAgent_MultiStartStop/#04.server: shutting down server writer.go:29: 2021-01-29T19:31:44.020Z [DEBUG] TestAgent_MultiStartStop/#04.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:44.022Z [WARN] TestAgent_MultiStartStop/#04.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:44.022Z [DEBUG] TestAgent_MultiStartStop/#04.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:44.045Z [WARN] TestAgent_MultiStartStop/#04.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:44.048Z [INFO] TestAgent_MultiStartStop/#04.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:44.048Z [INFO] TestAgent_MultiStartStop/#04: consul server down writer.go:29: 2021-01-29T19:31:44.050Z [INFO] TestAgent_MultiStartStop/#04: shutdown complete writer.go:29: 2021-01-29T19:31:44.050Z [INFO] TestAgent_MultiStartStop/#04: Stopping server: protocol=DNS address=127.0.0.1:29736 network=tcp writer.go:29: 2021-01-29T19:31:44.052Z [INFO] TestAgent_MultiStartStop/#04: Stopping server: protocol=DNS address=127.0.0.1:29736 network=udp writer.go:29: 2021-01-29T19:31:44.052Z [INFO] TestAgent_MultiStartStop/#04: Stopping server: protocol=HTTP address=127.0.0.1:29737 network=tcp === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:44.161Z [INFO] TestAgent_MultiStartStop/#07: Requesting shutdown writer.go:29: 2021-01-29T19:31:44.163Z [INFO] TestAgent_MultiStartStop/#07.server: shutting down server writer.go:29: 2021-01-29T19:31:44.165Z [DEBUG] TestAgent_MultiStartStop/#07.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:44.167Z [WARN] TestAgent_MultiStartStop/#07.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:44.169Z [DEBUG] TestAgent_MultiStartStop/#07.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:44.173Z [WARN] TestAgent_MultiStartStop/#07.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:44.178Z [INFO] TestAgent_MultiStartStop/#07.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:44.180Z [INFO] TestAgent_MultiStartStop/#07: consul server down writer.go:29: 2021-01-29T19:31:44.188Z [INFO] TestAgent_MultiStartStop/#07: shutdown complete writer.go:29: 2021-01-29T19:31:44.190Z [INFO] TestAgent_MultiStartStop/#07: Stopping server: protocol=DNS address=127.0.0.1:29748 network=tcp writer.go:29: 2021-01-29T19:31:44.192Z [INFO] TestAgent_MultiStartStop/#07: Stopping server: protocol=DNS address=127.0.0.1:29748 network=udp writer.go:29: 2021-01-29T19:31:44.194Z [INFO] TestAgent_MultiStartStop/#07: Stopping server: protocol=HTTP address=127.0.0.1:29749 network=tcp === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:44.499Z [INFO] TestAgent_MultiStartStop/#02: Waiting for endpoints to shut down === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:44.497Z [INFO] TestAgent_MultiStartStop/#09: Waiting for endpoints to shut down === CONT TestAgent_MultiStartStop/#02 writer.go:29: 2021-01-29T19:31:44.501Z [INFO] TestAgent_MultiStartStop/#02: Endpoints down === CONT TestAgent_MultiStartStop/#01 === CONT TestAgent_MultiStartStop/#09 writer.go:29: 2021-01-29T19:31:44.501Z [INFO] TestAgent_MultiStartStop/#09: Endpoints down === CONT TestAgent_MultiStartStop/#06 === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.509Z [WARN] TestAgent_MultiStartStop/#01: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:44.510Z [DEBUG] TestAgent_MultiStartStop/#01.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:44.511Z [DEBUG] TestAgent_MultiStartStop/#01.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.524Z [WARN] TestAgent_MultiStartStop/#06: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:44.525Z [DEBUG] TestAgent_MultiStartStop/#06.tlsutil: Update: version=1 === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.526Z [INFO] TestAgent_MultiStartStop/#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:309720a7-ec0b-3f15-292a-546f190dd294 Address:127.0.0.1:29765}]" === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.526Z [DEBUG] TestAgent_MultiStartStop/#06.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.528Z [INFO] TestAgent_MultiStartStop/#01.server.serf.wan: serf: EventMemberJoin: Node-309720a7-ec0b-3f15-292a-546f190dd294.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:44.528Z [INFO] TestAgent_MultiStartStop/#01.server.raft: entering follower state: follower="Node at 127.0.0.1:29765 [Follower]" leader= writer.go:29: 2021-01-29T19:31:44.532Z [INFO] TestAgent_MultiStartStop/#01.server.serf.lan: serf: EventMemberJoin: Node-309720a7-ec0b-3f15-292a-546f190dd294 127.0.0.1 === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.533Z [INFO] TestAgent_MultiStartStop/#06.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:018bf72c-91d6-5844-c8c9-845a1857d991 Address:127.0.0.1:29771}]" === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.534Z [INFO] TestAgent_MultiStartStop/#01.server: Handled event for server in area: event=member-join server=Node-309720a7-ec0b-3f15-292a-546f190dd294.dc1 area=wan writer.go:29: 2021-01-29T19:31:44.534Z [INFO] TestAgent_MultiStartStop/#01.server: Adding LAN server: server="Node-309720a7-ec0b-3f15-292a-546f190dd294 (Addr: tcp/127.0.0.1:29765) (DC: dc1)" === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.534Z [INFO] TestAgent_MultiStartStop/#06.server.raft: entering follower state: follower="Node at 127.0.0.1:29771 [Follower]" leader= writer.go:29: 2021-01-29T19:31:44.535Z [INFO] TestAgent_MultiStartStop/#06.server.serf.wan: serf: EventMemberJoin: Node-018bf72c-91d6-5844-c8c9-845a1857d991.dc1 127.0.0.1 === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.535Z [INFO] TestAgent_MultiStartStop/#01: Started DNS server: address=127.0.0.1:29760 network=udp writer.go:29: 2021-01-29T19:31:44.538Z [INFO] TestAgent_MultiStartStop/#01: Started DNS server: address=127.0.0.1:29760 network=tcp writer.go:29: 2021-01-29T19:31:44.540Z [INFO] TestAgent_MultiStartStop/#01: Started HTTP server: address=127.0.0.1:29761 network=tcp === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.539Z [INFO] TestAgent_MultiStartStop/#06.server.serf.lan: serf: EventMemberJoin: Node-018bf72c-91d6-5844-c8c9-845a1857d991 127.0.0.1 === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.541Z [INFO] TestAgent_MultiStartStop/#01: started state syncer === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.542Z [INFO] TestAgent_MultiStartStop/#06.server: Handled event for server in area: event=member-join server=Node-018bf72c-91d6-5844-c8c9-845a1857d991.dc1 area=wan writer.go:29: 2021-01-29T19:31:44.542Z [INFO] TestAgent_MultiStartStop/#06.server: Adding LAN server: server="Node-018bf72c-91d6-5844-c8c9-845a1857d991 (Addr: tcp/127.0.0.1:29771) (DC: dc1)" writer.go:29: 2021-01-29T19:31:44.546Z [INFO] TestAgent_MultiStartStop/#06: Started DNS server: address=127.0.0.1:29766 network=udp writer.go:29: 2021-01-29T19:31:44.547Z [INFO] TestAgent_MultiStartStop/#06: Started DNS server: address=127.0.0.1:29766 network=tcp writer.go:29: 2021-01-29T19:31:44.548Z [INFO] TestAgent_MultiStartStop/#06: Started HTTP server: address=127.0.0.1:29767 network=tcp writer.go:29: 2021-01-29T19:31:44.549Z [INFO] TestAgent_MultiStartStop/#06: started state syncer === CONT TestAgent_MultiStartStop/#04 writer.go:29: 2021-01-29T19:31:44.554Z [INFO] TestAgent_MultiStartStop/#04: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:44.555Z [INFO] TestAgent_MultiStartStop/#04: Endpoints down === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.587Z [WARN] TestAgent_MultiStartStop/#01.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.587Z [WARN] TestAgent_MultiStartStop/#06.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.589Z [INFO] TestAgent_MultiStartStop/#01.server.raft: entering candidate state: node="Node at 127.0.0.1:29765 [Candidate]" term=2 === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.589Z [INFO] TestAgent_MultiStartStop/#06.server.raft: entering candidate state: node="Node at 127.0.0.1:29771 [Candidate]" term=2 === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.591Z [DEBUG] TestAgent_MultiStartStop/#01.server.raft: votes: needed=1 === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.592Z [DEBUG] TestAgent_MultiStartStop/#06.server.raft: votes: needed=1 === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.593Z [DEBUG] TestAgent_MultiStartStop/#01.server.raft: vote granted: from=309720a7-ec0b-3f15-292a-546f190dd294 term=2 tally=1 === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.593Z [DEBUG] TestAgent_MultiStartStop/#06.server.raft: vote granted: from=018bf72c-91d6-5844-c8c9-845a1857d991 term=2 tally=1 === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.594Z [INFO] TestAgent_MultiStartStop/#01.server.raft: election won: tally=1 === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.595Z [INFO] TestAgent_MultiStartStop/#06.server.raft: election won: tally=1 === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.596Z [INFO] TestAgent_MultiStartStop/#01.server.raft: entering leader state: leader="Node at 127.0.0.1:29765 [Leader]" writer.go:29: 2021-01-29T19:31:44.597Z [INFO] TestAgent_MultiStartStop/#01.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:44.598Z [INFO] TestAgent_MultiStartStop/#01.server: New leader elected: payload=Node-309720a7-ec0b-3f15-292a-546f190dd294 writer.go:29: 2021-01-29T19:31:44.599Z [DEBUG] TestAgent_MultiStartStop/#01.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29765 writer.go:29: 2021-01-29T19:31:44.604Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.596Z [INFO] TestAgent_MultiStartStop/#06.server.raft: entering leader state: leader="Node at 127.0.0.1:29771 [Leader]" writer.go:29: 2021-01-29T19:31:44.608Z [INFO] TestAgent_MultiStartStop/#06.server: cluster leadership acquired === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.611Z [INFO] TestAgent_MultiStartStop/#01.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.608Z [INFO] TestAgent_MultiStartStop/#06.server: New leader elected: payload=Node-018bf72c-91d6-5844-c8c9-845a1857d991 writer.go:29: 2021-01-29T19:31:44.613Z [DEBUG] TestAgent_MultiStartStop/#06.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29771 writer.go:29: 2021-01-29T19:31:44.626Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.612Z [INFO] TestAgent_MultiStartStop/#01.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:44.632Z [DEBUG] TestAgent_MultiStartStop/#01.server: Skipping self join check for node since the cluster is too small: node=Node-309720a7-ec0b-3f15-292a-546f190dd294 writer.go:29: 2021-01-29T19:31:44.633Z [INFO] TestAgent_MultiStartStop/#01.server: member joined, marking health alive: member=Node-309720a7-ec0b-3f15-292a-546f190dd294 === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:44.637Z [INFO] TestAgent_MultiStartStop/#06.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:44.638Z [INFO] TestAgent_MultiStartStop/#06.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:44.640Z [DEBUG] TestAgent_MultiStartStop/#06.server: Skipping self join check for node since the cluster is too small: node=Node-018bf72c-91d6-5844-c8c9-845a1857d991 writer.go:29: 2021-01-29T19:31:44.641Z [INFO] TestAgent_MultiStartStop/#06.server: member joined, marking health alive: member=Node-018bf72c-91d6-5844-c8c9-845a1857d991 writer.go:29: 2021-01-29T19:31:44.690Z [DEBUG] TestAgent_MultiStartStop/#06: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:44.692Z [INFO] TestAgent_MultiStartStop/#06: Synced node info === CONT TestAgent_MultiStartStop/#07 writer.go:29: 2021-01-29T19:31:44.696Z [INFO] TestAgent_MultiStartStop/#07: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:44.697Z [INFO] TestAgent_MultiStartStop/#07: Endpoints down === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:44.698Z [DEBUG] TestAgent_MultiStartStop/#01: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:44.700Z [INFO] TestAgent_MultiStartStop/#01: Synced node info writer.go:29: 2021-01-29T19:31:45.037Z [INFO] TestAgent_MultiStartStop/#01: Requesting shutdown writer.go:29: 2021-01-29T19:31:45.038Z [INFO] TestAgent_MultiStartStop/#01.server: shutting down server writer.go:29: 2021-01-29T19:31:45.039Z [DEBUG] TestAgent_MultiStartStop/#01.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:45.040Z [WARN] TestAgent_MultiStartStop/#01.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:45.040Z [DEBUG] TestAgent_MultiStartStop/#01.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:45.042Z [WARN] TestAgent_MultiStartStop/#01.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:45.044Z [INFO] TestAgent_MultiStartStop/#01.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:45.044Z [INFO] TestAgent_MultiStartStop/#01: consul server down writer.go:29: 2021-01-29T19:31:45.046Z [INFO] TestAgent_MultiStartStop/#01: shutdown complete writer.go:29: 2021-01-29T19:31:45.046Z [INFO] TestAgent_MultiStartStop/#01: Stopping server: protocol=DNS address=127.0.0.1:29760 network=tcp writer.go:29: 2021-01-29T19:31:45.047Z [INFO] TestAgent_MultiStartStop/#01: Stopping server: protocol=DNS address=127.0.0.1:29760 network=udp writer.go:29: 2021-01-29T19:31:45.048Z [INFO] TestAgent_MultiStartStop/#01: Stopping server: protocol=HTTP address=127.0.0.1:29761 network=tcp === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:45.107Z [INFO] TestAgent_MultiStartStop/#06: Requesting shutdown writer.go:29: 2021-01-29T19:31:45.110Z [INFO] TestAgent_MultiStartStop/#06.server: shutting down server writer.go:29: 2021-01-29T19:31:45.111Z [DEBUG] TestAgent_MultiStartStop/#06.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:45.112Z [WARN] TestAgent_MultiStartStop/#06.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:45.112Z [DEBUG] TestAgent_MultiStartStop/#06.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:45.115Z [WARN] TestAgent_MultiStartStop/#06.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:45.119Z [INFO] TestAgent_MultiStartStop/#06.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:45.119Z [INFO] TestAgent_MultiStartStop/#06: consul server down writer.go:29: 2021-01-29T19:31:45.121Z [INFO] TestAgent_MultiStartStop/#06: shutdown complete writer.go:29: 2021-01-29T19:31:45.122Z [INFO] TestAgent_MultiStartStop/#06: Stopping server: protocol=DNS address=127.0.0.1:29766 network=tcp writer.go:29: 2021-01-29T19:31:45.133Z [INFO] TestAgent_MultiStartStop/#06: Stopping server: protocol=DNS address=127.0.0.1:29766 network=udp writer.go:29: 2021-01-29T19:31:45.134Z [INFO] TestAgent_MultiStartStop/#06: Stopping server: protocol=HTTP address=127.0.0.1:29767 network=tcp === CONT TestAgent_MultiStartStop/#01 writer.go:29: 2021-01-29T19:31:45.549Z [INFO] TestAgent_MultiStartStop/#01: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:45.551Z [INFO] TestAgent_MultiStartStop/#01: Endpoints down === CONT TestAgent_MultiStartStop/#06 writer.go:29: 2021-01-29T19:31:45.635Z [INFO] TestAgent_MultiStartStop/#06: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:45.636Z [INFO] TestAgent_MultiStartStop/#06: Endpoints down --- PASS: TestAgent_MultiStartStop (0.02s) --- PASS: TestAgent_MultiStartStop/#08 (0.96s) --- PASS: TestAgent_MultiStartStop/#05 (1.02s) --- PASS: TestAgent_MultiStartStop/#03 (1.10s) --- PASS: TestAgent_MultiStartStop/#00 (1.18s) --- PASS: TestAgent_MultiStartStop/#02 (1.10s) --- PASS: TestAgent_MultiStartStop/#09 (0.95s) --- PASS: TestAgent_MultiStartStop/#04 (1.23s) --- PASS: TestAgent_MultiStartStop/#07 (1.22s) --- PASS: TestAgent_MultiStartStop/#01 (1.05s) --- PASS: TestAgent_MultiStartStop/#06 (1.13s) === RUN TestAgent_ConnectClusterIDConfig === RUN TestAgent_ConnectClusterIDConfig/default_TestAgent_has_fixed_cluster_id writer.go:29: 2021-01-29T19:31:45.651Z [WARN] default TestAgent has fixed cluster id: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:45.653Z [DEBUG] default TestAgent has fixed cluster id.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:45.662Z [DEBUG] default TestAgent has fixed cluster id.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:45.719Z [INFO] default TestAgent has fixed cluster id.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3c13fd97-b137-12c7-4001-8b0e57775120 Address:127.0.0.1:29777}]" writer.go:29: 2021-01-29T19:31:45.729Z [INFO] default TestAgent has fixed cluster id.server.raft: entering follower state: follower="Node at 127.0.0.1:29777 [Follower]" leader= writer.go:29: 2021-01-29T19:31:45.731Z [INFO] default TestAgent has fixed cluster id.server.serf.wan: serf: EventMemberJoin: Node-3c13fd97-b137-12c7-4001-8b0e57775120.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:45.735Z [INFO] default TestAgent has fixed cluster id.server.serf.lan: serf: EventMemberJoin: Node-3c13fd97-b137-12c7-4001-8b0e57775120 127.0.0.1 writer.go:29: 2021-01-29T19:31:45.738Z [INFO] default TestAgent has fixed cluster id: Started DNS server: address=127.0.0.1:29772 network=udp writer.go:29: 2021-01-29T19:31:45.741Z [INFO] default TestAgent has fixed cluster id.server: Adding LAN server: server="Node-3c13fd97-b137-12c7-4001-8b0e57775120 (Addr: tcp/127.0.0.1:29777) (DC: dc1)" writer.go:29: 2021-01-29T19:31:45.743Z [INFO] default TestAgent has fixed cluster id: Started DNS server: address=127.0.0.1:29772 network=tcp writer.go:29: 2021-01-29T19:31:45.746Z [INFO] default TestAgent has fixed cluster id: Started HTTP server: address=127.0.0.1:29773 network=tcp writer.go:29: 2021-01-29T19:31:45.747Z [INFO] default TestAgent has fixed cluster id: started state syncer writer.go:29: 2021-01-29T19:31:45.751Z [INFO] default TestAgent has fixed cluster id.server: Handled event for server in area: event=member-join server=Node-3c13fd97-b137-12c7-4001-8b0e57775120.dc1 area=wan writer.go:29: 2021-01-29T19:31:45.772Z [WARN] default TestAgent has fixed cluster id.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:45.772Z [INFO] default TestAgent has fixed cluster id.server.raft: entering candidate state: node="Node at 127.0.0.1:29777 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:45.792Z [DEBUG] default TestAgent has fixed cluster id.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:45.793Z [DEBUG] default TestAgent has fixed cluster id.server.raft: vote granted: from=3c13fd97-b137-12c7-4001-8b0e57775120 term=2 tally=1 writer.go:29: 2021-01-29T19:31:45.794Z [INFO] default TestAgent has fixed cluster id.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:45.795Z [INFO] default TestAgent has fixed cluster id.server.raft: entering leader state: leader="Node at 127.0.0.1:29777 [Leader]" writer.go:29: 2021-01-29T19:31:45.796Z [INFO] default TestAgent has fixed cluster id.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:45.796Z [INFO] default TestAgent has fixed cluster id.server: New leader elected: payload=Node-3c13fd97-b137-12c7-4001-8b0e57775120 writer.go:29: 2021-01-29T19:31:45.797Z [DEBUG] default TestAgent has fixed cluster id.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29777 writer.go:29: 2021-01-29T19:31:45.804Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:45.813Z [INFO] default TestAgent has fixed cluster id.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:45.816Z [INFO] default TestAgent has fixed cluster id.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:45.819Z [DEBUG] default TestAgent has fixed cluster id.server: Skipping self join check for node since the cluster is too small: node=Node-3c13fd97-b137-12c7-4001-8b0e57775120 writer.go:29: 2021-01-29T19:31:45.824Z [INFO] default TestAgent has fixed cluster id.server: member joined, marking health alive: member=Node-3c13fd97-b137-12c7-4001-8b0e57775120 writer.go:29: 2021-01-29T19:31:45.947Z [INFO] default TestAgent has fixed cluster id: Requesting shutdown writer.go:29: 2021-01-29T19:31:45.948Z [INFO] default TestAgent has fixed cluster id.server: shutting down server writer.go:29: 2021-01-29T19:31:45.950Z [DEBUG] default TestAgent has fixed cluster id.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:45.950Z [ERROR] default TestAgent has fixed cluster id.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:45.951Z [WARN] default TestAgent has fixed cluster id.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:45.951Z [DEBUG] default TestAgent has fixed cluster id.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:45.957Z [WARN] default TestAgent has fixed cluster id.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:45.960Z [INFO] default TestAgent has fixed cluster id.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:45.961Z [INFO] default TestAgent has fixed cluster id: consul server down writer.go:29: 2021-01-29T19:31:45.964Z [INFO] default TestAgent has fixed cluster id: shutdown complete writer.go:29: 2021-01-29T19:31:45.965Z [INFO] default TestAgent has fixed cluster id: Stopping server: protocol=DNS address=127.0.0.1:29772 network=tcp writer.go:29: 2021-01-29T19:31:45.967Z [INFO] default TestAgent has fixed cluster id: Stopping server: protocol=DNS address=127.0.0.1:29772 network=udp writer.go:29: 2021-01-29T19:31:45.968Z [INFO] default TestAgent has fixed cluster id: Stopping server: protocol=HTTP address=127.0.0.1:29773 network=tcp writer.go:29: 2021-01-29T19:31:46.470Z [INFO] default TestAgent has fixed cluster id: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:46.472Z [INFO] default TestAgent has fixed cluster id: Endpoints down === RUN TestAgent_ConnectClusterIDConfig/no_cluster_ID_specified_sets_to_test_ID writer.go:29: 2021-01-29T19:31:46.489Z [WARN] no cluster ID specified sets to test ID: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:46.491Z [DEBUG] no cluster ID specified sets to test ID.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:46.494Z [DEBUG] no cluster ID specified sets to test ID.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:46.501Z [INFO] no cluster ID specified sets to test ID.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:606c6595-4149-1eb8-3eff-0a7a8c819667 Address:127.0.0.1:29783}]" writer.go:29: 2021-01-29T19:31:46.503Z [INFO] no cluster ID specified sets to test ID.server.raft: entering follower state: follower="Node at 127.0.0.1:29783 [Follower]" leader= writer.go:29: 2021-01-29T19:31:46.505Z [INFO] no cluster ID specified sets to test ID.server.serf.wan: serf: EventMemberJoin: Node-606c6595-4149-1eb8-3eff-0a7a8c819667.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:46.511Z [INFO] no cluster ID specified sets to test ID.server.serf.lan: serf: EventMemberJoin: Node-606c6595-4149-1eb8-3eff-0a7a8c819667 127.0.0.1 writer.go:29: 2021-01-29T19:31:46.515Z [INFO] no cluster ID specified sets to test ID.server: Adding LAN server: server="Node-606c6595-4149-1eb8-3eff-0a7a8c819667 (Addr: tcp/127.0.0.1:29783) (DC: dc1)" writer.go:29: 2021-01-29T19:31:46.515Z [INFO] no cluster ID specified sets to test ID.server: Handled event for server in area: event=member-join server=Node-606c6595-4149-1eb8-3eff-0a7a8c819667.dc1 area=wan writer.go:29: 2021-01-29T19:31:46.516Z [INFO] no cluster ID specified sets to test ID: Started DNS server: address=127.0.0.1:29778 network=tcp writer.go:29: 2021-01-29T19:31:46.517Z [INFO] no cluster ID specified sets to test ID: Started DNS server: address=127.0.0.1:29778 network=udp writer.go:29: 2021-01-29T19:31:46.518Z [INFO] no cluster ID specified sets to test ID: Started HTTP server: address=127.0.0.1:29779 network=tcp writer.go:29: 2021-01-29T19:31:46.520Z [INFO] no cluster ID specified sets to test ID: started state syncer writer.go:29: 2021-01-29T19:31:46.549Z [WARN] no cluster ID specified sets to test ID.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:46.550Z [INFO] no cluster ID specified sets to test ID.server.raft: entering candidate state: node="Node at 127.0.0.1:29783 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:46.552Z [DEBUG] no cluster ID specified sets to test ID.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:46.553Z [DEBUG] no cluster ID specified sets to test ID.server.raft: vote granted: from=606c6595-4149-1eb8-3eff-0a7a8c819667 term=2 tally=1 writer.go:29: 2021-01-29T19:31:46.555Z [INFO] no cluster ID specified sets to test ID.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:46.556Z [INFO] no cluster ID specified sets to test ID.server.raft: entering leader state: leader="Node at 127.0.0.1:29783 [Leader]" writer.go:29: 2021-01-29T19:31:46.558Z [INFO] no cluster ID specified sets to test ID.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:46.558Z [INFO] no cluster ID specified sets to test ID.server: New leader elected: payload=Node-606c6595-4149-1eb8-3eff-0a7a8c819667 writer.go:29: 2021-01-29T19:31:46.559Z [DEBUG] no cluster ID specified sets to test ID.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29783 writer.go:29: 2021-01-29T19:31:46.566Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:46.572Z [INFO] no cluster ID specified sets to test ID.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:46.573Z [INFO] no cluster ID specified sets to test ID.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:46.574Z [DEBUG] no cluster ID specified sets to test ID.server: Skipping self join check for node since the cluster is too small: node=Node-606c6595-4149-1eb8-3eff-0a7a8c819667 writer.go:29: 2021-01-29T19:31:46.575Z [INFO] no cluster ID specified sets to test ID.server: member joined, marking health alive: member=Node-606c6595-4149-1eb8-3eff-0a7a8c819667 writer.go:29: 2021-01-29T19:31:46.862Z [INFO] no cluster ID specified sets to test ID: Requesting shutdown writer.go:29: 2021-01-29T19:31:46.863Z [INFO] no cluster ID specified sets to test ID.server: shutting down server writer.go:29: 2021-01-29T19:31:46.864Z [DEBUG] no cluster ID specified sets to test ID.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:46.865Z [WARN] no cluster ID specified sets to test ID.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:46.864Z [ERROR] no cluster ID specified sets to test ID.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:46.865Z [DEBUG] no cluster ID specified sets to test ID.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:46.870Z [WARN] no cluster ID specified sets to test ID.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:46.880Z [INFO] no cluster ID specified sets to test ID.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:46.880Z [INFO] no cluster ID specified sets to test ID: consul server down writer.go:29: 2021-01-29T19:31:46.882Z [INFO] no cluster ID specified sets to test ID: shutdown complete writer.go:29: 2021-01-29T19:31:46.882Z [INFO] no cluster ID specified sets to test ID: Stopping server: protocol=DNS address=127.0.0.1:29778 network=tcp writer.go:29: 2021-01-29T19:31:46.883Z [INFO] no cluster ID specified sets to test ID: Stopping server: protocol=DNS address=127.0.0.1:29778 network=udp writer.go:29: 2021-01-29T19:31:46.885Z [INFO] no cluster ID specified sets to test ID: Stopping server: protocol=HTTP address=127.0.0.1:29779 network=tcp writer.go:29: 2021-01-29T19:31:47.386Z [INFO] no cluster ID specified sets to test ID: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:47.388Z [INFO] no cluster ID specified sets to test ID: Endpoints down === RUN TestAgent_ConnectClusterIDConfig/non-UUID_cluster_id_is_fatal writer.go:29: 2021-01-29T19:31:47.400Z [WARN] non-UUID cluster_id is fatal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:47.402Z [ERROR] non-UUID cluster_id is fatal: connect CA config cluster_id specified but is not a valid UUID, aborting startup writer.go:29: 2021-01-29T19:31:47.404Z [INFO] non-UUID cluster_id is fatal: Requesting shutdown writer.go:29: 2021-01-29T19:31:47.406Z [INFO] non-UUID cluster_id is fatal: shutdown complete --- PASS: TestAgent_ConnectClusterIDConfig (1.77s) --- PASS: TestAgent_ConnectClusterIDConfig/default_TestAgent_has_fixed_cluster_id (0.84s) --- PASS: TestAgent_ConnectClusterIDConfig/no_cluster_ID_specified_sets_to_test_ID (0.91s) --- PASS: TestAgent_ConnectClusterIDConfig/non-UUID_cluster_id_is_fatal (0.02s) === RUN TestAgent_StartStop === PAUSE TestAgent_StartStop === RUN TestAgent_RPCPing === PAUSE TestAgent_RPCPing === RUN TestAgent_TokenStore === PAUSE TestAgent_TokenStore === RUN TestAgent_ReconnectConfigSettings === PAUSE TestAgent_ReconnectConfigSettings === RUN TestAgent_ReconnectConfigWanDisabled === PAUSE TestAgent_ReconnectConfigWanDisabled === RUN TestAgent_setupNodeID === PAUSE TestAgent_setupNodeID === RUN TestAgent_makeNodeID === PAUSE TestAgent_makeNodeID === RUN TestAgent_AddService === RUN TestAgent_AddService/normal === PAUSE TestAgent_AddService/normal === RUN TestAgent_AddService/service_manager === PAUSE TestAgent_AddService/service_manager === CONT TestAgent_AddService/normal === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:47.492Z [WARN] TestAgent_AddService/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:47.493Z [DEBUG] TestAgent_AddService/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:47.497Z [DEBUG] TestAgent_AddService/service_manager.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:47.520Z [INFO] TestAgent_AddService/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:442d1b0c-a08a-a146-1fa0-ddcccb04abdd Address:127.0.0.1:29795}]" === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:47.525Z [WARN] TestAgent_AddService/normal: bootstrap = true: do not enable unless necessary === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:47.523Z [INFO] TestAgent_AddService/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29795 [Follower]" leader= === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:47.526Z [DEBUG] TestAgent_AddService/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:47.528Z [DEBUG] TestAgent_AddService/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:47.524Z [INFO] TestAgent_AddService/service_manager.server.serf.wan: serf: EventMemberJoin: node1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:47.531Z [INFO] TestAgent_AddService/service_manager.server.serf.lan: serf: EventMemberJoin: node1 127.0.0.1 === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:47.543Z [INFO] TestAgent_AddService/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b75c2bf6-5411-5cdb-b379-8d98d07c344c Address:127.0.0.1:29801}]" writer.go:29: 2021-01-29T19:31:47.545Z [INFO] TestAgent_AddService/normal.server.serf.wan: serf: EventMemberJoin: node1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:47.546Z [INFO] TestAgent_AddService/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29801 [Follower]" leader= writer.go:29: 2021-01-29T19:31:47.548Z [INFO] TestAgent_AddService/normal.server.serf.lan: serf: EventMemberJoin: node1 127.0.0.1 === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:47.551Z [INFO] TestAgent_AddService/service_manager.server: Adding LAN server: server="node1 (Addr: tcp/127.0.0.1:29795) (DC: dc1)" === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:47.552Z [INFO] TestAgent_AddService/normal.server: Handled event for server in area: event=member-join server=node1.dc1 area=wan === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:47.552Z [INFO] TestAgent_AddService/service_manager.server: Handled event for server in area: event=member-join server=node1.dc1 area=wan === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:47.553Z [INFO] TestAgent_AddService/normal.server: Adding LAN server: server="node1 (Addr: tcp/127.0.0.1:29801) (DC: dc1)" writer.go:29: 2021-01-29T19:31:47.554Z [INFO] TestAgent_AddService/normal: Started DNS server: address=127.0.0.1:29796 network=udp writer.go:29: 2021-01-29T19:31:47.555Z [INFO] TestAgent_AddService/normal: Started DNS server: address=127.0.0.1:29796 network=tcp writer.go:29: 2021-01-29T19:31:47.557Z [INFO] TestAgent_AddService/normal: Started HTTP server: address=127.0.0.1:29797 network=tcp writer.go:29: 2021-01-29T19:31:47.558Z [INFO] TestAgent_AddService/normal: started state syncer === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:47.564Z [INFO] TestAgent_AddService/service_manager: Started DNS server: address=127.0.0.1:29790 network=udp writer.go:29: 2021-01-29T19:31:47.567Z [INFO] TestAgent_AddService/service_manager: Started DNS server: address=127.0.0.1:29790 network=tcp writer.go:29: 2021-01-29T19:31:47.571Z [INFO] TestAgent_AddService/service_manager: Started HTTP server: address=127.0.0.1:29791 network=tcp writer.go:29: 2021-01-29T19:31:47.572Z [INFO] TestAgent_AddService/service_manager: started state syncer writer.go:29: 2021-01-29T19:31:47.577Z [WARN] TestAgent_AddService/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:47.578Z [INFO] TestAgent_AddService/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29795 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:47.581Z [DEBUG] TestAgent_AddService/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:47.583Z [DEBUG] TestAgent_AddService/service_manager.server.raft: vote granted: from=442d1b0c-a08a-a146-1fa0-ddcccb04abdd term=2 tally=1 writer.go:29: 2021-01-29T19:31:47.585Z [INFO] TestAgent_AddService/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:47.587Z [INFO] TestAgent_AddService/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29795 [Leader]" writer.go:29: 2021-01-29T19:31:47.590Z [INFO] TestAgent_AddService/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:47.592Z [INFO] TestAgent_AddService/service_manager.server: New leader elected: payload=node1 === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:47.595Z [WARN] TestAgent_AddService/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:47.596Z [INFO] TestAgent_AddService/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29801 [Candidate]" term=2 === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:47.598Z [DEBUG] TestAgent_AddService/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29795 === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:47.600Z [DEBUG] TestAgent_AddService/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:47.610Z [DEBUG] TestAgent_AddService/normal.server.raft: vote granted: from=b75c2bf6-5411-5cdb-b379-8d98d07c344c term=2 tally=1 === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:47.614Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:47.615Z [INFO] TestAgent_AddService/normal.server.raft: election won: tally=1 === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:47.619Z [INFO] TestAgent_AddService/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:47.620Z [INFO] TestAgent_AddService/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:47.621Z [DEBUG] TestAgent_AddService/service_manager.server: Skipping self join check for node since the cluster is too small: node=node1 writer.go:29: 2021-01-29T19:31:47.622Z [INFO] TestAgent_AddService/service_manager.server: member joined, marking health alive: member=node1 === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:47.620Z [INFO] TestAgent_AddService/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29801 [Leader]" writer.go:29: 2021-01-29T19:31:47.627Z [INFO] TestAgent_AddService/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:47.628Z [INFO] TestAgent_AddService/normal.server: New leader elected: payload=node1 writer.go:29: 2021-01-29T19:31:47.631Z [DEBUG] TestAgent_AddService/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29801 writer.go:29: 2021-01-29T19:31:47.649Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:47.654Z [INFO] TestAgent_AddService/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:47.654Z [INFO] TestAgent_AddService/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:47.655Z [DEBUG] TestAgent_AddService/normal.server: Skipping self join check for node since the cluster is too small: node=node1 writer.go:29: 2021-01-29T19:31:47.656Z [INFO] TestAgent_AddService/normal.server: member joined, marking health alive: member=node1 === RUN TestAgent_AddService/normal/one_check === RUN TestAgent_AddService/normal/one_check/svcid1 === RUN TestAgent_AddService/normal/one_check/check1 === RUN TestAgent_AddService/normal/one_check/check1_ttl === RUN TestAgent_AddService/normal/multiple_checks === RUN TestAgent_AddService/normal/multiple_checks/svcid2 === RUN TestAgent_AddService/normal/multiple_checks/check1 === RUN TestAgent_AddService/normal/multiple_checks/check-noname === RUN TestAgent_AddService/normal/multiple_checks/service:svcid2:3 === RUN TestAgent_AddService/normal/multiple_checks/service:svcid2:4 === RUN TestAgent_AddService/normal/multiple_checks/check1_ttl === RUN TestAgent_AddService/normal/multiple_checks/check-noname_ttl === RUN TestAgent_AddService/normal/multiple_checks/service:svcid2:3_ttl === RUN TestAgent_AddService/normal/multiple_checks/service:svcid2:4_ttl === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:47.726Z [INFO] TestAgent_AddService/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:47.730Z [INFO] TestAgent_AddService/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:47.733Z [DEBUG] TestAgent_AddService/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:47.736Z [WARN] TestAgent_AddService/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:47.733Z [ERROR] TestAgent_AddService/normal.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:47.736Z [DEBUG] TestAgent_AddService/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:47.742Z [WARN] TestAgent_AddService/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:47.753Z [INFO] TestAgent_AddService/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:47.753Z [INFO] TestAgent_AddService/normal: consul server down writer.go:29: 2021-01-29T19:31:47.756Z [INFO] TestAgent_AddService/normal: shutdown complete writer.go:29: 2021-01-29T19:31:47.757Z [INFO] TestAgent_AddService/normal: Stopping server: protocol=DNS address=127.0.0.1:29796 network=tcp writer.go:29: 2021-01-29T19:31:47.759Z [INFO] TestAgent_AddService/normal: Stopping server: protocol=DNS address=127.0.0.1:29796 network=udp writer.go:29: 2021-01-29T19:31:47.760Z [INFO] TestAgent_AddService/normal: Stopping server: protocol=HTTP address=127.0.0.1:29797 network=tcp === RUN TestAgent_AddService/service_manager/one_check === RUN TestAgent_AddService/service_manager/one_check/svcid1 === RUN TestAgent_AddService/service_manager/one_check/check1 === RUN TestAgent_AddService/service_manager/one_check/check1_ttl === RUN TestAgent_AddService/service_manager/multiple_checks === RUN TestAgent_AddService/service_manager/multiple_checks/svcid2 === RUN TestAgent_AddService/service_manager/multiple_checks/service:svcid2:4 === RUN TestAgent_AddService/service_manager/multiple_checks/check1 === RUN TestAgent_AddService/service_manager/multiple_checks/check-noname === RUN TestAgent_AddService/service_manager/multiple_checks/service:svcid2:3 === RUN TestAgent_AddService/service_manager/multiple_checks/service:svcid2:4_ttl === RUN TestAgent_AddService/service_manager/multiple_checks/check1_ttl === RUN TestAgent_AddService/service_manager/multiple_checks/check-noname_ttl === RUN TestAgent_AddService/service_manager/multiple_checks/service:svcid2:3_ttl === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:47.785Z [INFO] TestAgent_AddService/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:47.787Z [INFO] TestAgent_AddService/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:47.788Z [DEBUG] TestAgent_AddService/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:47.789Z [WARN] TestAgent_AddService/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:47.789Z [DEBUG] TestAgent_AddService/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:47.788Z [ERROR] TestAgent_AddService/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:47.792Z [WARN] TestAgent_AddService/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:47.796Z [INFO] TestAgent_AddService/service_manager: consul server down writer.go:29: 2021-01-29T19:31:47.796Z [INFO] TestAgent_AddService/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:47.797Z [INFO] TestAgent_AddService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29790 network=tcp writer.go:29: 2021-01-29T19:31:47.798Z [INFO] TestAgent_AddService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29790 network=udp writer.go:29: 2021-01-29T19:31:47.799Z [INFO] TestAgent_AddService/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29791 network=tcp writer.go:29: 2021-01-29T19:31:47.800Z [INFO] TestAgent_AddService/service_manager.server.router.manager: shutting down === CONT TestAgent_AddService/normal writer.go:29: 2021-01-29T19:31:48.262Z [INFO] TestAgent_AddService/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:48.263Z [INFO] TestAgent_AddService/normal: Endpoints down === CONT TestAgent_AddService/service_manager writer.go:29: 2021-01-29T19:31:48.300Z [INFO] TestAgent_AddService/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:48.301Z [INFO] TestAgent_AddService/service_manager: Endpoints down --- PASS: TestAgent_AddService (0.00s) --- PASS: TestAgent_AddService/normal (0.80s) --- PASS: TestAgent_AddService/normal/one_check (0.00s) --- PASS: TestAgent_AddService/normal/one_check/svcid1 (0.00s) --- PASS: TestAgent_AddService/normal/one_check/check1 (0.00s) --- PASS: TestAgent_AddService/normal/one_check/check1_ttl (0.00s) --- PASS: TestAgent_AddService/normal/multiple_checks (0.03s) --- PASS: TestAgent_AddService/normal/multiple_checks/svcid2 (0.00s) --- PASS: TestAgent_AddService/normal/multiple_checks/check1 (0.00s) --- PASS: TestAgent_AddService/normal/multiple_checks/check-noname (0.00s) --- PASS: TestAgent_AddService/normal/multiple_checks/service:svcid2:3 (0.00s) --- PASS: TestAgent_AddService/normal/multiple_checks/service:svcid2:4 (0.00s) --- PASS: TestAgent_AddService/normal/multiple_checks/check1_ttl (0.00s) --- PASS: TestAgent_AddService/normal/multiple_checks/check-noname_ttl (0.00s) --- PASS: TestAgent_AddService/normal/multiple_checks/service:svcid2:3_ttl (0.00s) --- PASS: TestAgent_AddService/normal/multiple_checks/service:svcid2:4_ttl (0.00s) --- PASS: TestAgent_AddService/service_manager (0.84s) --- PASS: TestAgent_AddService/service_manager/one_check (0.01s) --- PASS: TestAgent_AddService/service_manager/one_check/svcid1 (0.00s) --- PASS: TestAgent_AddService/service_manager/one_check/check1 (0.00s) --- PASS: TestAgent_AddService/service_manager/one_check/check1_ttl (0.00s) --- PASS: TestAgent_AddService/service_manager/multiple_checks (0.01s) --- PASS: TestAgent_AddService/service_manager/multiple_checks/svcid2 (0.00s) --- PASS: TestAgent_AddService/service_manager/multiple_checks/service:svcid2:4 (0.00s) --- PASS: TestAgent_AddService/service_manager/multiple_checks/check1 (0.00s) --- PASS: TestAgent_AddService/service_manager/multiple_checks/check-noname (0.00s) --- PASS: TestAgent_AddService/service_manager/multiple_checks/service:svcid2:3 (0.00s) --- PASS: TestAgent_AddService/service_manager/multiple_checks/service:svcid2:4_ttl (0.00s) --- PASS: TestAgent_AddService/service_manager/multiple_checks/check1_ttl (0.00s) --- PASS: TestAgent_AddService/service_manager/multiple_checks/check-noname_ttl (0.00s) --- PASS: TestAgent_AddService/service_manager/multiple_checks/service:svcid2:3_ttl (0.00s) === RUN TestAgent_AddServices_AliasUpdateCheckNotReverted === RUN TestAgent_AddServices_AliasUpdateCheckNotReverted/normal === PAUSE TestAgent_AddServices_AliasUpdateCheckNotReverted/normal === RUN TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager === PAUSE TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.451Z [WARN] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:48.455Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:48.466Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.470Z [WARN] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:48.471Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.tlsutil: Update: version=1 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.473Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c158c241-e8ab-235b-2bc4-16acfd6a49d1 Address:127.0.0.1:29813}]" === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.474Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.475Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29813 [Follower]" leader= === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.479Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:06310efb-e7e6-95c8-c330-bebdadd3a659 Address:127.0.0.1:29807}]" writer.go:29: 2021-01-29T19:31:48.481Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29807 [Follower]" leader= === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.487Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.serf.wan: serf: EventMemberJoin: node1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:48.491Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.serf.lan: serf: EventMemberJoin: node1 127.0.0.1 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.491Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.serf.wan: serf: EventMemberJoin: node1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:48.495Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.serf.lan: serf: EventMemberJoin: node1 127.0.0.1 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.494Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server: Handled event for server in area: event=member-join server=node1.dc1 area=wan === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.498Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server: Adding LAN server: server="node1 (Addr: tcp/127.0.0.1:29807) (DC: dc1)" === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.494Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server: Adding LAN server: server="node1 (Addr: tcp/127.0.0.1:29813) (DC: dc1)" === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.499Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server: Handled event for server in area: event=member-join server=node1.dc1 area=wan writer.go:29: 2021-01-29T19:31:48.500Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Started DNS server: address=127.0.0.1:29802 network=udp === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.502Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: Started DNS server: address=127.0.0.1:29808 network=tcp === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.501Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Started DNS server: address=127.0.0.1:29802 network=tcp === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.503Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: Started DNS server: address=127.0.0.1:29808 network=udp writer.go:29: 2021-01-29T19:31:48.505Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: Started HTTP server: address=127.0.0.1:29809 network=tcp === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.504Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Started HTTP server: address=127.0.0.1:29803 network=tcp === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.506Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: started state syncer === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.509Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: started state syncer === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.514Z [WARN] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:48.515Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29813 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:48.517Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:48.519Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.raft: vote granted: from=c158c241-e8ab-235b-2bc4-16acfd6a49d1 term=2 tally=1 writer.go:29: 2021-01-29T19:31:48.520Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:48.522Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29813 [Leader]" writer.go:29: 2021-01-29T19:31:48.524Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:48.524Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server: New leader elected: payload=node1 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.526Z [WARN] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:48.527Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29807 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:48.529Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:48.530Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.raft: vote granted: from=06310efb-e7e6-95c8-c330-bebdadd3a659 term=2 tally=1 writer.go:29: 2021-01-29T19:31:48.531Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:48.533Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29807 [Leader]" writer.go:29: 2021-01-29T19:31:48.535Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:48.536Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server: New leader elected: payload=node1 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.532Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29813 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.536Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29807 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.543Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.548Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.550Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:48.551Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:48.552Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server: Skipping self join check for node since the cluster is too small: node=node1 writer.go:29: 2021-01-29T19:31:48.553Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server: member joined, marking health alive: member=node1 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.557Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:48.558Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:48.559Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server: Skipping self join check for node since the cluster is too small: node=node1 writer.go:29: 2021-01-29T19:31:48.560Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server: member joined, marking health alive: member=node1 writer.go:29: 2021-01-29T19:31:48.582Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:48.584Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced node info === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:48.661Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:48.662Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:48.663Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:48.663Z [ERROR] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:48.664Z [WARN] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:48.664Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:48.669Z [WARN] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:48.672Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:48.672Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: consul server down writer.go:29: 2021-01-29T19:31:48.675Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:48.675Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29808 network=tcp writer.go:29: 2021-01-29T19:31:48.676Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29808 network=udp writer.go:29: 2021-01-29T19:31:48.677Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29809 network=tcp === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:48.845Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:48.847Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:48.849Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:48.896Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:48.899Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced service: service=web-0 writer.go:29: 2021-01-29T19:31:48.900Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:48.901Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:48.903Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:48.914Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced service: service=web-1 writer.go:29: 2021-01-29T19:31:48.918Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced check: check=service:web-0 writer.go:29: 2021-01-29T19:31:48.920Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:48.922Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:48.926Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:48.937Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced service: service=web-2 writer.go:29: 2021-01-29T19:31:48.939Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:48.940Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:48.943Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced check: check=service:web-1 writer.go:29: 2021-01-29T19:31:48.945Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:48.946Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:48.948Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:48.949Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:48.951Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:48.953Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:48.954Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:48.956Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:48.959Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:48.963Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:48.964Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:48.966Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:48.970Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced service: service=web-3 writer.go:29: 2021-01-29T19:31:48.972Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:48.974Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:48.975Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:48.978Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced check: check=service:web-2 writer.go:29: 2021-01-29T19:31:48.980Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:48.982Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:48.984Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:48.986Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:48.988Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:48.990Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:48.992Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:48.993Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:48.996Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:48.998Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.001Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.002Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:49.003Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.004Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:49.006Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced service: service=web-4 writer.go:29: 2021-01-29T19:31:49.008Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.009Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.010Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.011Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.012Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.013Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.015Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.016Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:49.017Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.018Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:49.019Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-4 writer.go:29: 2021-01-29T19:31:49.021Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced service: service=web-5 writer.go:29: 2021-01-29T19:31:49.023Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.024Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.025Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.026Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.027Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.028Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-5 writer.go:29: 2021-01-29T19:31:49.029Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.030Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.031Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:49.031Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-4 writer.go:29: 2021-01-29T19:31:49.032Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-5 writer.go:29: 2021-01-29T19:31:49.033Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.034Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:49.035Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.036Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.036Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.037Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-5 writer.go:29: 2021-01-29T19:31:49.038Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.039Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.040Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.043Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.045Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:49.047Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.050Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:49.053Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-4 writer.go:29: 2021-01-29T19:31:49.068Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-5 writer.go:29: 2021-01-29T19:31:49.073Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced service: service=web-6 writer.go:29: 2021-01-29T19:31:49.079Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced check: check=service:web-5 writer.go:29: 2021-01-29T19:31:49.082Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-6 writer.go:29: 2021-01-29T19:31:49.084Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.085Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.087Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.089Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.092Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced check: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.094Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.095Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-6 writer.go:29: 2021-01-29T19:31:49.095Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.096Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:49.097Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.098Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:49.099Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-4 writer.go:29: 2021-01-29T19:31:49.100Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-5 writer.go:29: 2021-01-29T19:31:49.101Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.102Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.102Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.104Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-5 writer.go:29: 2021-01-29T19:31:49.106Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-6 writer.go:29: 2021-01-29T19:31:49.108Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.110Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.112Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.116Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-4 writer.go:29: 2021-01-29T19:31:49.117Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-5 writer.go:29: 2021-01-29T19:31:49.119Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-6 writer.go:29: 2021-01-29T19:31:49.121Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced service: service=web-7 writer.go:29: 2021-01-29T19:31:49.124Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.128Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:49.133Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.137Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:49.141Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.144Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.148Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.152Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.155Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.158Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-5 writer.go:29: 2021-01-29T19:31:49.162Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-6 writer.go:29: 2021-01-29T19:31:49.165Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-7 writer.go:29: 2021-01-29T19:31:49.172Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.173Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-6 writer.go:29: 2021-01-29T19:31:49.173Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-7 writer.go:29: 2021-01-29T19:31:49.175Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced service: service=web-8 writer.go:29: 2021-01-29T19:31:49.177Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced service: service=web-9 writer.go:29: 2021-01-29T19:31:49.178Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.178Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.179Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:49.178Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: Waiting for endpoints to shut down === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:49.180Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-5 writer.go:29: 2021-01-29T19:31:49.181Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager writer.go:29: 2021-01-29T19:31:49.180Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager: Endpoints down === CONT TestAgent_AddServices_AliasUpdateCheckNotReverted/normal writer.go:29: 2021-01-29T19:31:49.182Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-4 writer.go:29: 2021-01-29T19:31:49.183Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.183Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-5 writer.go:29: 2021-01-29T19:31:49.184Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-7 writer.go:29: 2021-01-29T19:31:49.185Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-8 writer.go:29: 2021-01-29T19:31:49.186Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-9 writer.go:29: 2021-01-29T19:31:49.187Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.187Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.188Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.189Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.202Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced check: check=service:web-6 writer.go:29: 2021-01-29T19:31:49.204Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.205Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.206Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:49.207Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-5 writer.go:29: 2021-01-29T19:31:49.208Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-6 writer.go:29: 2021-01-29T19:31:49.210Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-7 writer.go:29: 2021-01-29T19:31:49.212Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-8 writer.go:29: 2021-01-29T19:31:49.213Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-9 writer.go:29: 2021-01-29T19:31:49.216Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.217Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-4 writer.go:29: 2021-01-29T19:31:49.219Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:49.220Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.221Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-5 writer.go:29: 2021-01-29T19:31:49.225Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-7 writer.go:29: 2021-01-29T19:31:49.227Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-8 writer.go:29: 2021-01-29T19:31:49.229Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-6 writer.go:29: 2021-01-29T19:31:49.232Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-9 writer.go:29: 2021-01-29T19:31:49.235Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.236Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.238Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.240Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.243Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.244Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:49.245Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-4 writer.go:29: 2021-01-29T19:31:49.246Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:49.247Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-5 writer.go:29: 2021-01-29T19:31:49.247Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-6 writer.go:29: 2021-01-29T19:31:49.248Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-7 writer.go:29: 2021-01-29T19:31:49.249Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-8 writer.go:29: 2021-01-29T19:31:49.250Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-9 writer.go:29: 2021-01-29T19:31:49.251Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.251Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.252Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.253Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.254Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-6 writer.go:29: 2021-01-29T19:31:49.258Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced check: check=service:web-9 writer.go:29: 2021-01-29T19:31:49.259Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.260Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.261Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-7 writer.go:29: 2021-01-29T19:31:49.262Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-8 writer.go:29: 2021-01-29T19:31:49.263Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.264Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-5 writer.go:29: 2021-01-29T19:31:49.265Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.266Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:49.267Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-4 writer.go:29: 2021-01-29T19:31:49.267Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-5 writer.go:29: 2021-01-29T19:31:49.268Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-6 writer.go:29: 2021-01-29T19:31:49.269Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-7 writer.go:29: 2021-01-29T19:31:49.270Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-8 writer.go:29: 2021-01-29T19:31:49.271Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-9 writer.go:29: 2021-01-29T19:31:49.282Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.283Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.284Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:49.285Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-6 writer.go:29: 2021-01-29T19:31:49.288Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-9 writer.go:29: 2021-01-29T19:31:49.291Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.295Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.299Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.302Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.305Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.308Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-5 writer.go:29: 2021-01-29T19:31:49.313Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced check: check=service:web-7 writer.go:29: 2021-01-29T19:31:49.314Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Synced check: check=service:web-8 writer.go:29: 2021-01-29T19:31:49.316Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:49.317Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Node info in sync writer.go:29: 2021-01-29T19:31:49.318Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-1 writer.go:29: 2021-01-29T19:31:49.318Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-4 writer.go:29: 2021-01-29T19:31:49.319Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-8 writer.go:29: 2021-01-29T19:31:49.320Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-9 writer.go:29: 2021-01-29T19:31:49.321Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-0 writer.go:29: 2021-01-29T19:31:49.322Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-2 writer.go:29: 2021-01-29T19:31:49.323Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-3 writer.go:29: 2021-01-29T19:31:49.326Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-5 writer.go:29: 2021-01-29T19:31:49.328Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-6 writer.go:29: 2021-01-29T19:31:49.331Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Service in sync: service=web-7 writer.go:29: 2021-01-29T19:31:49.333Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-7 writer.go:29: 2021-01-29T19:31:49.335Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-8 writer.go:29: 2021-01-29T19:31:49.337Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-1 writer.go:29: 2021-01-29T19:31:49.339Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-5 writer.go:29: 2021-01-29T19:31:49.341Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-3 writer.go:29: 2021-01-29T19:31:49.343Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-4 writer.go:29: 2021-01-29T19:31:49.345Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-6 writer.go:29: 2021-01-29T19:31:49.348Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-9 writer.go:29: 2021-01-29T19:31:49.349Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-0 writer.go:29: 2021-01-29T19:31:49.350Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Check in sync: check=service:web-2 writer.go:29: 2021-01-29T19:31:49.351Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:49.352Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:49.353Z [WARN] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:49.353Z [DEBUG] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:49.360Z [WARN] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:49.362Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:49.362Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: consul server down writer.go:29: 2021-01-29T19:31:49.364Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: shutdown complete writer.go:29: 2021-01-29T19:31:49.365Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Stopping server: protocol=DNS address=127.0.0.1:29802 network=tcp writer.go:29: 2021-01-29T19:31:49.365Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Stopping server: protocol=DNS address=127.0.0.1:29802 network=udp writer.go:29: 2021-01-29T19:31:49.367Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Stopping server: protocol=HTTP address=127.0.0.1:29803 network=tcp writer.go:29: 2021-01-29T19:31:49.868Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:49.870Z [INFO] TestAgent_AddServices_AliasUpdateCheckNotReverted/normal: Endpoints down --- PASS: TestAgent_AddServices_AliasUpdateCheckNotReverted (0.00s) --- PASS: TestAgent_AddServices_AliasUpdateCheckNotReverted/service_manager (0.88s) --- PASS: TestAgent_AddServices_AliasUpdateCheckNotReverted/normal (1.57s) === RUN TestAgent_AddServiceNoExec === RUN TestAgent_AddServiceNoExec/normal === PAUSE TestAgent_AddServiceNoExec/normal === RUN TestAgent_AddServiceNoExec/service_manager === PAUSE TestAgent_AddServiceNoExec/service_manager === CONT TestAgent_AddServiceNoExec/normal === CONT TestAgent_AddServiceNoExec/service_manager === CONT TestAgent_AddServiceNoExec/normal writer.go:29: 2021-01-29T19:31:49.886Z [WARN] TestAgent_AddServiceNoExec/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:49.888Z [DEBUG] TestAgent_AddServiceNoExec/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:49.893Z [DEBUG] TestAgent_AddServiceNoExec/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddServiceNoExec/service_manager writer.go:29: 2021-01-29T19:31:49.897Z [WARN] TestAgent_AddServiceNoExec/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:49.899Z [DEBUG] TestAgent_AddServiceNoExec/service_manager.tlsutil: Update: version=1 === CONT TestAgent_AddServiceNoExec/normal writer.go:29: 2021-01-29T19:31:49.898Z [INFO] TestAgent_AddServiceNoExec/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:de509112-0340-dd2e-8a9c-eb23cccaf664 Address:127.0.0.1:29819}]" === CONT TestAgent_AddServiceNoExec/service_manager writer.go:29: 2021-01-29T19:31:49.901Z [DEBUG] TestAgent_AddServiceNoExec/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddServiceNoExec/normal writer.go:29: 2021-01-29T19:31:49.901Z [INFO] TestAgent_AddServiceNoExec/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29819 [Follower]" leader= writer.go:29: 2021-01-29T19:31:49.902Z [INFO] TestAgent_AddServiceNoExec/normal.server.serf.wan: serf: EventMemberJoin: node1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:49.905Z [INFO] TestAgent_AddServiceNoExec/normal.server.serf.lan: serf: EventMemberJoin: node1 127.0.0.1 === CONT TestAgent_AddServiceNoExec/service_manager writer.go:29: 2021-01-29T19:31:49.905Z [INFO] TestAgent_AddServiceNoExec/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2a7dc34f-25fd-73c0-4233-9edfb07d1c62 Address:127.0.0.1:29825}]" === CONT TestAgent_AddServiceNoExec/normal writer.go:29: 2021-01-29T19:31:49.908Z [INFO] TestAgent_AddServiceNoExec/normal: Started DNS server: address=127.0.0.1:29814 network=udp writer.go:29: 2021-01-29T19:31:49.909Z [INFO] TestAgent_AddServiceNoExec/normal.server: Handled event for server in area: event=member-join server=node1.dc1 area=wan writer.go:29: 2021-01-29T19:31:49.910Z [INFO] TestAgent_AddServiceNoExec/normal.server: Adding LAN server: server="node1 (Addr: tcp/127.0.0.1:29819) (DC: dc1)" === CONT TestAgent_AddServiceNoExec/service_manager writer.go:29: 2021-01-29T19:31:49.909Z [INFO] TestAgent_AddServiceNoExec/service_manager.server.serf.wan: serf: EventMemberJoin: node1.dc1 127.0.0.1 === CONT TestAgent_AddServiceNoExec/normal writer.go:29: 2021-01-29T19:31:49.911Z [INFO] TestAgent_AddServiceNoExec/normal: Started DNS server: address=127.0.0.1:29814 network=tcp writer.go:29: 2021-01-29T19:31:49.913Z [INFO] TestAgent_AddServiceNoExec/normal: Started HTTP server: address=127.0.0.1:29815 network=tcp writer.go:29: 2021-01-29T19:31:49.915Z [INFO] TestAgent_AddServiceNoExec/normal: started state syncer === CONT TestAgent_AddServiceNoExec/service_manager writer.go:29: 2021-01-29T19:31:49.914Z [INFO] TestAgent_AddServiceNoExec/service_manager.server.serf.lan: serf: EventMemberJoin: node1 127.0.0.1 writer.go:29: 2021-01-29T19:31:49.917Z [INFO] TestAgent_AddServiceNoExec/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29825 [Follower]" leader= writer.go:29: 2021-01-29T19:31:49.918Z [INFO] TestAgent_AddServiceNoExec/service_manager.server: Adding LAN server: server="node1 (Addr: tcp/127.0.0.1:29825) (DC: dc1)" writer.go:29: 2021-01-29T19:31:49.918Z [INFO] TestAgent_AddServiceNoExec/service_manager.server: Handled event for server in area: event=member-join server=node1.dc1 area=wan writer.go:29: 2021-01-29T19:31:49.919Z [INFO] TestAgent_AddServiceNoExec/service_manager: Started DNS server: address=127.0.0.1:29820 network=tcp writer.go:29: 2021-01-29T19:31:49.930Z [INFO] TestAgent_AddServiceNoExec/service_manager: Started DNS server: address=127.0.0.1:29820 network=udp === CONT TestAgent_AddServiceNoExec/normal writer.go:29: 2021-01-29T19:31:49.951Z [WARN] TestAgent_AddServiceNoExec/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:49.953Z [INFO] TestAgent_AddServiceNoExec/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29819 [Candidate]" term=2 === CONT TestAgent_AddServiceNoExec/service_manager writer.go:29: 2021-01-29T19:31:49.955Z [INFO] TestAgent_AddServiceNoExec/service_manager: Started HTTP server: address=127.0.0.1:29821 network=tcp writer.go:29: 2021-01-29T19:31:49.957Z [INFO] TestAgent_AddServiceNoExec/service_manager: started state syncer === CONT TestAgent_AddServiceNoExec/normal writer.go:29: 2021-01-29T19:31:49.968Z [DEBUG] TestAgent_AddServiceNoExec/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:49.974Z [DEBUG] TestAgent_AddServiceNoExec/normal.server.raft: vote granted: from=de509112-0340-dd2e-8a9c-eb23cccaf664 term=2 tally=1 writer.go:29: 2021-01-29T19:31:49.975Z [INFO] TestAgent_AddServiceNoExec/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:49.976Z [INFO] TestAgent_AddServiceNoExec/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29819 [Leader]" writer.go:29: 2021-01-29T19:31:49.977Z [INFO] TestAgent_AddServiceNoExec/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:49.978Z [INFO] TestAgent_AddServiceNoExec/normal.server: New leader elected: payload=node1 === CONT TestAgent_AddServiceNoExec/service_manager writer.go:29: 2021-01-29T19:31:49.989Z [WARN] TestAgent_AddServiceNoExec/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:49.990Z [INFO] TestAgent_AddServiceNoExec/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29825 [Candidate]" term=2 === CONT TestAgent_AddServiceNoExec/normal writer.go:29: 2021-01-29T19:31:49.982Z [DEBUG] TestAgent_AddServiceNoExec/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29819 writer.go:29: 2021-01-29T19:31:49.998Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:50.004Z [INFO] TestAgent_AddServiceNoExec/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:50.004Z [INFO] TestAgent_AddServiceNoExec/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:50.005Z [DEBUG] TestAgent_AddServiceNoExec/normal.server: Skipping self join check for node since the cluster is too small: node=node1 writer.go:29: 2021-01-29T19:31:50.006Z [INFO] TestAgent_AddServiceNoExec/normal.server: member joined, marking health alive: member=node1 === CONT TestAgent_AddServiceNoExec/service_manager writer.go:29: 2021-01-29T19:31:49.995Z [DEBUG] TestAgent_AddServiceNoExec/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:50.032Z [DEBUG] TestAgent_AddServiceNoExec/service_manager.server.raft: vote granted: from=2a7dc34f-25fd-73c0-4233-9edfb07d1c62 term=2 tally=1 writer.go:29: 2021-01-29T19:31:50.033Z [INFO] TestAgent_AddServiceNoExec/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:50.034Z [INFO] TestAgent_AddServiceNoExec/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29825 [Leader]" writer.go:29: 2021-01-29T19:31:50.036Z [INFO] TestAgent_AddServiceNoExec/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:50.039Z [INFO] TestAgent_AddServiceNoExec/service_manager.server: New leader elected: payload=node1 writer.go:29: 2021-01-29T19:31:50.042Z [DEBUG] TestAgent_AddServiceNoExec/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29825 writer.go:29: 2021-01-29T19:31:50.061Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:50.098Z [INFO] TestAgent_AddServiceNoExec/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:50.099Z [INFO] TestAgent_AddServiceNoExec/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:50.100Z [DEBUG] TestAgent_AddServiceNoExec/service_manager.server: Skipping self join check for node since the cluster is too small: node=node1 writer.go:29: 2021-01-29T19:31:50.101Z [INFO] TestAgent_AddServiceNoExec/service_manager.server: member joined, marking health alive: member=node1 writer.go:29: 2021-01-29T19:31:50.152Z [DEBUG] TestAgent_AddServiceNoExec/service_manager: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:50.161Z [INFO] TestAgent_AddServiceNoExec/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:50.162Z [DEBUG] TestAgent_AddServiceNoExec/service_manager: Node info in sync writer.go:29: 2021-01-29T19:31:50.213Z [INFO] TestAgent_AddServiceNoExec/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:50.214Z [INFO] TestAgent_AddServiceNoExec/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:50.216Z [DEBUG] TestAgent_AddServiceNoExec/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:50.217Z [WARN] TestAgent_AddServiceNoExec/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:50.217Z [DEBUG] TestAgent_AddServiceNoExec/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:50.219Z [WARN] TestAgent_AddServiceNoExec/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:50.221Z [INFO] TestAgent_AddServiceNoExec/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:50.222Z [INFO] TestAgent_AddServiceNoExec/service_manager: consul server down writer.go:29: 2021-01-29T19:31:50.223Z [INFO] TestAgent_AddServiceNoExec/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:50.224Z [INFO] TestAgent_AddServiceNoExec/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29820 network=tcp writer.go:29: 2021-01-29T19:31:50.225Z [INFO] TestAgent_AddServiceNoExec/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29820 network=udp writer.go:29: 2021-01-29T19:31:50.226Z [INFO] TestAgent_AddServiceNoExec/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29821 network=tcp === CONT TestAgent_AddServiceNoExec/normal writer.go:29: 2021-01-29T19:31:50.245Z [DEBUG] TestAgent_AddServiceNoExec/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:50.249Z [INFO] TestAgent_AddServiceNoExec/normal: Synced node info writer.go:29: 2021-01-29T19:31:50.251Z [DEBUG] TestAgent_AddServiceNoExec/normal: Node info in sync writer.go:29: 2021-01-29T19:31:50.319Z [INFO] TestAgent_AddServiceNoExec/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:50.320Z [INFO] TestAgent_AddServiceNoExec/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:50.321Z [DEBUG] TestAgent_AddServiceNoExec/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:50.322Z [WARN] TestAgent_AddServiceNoExec/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:50.322Z [DEBUG] TestAgent_AddServiceNoExec/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:50.325Z [WARN] TestAgent_AddServiceNoExec/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:50.328Z [INFO] TestAgent_AddServiceNoExec/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:50.328Z [INFO] TestAgent_AddServiceNoExec/normal: consul server down writer.go:29: 2021-01-29T19:31:50.331Z [INFO] TestAgent_AddServiceNoExec/normal: shutdown complete writer.go:29: 2021-01-29T19:31:50.333Z [INFO] TestAgent_AddServiceNoExec/normal: Stopping server: protocol=DNS address=127.0.0.1:29814 network=tcp writer.go:29: 2021-01-29T19:31:50.334Z [INFO] TestAgent_AddServiceNoExec/normal: Stopping server: protocol=DNS address=127.0.0.1:29814 network=udp writer.go:29: 2021-01-29T19:31:50.336Z [INFO] TestAgent_AddServiceNoExec/normal: Stopping server: protocol=HTTP address=127.0.0.1:29815 network=tcp === CONT TestAgent_AddServiceNoExec/service_manager writer.go:29: 2021-01-29T19:31:50.728Z [INFO] TestAgent_AddServiceNoExec/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:50.730Z [INFO] TestAgent_AddServiceNoExec/service_manager: Endpoints down === CONT TestAgent_AddServiceNoExec/normal writer.go:29: 2021-01-29T19:31:50.841Z [INFO] TestAgent_AddServiceNoExec/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:50.843Z [INFO] TestAgent_AddServiceNoExec/normal: Endpoints down --- PASS: TestAgent_AddServiceNoExec (0.00s) --- PASS: TestAgent_AddServiceNoExec/service_manager (0.86s) --- PASS: TestAgent_AddServiceNoExec/normal (0.97s) === RUN TestAgent_AddServiceNoRemoteExec === RUN TestAgent_AddServiceNoRemoteExec/normal === PAUSE TestAgent_AddServiceNoRemoteExec/normal === RUN TestAgent_AddServiceNoRemoteExec/service_manager === PAUSE TestAgent_AddServiceNoRemoteExec/service_manager === CONT TestAgent_AddServiceNoRemoteExec/normal === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:50.867Z [WARN] TestAgent_AddServiceNoRemoteExec/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:50.868Z [DEBUG] TestAgent_AddServiceNoRemoteExec/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:50.869Z [DEBUG] TestAgent_AddServiceNoRemoteExec/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:50.870Z [WARN] TestAgent_AddServiceNoRemoteExec/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:50.875Z [DEBUG] TestAgent_AddServiceNoRemoteExec/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:50.879Z [DEBUG] TestAgent_AddServiceNoRemoteExec/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:50.969Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:479b8c93-58ac-dc39-89c2-73b5f52c360e Address:127.0.0.1:29837}]" === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:50.968Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ed301089-9a66-677d-86e9-70037c5d952f Address:127.0.0.1:29831}]" === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:50.971Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server.serf.wan: serf: EventMemberJoin: node1.dc1 127.0.0.1 === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:50.971Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server.serf.wan: serf: EventMemberJoin: node1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:50.972Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29831 [Follower]" leader= === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:50.974Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server.serf.lan: serf: EventMemberJoin: node1 127.0.0.1 === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:50.975Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server.serf.lan: serf: EventMemberJoin: node1 127.0.0.1 === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:50.976Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server: Adding LAN server: server="node1 (Addr: tcp/127.0.0.1:29837) (DC: dc1)" === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:50.977Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server: Adding LAN server: server="node1 (Addr: tcp/127.0.0.1:29831) (DC: dc1)" === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:50.977Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: Started DNS server: address=127.0.0.1:29832 network=udp === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:50.977Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: Started DNS server: address=127.0.0.1:29826 network=udp === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:50.979Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server: Handled event for server in area: event=member-join server=node1.dc1 area=wan === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:50.979Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: Started DNS server: address=127.0.0.1:29826 network=tcp === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:50.980Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: Started DNS server: address=127.0.0.1:29832 network=tcp === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:50.977Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server: Handled event for server in area: event=member-join server=node1.dc1 area=wan writer.go:29: 2021-01-29T19:31:50.982Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: Started HTTP server: address=127.0.0.1:29827 network=tcp writer.go:29: 2021-01-29T19:31:50.983Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: started state syncer === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:50.984Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: Started HTTP server: address=127.0.0.1:29833 network=tcp writer.go:29: 2021-01-29T19:31:50.981Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29837 [Follower]" leader= writer.go:29: 2021-01-29T19:31:50.986Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: started state syncer === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:51.018Z [WARN] TestAgent_AddServiceNoRemoteExec/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:51.020Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29831 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:51.022Z [DEBUG] TestAgent_AddServiceNoRemoteExec/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:51.024Z [DEBUG] TestAgent_AddServiceNoRemoteExec/normal.server.raft: vote granted: from=ed301089-9a66-677d-86e9-70037c5d952f term=2 tally=1 writer.go:29: 2021-01-29T19:31:51.027Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:51.029Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29831 [Leader]" writer.go:29: 2021-01-29T19:31:51.031Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:51.031Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server: New leader elected: payload=node1 === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:51.032Z [WARN] TestAgent_AddServiceNoRemoteExec/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:51.034Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29837 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:51.037Z [DEBUG] TestAgent_AddServiceNoRemoteExec/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:51.038Z [DEBUG] TestAgent_AddServiceNoRemoteExec/service_manager.server.raft: vote granted: from=479b8c93-58ac-dc39-89c2-73b5f52c360e term=2 tally=1 writer.go:29: 2021-01-29T19:31:51.040Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:51.042Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29837 [Leader]" writer.go:29: 2021-01-29T19:31:51.043Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server: cluster leadership acquired === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:51.043Z [DEBUG] TestAgent_AddServiceNoRemoteExec/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29831 === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:51.043Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server: New leader elected: payload=node1 writer.go:29: 2021-01-29T19:31:51.049Z [DEBUG] TestAgent_AddServiceNoRemoteExec/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29837 === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:51.090Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:51.092Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:51.097Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:51.098Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:51.099Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:51.099Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.leader: started routine: routine="CA root pruning" === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:51.100Z [DEBUG] TestAgent_AddServiceNoRemoteExec/normal.server: Skipping self join check for node since the cluster is too small: node=node1 === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:51.101Z [DEBUG] TestAgent_AddServiceNoRemoteExec/service_manager.server: Skipping self join check for node since the cluster is too small: node=node1 writer.go:29: 2021-01-29T19:31:51.102Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server: member joined, marking health alive: member=node1 === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:51.101Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server: member joined, marking health alive: member=node1 writer.go:29: 2021-01-29T19:31:51.109Z [DEBUG] TestAgent_AddServiceNoRemoteExec/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:51.112Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: Synced node info writer.go:29: 2021-01-29T19:31:51.113Z [DEBUG] TestAgent_AddServiceNoRemoteExec/normal: Node info in sync === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:51.204Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:51.205Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:51.206Z [DEBUG] TestAgent_AddServiceNoRemoteExec/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:51.207Z [WARN] TestAgent_AddServiceNoRemoteExec/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:51.208Z [DEBUG] TestAgent_AddServiceNoRemoteExec/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:51.208Z [ERROR] TestAgent_AddServiceNoRemoteExec/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:51.209Z [WARN] TestAgent_AddServiceNoRemoteExec/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:51.213Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:51.213Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: consul server down writer.go:29: 2021-01-29T19:31:51.215Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:51.216Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29832 network=tcp writer.go:29: 2021-01-29T19:31:51.217Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29832 network=udp writer.go:29: 2021-01-29T19:31:51.218Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29833 network=tcp === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:51.308Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:51.310Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:51.311Z [DEBUG] TestAgent_AddServiceNoRemoteExec/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:51.313Z [WARN] TestAgent_AddServiceNoRemoteExec/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:51.313Z [DEBUG] TestAgent_AddServiceNoRemoteExec/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:51.316Z [WARN] TestAgent_AddServiceNoRemoteExec/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:51.319Z [INFO] TestAgent_AddServiceNoRemoteExec/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:51.319Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: consul server down writer.go:29: 2021-01-29T19:31:51.321Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: shutdown complete writer.go:29: 2021-01-29T19:31:51.322Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: Stopping server: protocol=DNS address=127.0.0.1:29826 network=tcp writer.go:29: 2021-01-29T19:31:51.323Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: Stopping server: protocol=DNS address=127.0.0.1:29826 network=udp writer.go:29: 2021-01-29T19:31:51.324Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: Stopping server: protocol=HTTP address=127.0.0.1:29827 network=tcp === CONT TestAgent_AddServiceNoRemoteExec/service_manager writer.go:29: 2021-01-29T19:31:51.719Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:51.722Z [INFO] TestAgent_AddServiceNoRemoteExec/service_manager: Endpoints down === CONT TestAgent_AddServiceNoRemoteExec/normal writer.go:29: 2021-01-29T19:31:51.825Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:51.827Z [INFO] TestAgent_AddServiceNoRemoteExec/normal: Endpoints down --- PASS: TestAgent_AddServiceNoRemoteExec (0.00s) --- PASS: TestAgent_AddServiceNoRemoteExec/service_manager (0.87s) --- PASS: TestAgent_AddServiceNoRemoteExec/normal (0.98s) === RUN TestAddServiceIPv4TaggedDefault writer.go:29: 2021-01-29T19:31:51.923Z [WARN] TestAddServiceIPv4TaggedDefault: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:51.929Z [DEBUG] TestAddServiceIPv4TaggedDefault.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:51.933Z [DEBUG] TestAddServiceIPv4TaggedDefault.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:51.941Z [INFO] TestAddServiceIPv4TaggedDefault.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8fbacfa2-0de5-1cba-a398-e3c1b8cca8fa Address:127.0.0.1:29843}]" writer.go:29: 2021-01-29T19:31:51.943Z [INFO] TestAddServiceIPv4TaggedDefault.server.raft: entering follower state: follower="Node at 127.0.0.1:29843 [Follower]" leader= writer.go:29: 2021-01-29T19:31:51.944Z [INFO] TestAddServiceIPv4TaggedDefault.server.serf.wan: serf: EventMemberJoin: Node-8fbacfa2-0de5-1cba-a398-e3c1b8cca8fa.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:51.999Z [INFO] TestAddServiceIPv4TaggedDefault.server.serf.lan: serf: EventMemberJoin: Node-8fbacfa2-0de5-1cba-a398-e3c1b8cca8fa 127.0.0.1 writer.go:29: 2021-01-29T19:31:52.010Z [INFO] TestAddServiceIPv4TaggedDefault: Started DNS server: address=127.0.0.1:29838 network=udp writer.go:29: 2021-01-29T19:31:52.011Z [INFO] TestAddServiceIPv4TaggedDefault.server: Handled event for server in area: event=member-join server=Node-8fbacfa2-0de5-1cba-a398-e3c1b8cca8fa.dc1 area=wan writer.go:29: 2021-01-29T19:31:52.011Z [INFO] TestAddServiceIPv4TaggedDefault.server: Adding LAN server: server="Node-8fbacfa2-0de5-1cba-a398-e3c1b8cca8fa (Addr: tcp/127.0.0.1:29843) (DC: dc1)" writer.go:29: 2021-01-29T19:31:52.014Z [INFO] TestAddServiceIPv4TaggedDefault: Started DNS server: address=127.0.0.1:29838 network=tcp writer.go:29: 2021-01-29T19:31:52.018Z [INFO] TestAddServiceIPv4TaggedDefault: Started HTTP server: address=127.0.0.1:29839 network=tcp writer.go:29: 2021-01-29T19:31:52.019Z [INFO] TestAddServiceIPv4TaggedDefault: started state syncer writer.go:29: 2021-01-29T19:31:52.019Z [WARN] TestAddServiceIPv4TaggedDefault.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:52.021Z [INFO] TestAddServiceIPv4TaggedDefault.server.raft: entering candidate state: node="Node at 127.0.0.1:29843 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:52.023Z [DEBUG] TestAddServiceIPv4TaggedDefault.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:52.024Z [DEBUG] TestAddServiceIPv4TaggedDefault.server.raft: vote granted: from=8fbacfa2-0de5-1cba-a398-e3c1b8cca8fa term=2 tally=1 writer.go:29: 2021-01-29T19:31:52.025Z [INFO] TestAddServiceIPv4TaggedDefault.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:52.026Z [INFO] TestAddServiceIPv4TaggedDefault.server.raft: entering leader state: leader="Node at 127.0.0.1:29843 [Leader]" writer.go:29: 2021-01-29T19:31:52.027Z [INFO] TestAddServiceIPv4TaggedDefault.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:52.027Z [INFO] TestAddServiceIPv4TaggedDefault.server: New leader elected: payload=Node-8fbacfa2-0de5-1cba-a398-e3c1b8cca8fa writer.go:29: 2021-01-29T19:31:52.029Z [DEBUG] TestAddServiceIPv4TaggedDefault.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29843 writer.go:29: 2021-01-29T19:31:52.032Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:52.036Z [INFO] TestAddServiceIPv4TaggedDefault.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:52.037Z [INFO] TestAddServiceIPv4TaggedDefault.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:52.038Z [DEBUG] TestAddServiceIPv4TaggedDefault.server: Skipping self join check for node since the cluster is too small: node=Node-8fbacfa2-0de5-1cba-a398-e3c1b8cca8fa writer.go:29: 2021-01-29T19:31:52.039Z [INFO] TestAddServiceIPv4TaggedDefault.server: member joined, marking health alive: member=Node-8fbacfa2-0de5-1cba-a398-e3c1b8cca8fa writer.go:29: 2021-01-29T19:31:52.079Z [WARN] TestAddServiceIPv4TaggedDefault: Service name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: service=my_service writer.go:29: 2021-01-29T19:31:52.111Z [INFO] TestAddServiceIPv4TaggedDefault: Requesting shutdown writer.go:29: 2021-01-29T19:31:52.117Z [INFO] TestAddServiceIPv4TaggedDefault.server: shutting down server writer.go:29: 2021-01-29T19:31:52.118Z [DEBUG] TestAddServiceIPv4TaggedDefault.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:52.119Z [WARN] TestAddServiceIPv4TaggedDefault.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:52.118Z [ERROR] TestAddServiceIPv4TaggedDefault.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:52.119Z [DEBUG] TestAddServiceIPv4TaggedDefault.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:52.121Z [WARN] TestAddServiceIPv4TaggedDefault.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:52.123Z [INFO] TestAddServiceIPv4TaggedDefault.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:52.123Z [INFO] TestAddServiceIPv4TaggedDefault: consul server down writer.go:29: 2021-01-29T19:31:52.125Z [INFO] TestAddServiceIPv4TaggedDefault: shutdown complete writer.go:29: 2021-01-29T19:31:52.126Z [INFO] TestAddServiceIPv4TaggedDefault: Stopping server: protocol=DNS address=127.0.0.1:29838 network=tcp writer.go:29: 2021-01-29T19:31:52.126Z [INFO] TestAddServiceIPv4TaggedDefault: Stopping server: protocol=DNS address=127.0.0.1:29838 network=udp writer.go:29: 2021-01-29T19:31:52.127Z [INFO] TestAddServiceIPv4TaggedDefault: Stopping server: protocol=HTTP address=127.0.0.1:29839 network=tcp writer.go:29: 2021-01-29T19:31:52.628Z [INFO] TestAddServiceIPv4TaggedDefault: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:52.630Z [INFO] TestAddServiceIPv4TaggedDefault: Endpoints down --- PASS: TestAddServiceIPv4TaggedDefault (0.80s) === RUN TestAddServiceIPv6TaggedDefault writer.go:29: 2021-01-29T19:31:52.667Z [WARN] TestAddServiceIPv6TaggedDefault: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:52.669Z [DEBUG] TestAddServiceIPv6TaggedDefault.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:52.671Z [DEBUG] TestAddServiceIPv6TaggedDefault.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:52.706Z [INFO] TestAddServiceIPv6TaggedDefault.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fd7c2b54-ce7d-2bb7-92c0-643bea7c7720 Address:127.0.0.1:29849}]" writer.go:29: 2021-01-29T19:31:52.708Z [INFO] TestAddServiceIPv6TaggedDefault.server.serf.wan: serf: EventMemberJoin: Node-fd7c2b54-ce7d-2bb7-92c0-643bea7c7720.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:52.708Z [INFO] TestAddServiceIPv6TaggedDefault.server.raft: entering follower state: follower="Node at 127.0.0.1:29849 [Follower]" leader= writer.go:29: 2021-01-29T19:31:52.710Z [INFO] TestAddServiceIPv6TaggedDefault.server.serf.lan: serf: EventMemberJoin: Node-fd7c2b54-ce7d-2bb7-92c0-643bea7c7720 127.0.0.1 writer.go:29: 2021-01-29T19:31:52.713Z [INFO] TestAddServiceIPv6TaggedDefault.server: Adding LAN server: server="Node-fd7c2b54-ce7d-2bb7-92c0-643bea7c7720 (Addr: tcp/127.0.0.1:29849) (DC: dc1)" writer.go:29: 2021-01-29T19:31:52.714Z [INFO] TestAddServiceIPv6TaggedDefault.server: Handled event for server in area: event=member-join server=Node-fd7c2b54-ce7d-2bb7-92c0-643bea7c7720.dc1 area=wan writer.go:29: 2021-01-29T19:31:52.716Z [INFO] TestAddServiceIPv6TaggedDefault: Started DNS server: address=127.0.0.1:29844 network=tcp writer.go:29: 2021-01-29T19:31:52.717Z [INFO] TestAddServiceIPv6TaggedDefault: Started DNS server: address=127.0.0.1:29844 network=udp writer.go:29: 2021-01-29T19:31:52.720Z [INFO] TestAddServiceIPv6TaggedDefault: Started HTTP server: address=127.0.0.1:29845 network=tcp writer.go:29: 2021-01-29T19:31:52.722Z [INFO] TestAddServiceIPv6TaggedDefault: started state syncer writer.go:29: 2021-01-29T19:31:52.773Z [WARN] TestAddServiceIPv6TaggedDefault.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:52.776Z [INFO] TestAddServiceIPv6TaggedDefault.server.raft: entering candidate state: node="Node at 127.0.0.1:29849 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:52.780Z [DEBUG] TestAddServiceIPv6TaggedDefault.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:52.781Z [DEBUG] TestAddServiceIPv6TaggedDefault.server.raft: vote granted: from=fd7c2b54-ce7d-2bb7-92c0-643bea7c7720 term=2 tally=1 writer.go:29: 2021-01-29T19:31:52.783Z [INFO] TestAddServiceIPv6TaggedDefault.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:52.785Z [INFO] TestAddServiceIPv6TaggedDefault.server.raft: entering leader state: leader="Node at 127.0.0.1:29849 [Leader]" writer.go:29: 2021-01-29T19:31:52.787Z [INFO] TestAddServiceIPv6TaggedDefault.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:52.789Z [INFO] TestAddServiceIPv6TaggedDefault.server: New leader elected: payload=Node-fd7c2b54-ce7d-2bb7-92c0-643bea7c7720 writer.go:29: 2021-01-29T19:31:52.790Z [DEBUG] TestAddServiceIPv6TaggedDefault.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29849 writer.go:29: 2021-01-29T19:31:52.808Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:52.814Z [INFO] TestAddServiceIPv6TaggedDefault.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:52.816Z [INFO] TestAddServiceIPv6TaggedDefault.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:52.818Z [DEBUG] TestAddServiceIPv6TaggedDefault.server: Skipping self join check for node since the cluster is too small: node=Node-fd7c2b54-ce7d-2bb7-92c0-643bea7c7720 writer.go:29: 2021-01-29T19:31:52.819Z [INFO] TestAddServiceIPv6TaggedDefault.server: member joined, marking health alive: member=Node-fd7c2b54-ce7d-2bb7-92c0-643bea7c7720 writer.go:29: 2021-01-29T19:31:52.840Z [WARN] TestAddServiceIPv6TaggedDefault: Service name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: service=my_service writer.go:29: 2021-01-29T19:31:52.849Z [INFO] TestAddServiceIPv6TaggedDefault: Requesting shutdown writer.go:29: 2021-01-29T19:31:52.851Z [INFO] TestAddServiceIPv6TaggedDefault.server: shutting down server writer.go:29: 2021-01-29T19:31:52.852Z [DEBUG] TestAddServiceIPv6TaggedDefault.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:52.854Z [WARN] TestAddServiceIPv6TaggedDefault.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:52.853Z [ERROR] TestAddServiceIPv6TaggedDefault.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:52.854Z [DEBUG] TestAddServiceIPv6TaggedDefault.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:52.857Z [WARN] TestAddServiceIPv6TaggedDefault.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:52.862Z [INFO] TestAddServiceIPv6TaggedDefault.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:52.863Z [INFO] TestAddServiceIPv6TaggedDefault: consul server down writer.go:29: 2021-01-29T19:31:52.864Z [INFO] TestAddServiceIPv6TaggedDefault: shutdown complete writer.go:29: 2021-01-29T19:31:52.864Z [INFO] TestAddServiceIPv6TaggedDefault: Stopping server: protocol=DNS address=127.0.0.1:29844 network=tcp writer.go:29: 2021-01-29T19:31:52.865Z [INFO] TestAddServiceIPv6TaggedDefault: Stopping server: protocol=DNS address=127.0.0.1:29844 network=udp writer.go:29: 2021-01-29T19:31:52.866Z [INFO] TestAddServiceIPv6TaggedDefault: Stopping server: protocol=HTTP address=127.0.0.1:29845 network=tcp writer.go:29: 2021-01-29T19:31:53.368Z [INFO] TestAddServiceIPv6TaggedDefault: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:53.369Z [INFO] TestAddServiceIPv6TaggedDefault: Endpoints down --- PASS: TestAddServiceIPv6TaggedDefault (0.74s) === RUN TestAddServiceIPv4TaggedSet writer.go:29: 2021-01-29T19:31:53.380Z [WARN] TestAddServiceIPv4TaggedSet: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:53.382Z [DEBUG] TestAddServiceIPv4TaggedSet.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:53.387Z [DEBUG] TestAddServiceIPv4TaggedSet.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:53.394Z [INFO] TestAddServiceIPv4TaggedSet.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5f41892a-0c9a-b1e1-d15e-1fb4e76830a2 Address:127.0.0.1:29855}]" writer.go:29: 2021-01-29T19:31:53.398Z [INFO] TestAddServiceIPv4TaggedSet.server.raft: entering follower state: follower="Node at 127.0.0.1:29855 [Follower]" leader= writer.go:29: 2021-01-29T19:31:53.399Z [INFO] TestAddServiceIPv4TaggedSet.server.serf.wan: serf: EventMemberJoin: Node-5f41892a-0c9a-b1e1-d15e-1fb4e76830a2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:53.403Z [INFO] TestAddServiceIPv4TaggedSet.server.serf.lan: serf: EventMemberJoin: Node-5f41892a-0c9a-b1e1-d15e-1fb4e76830a2 127.0.0.1 writer.go:29: 2021-01-29T19:31:53.405Z [INFO] TestAddServiceIPv4TaggedSet.server: Handled event for server in area: event=member-join server=Node-5f41892a-0c9a-b1e1-d15e-1fb4e76830a2.dc1 area=wan writer.go:29: 2021-01-29T19:31:53.405Z [INFO] TestAddServiceIPv4TaggedSet: Started DNS server: address=127.0.0.1:29850 network=udp writer.go:29: 2021-01-29T19:31:53.408Z [INFO] TestAddServiceIPv4TaggedSet: Started DNS server: address=127.0.0.1:29850 network=tcp writer.go:29: 2021-01-29T19:31:53.406Z [INFO] TestAddServiceIPv4TaggedSet.server: Adding LAN server: server="Node-5f41892a-0c9a-b1e1-d15e-1fb4e76830a2 (Addr: tcp/127.0.0.1:29855) (DC: dc1)" writer.go:29: 2021-01-29T19:31:53.411Z [INFO] TestAddServiceIPv4TaggedSet: Started HTTP server: address=127.0.0.1:29851 network=tcp writer.go:29: 2021-01-29T19:31:53.415Z [INFO] TestAddServiceIPv4TaggedSet: started state syncer writer.go:29: 2021-01-29T19:31:53.450Z [WARN] TestAddServiceIPv4TaggedSet.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:53.452Z [INFO] TestAddServiceIPv4TaggedSet.server.raft: entering candidate state: node="Node at 127.0.0.1:29855 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:53.455Z [DEBUG] TestAddServiceIPv4TaggedSet.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:53.456Z [DEBUG] TestAddServiceIPv4TaggedSet.server.raft: vote granted: from=5f41892a-0c9a-b1e1-d15e-1fb4e76830a2 term=2 tally=1 writer.go:29: 2021-01-29T19:31:53.457Z [INFO] TestAddServiceIPv4TaggedSet.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:53.458Z [INFO] TestAddServiceIPv4TaggedSet.server.raft: entering leader state: leader="Node at 127.0.0.1:29855 [Leader]" writer.go:29: 2021-01-29T19:31:53.459Z [INFO] TestAddServiceIPv4TaggedSet.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:53.460Z [INFO] TestAddServiceIPv4TaggedSet.server: New leader elected: payload=Node-5f41892a-0c9a-b1e1-d15e-1fb4e76830a2 writer.go:29: 2021-01-29T19:31:53.460Z [DEBUG] TestAddServiceIPv4TaggedSet.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29855 writer.go:29: 2021-01-29T19:31:53.480Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:53.487Z [INFO] TestAddServiceIPv4TaggedSet.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:53.489Z [INFO] TestAddServiceIPv4TaggedSet.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:53.502Z [DEBUG] TestAddServiceIPv4TaggedSet.server: Skipping self join check for node since the cluster is too small: node=Node-5f41892a-0c9a-b1e1-d15e-1fb4e76830a2 writer.go:29: 2021-01-29T19:31:53.505Z [INFO] TestAddServiceIPv4TaggedSet.server: member joined, marking health alive: member=Node-5f41892a-0c9a-b1e1-d15e-1fb4e76830a2 writer.go:29: 2021-01-29T19:31:53.587Z [DEBUG] TestAddServiceIPv4TaggedSet: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:53.591Z [INFO] TestAddServiceIPv4TaggedSet: Synced node info writer.go:29: 2021-01-29T19:31:53.831Z [WARN] TestAddServiceIPv4TaggedSet: Service name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: service=my_service writer.go:29: 2021-01-29T19:31:53.841Z [INFO] TestAddServiceIPv4TaggedSet: Requesting shutdown writer.go:29: 2021-01-29T19:31:53.859Z [INFO] TestAddServiceIPv4TaggedSet.server: shutting down server writer.go:29: 2021-01-29T19:31:53.860Z [DEBUG] TestAddServiceIPv4TaggedSet.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:53.862Z [WARN] TestAddServiceIPv4TaggedSet.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:53.862Z [DEBUG] TestAddServiceIPv4TaggedSet.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:53.866Z [WARN] TestAddServiceIPv4TaggedSet.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:53.869Z [INFO] TestAddServiceIPv4TaggedSet.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:53.870Z [INFO] TestAddServiceIPv4TaggedSet: consul server down writer.go:29: 2021-01-29T19:31:53.873Z [INFO] TestAddServiceIPv4TaggedSet: shutdown complete writer.go:29: 2021-01-29T19:31:53.874Z [INFO] TestAddServiceIPv4TaggedSet: Stopping server: protocol=DNS address=127.0.0.1:29850 network=tcp writer.go:29: 2021-01-29T19:31:53.875Z [INFO] TestAddServiceIPv4TaggedSet: Stopping server: protocol=DNS address=127.0.0.1:29850 network=udp writer.go:29: 2021-01-29T19:31:53.876Z [INFO] TestAddServiceIPv4TaggedSet: Stopping server: protocol=HTTP address=127.0.0.1:29851 network=tcp writer.go:29: 2021-01-29T19:31:54.378Z [INFO] TestAddServiceIPv4TaggedSet: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:54.379Z [INFO] TestAddServiceIPv4TaggedSet: Endpoints down --- PASS: TestAddServiceIPv4TaggedSet (1.01s) === RUN TestAddServiceIPv6TaggedSet writer.go:29: 2021-01-29T19:31:54.393Z [WARN] TestAddServiceIPv6TaggedSet: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:54.394Z [DEBUG] TestAddServiceIPv6TaggedSet.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:54.395Z [DEBUG] TestAddServiceIPv6TaggedSet.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:54.407Z [INFO] TestAddServiceIPv6TaggedSet.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:61846677-1365-9374-59ef-9309db975c82 Address:127.0.0.1:29861}]" writer.go:29: 2021-01-29T19:31:54.413Z [INFO] TestAddServiceIPv6TaggedSet.server.raft: entering follower state: follower="Node at 127.0.0.1:29861 [Follower]" leader= writer.go:29: 2021-01-29T19:31:54.424Z [INFO] TestAddServiceIPv6TaggedSet.server.serf.wan: serf: EventMemberJoin: Node-61846677-1365-9374-59ef-9309db975c82.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:54.433Z [INFO] TestAddServiceIPv6TaggedSet.server.serf.lan: serf: EventMemberJoin: Node-61846677-1365-9374-59ef-9309db975c82 127.0.0.1 writer.go:29: 2021-01-29T19:31:54.440Z [INFO] TestAddServiceIPv6TaggedSet.server: Adding LAN server: server="Node-61846677-1365-9374-59ef-9309db975c82 (Addr: tcp/127.0.0.1:29861) (DC: dc1)" writer.go:29: 2021-01-29T19:31:54.441Z [INFO] TestAddServiceIPv6TaggedSet.server: Handled event for server in area: event=member-join server=Node-61846677-1365-9374-59ef-9309db975c82.dc1 area=wan writer.go:29: 2021-01-29T19:31:54.442Z [INFO] TestAddServiceIPv6TaggedSet: Started DNS server: address=127.0.0.1:29856 network=tcp writer.go:29: 2021-01-29T19:31:54.444Z [INFO] TestAddServiceIPv6TaggedSet: Started DNS server: address=127.0.0.1:29856 network=udp writer.go:29: 2021-01-29T19:31:54.447Z [INFO] TestAddServiceIPv6TaggedSet: Started HTTP server: address=127.0.0.1:29857 network=tcp writer.go:29: 2021-01-29T19:31:54.448Z [INFO] TestAddServiceIPv6TaggedSet: started state syncer writer.go:29: 2021-01-29T19:31:54.470Z [WARN] TestAddServiceIPv6TaggedSet.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:54.471Z [INFO] TestAddServiceIPv6TaggedSet.server.raft: entering candidate state: node="Node at 127.0.0.1:29861 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:54.473Z [DEBUG] TestAddServiceIPv6TaggedSet.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:54.474Z [DEBUG] TestAddServiceIPv6TaggedSet.server.raft: vote granted: from=61846677-1365-9374-59ef-9309db975c82 term=2 tally=1 writer.go:29: 2021-01-29T19:31:54.475Z [INFO] TestAddServiceIPv6TaggedSet.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:54.476Z [INFO] TestAddServiceIPv6TaggedSet.server.raft: entering leader state: leader="Node at 127.0.0.1:29861 [Leader]" writer.go:29: 2021-01-29T19:31:54.477Z [INFO] TestAddServiceIPv6TaggedSet.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:54.477Z [INFO] TestAddServiceIPv6TaggedSet.server: New leader elected: payload=Node-61846677-1365-9374-59ef-9309db975c82 writer.go:29: 2021-01-29T19:31:54.479Z [DEBUG] TestAddServiceIPv6TaggedSet.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29861 writer.go:29: 2021-01-29T19:31:54.492Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:54.496Z [INFO] TestAddServiceIPv6TaggedSet.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:54.497Z [INFO] TestAddServiceIPv6TaggedSet.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:54.498Z [DEBUG] TestAddServiceIPv6TaggedSet.server: Skipping self join check for node since the cluster is too small: node=Node-61846677-1365-9374-59ef-9309db975c82 writer.go:29: 2021-01-29T19:31:54.499Z [INFO] TestAddServiceIPv6TaggedSet.server: member joined, marking health alive: member=Node-61846677-1365-9374-59ef-9309db975c82 writer.go:29: 2021-01-29T19:31:54.691Z [WARN] TestAddServiceIPv6TaggedSet: Service name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: service=my_service writer.go:29: 2021-01-29T19:31:54.711Z [INFO] TestAddServiceIPv6TaggedSet: Requesting shutdown writer.go:29: 2021-01-29T19:31:54.713Z [INFO] TestAddServiceIPv6TaggedSet.server: shutting down server writer.go:29: 2021-01-29T19:31:54.714Z [DEBUG] TestAddServiceIPv6TaggedSet.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:54.714Z [ERROR] TestAddServiceIPv6TaggedSet.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:54.716Z [WARN] TestAddServiceIPv6TaggedSet.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:54.716Z [DEBUG] TestAddServiceIPv6TaggedSet.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:54.721Z [WARN] TestAddServiceIPv6TaggedSet.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:54.724Z [INFO] TestAddServiceIPv6TaggedSet.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:54.724Z [INFO] TestAddServiceIPv6TaggedSet: consul server down writer.go:29: 2021-01-29T19:31:54.727Z [INFO] TestAddServiceIPv6TaggedSet: shutdown complete writer.go:29: 2021-01-29T19:31:54.729Z [INFO] TestAddServiceIPv6TaggedSet: Stopping server: protocol=DNS address=127.0.0.1:29856 network=tcp writer.go:29: 2021-01-29T19:31:54.730Z [INFO] TestAddServiceIPv6TaggedSet: Stopping server: protocol=DNS address=127.0.0.1:29856 network=udp writer.go:29: 2021-01-29T19:31:54.733Z [INFO] TestAddServiceIPv6TaggedSet: Stopping server: protocol=HTTP address=127.0.0.1:29857 network=tcp writer.go:29: 2021-01-29T19:31:55.236Z [INFO] TestAddServiceIPv6TaggedSet: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:55.237Z [INFO] TestAddServiceIPv6TaggedSet: Endpoints down --- PASS: TestAddServiceIPv6TaggedSet (0.86s) === RUN TestAgent_RemoveService === RUN TestAgent_RemoveService/normal === PAUSE TestAgent_RemoveService/normal === RUN TestAgent_RemoveService/service_manager === PAUSE TestAgent_RemoveService/service_manager === CONT TestAgent_RemoveService/normal === CONT TestAgent_RemoveService/service_manager === CONT TestAgent_RemoveService/normal writer.go:29: 2021-01-29T19:31:55.254Z [WARN] TestAgent_RemoveService/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:55.256Z [DEBUG] TestAgent_RemoveService/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:55.259Z [DEBUG] TestAgent_RemoveService/normal.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:55.265Z [INFO] TestAgent_RemoveService/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:34c06649-452d-b3f1-5021-b2ee4c6d4d97 Address:127.0.0.1:29867}]" writer.go:29: 2021-01-29T19:31:55.267Z [INFO] TestAgent_RemoveService/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29867 [Follower]" leader= writer.go:29: 2021-01-29T19:31:55.268Z [INFO] TestAgent_RemoveService/normal.server.serf.wan: serf: EventMemberJoin: Node-34c06649-452d-b3f1-5021-b2ee4c6d4d97.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:55.276Z [INFO] TestAgent_RemoveService/normal.server.serf.lan: serf: EventMemberJoin: Node-34c06649-452d-b3f1-5021-b2ee4c6d4d97 127.0.0.1 writer.go:29: 2021-01-29T19:31:55.289Z [INFO] TestAgent_RemoveService/normal.server: Handled event for server in area: event=member-join server=Node-34c06649-452d-b3f1-5021-b2ee4c6d4d97.dc1 area=wan writer.go:29: 2021-01-29T19:31:55.289Z [INFO] TestAgent_RemoveService/normal: Started DNS server: address=127.0.0.1:29862 network=udp writer.go:29: 2021-01-29T19:31:55.291Z [INFO] TestAgent_RemoveService/normal: Started DNS server: address=127.0.0.1:29862 network=tcp writer.go:29: 2021-01-29T19:31:55.291Z [INFO] TestAgent_RemoveService/normal.server: Adding LAN server: server="Node-34c06649-452d-b3f1-5021-b2ee4c6d4d97 (Addr: tcp/127.0.0.1:29867) (DC: dc1)" writer.go:29: 2021-01-29T19:31:55.294Z [INFO] TestAgent_RemoveService/normal: Started HTTP server: address=127.0.0.1:29863 network=tcp === CONT TestAgent_RemoveService/service_manager writer.go:29: 2021-01-29T19:31:55.251Z [WARN] TestAgent_RemoveService/service_manager: bootstrap = true: do not enable unless necessary === CONT TestAgent_RemoveService/normal writer.go:29: 2021-01-29T19:31:55.299Z [INFO] TestAgent_RemoveService/normal: started state syncer === CONT TestAgent_RemoveService/service_manager writer.go:29: 2021-01-29T19:31:55.305Z [DEBUG] TestAgent_RemoveService/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:55.342Z [DEBUG] TestAgent_RemoveService/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RemoveService/normal writer.go:29: 2021-01-29T19:31:55.342Z [WARN] TestAgent_RemoveService/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:55.347Z [INFO] TestAgent_RemoveService/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29867 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:55.355Z [DEBUG] TestAgent_RemoveService/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:55.355Z [DEBUG] TestAgent_RemoveService/normal.server.raft: vote granted: from=34c06649-452d-b3f1-5021-b2ee4c6d4d97 term=2 tally=1 writer.go:29: 2021-01-29T19:31:55.356Z [INFO] TestAgent_RemoveService/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:55.357Z [INFO] TestAgent_RemoveService/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29867 [Leader]" writer.go:29: 2021-01-29T19:31:55.359Z [INFO] TestAgent_RemoveService/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:55.361Z [INFO] TestAgent_RemoveService/normal.server: New leader elected: payload=Node-34c06649-452d-b3f1-5021-b2ee4c6d4d97 === CONT TestAgent_RemoveService/service_manager writer.go:29: 2021-01-29T19:31:55.363Z [INFO] TestAgent_RemoveService/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:43faed93-bc6e-1eb5-eb72-3caf413a164b Address:127.0.0.1:29873}]" writer.go:29: 2021-01-29T19:31:55.368Z [INFO] TestAgent_RemoveService/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29873 [Follower]" leader= writer.go:29: 2021-01-29T19:31:55.373Z [INFO] TestAgent_RemoveService/service_manager.server.serf.wan: serf: EventMemberJoin: Node-43faed93-bc6e-1eb5-eb72-3caf413a164b.dc1 127.0.0.1 === CONT TestAgent_RemoveService/normal writer.go:29: 2021-01-29T19:31:55.394Z [WARN] TestAgent_RemoveService/normal: Failed to deregister service: service=redis error="Service {"redis" {}} does not exist" writer.go:29: 2021-01-29T19:31:55.395Z [DEBUG] TestAgent_RemoveService/normal: removed check: check=service:memcache writer.go:29: 2021-01-29T19:31:55.396Z [DEBUG] TestAgent_RemoveService/normal: removed check: check=check2 writer.go:29: 2021-01-29T19:31:55.398Z [DEBUG] TestAgent_RemoveService/normal: removed service: service=memcache writer.go:29: 2021-01-29T19:31:55.400Z [DEBUG] TestAgent_RemoveService/normal: removed check: check=service:redis:2 writer.go:29: 2021-01-29T19:31:55.400Z [DEBUG] TestAgent_RemoveService/normal: removed check: check=service:redis:1 === CONT TestAgent_RemoveService/service_manager writer.go:29: 2021-01-29T19:31:55.385Z [INFO] TestAgent_RemoveService/service_manager.server.serf.lan: serf: EventMemberJoin: Node-43faed93-bc6e-1eb5-eb72-3caf413a164b 127.0.0.1 writer.go:29: 2021-01-29T19:31:55.404Z [INFO] TestAgent_RemoveService/service_manager: Started DNS server: address=127.0.0.1:29868 network=udp writer.go:29: 2021-01-29T19:31:55.404Z [INFO] TestAgent_RemoveService/service_manager.server: Adding LAN server: server="Node-43faed93-bc6e-1eb5-eb72-3caf413a164b (Addr: tcp/127.0.0.1:29873) (DC: dc1)" writer.go:29: 2021-01-29T19:31:55.406Z [INFO] TestAgent_RemoveService/service_manager: Started DNS server: address=127.0.0.1:29868 network=tcp === CONT TestAgent_RemoveService/normal writer.go:29: 2021-01-29T19:31:55.401Z [DEBUG] TestAgent_RemoveService/normal: removed service: service=redis writer.go:29: 2021-01-29T19:31:55.417Z [INFO] TestAgent_RemoveService/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:55.404Z [DEBUG] TestAgent_RemoveService/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29867 writer.go:29: 2021-01-29T19:31:55.420Z [INFO] TestAgent_RemoveService/normal.server: shutting down server === CONT TestAgent_RemoveService/service_manager writer.go:29: 2021-01-29T19:31:55.411Z [INFO] TestAgent_RemoveService/service_manager: Started HTTP server: address=127.0.0.1:29869 network=tcp writer.go:29: 2021-01-29T19:31:55.405Z [INFO] TestAgent_RemoveService/service_manager.server: Handled event for server in area: event=member-join server=Node-43faed93-bc6e-1eb5-eb72-3caf413a164b.dc1 area=wan writer.go:29: 2021-01-29T19:31:55.428Z [WARN] TestAgent_RemoveService/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:55.434Z [INFO] TestAgent_RemoveService/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29873 [Candidate]" term=2 === CONT TestAgent_RemoveService/normal writer.go:29: 2021-01-29T19:31:55.421Z [WARN] TestAgent_RemoveService/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:55.428Z [ERROR] TestAgent_RemoveService/normal.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_RemoveService/service_manager writer.go:29: 2021-01-29T19:31:55.434Z [INFO] TestAgent_RemoveService/service_manager: started state syncer === CONT TestAgent_RemoveService/normal writer.go:29: 2021-01-29T19:31:55.446Z [WARN] TestAgent_RemoveService/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:55.451Z [INFO] TestAgent_RemoveService/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:55.458Z [INFO] TestAgent_RemoveService/normal: consul server down writer.go:29: 2021-01-29T19:31:55.459Z [INFO] TestAgent_RemoveService/normal: shutdown complete writer.go:29: 2021-01-29T19:31:55.459Z [INFO] TestAgent_RemoveService/normal: Stopping server: protocol=DNS address=127.0.0.1:29862 network=tcp writer.go:29: 2021-01-29T19:31:55.461Z [INFO] TestAgent_RemoveService/normal: Stopping server: protocol=DNS address=127.0.0.1:29862 network=udp writer.go:29: 2021-01-29T19:31:55.462Z [ERROR] TestAgent_RemoveService/normal.server: failed to establish leadership: error="error configuring provider: leadership lost while committing log" writer.go:29: 2021-01-29T19:31:55.478Z [INFO] TestAgent_RemoveService/normal: Stopping server: protocol=HTTP address=127.0.0.1:29863 network=tcp === CONT TestAgent_RemoveService/service_manager writer.go:29: 2021-01-29T19:31:55.463Z [DEBUG] TestAgent_RemoveService/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:55.493Z [DEBUG] TestAgent_RemoveService/service_manager.server.raft: vote granted: from=43faed93-bc6e-1eb5-eb72-3caf413a164b term=2 tally=1 writer.go:29: 2021-01-29T19:31:55.497Z [INFO] TestAgent_RemoveService/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:55.501Z [INFO] TestAgent_RemoveService/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29873 [Leader]" writer.go:29: 2021-01-29T19:31:55.523Z [INFO] TestAgent_RemoveService/service_manager.server: New leader elected: payload=Node-43faed93-bc6e-1eb5-eb72-3caf413a164b writer.go:29: 2021-01-29T19:31:55.531Z [INFO] TestAgent_RemoveService/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:55.534Z [DEBUG] TestAgent_RemoveService/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29873 writer.go:29: 2021-01-29T19:31:55.547Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:55.553Z [INFO] TestAgent_RemoveService/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:55.554Z [INFO] TestAgent_RemoveService/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:55.555Z [DEBUG] TestAgent_RemoveService/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-43faed93-bc6e-1eb5-eb72-3caf413a164b writer.go:29: 2021-01-29T19:31:55.556Z [INFO] TestAgent_RemoveService/service_manager.server: member joined, marking health alive: member=Node-43faed93-bc6e-1eb5-eb72-3caf413a164b writer.go:29: 2021-01-29T19:31:55.738Z [WARN] TestAgent_RemoveService/service_manager: Failed to deregister service: service=redis error="Service {"redis" {}} does not exist" writer.go:29: 2021-01-29T19:31:55.740Z [DEBUG] TestAgent_RemoveService/service_manager: removed check: check=service:memcache writer.go:29: 2021-01-29T19:31:55.742Z [DEBUG] TestAgent_RemoveService/service_manager: removed check: check=check2 writer.go:29: 2021-01-29T19:31:55.745Z [DEBUG] TestAgent_RemoveService/service_manager: removed service: service=memcache writer.go:29: 2021-01-29T19:31:55.750Z [DEBUG] TestAgent_RemoveService/service_manager: removed check: check=service:redis:1 writer.go:29: 2021-01-29T19:31:55.752Z [DEBUG] TestAgent_RemoveService/service_manager: removed check: check=service:redis:2 writer.go:29: 2021-01-29T19:31:55.754Z [DEBUG] TestAgent_RemoveService/service_manager: removed service: service=redis writer.go:29: 2021-01-29T19:31:55.754Z [DEBUG] TestAgent_RemoveService/service_manager: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:55.758Z [INFO] TestAgent_RemoveService/service_manager: Synced node info writer.go:29: 2021-01-29T19:31:55.761Z [INFO] TestAgent_RemoveService/service_manager: Deregistered service: service=memcache writer.go:29: 2021-01-29T19:31:55.764Z [INFO] TestAgent_RemoveService/service_manager: Deregistered service: service=redis writer.go:29: 2021-01-29T19:31:55.768Z [INFO] TestAgent_RemoveService/service_manager: Synced service: service=mysql writer.go:29: 2021-01-29T19:31:55.770Z [DEBUG] TestAgent_RemoveService/service_manager: Check in sync: check=service:mysql:1 writer.go:29: 2021-01-29T19:31:55.772Z [DEBUG] TestAgent_RemoveService/service_manager: Check in sync: check=service:mysql:2 writer.go:29: 2021-01-29T19:31:55.782Z [INFO] TestAgent_RemoveService/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:55.784Z [INFO] TestAgent_RemoveService/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:55.786Z [DEBUG] TestAgent_RemoveService/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:55.789Z [WARN] TestAgent_RemoveService/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:55.789Z [DEBUG] TestAgent_RemoveService/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:55.796Z [WARN] TestAgent_RemoveService/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:55.798Z [INFO] TestAgent_RemoveService/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:55.799Z [INFO] TestAgent_RemoveService/service_manager: consul server down writer.go:29: 2021-01-29T19:31:55.800Z [INFO] TestAgent_RemoveService/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:55.801Z [INFO] TestAgent_RemoveService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29868 network=tcp writer.go:29: 2021-01-29T19:31:55.802Z [INFO] TestAgent_RemoveService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29868 network=udp writer.go:29: 2021-01-29T19:31:55.804Z [INFO] TestAgent_RemoveService/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29869 network=tcp === CONT TestAgent_RemoveService/normal writer.go:29: 2021-01-29T19:31:55.982Z [INFO] TestAgent_RemoveService/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:55.984Z [INFO] TestAgent_RemoveService/normal: Endpoints down === CONT TestAgent_RemoveService/service_manager writer.go:29: 2021-01-29T19:31:56.311Z [INFO] TestAgent_RemoveService/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:56.314Z [INFO] TestAgent_RemoveService/service_manager: Endpoints down --- PASS: TestAgent_RemoveService (0.00s) --- PASS: TestAgent_RemoveService/normal (0.74s) --- PASS: TestAgent_RemoveService/service_manager (1.08s) === RUN TestAgent_RemoveServiceRemovesAllChecks === RUN TestAgent_RemoveServiceRemovesAllChecks/normal === PAUSE TestAgent_RemoveServiceRemovesAllChecks/normal === RUN TestAgent_RemoveServiceRemovesAllChecks/service_manager === PAUSE TestAgent_RemoveServiceRemovesAllChecks/service_manager === CONT TestAgent_RemoveServiceRemovesAllChecks/normal === CONT TestAgent_RemoveServiceRemovesAllChecks/service_manager writer.go:29: 2021-01-29T19:31:56.350Z [WARN] TestAgent_RemoveServiceRemovesAllChecks/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:56.352Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:56.353Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:56.360Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fd1daf08-c0c1-a0d1-ca77-47ed154719ee Address:127.0.0.1:29879}]" writer.go:29: 2021-01-29T19:31:56.362Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29879 [Follower]" leader= writer.go:29: 2021-01-29T19:31:56.362Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.serf.wan: serf: EventMemberJoin: node1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:56.368Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.serf.lan: serf: EventMemberJoin: node1 127.0.0.1 writer.go:29: 2021-01-29T19:31:56.372Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server: Adding LAN server: server="node1 (Addr: tcp/127.0.0.1:29879) (DC: dc1)" writer.go:29: 2021-01-29T19:31:56.374Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server: Handled event for server in area: event=member-join server=node1.dc1 area=wan writer.go:29: 2021-01-29T19:31:56.374Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: Started DNS server: address=127.0.0.1:29874 network=tcp === CONT TestAgent_RemoveServiceRemovesAllChecks/normal writer.go:29: 2021-01-29T19:31:56.371Z [WARN] TestAgent_RemoveServiceRemovesAllChecks/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:56.376Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:56.378Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RemoveServiceRemovesAllChecks/service_manager writer.go:29: 2021-01-29T19:31:56.377Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: Started DNS server: address=127.0.0.1:29874 network=udp === CONT TestAgent_RemoveServiceRemovesAllChecks/normal writer.go:29: 2021-01-29T19:31:56.381Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:995f7dda-65d3-05cf-46d2-8afe74f6ae03 Address:127.0.0.1:29885}]" === CONT TestAgent_RemoveServiceRemovesAllChecks/service_manager writer.go:29: 2021-01-29T19:31:56.383Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: Started HTTP server: address=127.0.0.1:29875 network=tcp === CONT TestAgent_RemoveServiceRemovesAllChecks/normal writer.go:29: 2021-01-29T19:31:56.383Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29885 [Follower]" leader= === CONT TestAgent_RemoveServiceRemovesAllChecks/service_manager writer.go:29: 2021-01-29T19:31:56.385Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: started state syncer === CONT TestAgent_RemoveServiceRemovesAllChecks/normal writer.go:29: 2021-01-29T19:31:56.387Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server.serf.wan: serf: EventMemberJoin: node1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:56.396Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server.serf.lan: serf: EventMemberJoin: node1 127.0.0.1 writer.go:29: 2021-01-29T19:31:56.404Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server: Handled event for server in area: event=member-join server=node1.dc1 area=wan writer.go:29: 2021-01-29T19:31:56.405Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: Started DNS server: address=127.0.0.1:29880 network=udp writer.go:29: 2021-01-29T19:31:56.406Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: Started DNS server: address=127.0.0.1:29880 network=tcp writer.go:29: 2021-01-29T19:31:56.408Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server: Adding LAN server: server="node1 (Addr: tcp/127.0.0.1:29885) (DC: dc1)" writer.go:29: 2021-01-29T19:31:56.408Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: Started HTTP server: address=127.0.0.1:29881 network=tcp writer.go:29: 2021-01-29T19:31:56.411Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: started state syncer === CONT TestAgent_RemoveServiceRemovesAllChecks/service_manager writer.go:29: 2021-01-29T19:31:56.429Z [WARN] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:56.430Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29879 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:56.432Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:56.433Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.raft: vote granted: from=fd1daf08-c0c1-a0d1-ca77-47ed154719ee term=2 tally=1 writer.go:29: 2021-01-29T19:31:56.433Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:56.434Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29879 [Leader]" writer.go:29: 2021-01-29T19:31:56.435Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:56.436Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server: New leader elected: payload=node1 writer.go:29: 2021-01-29T19:31:56.438Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29879 === CONT TestAgent_RemoveServiceRemovesAllChecks/normal writer.go:29: 2021-01-29T19:31:56.452Z [WARN] TestAgent_RemoveServiceRemovesAllChecks/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:56.453Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29885 [Candidate]" term=2 === CONT TestAgent_RemoveServiceRemovesAllChecks/service_manager writer.go:29: 2021-01-29T19:31:56.454Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RemoveServiceRemovesAllChecks/normal writer.go:29: 2021-01-29T19:31:56.455Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal.server.raft: votes: needed=1 === CONT TestAgent_RemoveServiceRemovesAllChecks/service_manager writer.go:29: 2021-01-29T19:31:56.458Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:56.459Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:56.459Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server: Skipping self join check for node since the cluster is too small: node=node1 writer.go:29: 2021-01-29T19:31:56.460Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server: member joined, marking health alive: member=node1 === CONT TestAgent_RemoveServiceRemovesAllChecks/normal writer.go:29: 2021-01-29T19:31:56.456Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal.server.raft: vote granted: from=995f7dda-65d3-05cf-46d2-8afe74f6ae03 term=2 tally=1 writer.go:29: 2021-01-29T19:31:56.463Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:56.464Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29885 [Leader]" writer.go:29: 2021-01-29T19:31:56.465Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:56.465Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server: New leader elected: payload=node1 writer.go:29: 2021-01-29T19:31:56.468Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29885 writer.go:29: 2021-01-29T19:31:56.473Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:56.478Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:56.479Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:56.481Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal.server: Skipping self join check for node since the cluster is too small: node=node1 writer.go:29: 2021-01-29T19:31:56.482Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server: member joined, marking health alive: member=node1 === CONT TestAgent_RemoveServiceRemovesAllChecks/service_manager writer.go:29: 2021-01-29T19:31:56.560Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager: removed check: check=chk1 writer.go:29: 2021-01-29T19:31:56.562Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager: removed check: check=chk2 writer.go:29: 2021-01-29T19:31:56.564Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager: removed service: service=redis writer.go:29: 2021-01-29T19:31:56.568Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:56.570Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:56.571Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:56.571Z [ERROR] TestAgent_RemoveServiceRemovesAllChecks/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:56.573Z [WARN] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:56.573Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:56.577Z [WARN] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:56.583Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:56.583Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: consul server down writer.go:29: 2021-01-29T19:31:56.587Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:56.589Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29874 network=tcp writer.go:29: 2021-01-29T19:31:56.591Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29874 network=udp writer.go:29: 2021-01-29T19:31:56.593Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29875 network=tcp === CONT TestAgent_RemoveServiceRemovesAllChecks/normal writer.go:29: 2021-01-29T19:31:56.635Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:56.640Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: Synced node info writer.go:29: 2021-01-29T19:31:56.642Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal: Node info in sync writer.go:29: 2021-01-29T19:31:56.723Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal: removed check: check=chk1 writer.go:29: 2021-01-29T19:31:56.725Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal: removed check: check=chk2 writer.go:29: 2021-01-29T19:31:56.727Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal: removed service: service=redis writer.go:29: 2021-01-29T19:31:56.745Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:56.748Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:56.750Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:56.752Z [WARN] TestAgent_RemoveServiceRemovesAllChecks/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:56.752Z [DEBUG] TestAgent_RemoveServiceRemovesAllChecks/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:56.759Z [WARN] TestAgent_RemoveServiceRemovesAllChecks/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:56.766Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:56.766Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: consul server down writer.go:29: 2021-01-29T19:31:56.770Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: shutdown complete writer.go:29: 2021-01-29T19:31:56.771Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: Stopping server: protocol=DNS address=127.0.0.1:29880 network=tcp writer.go:29: 2021-01-29T19:31:56.773Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: Stopping server: protocol=DNS address=127.0.0.1:29880 network=udp writer.go:29: 2021-01-29T19:31:56.775Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: Stopping server: protocol=HTTP address=127.0.0.1:29881 network=tcp === CONT TestAgent_RemoveServiceRemovesAllChecks/service_manager writer.go:29: 2021-01-29T19:31:57.097Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:57.098Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/service_manager: Endpoints down === CONT TestAgent_RemoveServiceRemovesAllChecks/normal writer.go:29: 2021-01-29T19:31:57.277Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:57.280Z [INFO] TestAgent_RemoveServiceRemovesAllChecks/normal: Endpoints down --- PASS: TestAgent_RemoveServiceRemovesAllChecks (0.00s) --- PASS: TestAgent_RemoveServiceRemovesAllChecks/service_manager (0.77s) --- PASS: TestAgent_RemoveServiceRemovesAllChecks/normal (0.95s) === RUN TestAgent_IndexChurn === PAUSE TestAgent_IndexChurn === RUN TestAgent_AddCheck === PAUSE TestAgent_AddCheck === RUN TestAgent_AddCheck_StartPassing === PAUSE TestAgent_AddCheck_StartPassing === RUN TestAgent_AddCheck_MinInterval === PAUSE TestAgent_AddCheck_MinInterval === RUN TestAgent_AddCheck_MissingService === PAUSE TestAgent_AddCheck_MissingService === RUN TestAgent_AddCheck_RestoreState === PAUSE TestAgent_AddCheck_RestoreState === RUN TestAgent_AddCheck_ExecDisable === PAUSE TestAgent_AddCheck_ExecDisable === RUN TestAgent_AddCheck_ExecRemoteDisable === PAUSE TestAgent_AddCheck_ExecRemoteDisable === RUN TestAgent_AddCheck_GRPC === PAUSE TestAgent_AddCheck_GRPC === RUN TestAgent_RestoreServiceWithAliasCheck agent_test.go:1412: skipping slow test; set SLOWTEST=1 to run --- SKIP: TestAgent_RestoreServiceWithAliasCheck (0.01s) === RUN TestAgent_AddCheck_Alias === PAUSE TestAgent_AddCheck_Alias === RUN TestAgent_AddCheck_Alias_setToken === PAUSE TestAgent_AddCheck_Alias_setToken === RUN TestAgent_AddCheck_Alias_userToken === PAUSE TestAgent_AddCheck_Alias_userToken === RUN TestAgent_AddCheck_Alias_userAndSetToken === PAUSE TestAgent_AddCheck_Alias_userAndSetToken === RUN TestAgent_RemoveCheck === PAUSE TestAgent_RemoveCheck === RUN TestAgent_HTTPCheck_TLSSkipVerify === PAUSE TestAgent_HTTPCheck_TLSSkipVerify === RUN TestAgent_HTTPCheck_EnableAgentTLSForChecks agent_test.go:1774: DM-skipped --- SKIP: TestAgent_HTTPCheck_EnableAgentTLSForChecks (0.00s) === RUN TestAgent_updateTTLCheck === PAUSE TestAgent_updateTTLCheck === RUN TestAgent_PersistService === RUN TestAgent_PersistService/normal === PAUSE TestAgent_PersistService/normal === RUN TestAgent_PersistService/service_manager === PAUSE TestAgent_PersistService/service_manager === CONT TestAgent_PersistService/normal === CONT TestAgent_PersistService/service_manager === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:57.399Z [DEBUG] TestAgent_PersistService/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:57.406Z [INFO] TestAgent_PersistService/normal.client.serf.lan: serf: EventMemberJoin: Node-395ade5f-01c5-b75d-eef9-58a3fb72cbcc 127.0.0.1 === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:57.413Z [DEBUG] TestAgent_PersistService/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:57.415Z [INFO] TestAgent_PersistService/service_manager.client.serf.lan: serf: EventMemberJoin: Node-4134cb0c-2d02-20c0-4c5b-4331588944a9 127.0.0.1 writer.go:29: 2021-01-29T19:31:57.418Z [INFO] TestAgent_PersistService/service_manager: Started DNS server: address=127.0.0.1:29892 network=udp writer.go:29: 2021-01-29T19:31:57.419Z [INFO] TestAgent_PersistService/service_manager: Started DNS server: address=127.0.0.1:29892 network=tcp === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:57.422Z [INFO] TestAgent_PersistService/normal: Started DNS server: address=127.0.0.1:29886 network=tcp === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:57.421Z [INFO] TestAgent_PersistService/service_manager: Started HTTP server: address=127.0.0.1:29893 network=tcp === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:57.425Z [INFO] TestAgent_PersistService/normal: Started DNS server: address=127.0.0.1:29886 network=udp === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:57.427Z [INFO] TestAgent_PersistService/service_manager: started state syncer writer.go:29: 2021-01-29T19:31:57.428Z [WARN] TestAgent_PersistService/service_manager.client.manager: No servers available writer.go:29: 2021-01-29T19:31:57.430Z [ERROR] TestAgent_PersistService/service_manager.anti_entropy: failed to sync remote state: error="No known Consul servers" === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:57.430Z [INFO] TestAgent_PersistService/normal: Started HTTP server: address=127.0.0.1:29887 network=tcp writer.go:29: 2021-01-29T19:31:57.443Z [INFO] TestAgent_PersistService/normal: started state syncer writer.go:29: 2021-01-29T19:31:57.443Z [WARN] TestAgent_PersistService/normal.client.manager: No servers available === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:57.445Z [INFO] TestAgent_PersistService/service_manager: Requesting shutdown === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:57.447Z [ERROR] TestAgent_PersistService/normal.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:31:57.449Z [INFO] TestAgent_PersistService/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:57.450Z [INFO] TestAgent_PersistService/normal.client: shutting down client writer.go:29: 2021-01-29T19:31:57.451Z [WARN] TestAgent_PersistService/normal.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:57.455Z [INFO] TestAgent_PersistService/normal.client.manager: shutting down writer.go:29: 2021-01-29T19:31:57.460Z [INFO] TestAgent_PersistService/normal: consul client down writer.go:29: 2021-01-29T19:31:57.463Z [INFO] TestAgent_PersistService/normal: shutdown complete === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:57.448Z [INFO] TestAgent_PersistService/service_manager.client: shutting down client === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:57.465Z [INFO] TestAgent_PersistService/normal: Stopping server: protocol=DNS address=127.0.0.1:29886 network=tcp writer.go:29: 2021-01-29T19:31:57.469Z [INFO] TestAgent_PersistService/normal: Stopping server: protocol=DNS address=127.0.0.1:29886 network=udp === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:57.468Z [WARN] TestAgent_PersistService/service_manager.client.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:57.471Z [INFO] TestAgent_PersistService/normal: Stopping server: protocol=HTTP address=127.0.0.1:29887 network=tcp === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:57.468Z [INFO] TestAgent_PersistService/service_manager.client.manager: shutting down writer.go:29: 2021-01-29T19:31:57.481Z [INFO] TestAgent_PersistService/service_manager: consul client down writer.go:29: 2021-01-29T19:31:57.483Z [INFO] TestAgent_PersistService/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:57.485Z [INFO] TestAgent_PersistService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29892 network=tcp writer.go:29: 2021-01-29T19:31:57.487Z [INFO] TestAgent_PersistService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29892 network=udp writer.go:29: 2021-01-29T19:31:57.489Z [INFO] TestAgent_PersistService/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29893 network=tcp === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:57.980Z [INFO] TestAgent_PersistService/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:57.982Z [INFO] TestAgent_PersistService/normal: Endpoints down === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:58.001Z [INFO] TestAgent_PersistService/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:58.003Z [INFO] TestAgent_PersistService/service_manager: Endpoints down === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:58.126Z [DEBUG] TestAgent_PersistService/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:58.128Z [INFO] TestAgent_PersistService/normal.client.serf.lan: serf: EventMemberJoin: Node-b3e7c326-042d-c158-0588-8f7534aefc0c 127.0.0.1 writer.go:29: 2021-01-29T19:31:58.143Z [DEBUG] TestAgent_PersistService/normal: restored service definition from file: service=redis file=/tmp/consul-test/TestAgent_PersistService_normal-agent074823670/services/86a1b907d54bf7010394bf316e183e67 === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:58.140Z [DEBUG] TestAgent_PersistService/service_manager.tlsutil: Update: version=1 === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:58.146Z [INFO] TestAgent_PersistService/normal: Started DNS server: address=127.0.0.1:29898 network=udp writer.go:29: 2021-01-29T19:31:58.148Z [INFO] TestAgent_PersistService/normal: Started DNS server: address=127.0.0.1:29898 network=tcp writer.go:29: 2021-01-29T19:31:58.149Z [INFO] TestAgent_PersistService/normal: Started HTTP server: address=127.0.0.1:29899 network=tcp writer.go:29: 2021-01-29T19:31:58.150Z [INFO] TestAgent_PersistService/normal: started state syncer writer.go:29: 2021-01-29T19:31:58.151Z [INFO] TestAgent_PersistService/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:58.153Z [WARN] TestAgent_PersistService/normal.client.manager: No servers available writer.go:29: 2021-01-29T19:31:58.154Z [ERROR] TestAgent_PersistService/normal.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:31:58.156Z [INFO] TestAgent_PersistService/normal.client.serf.lan: serf: Attempting re-join to previously known node: Node-395ade5f-01c5-b75d-eef9-58a3fb72cbcc: 127.0.0.1:29889 writer.go:29: 2021-01-29T19:31:58.157Z [DEBUG] TestAgent_PersistService/normal.client.memberlist.lan: memberlist: Failed to join 127.0.0.1: dial tcp 127.0.0.1:29889: connect: connection refused writer.go:29: 2021-01-29T19:31:58.158Z [WARN] TestAgent_PersistService/normal.client.serf.lan: serf: Failed to re-join any previously known node writer.go:29: 2021-01-29T19:31:58.159Z [INFO] TestAgent_PersistService/normal.client: shutting down client === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:58.162Z [INFO] TestAgent_PersistService/service_manager.client.serf.lan: serf: EventMemberJoin: Node-20bf75e4-a6a3-d095-bf72-97b8b63c8980 127.0.0.1 writer.go:29: 2021-01-29T19:31:58.165Z [DEBUG] TestAgent_PersistService/service_manager: restored service definition from file: service=redis file=/tmp/consul-test/TestAgent_PersistService_service_manager-agent421561565/services/86a1b907d54bf7010394bf316e183e67 writer.go:29: 2021-01-29T19:31:58.167Z [INFO] TestAgent_PersistService/service_manager.client.serf.lan: serf: Attempting re-join to previously known node: Node-4134cb0c-2d02-20c0-4c5b-4331588944a9: 127.0.0.1:29895 writer.go:29: 2021-01-29T19:31:58.168Z [INFO] TestAgent_PersistService/service_manager: Started DNS server: address=127.0.0.1:29904 network=udp writer.go:29: 2021-01-29T19:31:58.170Z [DEBUG] TestAgent_PersistService/service_manager.client.memberlist.lan: memberlist: Failed to join 127.0.0.1: dial tcp 127.0.0.1:29895: connect: connection refused writer.go:29: 2021-01-29T19:31:58.171Z [WARN] TestAgent_PersistService/service_manager.client.serf.lan: serf: Failed to re-join any previously known node === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:58.160Z [WARN] TestAgent_PersistService/normal.client.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:58.171Z [INFO] TestAgent_PersistService/service_manager: Started DNS server: address=127.0.0.1:29904 network=tcp === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:58.161Z [ERROR] TestAgent_PersistService/normal: proxy config manager exited with error: error="manager stopped" writer.go:29: 2021-01-29T19:31:58.161Z [INFO] TestAgent_PersistService/normal.client.manager: shutting down === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:58.173Z [INFO] TestAgent_PersistService/service_manager: Started HTTP server: address=127.0.0.1:29905 network=tcp writer.go:29: 2021-01-29T19:31:58.175Z [INFO] TestAgent_PersistService/service_manager: started state syncer writer.go:29: 2021-01-29T19:31:58.178Z [INFO] TestAgent_PersistService/service_manager: Requesting shutdown === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:58.180Z [INFO] TestAgent_PersistService/normal: consul client down writer.go:29: 2021-01-29T19:31:58.182Z [INFO] TestAgent_PersistService/normal: shutdown complete === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:58.177Z [WARN] TestAgent_PersistService/service_manager.client.manager: No servers available === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:58.184Z [INFO] TestAgent_PersistService/normal: Stopping server: protocol=DNS address=127.0.0.1:29898 network=tcp writer.go:29: 2021-01-29T19:31:58.186Z [INFO] TestAgent_PersistService/normal: Stopping server: protocol=DNS address=127.0.0.1:29898 network=udp === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:58.185Z [ERROR] TestAgent_PersistService/service_manager.anti_entropy: failed to sync remote state: error="No known Consul servers" === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:58.188Z [INFO] TestAgent_PersistService/normal: Stopping server: protocol=HTTP address=127.0.0.1:29899 network=tcp === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:58.181Z [INFO] TestAgent_PersistService/service_manager.client: shutting down client writer.go:29: 2021-01-29T19:31:58.190Z [WARN] TestAgent_PersistService/service_manager.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:58.190Z [INFO] TestAgent_PersistService/service_manager.client.manager: shutting down writer.go:29: 2021-01-29T19:31:58.193Z [INFO] TestAgent_PersistService/service_manager: consul client down writer.go:29: 2021-01-29T19:31:58.196Z [INFO] TestAgent_PersistService/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:58.198Z [INFO] TestAgent_PersistService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29904 network=tcp writer.go:29: 2021-01-29T19:31:58.200Z [INFO] TestAgent_PersistService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29904 network=udp writer.go:29: 2021-01-29T19:31:58.201Z [INFO] TestAgent_PersistService/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29905 network=tcp === CONT TestAgent_PersistService/normal writer.go:29: 2021-01-29T19:31:58.689Z [INFO] TestAgent_PersistService/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:58.691Z [INFO] TestAgent_PersistService/normal: Endpoints down === CONT TestAgent_PersistService/service_manager writer.go:29: 2021-01-29T19:31:58.702Z [INFO] TestAgent_PersistService/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:58.704Z [INFO] TestAgent_PersistService/service_manager: Endpoints down --- PASS: TestAgent_PersistService (0.00s) --- PASS: TestAgent_PersistService/normal (1.34s) --- PASS: TestAgent_PersistService/service_manager (1.35s) === RUN TestAgent_persistedService_compat === RUN TestAgent_persistedService_compat/normal === PAUSE TestAgent_persistedService_compat/normal === RUN TestAgent_persistedService_compat/service_manager === PAUSE TestAgent_persistedService_compat/service_manager === CONT TestAgent_persistedService_compat/normal === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.742Z [WARN] TestAgent_persistedService_compat/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:58.743Z [DEBUG] TestAgent_persistedService_compat/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:58.745Z [DEBUG] TestAgent_persistedService_compat/service_manager.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:58.752Z [INFO] TestAgent_persistedService_compat/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:90f2d70f-7184-2082-4e61-3116160b3a31 Address:127.0.0.1:29915}]" writer.go:29: 2021-01-29T19:31:58.753Z [INFO] TestAgent_persistedService_compat/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29915 [Follower]" leader= writer.go:29: 2021-01-29T19:31:58.756Z [INFO] TestAgent_persistedService_compat/service_manager.server.serf.wan: serf: EventMemberJoin: Node-90f2d70f-7184-2082-4e61-3116160b3a31.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:58.760Z [INFO] TestAgent_persistedService_compat/service_manager.server.serf.lan: serf: EventMemberJoin: Node-90f2d70f-7184-2082-4e61-3116160b3a31 127.0.0.1 === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.758Z [WARN] TestAgent_persistedService_compat/normal: bootstrap = true: do not enable unless necessary === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.762Z [INFO] TestAgent_persistedService_compat/service_manager.server: Handled event for server in area: event=member-join server=Node-90f2d70f-7184-2082-4e61-3116160b3a31.dc1 area=wan writer.go:29: 2021-01-29T19:31:58.763Z [INFO] TestAgent_persistedService_compat/service_manager.server: Adding LAN server: server="Node-90f2d70f-7184-2082-4e61-3116160b3a31 (Addr: tcp/127.0.0.1:29915) (DC: dc1)" writer.go:29: 2021-01-29T19:31:58.766Z [INFO] TestAgent_persistedService_compat/service_manager: Started DNS server: address=127.0.0.1:29910 network=tcp writer.go:29: 2021-01-29T19:31:58.768Z [INFO] TestAgent_persistedService_compat/service_manager: Started DNS server: address=127.0.0.1:29910 network=udp writer.go:29: 2021-01-29T19:31:58.770Z [INFO] TestAgent_persistedService_compat/service_manager: Started HTTP server: address=127.0.0.1:29911 network=tcp writer.go:29: 2021-01-29T19:31:58.771Z [INFO] TestAgent_persistedService_compat/service_manager: started state syncer === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.763Z [DEBUG] TestAgent_persistedService_compat/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:58.773Z [DEBUG] TestAgent_persistedService_compat/normal.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:58.782Z [INFO] TestAgent_persistedService_compat/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6cf15998-9838-c432-7661-b2577224c589 Address:127.0.0.1:29921}]" writer.go:29: 2021-01-29T19:31:58.784Z [INFO] TestAgent_persistedService_compat/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29921 [Follower]" leader= writer.go:29: 2021-01-29T19:31:58.786Z [INFO] TestAgent_persistedService_compat/normal.server.serf.wan: serf: EventMemberJoin: Node-6cf15998-9838-c432-7661-b2577224c589.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:58.794Z [INFO] TestAgent_persistedService_compat/normal.server.serf.lan: serf: EventMemberJoin: Node-6cf15998-9838-c432-7661-b2577224c589 127.0.0.1 writer.go:29: 2021-01-29T19:31:58.798Z [INFO] TestAgent_persistedService_compat/normal.server: Handled event for server in area: event=member-join server=Node-6cf15998-9838-c432-7661-b2577224c589.dc1 area=wan writer.go:29: 2021-01-29T19:31:58.798Z [INFO] TestAgent_persistedService_compat/normal.server: Adding LAN server: server="Node-6cf15998-9838-c432-7661-b2577224c589 (Addr: tcp/127.0.0.1:29921) (DC: dc1)" writer.go:29: 2021-01-29T19:31:58.801Z [INFO] TestAgent_persistedService_compat/normal: Started DNS server: address=127.0.0.1:29916 network=tcp writer.go:29: 2021-01-29T19:31:58.804Z [INFO] TestAgent_persistedService_compat/normal: Started DNS server: address=127.0.0.1:29916 network=udp writer.go:29: 2021-01-29T19:31:58.812Z [INFO] TestAgent_persistedService_compat/normal: Started HTTP server: address=127.0.0.1:29917 network=tcp writer.go:29: 2021-01-29T19:31:58.813Z [INFO] TestAgent_persistedService_compat/normal: started state syncer === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.813Z [WARN] TestAgent_persistedService_compat/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:58.815Z [INFO] TestAgent_persistedService_compat/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29915 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:58.819Z [DEBUG] TestAgent_persistedService_compat/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:58.822Z [DEBUG] TestAgent_persistedService_compat/service_manager.server.raft: vote granted: from=90f2d70f-7184-2082-4e61-3116160b3a31 term=2 tally=1 === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.823Z [WARN] TestAgent_persistedService_compat/normal.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.826Z [INFO] TestAgent_persistedService_compat/service_manager.server.raft: election won: tally=1 === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.826Z [INFO] TestAgent_persistedService_compat/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29921 [Candidate]" term=2 === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.829Z [INFO] TestAgent_persistedService_compat/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29915 [Leader]" === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.831Z [DEBUG] TestAgent_persistedService_compat/normal.server.raft: votes: needed=1 === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.833Z [INFO] TestAgent_persistedService_compat/service_manager.server: cluster leadership acquired === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.834Z [DEBUG] TestAgent_persistedService_compat/normal.server.raft: vote granted: from=6cf15998-9838-c432-7661-b2577224c589 term=2 tally=1 === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.834Z [INFO] TestAgent_persistedService_compat/service_manager.server: New leader elected: payload=Node-90f2d70f-7184-2082-4e61-3116160b3a31 === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.835Z [INFO] TestAgent_persistedService_compat/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:58.837Z [INFO] TestAgent_persistedService_compat/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29921 [Leader]" === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.835Z [DEBUG] TestAgent_persistedService_compat/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29915 === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.838Z [INFO] TestAgent_persistedService_compat/normal.server: cluster leadership acquired === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.841Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.839Z [INFO] TestAgent_persistedService_compat/normal.server: New leader elected: payload=Node-6cf15998-9838-c432-7661-b2577224c589 === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.847Z [INFO] TestAgent_persistedService_compat/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:58.849Z [INFO] TestAgent_persistedService_compat/service_manager.leader: started routine: routine="CA root pruning" === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.846Z [DEBUG] TestAgent_persistedService_compat/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29921 === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:58.851Z [DEBUG] TestAgent_persistedService_compat/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-90f2d70f-7184-2082-4e61-3116160b3a31 writer.go:29: 2021-01-29T19:31:58.853Z [INFO] TestAgent_persistedService_compat/service_manager.server: member joined, marking health alive: member=Node-90f2d70f-7184-2082-4e61-3116160b3a31 === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:58.854Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:58.862Z [INFO] TestAgent_persistedService_compat/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:58.866Z [INFO] TestAgent_persistedService_compat/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:58.872Z [DEBUG] TestAgent_persistedService_compat/normal.server: Skipping self join check for node since the cluster is too small: node=Node-6cf15998-9838-c432-7661-b2577224c589 writer.go:29: 2021-01-29T19:31:58.874Z [INFO] TestAgent_persistedService_compat/normal.server: member joined, marking health alive: member=Node-6cf15998-9838-c432-7661-b2577224c589 writer.go:29: 2021-01-29T19:31:58.931Z [DEBUG] TestAgent_persistedService_compat/normal: restored service definition from file: service=redis file=/tmp/TestAgent_persistedService_compat_normal-agent728830231/services/86a1b907d54bf7010394bf316e183e67 writer.go:29: 2021-01-29T19:31:58.934Z [INFO] TestAgent_persistedService_compat/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:58.936Z [DEBUG] TestAgent_persistedService_compat/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:58.936Z [INFO] TestAgent_persistedService_compat/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:58.964Z [DEBUG] TestAgent_persistedService_compat/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:58.938Z [INFO] TestAgent_persistedService_compat/normal: Synced node info writer.go:29: 2021-01-29T19:31:58.969Z [WARN] TestAgent_persistedService_compat/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:58.969Z [DEBUG] TestAgent_persistedService_compat/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:58.979Z [WARN] TestAgent_persistedService_compat/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:58.981Z [INFO] TestAgent_persistedService_compat/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:58.982Z [INFO] TestAgent_persistedService_compat/normal: Synced service: service=redis writer.go:29: 2021-01-29T19:31:58.984Z [DEBUG] TestAgent_persistedService_compat/normal: Node info in sync writer.go:29: 2021-01-29T19:31:58.987Z [DEBUG] TestAgent_persistedService_compat/normal: Service in sync: service=redis writer.go:29: 2021-01-29T19:31:58.982Z [INFO] TestAgent_persistedService_compat/normal: consul server down writer.go:29: 2021-01-29T19:31:58.995Z [INFO] TestAgent_persistedService_compat/normal: shutdown complete writer.go:29: 2021-01-29T19:31:59.001Z [INFO] TestAgent_persistedService_compat/normal: Stopping server: protocol=DNS address=127.0.0.1:29916 network=tcp writer.go:29: 2021-01-29T19:31:59.008Z [INFO] TestAgent_persistedService_compat/normal: Stopping server: protocol=DNS address=127.0.0.1:29916 network=udp writer.go:29: 2021-01-29T19:31:59.009Z [INFO] TestAgent_persistedService_compat/normal: Stopping server: protocol=HTTP address=127.0.0.1:29917 network=tcp === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:59.058Z [DEBUG] TestAgent_persistedService_compat/service_manager: restored service definition from file: service=redis file=/tmp/TestAgent_persistedService_compat_service_manager-agent740467864/services/86a1b907d54bf7010394bf316e183e67 writer.go:29: 2021-01-29T19:31:59.061Z [INFO] TestAgent_persistedService_compat/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:31:59.062Z [INFO] TestAgent_persistedService_compat/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:31:59.064Z [ERROR] TestAgent_persistedService_compat/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:59.065Z [DEBUG] TestAgent_persistedService_compat/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:59.067Z [WARN] TestAgent_persistedService_compat/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:59.067Z [DEBUG] TestAgent_persistedService_compat/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:59.072Z [WARN] TestAgent_persistedService_compat/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:59.076Z [INFO] TestAgent_persistedService_compat/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:59.082Z [INFO] TestAgent_persistedService_compat/service_manager: consul server down writer.go:29: 2021-01-29T19:31:59.086Z [INFO] TestAgent_persistedService_compat/service_manager: shutdown complete writer.go:29: 2021-01-29T19:31:59.088Z [INFO] TestAgent_persistedService_compat/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29910 network=tcp writer.go:29: 2021-01-29T19:31:59.091Z [INFO] TestAgent_persistedService_compat/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29910 network=udp writer.go:29: 2021-01-29T19:31:59.092Z [INFO] TestAgent_persistedService_compat/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29911 network=tcp === CONT TestAgent_persistedService_compat/normal writer.go:29: 2021-01-29T19:31:59.512Z [INFO] TestAgent_persistedService_compat/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:59.514Z [INFO] TestAgent_persistedService_compat/normal: Endpoints down === CONT TestAgent_persistedService_compat/service_manager writer.go:29: 2021-01-29T19:31:59.593Z [INFO] TestAgent_persistedService_compat/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:59.594Z [INFO] TestAgent_persistedService_compat/service_manager: Endpoints down --- PASS: TestAgent_persistedService_compat (0.00s) --- PASS: TestAgent_persistedService_compat/normal (0.80s) --- PASS: TestAgent_persistedService_compat/service_manager (0.88s) === RUN TestAgent_PurgeService === RUN TestAgent_PurgeService/normal === PAUSE TestAgent_PurgeService/normal === RUN TestAgent_PurgeService/service_manager === PAUSE TestAgent_PurgeService/service_manager === CONT TestAgent_PurgeService/normal === CONT TestAgent_PurgeService/service_manager === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.617Z [WARN] TestAgent_PurgeService/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:59.618Z [DEBUG] TestAgent_PurgeService/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:59.619Z [DEBUG] TestAgent_PurgeService/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.662Z [WARN] TestAgent_PurgeService/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:59.664Z [DEBUG] TestAgent_PurgeService/service_manager.tlsutil: Update: version=1 === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.665Z [INFO] TestAgent_PurgeService/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bc9594df-43b4-cb09-e555-d9b46506e14d Address:127.0.0.1:29927}]" writer.go:29: 2021-01-29T19:31:59.667Z [INFO] TestAgent_PurgeService/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29927 [Follower]" leader= writer.go:29: 2021-01-29T19:31:59.669Z [INFO] TestAgent_PurgeService/normal.server.serf.wan: serf: EventMemberJoin: Node-bc9594df-43b4-cb09-e555-d9b46506e14d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:59.673Z [INFO] TestAgent_PurgeService/normal.server.serf.lan: serf: EventMemberJoin: Node-bc9594df-43b4-cb09-e555-d9b46506e14d 127.0.0.1 === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.670Z [DEBUG] TestAgent_PurgeService/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.677Z [INFO] TestAgent_PurgeService/normal: Started DNS server: address=127.0.0.1:29922 network=udp writer.go:29: 2021-01-29T19:31:59.692Z [INFO] TestAgent_PurgeService/normal: Started DNS server: address=127.0.0.1:29922 network=tcp writer.go:29: 2021-01-29T19:31:59.695Z [INFO] TestAgent_PurgeService/normal: Started HTTP server: address=127.0.0.1:29923 network=tcp writer.go:29: 2021-01-29T19:31:59.696Z [INFO] TestAgent_PurgeService/normal: started state syncer writer.go:29: 2021-01-29T19:31:59.693Z [INFO] TestAgent_PurgeService/normal.server: Handled event for server in area: event=member-join server=Node-bc9594df-43b4-cb09-e555-d9b46506e14d.dc1 area=wan writer.go:29: 2021-01-29T19:31:59.693Z [INFO] TestAgent_PurgeService/normal.server: Adding LAN server: server="Node-bc9594df-43b4-cb09-e555-d9b46506e14d (Addr: tcp/127.0.0.1:29927) (DC: dc1)" === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.705Z [INFO] TestAgent_PurgeService/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ccadc8b6-7135-41e1-7f30-212ad56a4b7e Address:127.0.0.1:29933}]" writer.go:29: 2021-01-29T19:31:59.708Z [INFO] TestAgent_PurgeService/service_manager.server.serf.wan: serf: EventMemberJoin: Node-ccadc8b6-7135-41e1-7f30-212ad56a4b7e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:59.713Z [INFO] TestAgent_PurgeService/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29933 [Follower]" leader= writer.go:29: 2021-01-29T19:31:59.714Z [INFO] TestAgent_PurgeService/service_manager.server.serf.lan: serf: EventMemberJoin: Node-ccadc8b6-7135-41e1-7f30-212ad56a4b7e 127.0.0.1 === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.746Z [WARN] TestAgent_PurgeService/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:59.752Z [INFO] TestAgent_PurgeService/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29927 [Candidate]" term=2 === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.722Z [INFO] TestAgent_PurgeService/service_manager: Started DNS server: address=127.0.0.1:29928 network=udp writer.go:29: 2021-01-29T19:31:59.768Z [INFO] TestAgent_PurgeService/service_manager.server: Adding LAN server: server="Node-ccadc8b6-7135-41e1-7f30-212ad56a4b7e (Addr: tcp/127.0.0.1:29933) (DC: dc1)" writer.go:29: 2021-01-29T19:31:59.769Z [INFO] TestAgent_PurgeService/service_manager.server: Handled event for server in area: event=member-join server=Node-ccadc8b6-7135-41e1-7f30-212ad56a4b7e.dc1 area=wan === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.769Z [DEBUG] TestAgent_PurgeService/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:59.770Z [DEBUG] TestAgent_PurgeService/normal.server.raft: vote granted: from=bc9594df-43b4-cb09-e555-d9b46506e14d term=2 tally=1 === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.771Z [INFO] TestAgent_PurgeService/service_manager: Started DNS server: address=127.0.0.1:29928 network=tcp === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.772Z [INFO] TestAgent_PurgeService/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:59.773Z [INFO] TestAgent_PurgeService/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29927 [Leader]" === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.773Z [INFO] TestAgent_PurgeService/service_manager: Started HTTP server: address=127.0.0.1:29929 network=tcp writer.go:29: 2021-01-29T19:31:59.776Z [INFO] TestAgent_PurgeService/service_manager: started state syncer === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.776Z [INFO] TestAgent_PurgeService/normal.server: cluster leadership acquired === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.778Z [WARN] TestAgent_PurgeService/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:59.779Z [INFO] TestAgent_PurgeService/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29933 [Candidate]" term=2 === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.784Z [INFO] TestAgent_PurgeService/normal.server: New leader elected: payload=Node-bc9594df-43b4-cb09-e555-d9b46506e14d writer.go:29: 2021-01-29T19:31:59.785Z [DEBUG] TestAgent_PurgeService/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29927 writer.go:29: 2021-01-29T19:31:59.789Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.783Z [DEBUG] TestAgent_PurgeService/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:59.790Z [DEBUG] TestAgent_PurgeService/service_manager.server.raft: vote granted: from=ccadc8b6-7135-41e1-7f30-212ad56a4b7e term=2 tally=1 writer.go:29: 2021-01-29T19:31:59.793Z [INFO] TestAgent_PurgeService/service_manager.server.raft: election won: tally=1 === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.794Z [INFO] TestAgent_PurgeService/normal.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.794Z [INFO] TestAgent_PurgeService/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29933 [Leader]" === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.795Z [INFO] TestAgent_PurgeService/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.796Z [INFO] TestAgent_PurgeService/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:59.798Z [INFO] TestAgent_PurgeService/service_manager.server: New leader elected: payload=Node-ccadc8b6-7135-41e1-7f30-212ad56a4b7e === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.797Z [DEBUG] TestAgent_PurgeService/normal.server: Skipping self join check for node since the cluster is too small: node=Node-bc9594df-43b4-cb09-e555-d9b46506e14d writer.go:29: 2021-01-29T19:31:59.800Z [INFO] TestAgent_PurgeService/normal.server: member joined, marking health alive: member=Node-bc9594df-43b4-cb09-e555-d9b46506e14d writer.go:29: 2021-01-29T19:31:59.806Z [DEBUG] TestAgent_PurgeService/normal: removed service: service=redis writer.go:29: 2021-01-29T19:31:59.807Z [DEBUG] TestAgent_PurgeService/normal: removed service: service=redis writer.go:29: 2021-01-29T19:31:59.808Z [INFO] TestAgent_PurgeService/normal: Requesting shutdown writer.go:29: 2021-01-29T19:31:59.809Z [INFO] TestAgent_PurgeService/normal.server: shutting down server writer.go:29: 2021-01-29T19:31:59.810Z [DEBUG] TestAgent_PurgeService/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:59.811Z [WARN] TestAgent_PurgeService/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:59.810Z [ERROR] TestAgent_PurgeService/normal.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:59.811Z [DEBUG] TestAgent_PurgeService/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:59.813Z [WARN] TestAgent_PurgeService/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:59.816Z [INFO] TestAgent_PurgeService/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:59.818Z [INFO] TestAgent_PurgeService/normal: consul server down writer.go:29: 2021-01-29T19:31:59.822Z [INFO] TestAgent_PurgeService/normal: shutdown complete === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.801Z [DEBUG] TestAgent_PurgeService/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29933 === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.824Z [INFO] TestAgent_PurgeService/normal: Stopping server: protocol=DNS address=127.0.0.1:29922 network=tcp writer.go:29: 2021-01-29T19:31:59.830Z [INFO] TestAgent_PurgeService/normal: Stopping server: protocol=DNS address=127.0.0.1:29922 network=udp === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.834Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:59.840Z [INFO] TestAgent_PurgeService/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:59.842Z [INFO] TestAgent_PurgeService/service_manager.leader: started routine: routine="CA root pruning" === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:31:59.831Z [INFO] TestAgent_PurgeService/normal: Stopping server: protocol=HTTP address=127.0.0.1:29923 network=tcp === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:31:59.845Z [DEBUG] TestAgent_PurgeService/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-ccadc8b6-7135-41e1-7f30-212ad56a4b7e writer.go:29: 2021-01-29T19:31:59.848Z [INFO] TestAgent_PurgeService/service_manager.server: member joined, marking health alive: member=Node-ccadc8b6-7135-41e1-7f30-212ad56a4b7e writer.go:29: 2021-01-29T19:31:59.997Z [DEBUG] TestAgent_PurgeService/service_manager: removed service: service=redis writer.go:29: 2021-01-29T19:31:59.999Z [DEBUG] TestAgent_PurgeService/service_manager: removed service: service=redis writer.go:29: 2021-01-29T19:32:00.000Z [INFO] TestAgent_PurgeService/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:32:00.003Z [INFO] TestAgent_PurgeService/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:32:00.004Z [DEBUG] TestAgent_PurgeService/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:00.006Z [WARN] TestAgent_PurgeService/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:00.004Z [ERROR] TestAgent_PurgeService/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:00.006Z [DEBUG] TestAgent_PurgeService/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:00.009Z [WARN] TestAgent_PurgeService/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:00.018Z [INFO] TestAgent_PurgeService/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:00.018Z [INFO] TestAgent_PurgeService/service_manager: consul server down writer.go:29: 2021-01-29T19:32:00.021Z [INFO] TestAgent_PurgeService/service_manager: shutdown complete writer.go:29: 2021-01-29T19:32:00.023Z [INFO] TestAgent_PurgeService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29928 network=tcp writer.go:29: 2021-01-29T19:32:00.025Z [INFO] TestAgent_PurgeService/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29928 network=udp writer.go:29: 2021-01-29T19:32:00.027Z [INFO] TestAgent_PurgeService/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29929 network=tcp === CONT TestAgent_PurgeService/normal writer.go:29: 2021-01-29T19:32:00.347Z [INFO] TestAgent_PurgeService/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:00.349Z [INFO] TestAgent_PurgeService/normal: Endpoints down === CONT TestAgent_PurgeService/service_manager writer.go:29: 2021-01-29T19:32:00.529Z [INFO] TestAgent_PurgeService/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:00.531Z [INFO] TestAgent_PurgeService/service_manager: Endpoints down --- PASS: TestAgent_PurgeService (0.00s) --- PASS: TestAgent_PurgeService/normal (0.75s) --- PASS: TestAgent_PurgeService/service_manager (0.93s) === RUN TestAgent_PurgeServiceOnDuplicate === RUN TestAgent_PurgeServiceOnDuplicate/normal === PAUSE TestAgent_PurgeServiceOnDuplicate/normal === RUN TestAgent_PurgeServiceOnDuplicate/service_manager === PAUSE TestAgent_PurgeServiceOnDuplicate/service_manager === CONT TestAgent_PurgeServiceOnDuplicate/normal === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:00.547Z [DEBUG] TestAgent_PurgeServiceOnDuplicate/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:00.559Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager.client.serf.lan: serf: EventMemberJoin: Node-df6b6fd8-b773-4d91-34fd-b1d36f77fa83 127.0.0.1 writer.go:29: 2021-01-29T19:32:00.565Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: Started DNS server: address=127.0.0.1:29940 network=udp writer.go:29: 2021-01-29T19:32:00.566Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: Started DNS server: address=127.0.0.1:29940 network=tcp writer.go:29: 2021-01-29T19:32:00.569Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: Started HTTP server: address=127.0.0.1:29941 network=tcp writer.go:29: 2021-01-29T19:32:00.569Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: started state syncer writer.go:29: 2021-01-29T19:32:00.571Z [WARN] TestAgent_PurgeServiceOnDuplicate/service_manager.client.manager: No servers available writer.go:29: 2021-01-29T19:32:00.573Z [ERROR] TestAgent_PurgeServiceOnDuplicate/service_manager.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:32:00.579Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:32:00.585Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager.client: shutting down client writer.go:29: 2021-01-29T19:32:00.586Z [WARN] TestAgent_PurgeServiceOnDuplicate/service_manager.client.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:00.591Z [DEBUG] TestAgent_PurgeServiceOnDuplicate/normal.tlsutil: Update: version=1 === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:00.587Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager.client.manager: shutting down writer.go:29: 2021-01-29T19:32:00.589Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: consul client down writer.go:29: 2021-01-29T19:32:00.594Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: shutdown complete writer.go:29: 2021-01-29T19:32:00.594Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29940 network=tcp === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:00.594Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal.client.serf.lan: serf: EventMemberJoin: Node-feed0447-e846-ed2b-d5b0-7866e18fc4fb 127.0.0.1 === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:00.595Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29940 network=udp writer.go:29: 2021-01-29T19:32:00.596Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29941 network=tcp === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:00.599Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: Started DNS server: address=127.0.0.1:29934 network=udp writer.go:29: 2021-01-29T19:32:00.600Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: Started DNS server: address=127.0.0.1:29934 network=tcp writer.go:29: 2021-01-29T19:32:00.608Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: Started HTTP server: address=127.0.0.1:29935 network=tcp writer.go:29: 2021-01-29T19:32:00.610Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: started state syncer writer.go:29: 2021-01-29T19:32:00.610Z [WARN] TestAgent_PurgeServiceOnDuplicate/normal.client.manager: No servers available writer.go:29: 2021-01-29T19:32:00.613Z [ERROR] TestAgent_PurgeServiceOnDuplicate/normal.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:32:00.615Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: Requesting shutdown writer.go:29: 2021-01-29T19:32:00.616Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal.client: shutting down client writer.go:29: 2021-01-29T19:32:00.617Z [WARN] TestAgent_PurgeServiceOnDuplicate/normal.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:00.618Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal.client.manager: shutting down writer.go:29: 2021-01-29T19:32:00.620Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: consul client down writer.go:29: 2021-01-29T19:32:00.621Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: shutdown complete writer.go:29: 2021-01-29T19:32:00.623Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: Stopping server: protocol=DNS address=127.0.0.1:29934 network=tcp writer.go:29: 2021-01-29T19:32:00.626Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: Stopping server: protocol=DNS address=127.0.0.1:29934 network=udp writer.go:29: 2021-01-29T19:32:00.628Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: Stopping server: protocol=HTTP address=127.0.0.1:29935 network=tcp === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.097Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:01.099Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager: Endpoints down writer.go:29: 2021-01-29T19:32:01.112Z [DEBUG] TestAgent_PurgeServiceOnDuplicate/service_manager-a2.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:01.116Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2.client.serf.lan: serf: EventMemberJoin: Node-0c306f3b-eb8a-b5d4-11aa-f823aaa80f22 127.0.0.1 writer.go:29: 2021-01-29T19:32:01.119Z [DEBUG] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: service exists, not restoring from file: service=redis file=/tmp/consul-test/TestAgent_PurgeServiceOnDuplicate_service_manager-agent020158011/services/86a1b907d54bf7010394bf316e183e67 writer.go:29: 2021-01-29T19:32:01.121Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2.client.serf.lan: serf: Attempting re-join to previously known node: Node-df6b6fd8-b773-4d91-34fd-b1d36f77fa83: 127.0.0.1:29943 writer.go:29: 2021-01-29T19:32:01.126Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: Started DNS server: address=127.0.0.1:29946 network=udp === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.129Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:01.130Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal: Endpoints down === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.130Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: Started DNS server: address=127.0.0.1:29946 network=tcp === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.146Z [DEBUG] TestAgent_PurgeServiceOnDuplicate/normal-a2.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:01.151Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2.client.serf.lan: serf: EventMemberJoin: Node-d9bec1e6-2674-63c1-4d57-9d4b768145b2 127.0.0.1 === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.126Z [DEBUG] TestAgent_PurgeServiceOnDuplicate/service_manager-a2.client.memberlist.lan: memberlist: Failed to join 127.0.0.1: dial tcp 127.0.0.1:29943: connect: connection refused === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.155Z [DEBUG] TestAgent_PurgeServiceOnDuplicate/normal-a2: service exists, not restoring from file: service=redis file=/tmp/consul-test/TestAgent_PurgeServiceOnDuplicate_normal-agent419206476/services/86a1b907d54bf7010394bf316e183e67 writer.go:29: 2021-01-29T19:32:01.155Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2.client.serf.lan: serf: Attempting re-join to previously known node: Node-feed0447-e846-ed2b-d5b0-7866e18fc4fb: 127.0.0.1:29937 === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.162Z [WARN] TestAgent_PurgeServiceOnDuplicate/service_manager-a2.client.serf.lan: serf: Failed to re-join any previously known node writer.go:29: 2021-01-29T19:32:01.135Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: Started HTTP server: address=127.0.0.1:29947 network=tcp === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.163Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: Started DNS server: address=127.0.0.1:29952 network=udp === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.165Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: started state syncer === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.164Z [DEBUG] TestAgent_PurgeServiceOnDuplicate/normal-a2.client.memberlist.lan: memberlist: Failed to join 127.0.0.1: dial tcp 127.0.0.1:29937: connect: connection refused writer.go:29: 2021-01-29T19:32:01.167Z [WARN] TestAgent_PurgeServiceOnDuplicate/normal-a2.client.serf.lan: serf: Failed to re-join any previously known node === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.168Z [WARN] TestAgent_PurgeServiceOnDuplicate/service_manager-a2.client.manager: No servers available === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.166Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: Started DNS server: address=127.0.0.1:29952 network=tcp writer.go:29: 2021-01-29T19:32:01.174Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: Started HTTP server: address=127.0.0.1:29953 network=tcp writer.go:29: 2021-01-29T19:32:01.176Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: started state syncer writer.go:29: 2021-01-29T19:32:01.177Z [WARN] TestAgent_PurgeServiceOnDuplicate/normal-a2.client.manager: No servers available === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.171Z [ERROR] TestAgent_PurgeServiceOnDuplicate/service_manager-a2.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:32:01.177Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: Requesting shutdown === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.179Z [ERROR] TestAgent_PurgeServiceOnDuplicate/normal-a2.anti_entropy: failed to sync remote state: error="No known Consul servers" === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.192Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2.client: shutting down client writer.go:29: 2021-01-29T19:32:01.193Z [WARN] TestAgent_PurgeServiceOnDuplicate/service_manager-a2.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:01.193Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2.client.manager: shutting down writer.go:29: 2021-01-29T19:32:01.197Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: consul client down writer.go:29: 2021-01-29T19:32:01.199Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: shutdown complete === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.199Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: Requesting shutdown === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.200Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: Stopping server: protocol=DNS address=127.0.0.1:29946 network=tcp === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.200Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2.client: shutting down client writer.go:29: 2021-01-29T19:32:01.201Z [WARN] TestAgent_PurgeServiceOnDuplicate/normal-a2.client.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.201Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: Stopping server: protocol=DNS address=127.0.0.1:29946 network=udp writer.go:29: 2021-01-29T19:32:01.203Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: Stopping server: protocol=HTTP address=127.0.0.1:29947 network=tcp === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.201Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2.client.manager: shutting down writer.go:29: 2021-01-29T19:32:01.203Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: consul client down writer.go:29: 2021-01-29T19:32:01.206Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: shutdown complete writer.go:29: 2021-01-29T19:32:01.208Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: Stopping server: protocol=DNS address=127.0.0.1:29952 network=tcp writer.go:29: 2021-01-29T19:32:01.210Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: Stopping server: protocol=DNS address=127.0.0.1:29952 network=udp writer.go:29: 2021-01-29T19:32:01.211Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: Stopping server: protocol=HTTP address=127.0.0.1:29953 network=tcp === CONT TestAgent_PurgeServiceOnDuplicate/service_manager writer.go:29: 2021-01-29T19:32:01.704Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:01.705Z [INFO] TestAgent_PurgeServiceOnDuplicate/service_manager-a2: Endpoints down === CONT TestAgent_PurgeServiceOnDuplicate/normal writer.go:29: 2021-01-29T19:32:01.712Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:01.712Z [INFO] TestAgent_PurgeServiceOnDuplicate/normal-a2: Endpoints down --- PASS: TestAgent_PurgeServiceOnDuplicate (0.00s) --- PASS: TestAgent_PurgeServiceOnDuplicate/service_manager (1.17s) --- PASS: TestAgent_PurgeServiceOnDuplicate/normal (1.18s) === RUN TestAgent_PersistCheck === PAUSE TestAgent_PersistCheck === RUN TestAgent_PurgeCheck agent_test.go:2236: DM-skipped --- SKIP: TestAgent_PurgeCheck (0.00s) === RUN TestAgent_PurgeCheckOnDuplicate === PAUSE TestAgent_PurgeCheckOnDuplicate === RUN TestAgent_loadChecks_token === PAUSE TestAgent_loadChecks_token === RUN TestAgent_unloadChecks === PAUSE TestAgent_unloadChecks === RUN TestAgent_loadServices_token === RUN TestAgent_loadServices_token/normal === PAUSE TestAgent_loadServices_token/normal === RUN TestAgent_loadServices_token/service_manager === PAUSE TestAgent_loadServices_token/service_manager === CONT TestAgent_loadServices_token/normal === CONT TestAgent_loadServices_token/service_manager writer.go:29: 2021-01-29T19:32:01.754Z [WARN] TestAgent_loadServices_token/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:01.756Z [DEBUG] TestAgent_loadServices_token/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:01.759Z [DEBUG] TestAgent_loadServices_token/service_manager.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:01.824Z [INFO] TestAgent_loadServices_token/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b28e793d-a7ed-a895-dad5-9f1139223a3f Address:127.0.0.1:29963}]" writer.go:29: 2021-01-29T19:32:01.827Z [INFO] TestAgent_loadServices_token/service_manager.server.serf.wan: serf: EventMemberJoin: Node-b28e793d-a7ed-a895-dad5-9f1139223a3f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:01.830Z [INFO] TestAgent_loadServices_token/service_manager.server.serf.lan: serf: EventMemberJoin: Node-b28e793d-a7ed-a895-dad5-9f1139223a3f 127.0.0.1 writer.go:29: 2021-01-29T19:32:01.837Z [INFO] TestAgent_loadServices_token/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29963 [Follower]" leader= writer.go:29: 2021-01-29T19:32:01.857Z [INFO] TestAgent_loadServices_token/service_manager: Started DNS server: address=127.0.0.1:29958 network=udp writer.go:29: 2021-01-29T19:32:01.858Z [INFO] TestAgent_loadServices_token/service_manager.server: Adding LAN server: server="Node-b28e793d-a7ed-a895-dad5-9f1139223a3f (Addr: tcp/127.0.0.1:29963) (DC: dc1)" writer.go:29: 2021-01-29T19:32:01.858Z [INFO] TestAgent_loadServices_token/service_manager.server: Handled event for server in area: event=member-join server=Node-b28e793d-a7ed-a895-dad5-9f1139223a3f.dc1 area=wan === CONT TestAgent_loadServices_token/normal writer.go:29: 2021-01-29T19:32:01.862Z [WARN] TestAgent_loadServices_token/normal: bootstrap = true: do not enable unless necessary === CONT TestAgent_loadServices_token/service_manager writer.go:29: 2021-01-29T19:32:01.862Z [INFO] TestAgent_loadServices_token/service_manager: Started DNS server: address=127.0.0.1:29958 network=tcp === CONT TestAgent_loadServices_token/normal writer.go:29: 2021-01-29T19:32:01.863Z [DEBUG] TestAgent_loadServices_token/normal.tlsutil: Update: version=1 === CONT TestAgent_loadServices_token/service_manager writer.go:29: 2021-01-29T19:32:01.864Z [INFO] TestAgent_loadServices_token/service_manager: Started HTTP server: address=127.0.0.1:29959 network=tcp writer.go:29: 2021-01-29T19:32:01.865Z [INFO] TestAgent_loadServices_token/service_manager: started state syncer === CONT TestAgent_loadServices_token/normal writer.go:29: 2021-01-29T19:32:01.864Z [DEBUG] TestAgent_loadServices_token/normal.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:01.885Z [INFO] TestAgent_loadServices_token/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c98cb043-56bc-35cb-28eb-6b73f5fbaf76 Address:127.0.0.1:29969}]" writer.go:29: 2021-01-29T19:32:01.886Z [INFO] TestAgent_loadServices_token/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29969 [Follower]" leader= writer.go:29: 2021-01-29T19:32:01.887Z [INFO] TestAgent_loadServices_token/normal.server.serf.wan: serf: EventMemberJoin: Node-c98cb043-56bc-35cb-28eb-6b73f5fbaf76.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:01.890Z [INFO] TestAgent_loadServices_token/normal.server.serf.lan: serf: EventMemberJoin: Node-c98cb043-56bc-35cb-28eb-6b73f5fbaf76 127.0.0.1 writer.go:29: 2021-01-29T19:32:01.892Z [INFO] TestAgent_loadServices_token/normal.server: Handled event for server in area: event=member-join server=Node-c98cb043-56bc-35cb-28eb-6b73f5fbaf76.dc1 area=wan writer.go:29: 2021-01-29T19:32:01.893Z [INFO] TestAgent_loadServices_token/normal.server: Adding LAN server: server="Node-c98cb043-56bc-35cb-28eb-6b73f5fbaf76 (Addr: tcp/127.0.0.1:29969) (DC: dc1)" writer.go:29: 2021-01-29T19:32:01.894Z [INFO] TestAgent_loadServices_token/normal: Started DNS server: address=127.0.0.1:29964 network=tcp writer.go:29: 2021-01-29T19:32:01.895Z [INFO] TestAgent_loadServices_token/normal: Started DNS server: address=127.0.0.1:29964 network=udp writer.go:29: 2021-01-29T19:32:01.898Z [INFO] TestAgent_loadServices_token/normal: Started HTTP server: address=127.0.0.1:29965 network=tcp writer.go:29: 2021-01-29T19:32:01.899Z [INFO] TestAgent_loadServices_token/normal: started state syncer === CONT TestAgent_loadServices_token/service_manager writer.go:29: 2021-01-29T19:32:01.902Z [WARN] TestAgent_loadServices_token/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:01.903Z [INFO] TestAgent_loadServices_token/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29963 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:01.906Z [DEBUG] TestAgent_loadServices_token/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:01.907Z [DEBUG] TestAgent_loadServices_token/service_manager.server.raft: vote granted: from=b28e793d-a7ed-a895-dad5-9f1139223a3f term=2 tally=1 writer.go:29: 2021-01-29T19:32:01.909Z [INFO] TestAgent_loadServices_token/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:01.910Z [INFO] TestAgent_loadServices_token/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29963 [Leader]" writer.go:29: 2021-01-29T19:32:01.913Z [INFO] TestAgent_loadServices_token/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:01.913Z [INFO] TestAgent_loadServices_token/service_manager.server: New leader elected: payload=Node-b28e793d-a7ed-a895-dad5-9f1139223a3f writer.go:29: 2021-01-29T19:32:01.916Z [DEBUG] TestAgent_loadServices_token/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29963 writer.go:29: 2021-01-29T19:32:01.921Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:01.926Z [INFO] TestAgent_loadServices_token/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:01.927Z [INFO] TestAgent_loadServices_token/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:01.928Z [DEBUG] TestAgent_loadServices_token/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-b28e793d-a7ed-a895-dad5-9f1139223a3f writer.go:29: 2021-01-29T19:32:01.929Z [INFO] TestAgent_loadServices_token/service_manager.server: member joined, marking health alive: member=Node-b28e793d-a7ed-a895-dad5-9f1139223a3f === CONT TestAgent_loadServices_token/normal writer.go:29: 2021-01-29T19:32:01.943Z [WARN] TestAgent_loadServices_token/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:01.944Z [INFO] TestAgent_loadServices_token/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29969 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:01.946Z [DEBUG] TestAgent_loadServices_token/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:01.947Z [DEBUG] TestAgent_loadServices_token/normal.server.raft: vote granted: from=c98cb043-56bc-35cb-28eb-6b73f5fbaf76 term=2 tally=1 writer.go:29: 2021-01-29T19:32:01.947Z [INFO] TestAgent_loadServices_token/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:01.948Z [INFO] TestAgent_loadServices_token/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29969 [Leader]" writer.go:29: 2021-01-29T19:32:01.949Z [INFO] TestAgent_loadServices_token/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:01.949Z [INFO] TestAgent_loadServices_token/normal.server: New leader elected: payload=Node-c98cb043-56bc-35cb-28eb-6b73f5fbaf76 writer.go:29: 2021-01-29T19:32:01.952Z [DEBUG] TestAgent_loadServices_token/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29969 writer.go:29: 2021-01-29T19:32:01.957Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:01.962Z [INFO] TestAgent_loadServices_token/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:01.963Z [INFO] TestAgent_loadServices_token/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:01.964Z [DEBUG] TestAgent_loadServices_token/normal.server: Skipping self join check for node since the cluster is too small: node=Node-c98cb043-56bc-35cb-28eb-6b73f5fbaf76 writer.go:29: 2021-01-29T19:32:01.965Z [INFO] TestAgent_loadServices_token/normal.server: member joined, marking health alive: member=Node-c98cb043-56bc-35cb-28eb-6b73f5fbaf76 writer.go:29: 2021-01-29T19:32:02.015Z [INFO] TestAgent_loadServices_token/normal: Requesting shutdown writer.go:29: 2021-01-29T19:32:02.019Z [INFO] TestAgent_loadServices_token/normal.server: shutting down server writer.go:29: 2021-01-29T19:32:02.020Z [DEBUG] TestAgent_loadServices_token/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:02.021Z [WARN] TestAgent_loadServices_token/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:02.020Z [ERROR] TestAgent_loadServices_token/normal.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:02.021Z [DEBUG] TestAgent_loadServices_token/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:02.022Z [WARN] TestAgent_loadServices_token/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:02.031Z [INFO] TestAgent_loadServices_token/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:02.032Z [INFO] TestAgent_loadServices_token/normal: consul server down writer.go:29: 2021-01-29T19:32:02.033Z [INFO] TestAgent_loadServices_token/normal: shutdown complete writer.go:29: 2021-01-29T19:32:02.033Z [INFO] TestAgent_loadServices_token/normal: Stopping server: protocol=DNS address=127.0.0.1:29964 network=tcp writer.go:29: 2021-01-29T19:32:02.034Z [INFO] TestAgent_loadServices_token/normal: Stopping server: protocol=DNS address=127.0.0.1:29964 network=udp writer.go:29: 2021-01-29T19:32:02.035Z [INFO] TestAgent_loadServices_token/normal: Stopping server: protocol=HTTP address=127.0.0.1:29965 network=tcp === CONT TestAgent_loadServices_token/service_manager writer.go:29: 2021-01-29T19:32:02.122Z [DEBUG] TestAgent_loadServices_token/service_manager: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:02.124Z [INFO] TestAgent_loadServices_token/service_manager: Synced node info writer.go:29: 2021-01-29T19:32:02.126Z [INFO] TestAgent_loadServices_token/service_manager: Synced service: service=rabbitmq writer.go:29: 2021-01-29T19:32:02.127Z [DEBUG] TestAgent_loadServices_token/service_manager: Node info in sync writer.go:29: 2021-01-29T19:32:02.128Z [DEBUG] TestAgent_loadServices_token/service_manager: Service in sync: service=rabbitmq writer.go:29: 2021-01-29T19:32:02.214Z [INFO] TestAgent_loadServices_token/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:32:02.215Z [INFO] TestAgent_loadServices_token/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:32:02.216Z [DEBUG] TestAgent_loadServices_token/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:02.217Z [WARN] TestAgent_loadServices_token/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:02.218Z [DEBUG] TestAgent_loadServices_token/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:02.220Z [WARN] TestAgent_loadServices_token/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:02.223Z [INFO] TestAgent_loadServices_token/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:02.223Z [INFO] TestAgent_loadServices_token/service_manager: consul server down writer.go:29: 2021-01-29T19:32:02.224Z [INFO] TestAgent_loadServices_token/service_manager: shutdown complete writer.go:29: 2021-01-29T19:32:02.225Z [INFO] TestAgent_loadServices_token/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29958 network=tcp writer.go:29: 2021-01-29T19:32:02.226Z [INFO] TestAgent_loadServices_token/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29958 network=udp writer.go:29: 2021-01-29T19:32:02.227Z [INFO] TestAgent_loadServices_token/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29959 network=tcp === CONT TestAgent_loadServices_token/normal writer.go:29: 2021-01-29T19:32:02.536Z [INFO] TestAgent_loadServices_token/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:02.538Z [INFO] TestAgent_loadServices_token/normal: Endpoints down === CONT TestAgent_loadServices_token/service_manager writer.go:29: 2021-01-29T19:32:02.728Z [INFO] TestAgent_loadServices_token/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:02.729Z [INFO] TestAgent_loadServices_token/service_manager: Endpoints down --- PASS: TestAgent_loadServices_token (0.00s) --- PASS: TestAgent_loadServices_token/normal (0.82s) --- PASS: TestAgent_loadServices_token/service_manager (1.01s) === RUN TestAgent_loadServices_sidecar === RUN TestAgent_loadServices_sidecar/normal === PAUSE TestAgent_loadServices_sidecar/normal === RUN TestAgent_loadServices_sidecar/service_manager === PAUSE TestAgent_loadServices_sidecar/service_manager === CONT TestAgent_loadServices_sidecar/normal === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.748Z [WARN] TestAgent_loadServices_sidecar/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:02.750Z [DEBUG] TestAgent_loadServices_sidecar/service_manager.tlsutil: Update: version=1 === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.752Z [WARN] TestAgent_loadServices_sidecar/normal: bootstrap = true: do not enable unless necessary === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.752Z [DEBUG] TestAgent_loadServices_sidecar/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.754Z [DEBUG] TestAgent_loadServices_sidecar/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:02.757Z [DEBUG] TestAgent_loadServices_sidecar/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.760Z [INFO] TestAgent_loadServices_sidecar/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:14dfb91c-f043-03c4-2609-8b32c9a1dcce Address:127.0.0.1:29981}]" writer.go:29: 2021-01-29T19:32:02.762Z [INFO] TestAgent_loadServices_sidecar/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29981 [Follower]" leader= === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.765Z [INFO] TestAgent_loadServices_sidecar/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ee91016b-e868-36bc-15e1-92a224e68d9a Address:127.0.0.1:29975}]" writer.go:29: 2021-01-29T19:32:02.766Z [INFO] TestAgent_loadServices_sidecar/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29975 [Follower]" leader= writer.go:29: 2021-01-29T19:32:02.767Z [INFO] TestAgent_loadServices_sidecar/normal.server.serf.wan: serf: EventMemberJoin: Node-ee91016b-e868-36bc-15e1-92a224e68d9a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:02.770Z [INFO] TestAgent_loadServices_sidecar/normal.server.serf.lan: serf: EventMemberJoin: Node-ee91016b-e868-36bc-15e1-92a224e68d9a 127.0.0.1 === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.765Z [INFO] TestAgent_loadServices_sidecar/service_manager.server.serf.wan: serf: EventMemberJoin: Node-14dfb91c-f043-03c4-2609-8b32c9a1dcce.dc1 127.0.0.1 === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.778Z [INFO] TestAgent_loadServices_sidecar/normal: Started DNS server: address=127.0.0.1:29970 network=udp writer.go:29: 2021-01-29T19:32:02.785Z [INFO] TestAgent_loadServices_sidecar/normal: Started DNS server: address=127.0.0.1:29970 network=tcp === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.786Z [INFO] TestAgent_loadServices_sidecar/service_manager.server.serf.lan: serf: EventMemberJoin: Node-14dfb91c-f043-03c4-2609-8b32c9a1dcce 127.0.0.1 === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.779Z [INFO] TestAgent_loadServices_sidecar/normal.server: Adding LAN server: server="Node-ee91016b-e868-36bc-15e1-92a224e68d9a (Addr: tcp/127.0.0.1:29975) (DC: dc1)" writer.go:29: 2021-01-29T19:32:02.779Z [INFO] TestAgent_loadServices_sidecar/normal.server: Handled event for server in area: event=member-join server=Node-ee91016b-e868-36bc-15e1-92a224e68d9a.dc1 area=wan writer.go:29: 2021-01-29T19:32:02.793Z [INFO] TestAgent_loadServices_sidecar/normal: Started HTTP server: address=127.0.0.1:29971 network=tcp writer.go:29: 2021-01-29T19:32:02.799Z [INFO] TestAgent_loadServices_sidecar/normal: started state syncer writer.go:29: 2021-01-29T19:32:02.805Z [WARN] TestAgent_loadServices_sidecar/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:02.806Z [INFO] TestAgent_loadServices_sidecar/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29975 [Candidate]" term=2 === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.796Z [INFO] TestAgent_loadServices_sidecar/service_manager.server: Handled event for server in area: event=member-join server=Node-14dfb91c-f043-03c4-2609-8b32c9a1dcce.dc1 area=wan === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.809Z [DEBUG] TestAgent_loadServices_sidecar/normal.server.raft: votes: needed=1 === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.797Z [DEBUG] TestAgent_loadServices_sidecar/service_manager: added local registration for service: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:02.799Z [INFO] TestAgent_loadServices_sidecar/service_manager.server: Adding LAN server: server="Node-14dfb91c-f043-03c4-2609-8b32c9a1dcce (Addr: tcp/127.0.0.1:29981) (DC: dc1)" === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.811Z [DEBUG] TestAgent_loadServices_sidecar/normal.server.raft: vote granted: from=ee91016b-e868-36bc-15e1-92a224e68d9a term=2 tally=1 writer.go:29: 2021-01-29T19:32:02.816Z [INFO] TestAgent_loadServices_sidecar/normal.server.raft: election won: tally=1 === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.820Z [INFO] TestAgent_loadServices_sidecar/service_manager: Started DNS server: address=127.0.0.1:29976 network=udp writer.go:29: 2021-01-29T19:32:02.821Z [INFO] TestAgent_loadServices_sidecar/service_manager: Started DNS server: address=127.0.0.1:29976 network=tcp === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.821Z [INFO] TestAgent_loadServices_sidecar/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29975 [Leader]" writer.go:29: 2021-01-29T19:32:02.824Z [INFO] TestAgent_loadServices_sidecar/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:02.825Z [INFO] TestAgent_loadServices_sidecar/normal.server: New leader elected: payload=Node-ee91016b-e868-36bc-15e1-92a224e68d9a === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.822Z [INFO] TestAgent_loadServices_sidecar/service_manager: Started HTTP server: address=127.0.0.1:29977 network=tcp writer.go:29: 2021-01-29T19:32:02.827Z [INFO] TestAgent_loadServices_sidecar/service_manager: started state syncer === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.827Z [DEBUG] TestAgent_loadServices_sidecar/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29975 === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.832Z [WARN] TestAgent_loadServices_sidecar/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:02.833Z [INFO] TestAgent_loadServices_sidecar/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29981 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:02.835Z [DEBUG] TestAgent_loadServices_sidecar/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:02.835Z [DEBUG] TestAgent_loadServices_sidecar/service_manager.server.raft: vote granted: from=14dfb91c-f043-03c4-2609-8b32c9a1dcce term=2 tally=1 writer.go:29: 2021-01-29T19:32:02.836Z [INFO] TestAgent_loadServices_sidecar/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:02.837Z [INFO] TestAgent_loadServices_sidecar/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29981 [Leader]" === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.837Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.838Z [INFO] TestAgent_loadServices_sidecar/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:02.840Z [INFO] TestAgent_loadServices_sidecar/service_manager.server: New leader elected: payload=Node-14dfb91c-f043-03c4-2609-8b32c9a1dcce writer.go:29: 2021-01-29T19:32:02.847Z [DEBUG] TestAgent_loadServices_sidecar/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29981 === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.847Z [INFO] TestAgent_loadServices_sidecar/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:02.850Z [INFO] TestAgent_loadServices_sidecar/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:02.853Z [DEBUG] TestAgent_loadServices_sidecar/normal.server: Skipping self join check for node since the cluster is too small: node=Node-ee91016b-e868-36bc-15e1-92a224e68d9a === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.862Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.857Z [INFO] TestAgent_loadServices_sidecar/normal.server: member joined, marking health alive: member=Node-ee91016b-e868-36bc-15e1-92a224e68d9a === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:02.870Z [INFO] TestAgent_loadServices_sidecar/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:02.871Z [INFO] TestAgent_loadServices_sidecar/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:02.872Z [DEBUG] TestAgent_loadServices_sidecar/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-14dfb91c-f043-03c4-2609-8b32c9a1dcce writer.go:29: 2021-01-29T19:32:02.873Z [INFO] TestAgent_loadServices_sidecar/service_manager.server: member joined, marking health alive: member=Node-14dfb91c-f043-03c4-2609-8b32c9a1dcce === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:02.963Z [DEBUG] TestAgent_loadServices_sidecar/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:02.964Z [INFO] TestAgent_loadServices_sidecar/normal: Synced node info writer.go:29: 2021-01-29T19:32:02.966Z [INFO] TestAgent_loadServices_sidecar/normal: Synced service: service=rabbitmq writer.go:29: 2021-01-29T19:32:02.968Z [INFO] TestAgent_loadServices_sidecar/normal: Synced service: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:02.969Z [DEBUG] TestAgent_loadServices_sidecar/normal: Check in sync: check=service:rabbitmq-sidecar-proxy:1 writer.go:29: 2021-01-29T19:32:02.970Z [DEBUG] TestAgent_loadServices_sidecar/normal: Check in sync: check=service:rabbitmq-sidecar-proxy:2 === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:03.040Z [DEBUG] TestAgent_loadServices_sidecar/service_manager: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:03.041Z [INFO] TestAgent_loadServices_sidecar/service_manager: Synced node info writer.go:29: 2021-01-29T19:32:03.042Z [INFO] TestAgent_loadServices_sidecar/service_manager: Synced service: service=rabbitmq writer.go:29: 2021-01-29T19:32:03.044Z [INFO] TestAgent_loadServices_sidecar/service_manager: Synced service: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:03.045Z [DEBUG] TestAgent_loadServices_sidecar/service_manager: Check in sync: check=service:rabbitmq-sidecar-proxy:1 writer.go:29: 2021-01-29T19:32:03.046Z [DEBUG] TestAgent_loadServices_sidecar/service_manager: Check in sync: check=service:rabbitmq-sidecar-proxy:2 writer.go:29: 2021-01-29T19:32:03.047Z [DEBUG] TestAgent_loadServices_sidecar/service_manager: Node info in sync writer.go:29: 2021-01-29T19:32:03.048Z [DEBUG] TestAgent_loadServices_sidecar/service_manager: Service in sync: service=rabbitmq writer.go:29: 2021-01-29T19:32:03.049Z [DEBUG] TestAgent_loadServices_sidecar/service_manager: Service in sync: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:03.050Z [DEBUG] TestAgent_loadServices_sidecar/service_manager: Check in sync: check=service:rabbitmq-sidecar-proxy:1 writer.go:29: 2021-01-29T19:32:03.052Z [DEBUG] TestAgent_loadServices_sidecar/service_manager: Check in sync: check=service:rabbitmq-sidecar-proxy:2 writer.go:29: 2021-01-29T19:32:03.124Z [INFO] TestAgent_loadServices_sidecar/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:32:03.125Z [INFO] TestAgent_loadServices_sidecar/service_manager.server: shutting down server === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:03.129Z [INFO] TestAgent_loadServices_sidecar/normal: Requesting shutdown writer.go:29: 2021-01-29T19:32:03.130Z [INFO] TestAgent_loadServices_sidecar/normal.server: shutting down server writer.go:29: 2021-01-29T19:32:03.131Z [DEBUG] TestAgent_loadServices_sidecar/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:03.132Z [WARN] TestAgent_loadServices_sidecar/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:03.133Z [DEBUG] TestAgent_loadServices_sidecar/normal.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:03.126Z [DEBUG] TestAgent_loadServices_sidecar/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:03.134Z [DEBUG] TestAgent_loadServices_sidecar/service_manager.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:03.135Z [WARN] TestAgent_loadServices_sidecar/normal.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:03.135Z [WARN] TestAgent_loadServices_sidecar/service_manager.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:03.136Z [INFO] TestAgent_loadServices_sidecar/normal.server.router.manager: shutting down === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:03.137Z [WARN] TestAgent_loadServices_sidecar/service_manager.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:03.137Z [INFO] TestAgent_loadServices_sidecar/normal: consul server down writer.go:29: 2021-01-29T19:32:03.138Z [INFO] TestAgent_loadServices_sidecar/normal: shutdown complete writer.go:29: 2021-01-29T19:32:03.139Z [INFO] TestAgent_loadServices_sidecar/normal: Stopping server: protocol=DNS address=127.0.0.1:29970 network=tcp === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:03.139Z [INFO] TestAgent_loadServices_sidecar/service_manager.server.router.manager: shutting down === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:03.140Z [INFO] TestAgent_loadServices_sidecar/normal: Stopping server: protocol=DNS address=127.0.0.1:29970 network=udp === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:03.140Z [INFO] TestAgent_loadServices_sidecar/service_manager: consul server down writer.go:29: 2021-01-29T19:32:03.141Z [INFO] TestAgent_loadServices_sidecar/service_manager: shutdown complete writer.go:29: 2021-01-29T19:32:03.142Z [INFO] TestAgent_loadServices_sidecar/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29976 network=tcp === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:03.140Z [INFO] TestAgent_loadServices_sidecar/normal: Stopping server: protocol=HTTP address=127.0.0.1:29971 network=tcp === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:03.143Z [INFO] TestAgent_loadServices_sidecar/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29976 network=udp writer.go:29: 2021-01-29T19:32:03.144Z [INFO] TestAgent_loadServices_sidecar/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29977 network=tcp === CONT TestAgent_loadServices_sidecar/normal writer.go:29: 2021-01-29T19:32:03.643Z [INFO] TestAgent_loadServices_sidecar/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:03.644Z [INFO] TestAgent_loadServices_sidecar/normal: Endpoints down === CONT TestAgent_loadServices_sidecar/service_manager writer.go:29: 2021-01-29T19:32:03.645Z [INFO] TestAgent_loadServices_sidecar/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:03.646Z [INFO] TestAgent_loadServices_sidecar/service_manager: Endpoints down --- PASS: TestAgent_loadServices_sidecar (0.00s) --- PASS: TestAgent_loadServices_sidecar/normal (0.91s) --- PASS: TestAgent_loadServices_sidecar/service_manager (0.91s) === RUN TestAgent_loadServices_sidecarSeparateToken === RUN TestAgent_loadServices_sidecarSeparateToken/normal === PAUSE TestAgent_loadServices_sidecarSeparateToken/normal === RUN TestAgent_loadServices_sidecarSeparateToken/service_manager === PAUSE TestAgent_loadServices_sidecarSeparateToken/service_manager === CONT TestAgent_loadServices_sidecarSeparateToken/normal === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.661Z [WARN] TestAgent_loadServices_sidecarSeparateToken/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:03.662Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:03.664Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/service_manager.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:03.681Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d51778fb-1cc5-d273-3bfa-3bddb5c081ec Address:127.0.0.1:29987}]" writer.go:29: 2021-01-29T19:32:03.685Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.serf.wan: serf: EventMemberJoin: Node-d51778fb-1cc5-d273-3bfa-3bddb5c081ec.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:03.689Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.serf.lan: serf: EventMemberJoin: Node-d51778fb-1cc5-d273-3bfa-3bddb5c081ec 127.0.0.1 writer.go:29: 2021-01-29T19:32:03.692Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/service_manager: added local registration for service: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:03.696Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: Started DNS server: address=127.0.0.1:29982 network=udp writer.go:29: 2021-01-29T19:32:03.699Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:29987 [Follower]" leader= writer.go:29: 2021-01-29T19:32:03.704Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server: Adding LAN server: server="Node-d51778fb-1cc5-d273-3bfa-3bddb5c081ec (Addr: tcp/127.0.0.1:29987) (DC: dc1)" writer.go:29: 2021-01-29T19:32:03.706Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server: Handled event for server in area: event=member-join server=Node-d51778fb-1cc5-d273-3bfa-3bddb5c081ec.dc1 area=wan writer.go:29: 2021-01-29T19:32:03.711Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: Started DNS server: address=127.0.0.1:29982 network=tcp writer.go:29: 2021-01-29T19:32:03.717Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: Started HTTP server: address=127.0.0.1:29983 network=tcp writer.go:29: 2021-01-29T19:32:03.718Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: started state syncer writer.go:29: 2021-01-29T19:32:03.787Z [WARN] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:03.790Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:29987 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:03.796Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.raft: votes: needed=1 === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.793Z [WARN] TestAgent_loadServices_sidecarSeparateToken/normal: bootstrap = true: do not enable unless necessary === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.797Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.raft: vote granted: from=d51778fb-1cc5-d273-3bfa-3bddb5c081ec term=2 tally=1 writer.go:29: 2021-01-29T19:32:03.799Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:03.800Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:29987 [Leader]" === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.799Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal.tlsutil: Update: version=1 === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.802Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:03.803Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server: New leader elected: payload=Node-d51778fb-1cc5-d273-3bfa-3bddb5c081ec === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.841Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.842Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29987 writer.go:29: 2021-01-29T19:32:03.846Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: Requesting shutdown === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.846Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d169528c-e276-292c-0970-d10d8c4a784e Address:127.0.0.1:29993}]" === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.848Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:32:03.849Z [WARN] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.849Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29993 [Follower]" leader= === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.849Z [ERROR] TestAgent_loadServices_sidecarSeparateToken/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.851Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server.serf.wan: serf: EventMemberJoin: Node-d169528c-e276-292c-0970-d10d8c4a784e.dc1 127.0.0.1 === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.851Z [WARN] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.857Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server.serf.lan: serf: EventMemberJoin: Node-d169528c-e276-292c-0970-d10d8c4a784e 127.0.0.1 === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.858Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager.server.router.manager: shutting down === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.860Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server: Adding LAN server: server="Node-d169528c-e276-292c-0970-d10d8c4a784e (Addr: tcp/127.0.0.1:29993) (DC: dc1)" === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.869Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: consul server down === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.861Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server: Handled event for server in area: event=member-join server=Node-d169528c-e276-292c-0970-d10d8c4a784e.dc1 area=wan === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.891Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: shutdown complete writer.go:29: 2021-01-29T19:32:03.870Z [ERROR] TestAgent_loadServices_sidecarSeparateToken/service_manager.server: failed to establish leadership: error="error configuring provider: leadership lost while committing log" === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.871Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Started DNS server: address=127.0.0.1:29988 network=udp === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.892Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29982 network=tcp writer.go:29: 2021-01-29T19:32:03.894Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: Stopping server: protocol=DNS address=127.0.0.1:29982 network=udp writer.go:29: 2021-01-29T19:32:03.895Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:29983 network=tcp === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.894Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Started DNS server: address=127.0.0.1:29988 network=tcp === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:03.896Z [ERROR] TestAgent_loadServices_sidecarSeparateToken/service_manager.server: failed to transfer leadership attempt, will retry: attempt=0 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:32:03.897Z [ERROR] TestAgent_loadServices_sidecarSeparateToken/service_manager.server: failed to transfer leadership attempt, will retry: attempt=1 retry_limit=3 error="raft is already shutdown" === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:03.892Z [WARN] TestAgent_loadServices_sidecarSeparateToken/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:03.898Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29993 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:03.898Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Started HTTP server: address=127.0.0.1:29989 network=tcp writer.go:29: 2021-01-29T19:32:03.900Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: started state syncer writer.go:29: 2021-01-29T19:32:03.902Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:03.903Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal.server.raft: vote granted: from=d169528c-e276-292c-0970-d10d8c4a784e term=2 tally=1 writer.go:29: 2021-01-29T19:32:03.903Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:03.904Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29993 [Leader]" writer.go:29: 2021-01-29T19:32:03.905Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:03.906Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server: New leader elected: payload=Node-d169528c-e276-292c-0970-d10d8c4a784e writer.go:29: 2021-01-29T19:32:03.907Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29993 writer.go:29: 2021-01-29T19:32:03.910Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:03.915Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:03.916Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:03.918Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal.server: Skipping self join check for node since the cluster is too small: node=Node-d169528c-e276-292c-0970-d10d8c4a784e writer.go:29: 2021-01-29T19:32:03.919Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server: member joined, marking health alive: member=Node-d169528c-e276-292c-0970-d10d8c4a784e writer.go:29: 2021-01-29T19:32:03.929Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:03.931Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Synced node info writer.go:29: 2021-01-29T19:32:03.932Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Synced service: service=rabbitmq writer.go:29: 2021-01-29T19:32:03.935Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Synced service: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:03.936Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal: Check in sync: check=service:rabbitmq-sidecar-proxy:1 writer.go:29: 2021-01-29T19:32:03.937Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal: Check in sync: check=service:rabbitmq-sidecar-proxy:2 writer.go:29: 2021-01-29T19:32:04.284Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Requesting shutdown writer.go:29: 2021-01-29T19:32:04.287Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server: shutting down server writer.go:29: 2021-01-29T19:32:04.288Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:04.289Z [WARN] TestAgent_loadServices_sidecarSeparateToken/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:04.291Z [DEBUG] TestAgent_loadServices_sidecarSeparateToken/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:04.293Z [WARN] TestAgent_loadServices_sidecarSeparateToken/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:04.297Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:04.300Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: consul server down writer.go:29: 2021-01-29T19:32:04.301Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: shutdown complete writer.go:29: 2021-01-29T19:32:04.303Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Stopping server: protocol=DNS address=127.0.0.1:29988 network=tcp writer.go:29: 2021-01-29T19:32:04.304Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Stopping server: protocol=DNS address=127.0.0.1:29988 network=udp writer.go:29: 2021-01-29T19:32:04.306Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Stopping server: protocol=HTTP address=127.0.0.1:29989 network=tcp === CONT TestAgent_loadServices_sidecarSeparateToken/service_manager writer.go:29: 2021-01-29T19:32:04.396Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:04.398Z [INFO] TestAgent_loadServices_sidecarSeparateToken/service_manager: Endpoints down === CONT TestAgent_loadServices_sidecarSeparateToken/normal writer.go:29: 2021-01-29T19:32:04.807Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:04.809Z [INFO] TestAgent_loadServices_sidecarSeparateToken/normal: Endpoints down --- PASS: TestAgent_loadServices_sidecarSeparateToken (0.00s) --- PASS: TestAgent_loadServices_sidecarSeparateToken/service_manager (0.75s) --- PASS: TestAgent_loadServices_sidecarSeparateToken/normal (1.16s) === RUN TestAgent_loadServices_sidecarInheritMeta === RUN TestAgent_loadServices_sidecarInheritMeta/normal === PAUSE TestAgent_loadServices_sidecarInheritMeta/normal === RUN TestAgent_loadServices_sidecarInheritMeta/service_manager === PAUSE TestAgent_loadServices_sidecarInheritMeta/service_manager === CONT TestAgent_loadServices_sidecarInheritMeta/normal === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.839Z [WARN] TestAgent_loadServices_sidecarInheritMeta/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:04.840Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:04.842Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.844Z [WARN] TestAgent_loadServices_sidecarInheritMeta/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:04.846Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/normal.tlsutil: Update: version=1 === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.847Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:17beac53-875b-1e92-e601-22f67a83cb15 Address:127.0.0.1:30005}]" === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.848Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.850Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:30005 [Follower]" leader= === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.854Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5ff39722-10c5-bdbf-7572-af9c1a21004b Address:127.0.0.1:29999}]" === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.851Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.serf.wan: serf: EventMemberJoin: Node-17beac53-875b-1e92-e601-22f67a83cb15.dc1 127.0.0.1 === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.856Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server.serf.wan: serf: EventMemberJoin: Node-5ff39722-10c5-bdbf-7572-af9c1a21004b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:04.856Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:29999 [Follower]" leader= writer.go:29: 2021-01-29T19:32:04.858Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server.serf.lan: serf: EventMemberJoin: Node-5ff39722-10c5-bdbf-7572-af9c1a21004b 127.0.0.1 writer.go:29: 2021-01-29T19:32:04.862Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server: Adding LAN server: server="Node-5ff39722-10c5-bdbf-7572-af9c1a21004b (Addr: tcp/127.0.0.1:29999) (DC: dc1)" writer.go:29: 2021-01-29T19:32:04.862Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Started DNS server: address=127.0.0.1:29994 network=udp writer.go:29: 2021-01-29T19:32:04.869Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Started DNS server: address=127.0.0.1:29994 network=tcp writer.go:29: 2021-01-29T19:32:04.863Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server: Handled event for server in area: event=member-join server=Node-5ff39722-10c5-bdbf-7572-af9c1a21004b.dc1 area=wan === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.859Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.serf.lan: serf: EventMemberJoin: Node-17beac53-875b-1e92-e601-22f67a83cb15 127.0.0.1 === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.882Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Started HTTP server: address=127.0.0.1:29995 network=tcp writer.go:29: 2021-01-29T19:32:04.883Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: started state syncer === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.885Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server: Adding LAN server: server="Node-17beac53-875b-1e92-e601-22f67a83cb15 (Addr: tcp/127.0.0.1:30005) (DC: dc1)" writer.go:29: 2021-01-29T19:32:04.885Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server: Handled event for server in area: event=member-join server=Node-17beac53-875b-1e92-e601-22f67a83cb15.dc1 area=wan writer.go:29: 2021-01-29T19:32:04.888Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager: added local registration for service: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:04.893Z [WARN] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:04.894Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:30005 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:04.896Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:04.897Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.raft: vote granted: from=17beac53-875b-1e92-e601-22f67a83cb15 term=2 tally=1 writer.go:29: 2021-01-29T19:32:04.899Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.raft: election won: tally=1 === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.902Z [WARN] TestAgent_loadServices_sidecarInheritMeta/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:04.902Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:29999 [Candidate]" term=2 === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.897Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Started DNS server: address=127.0.0.1:30000 network=udp === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.904Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:04.905Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/normal.server.raft: vote granted: from=5ff39722-10c5-bdbf-7572-af9c1a21004b term=2 tally=1 === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.904Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Started DNS server: address=127.0.0.1:30000 network=tcp === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.906Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server.raft: election won: tally=1 === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.902Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:30005 [Leader]" === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.906Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:29999 [Leader]" === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.906Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Started HTTP server: address=127.0.0.1:30001 network=tcp === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.907Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server: cluster leadership acquired === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.908Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: started state syncer writer.go:29: 2021-01-29T19:32:04.907Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server: cluster leadership acquired === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.908Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server: New leader elected: payload=Node-5ff39722-10c5-bdbf-7572-af9c1a21004b === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.907Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server: New leader elected: payload=Node-17beac53-875b-1e92-e601-22f67a83cb15 writer.go:29: 2021-01-29T19:32:04.912Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30005 === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.909Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29999 === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.912Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Synced node info === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.912Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Synced node info === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.915Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Synced service: service=rabbitmq === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.917Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Synced service: service=rabbitmq === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.916Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.917Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.918Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Synced service: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:04.919Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager: Check in sync: check=service:rabbitmq-sidecar-proxy:1 writer.go:29: 2021-01-29T19:32:04.920Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager: Check in sync: check=service:rabbitmq-sidecar-proxy:2 === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.920Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Synced service: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:04.921Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/normal: Check in sync: check=service:rabbitmq-sidecar-proxy:2 === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.921Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager: Node info in sync === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.922Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/normal: Check in sync: check=service:rabbitmq-sidecar-proxy:1 writer.go:29: 2021-01-29T19:32:04.923Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:04.924Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.923Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager: Service in sync: service=rabbitmq === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.926Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/normal.server: Skipping self join check for node since the cluster is too small: node=Node-5ff39722-10c5-bdbf-7572-af9c1a21004b === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.927Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager: Service in sync: service=rabbitmq-sidecar-proxy === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:04.928Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server: member joined, marking health alive: member=Node-5ff39722-10c5-bdbf-7572-af9c1a21004b === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:04.929Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager: Check in sync: check=service:rabbitmq-sidecar-proxy:1 writer.go:29: 2021-01-29T19:32:04.931Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager: Check in sync: check=service:rabbitmq-sidecar-proxy:2 writer.go:29: 2021-01-29T19:32:04.930Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:04.935Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:04.939Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-17beac53-875b-1e92-e601-22f67a83cb15 writer.go:29: 2021-01-29T19:32:04.941Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server: member joined, marking health alive: member=Node-17beac53-875b-1e92-e601-22f67a83cb15 writer.go:29: 2021-01-29T19:32:04.943Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:32:04.945Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:32:04.946Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:04.947Z [WARN] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:04.948Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:04.949Z [WARN] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:04.952Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:04.952Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: consul server down writer.go:29: 2021-01-29T19:32:04.956Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: shutdown complete writer.go:29: 2021-01-29T19:32:04.957Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Stopping server: protocol=DNS address=127.0.0.1:30000 network=tcp writer.go:29: 2021-01-29T19:32:04.959Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Stopping server: protocol=DNS address=127.0.0.1:30000 network=udp writer.go:29: 2021-01-29T19:32:04.961Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:30001 network=tcp === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:05.015Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Requesting shutdown writer.go:29: 2021-01-29T19:32:05.018Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server: shutting down server writer.go:29: 2021-01-29T19:32:05.019Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.021Z [WARN] TestAgent_loadServices_sidecarInheritMeta/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:05.022Z [DEBUG] TestAgent_loadServices_sidecarInheritMeta/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.027Z [WARN] TestAgent_loadServices_sidecarInheritMeta/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:05.029Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:05.029Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: consul server down writer.go:29: 2021-01-29T19:32:05.031Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: shutdown complete writer.go:29: 2021-01-29T19:32:05.032Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Stopping server: protocol=DNS address=127.0.0.1:29994 network=tcp writer.go:29: 2021-01-29T19:32:05.033Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Stopping server: protocol=DNS address=127.0.0.1:29994 network=udp writer.go:29: 2021-01-29T19:32:05.033Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Stopping server: protocol=HTTP address=127.0.0.1:29995 network=tcp === CONT TestAgent_loadServices_sidecarInheritMeta/service_manager writer.go:29: 2021-01-29T19:32:05.462Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:05.464Z [INFO] TestAgent_loadServices_sidecarInheritMeta/service_manager: Endpoints down === CONT TestAgent_loadServices_sidecarInheritMeta/normal writer.go:29: 2021-01-29T19:32:05.534Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:05.536Z [INFO] TestAgent_loadServices_sidecarInheritMeta/normal: Endpoints down --- PASS: TestAgent_loadServices_sidecarInheritMeta (0.00s) --- PASS: TestAgent_loadServices_sidecarInheritMeta/service_manager (0.65s) --- PASS: TestAgent_loadServices_sidecarInheritMeta/normal (0.73s) === RUN TestAgent_loadServices_sidecarOverrideMeta === RUN TestAgent_loadServices_sidecarOverrideMeta/normal === PAUSE TestAgent_loadServices_sidecarOverrideMeta/normal === RUN TestAgent_loadServices_sidecarOverrideMeta/service_manager === PAUSE TestAgent_loadServices_sidecarOverrideMeta/service_manager === CONT TestAgent_loadServices_sidecarOverrideMeta/normal === CONT TestAgent_loadServices_sidecarOverrideMeta/service_manager writer.go:29: 2021-01-29T19:32:05.556Z [WARN] TestAgent_loadServices_sidecarOverrideMeta/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:05.561Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:05.566Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_loadServices_sidecarOverrideMeta/normal writer.go:29: 2021-01-29T19:32:05.574Z [WARN] TestAgent_loadServices_sidecarOverrideMeta/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:05.585Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:05.588Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/normal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_loadServices_sidecarOverrideMeta/service_manager writer.go:29: 2021-01-29T19:32:05.595Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:db0f6d8e-f12a-b207-9eef-36bbcc086d18 Address:127.0.0.1:30017}]" writer.go:29: 2021-01-29T19:32:05.602Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.serf.wan: serf: EventMemberJoin: Node-db0f6d8e-f12a-b207-9eef-36bbcc086d18.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:05.609Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.serf.lan: serf: EventMemberJoin: Node-db0f6d8e-f12a-b207-9eef-36bbcc086d18 127.0.0.1 writer.go:29: 2021-01-29T19:32:05.602Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:30017 [Follower]" leader= writer.go:29: 2021-01-29T19:32:05.616Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager: added local registration for service: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:05.618Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server: Adding LAN server: server="Node-db0f6d8e-f12a-b207-9eef-36bbcc086d18 (Addr: tcp/127.0.0.1:30017) (DC: dc1)" writer.go:29: 2021-01-29T19:32:05.618Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server: Handled event for server in area: event=member-join server=Node-db0f6d8e-f12a-b207-9eef-36bbcc086d18.dc1 area=wan writer.go:29: 2021-01-29T19:32:05.628Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Started DNS server: address=127.0.0.1:30012 network=tcp writer.go:29: 2021-01-29T19:32:05.633Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Started DNS server: address=127.0.0.1:30012 network=udp === CONT TestAgent_loadServices_sidecarOverrideMeta/normal writer.go:29: 2021-01-29T19:32:05.638Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4d8dcdf7-944d-5c6e-5574-5c358fff64a8 Address:127.0.0.1:30011}]" === CONT TestAgent_loadServices_sidecarOverrideMeta/service_manager writer.go:29: 2021-01-29T19:32:05.653Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Started HTTP server: address=127.0.0.1:30013 network=tcp writer.go:29: 2021-01-29T19:32:05.654Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: started state syncer === CONT TestAgent_loadServices_sidecarOverrideMeta/normal writer.go:29: 2021-01-29T19:32:05.653Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server.serf.wan: serf: EventMemberJoin: Node-4d8dcdf7-944d-5c6e-5574-5c358fff64a8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:05.656Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server.serf.lan: serf: EventMemberJoin: Node-4d8dcdf7-944d-5c6e-5574-5c358fff64a8 127.0.0.1 writer.go:29: 2021-01-29T19:32:05.660Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: Started DNS server: address=127.0.0.1:30006 network=udp writer.go:29: 2021-01-29T19:32:05.689Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: Started DNS server: address=127.0.0.1:30006 network=tcp writer.go:29: 2021-01-29T19:32:05.657Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:30011 [Follower]" leader= === CONT TestAgent_loadServices_sidecarOverrideMeta/service_manager writer.go:29: 2021-01-29T19:32:05.691Z [WARN] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:05.701Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:30017 [Candidate]" term=2 === CONT TestAgent_loadServices_sidecarOverrideMeta/normal writer.go:29: 2021-01-29T19:32:05.692Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: Started HTTP server: address=127.0.0.1:30007 network=tcp === CONT TestAgent_loadServices_sidecarOverrideMeta/service_manager writer.go:29: 2021-01-29T19:32:05.703Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.raft: votes: needed=1 === CONT TestAgent_loadServices_sidecarOverrideMeta/normal writer.go:29: 2021-01-29T19:32:05.703Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: started state syncer === CONT TestAgent_loadServices_sidecarOverrideMeta/service_manager writer.go:29: 2021-01-29T19:32:05.704Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.raft: vote granted: from=db0f6d8e-f12a-b207-9eef-36bbcc086d18 term=2 tally=1 === CONT TestAgent_loadServices_sidecarOverrideMeta/normal writer.go:29: 2021-01-29T19:32:05.692Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server: Handled event for server in area: event=member-join server=Node-4d8dcdf7-944d-5c6e-5574-5c358fff64a8.dc1 area=wan === CONT TestAgent_loadServices_sidecarOverrideMeta/service_manager writer.go:29: 2021-01-29T19:32:05.706Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.raft: election won: tally=1 === CONT TestAgent_loadServices_sidecarOverrideMeta/normal writer.go:29: 2021-01-29T19:32:05.694Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server: Adding LAN server: server="Node-4d8dcdf7-944d-5c6e-5574-5c358fff64a8 (Addr: tcp/127.0.0.1:30011) (DC: dc1)" === CONT TestAgent_loadServices_sidecarOverrideMeta/service_manager writer.go:29: 2021-01-29T19:32:05.708Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:30017 [Leader]" writer.go:29: 2021-01-29T19:32:05.710Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:05.710Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server: New leader elected: payload=Node-db0f6d8e-f12a-b207-9eef-36bbcc086d18 writer.go:29: 2021-01-29T19:32:05.711Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30017 writer.go:29: 2021-01-29T19:32:05.717Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:05.739Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:05.740Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.741Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-db0f6d8e-f12a-b207-9eef-36bbcc086d18 writer.go:29: 2021-01-29T19:32:05.742Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server: member joined, marking health alive: member=Node-db0f6d8e-f12a-b207-9eef-36bbcc086d18 === CONT TestAgent_loadServices_sidecarOverrideMeta/normal writer.go:29: 2021-01-29T19:32:05.767Z [WARN] TestAgent_loadServices_sidecarOverrideMeta/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:05.768Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:30011 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:05.772Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:05.773Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/normal.server.raft: vote granted: from=4d8dcdf7-944d-5c6e-5574-5c358fff64a8 term=2 tally=1 writer.go:29: 2021-01-29T19:32:05.774Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:05.775Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:30011 [Leader]" writer.go:29: 2021-01-29T19:32:05.776Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:05.776Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server: New leader elected: payload=Node-4d8dcdf7-944d-5c6e-5574-5c358fff64a8 writer.go:29: 2021-01-29T19:32:05.779Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30011 writer.go:29: 2021-01-29T19:32:05.783Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:05.789Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:05.790Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.791Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/normal.server: Skipping self join check for node since the cluster is too small: node=Node-4d8dcdf7-944d-5c6e-5574-5c358fff64a8 writer.go:29: 2021-01-29T19:32:05.792Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server: member joined, marking health alive: member=Node-4d8dcdf7-944d-5c6e-5574-5c358fff64a8 writer.go:29: 2021-01-29T19:32:05.797Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: Requesting shutdown writer.go:29: 2021-01-29T19:32:05.798Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server: shutting down server writer.go:29: 2021-01-29T19:32:05.799Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.800Z [WARN] TestAgent_loadServices_sidecarOverrideMeta/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:05.801Z [ERROR] TestAgent_loadServices_sidecarOverrideMeta/normal.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:05.802Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.803Z [WARN] TestAgent_loadServices_sidecarOverrideMeta/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:05.806Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:05.806Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: consul server down writer.go:29: 2021-01-29T19:32:05.809Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: shutdown complete writer.go:29: 2021-01-29T19:32:05.810Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: Stopping server: protocol=DNS address=127.0.0.1:30006 network=tcp writer.go:29: 2021-01-29T19:32:05.812Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: Stopping server: protocol=DNS address=127.0.0.1:30006 network=udp writer.go:29: 2021-01-29T19:32:05.813Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: Stopping server: protocol=HTTP address=127.0.0.1:30007 network=tcp === CONT TestAgent_loadServices_sidecarOverrideMeta/service_manager writer.go:29: 2021-01-29T19:32:05.941Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:05.945Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Synced node info writer.go:29: 2021-01-29T19:32:05.947Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Synced service: service=rabbitmq writer.go:29: 2021-01-29T19:32:05.955Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Synced service: service=rabbitmq-sidecar-proxy writer.go:29: 2021-01-29T19:32:05.959Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Check in sync: check=service:rabbitmq-sidecar-proxy:1 writer.go:29: 2021-01-29T19:32:05.961Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Check in sync: check=service:rabbitmq-sidecar-proxy:2 writer.go:29: 2021-01-29T19:32:06.071Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:32:06.072Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:32:06.073Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:06.074Z [WARN] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:06.074Z [DEBUG] TestAgent_loadServices_sidecarOverrideMeta/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:06.077Z [WARN] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:06.080Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:06.080Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: consul server down writer.go:29: 2021-01-29T19:32:06.093Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: shutdown complete writer.go:29: 2021-01-29T19:32:06.094Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Stopping server: protocol=DNS address=127.0.0.1:30012 network=tcp writer.go:29: 2021-01-29T19:32:06.097Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Stopping server: protocol=DNS address=127.0.0.1:30012 network=udp writer.go:29: 2021-01-29T19:32:06.098Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:30013 network=tcp === CONT TestAgent_loadServices_sidecarOverrideMeta/normal writer.go:29: 2021-01-29T19:32:06.315Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:06.316Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/normal: Endpoints down === CONT TestAgent_loadServices_sidecarOverrideMeta/service_manager writer.go:29: 2021-01-29T19:32:06.599Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:06.601Z [INFO] TestAgent_loadServices_sidecarOverrideMeta/service_manager: Endpoints down --- PASS: TestAgent_loadServices_sidecarOverrideMeta (0.00s) --- PASS: TestAgent_loadServices_sidecarOverrideMeta/normal (0.78s) --- PASS: TestAgent_loadServices_sidecarOverrideMeta/service_manager (1.06s) === RUN TestAgent_unloadServices === RUN TestAgent_unloadServices/normal === PAUSE TestAgent_unloadServices/normal === RUN TestAgent_unloadServices/service_manager === PAUSE TestAgent_unloadServices/service_manager === CONT TestAgent_unloadServices/normal === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.613Z [WARN] TestAgent_unloadServices/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:06.614Z [DEBUG] TestAgent_unloadServices/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:06.616Z [DEBUG] TestAgent_unloadServices/service_manager.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:06.622Z [INFO] TestAgent_unloadServices/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:58f611f4-7146-d369-0dd7-85cae64d1ef6 Address:127.0.0.1:30029}]" writer.go:29: 2021-01-29T19:32:06.624Z [INFO] TestAgent_unloadServices/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:30029 [Follower]" leader= writer.go:29: 2021-01-29T19:32:06.625Z [INFO] TestAgent_unloadServices/service_manager.server.serf.wan: serf: EventMemberJoin: Node-58f611f4-7146-d369-0dd7-85cae64d1ef6.dc1 127.0.0.1 === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.625Z [WARN] TestAgent_unloadServices/normal: bootstrap = true: do not enable unless necessary === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.628Z [INFO] TestAgent_unloadServices/service_manager.server.serf.lan: serf: EventMemberJoin: Node-58f611f4-7146-d369-0dd7-85cae64d1ef6 127.0.0.1 writer.go:29: 2021-01-29T19:32:06.630Z [INFO] TestAgent_unloadServices/service_manager.server: Handled event for server in area: event=member-join server=Node-58f611f4-7146-d369-0dd7-85cae64d1ef6.dc1 area=wan writer.go:29: 2021-01-29T19:32:06.630Z [INFO] TestAgent_unloadServices/service_manager: Started DNS server: address=127.0.0.1:30024 network=udp writer.go:29: 2021-01-29T19:32:06.632Z [INFO] TestAgent_unloadServices/service_manager: Started DNS server: address=127.0.0.1:30024 network=tcp writer.go:29: 2021-01-29T19:32:06.631Z [INFO] TestAgent_unloadServices/service_manager.server: Adding LAN server: server="Node-58f611f4-7146-d369-0dd7-85cae64d1ef6 (Addr: tcp/127.0.0.1:30029) (DC: dc1)" === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.628Z [DEBUG] TestAgent_unloadServices/normal.tlsutil: Update: version=1 === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.633Z [INFO] TestAgent_unloadServices/service_manager: Started HTTP server: address=127.0.0.1:30025 network=tcp writer.go:29: 2021-01-29T19:32:06.635Z [INFO] TestAgent_unloadServices/service_manager: started state syncer === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.637Z [DEBUG] TestAgent_unloadServices/normal.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:06.643Z [INFO] TestAgent_unloadServices/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:836dfdff-e534-0bc1-76d5-f26de7416136 Address:127.0.0.1:30023}]" writer.go:29: 2021-01-29T19:32:06.645Z [INFO] TestAgent_unloadServices/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:30023 [Follower]" leader= writer.go:29: 2021-01-29T19:32:06.647Z [INFO] TestAgent_unloadServices/normal.server.serf.wan: serf: EventMemberJoin: Node-836dfdff-e534-0bc1-76d5-f26de7416136.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:06.650Z [INFO] TestAgent_unloadServices/normal.server.serf.lan: serf: EventMemberJoin: Node-836dfdff-e534-0bc1-76d5-f26de7416136 127.0.0.1 writer.go:29: 2021-01-29T19:32:06.652Z [INFO] TestAgent_unloadServices/normal.server: Handled event for server in area: event=member-join server=Node-836dfdff-e534-0bc1-76d5-f26de7416136.dc1 area=wan writer.go:29: 2021-01-29T19:32:06.652Z [INFO] TestAgent_unloadServices/normal.server: Adding LAN server: server="Node-836dfdff-e534-0bc1-76d5-f26de7416136 (Addr: tcp/127.0.0.1:30023) (DC: dc1)" writer.go:29: 2021-01-29T19:32:06.653Z [INFO] TestAgent_unloadServices/normal: Started DNS server: address=127.0.0.1:30018 network=tcp writer.go:29: 2021-01-29T19:32:06.655Z [INFO] TestAgent_unloadServices/normal: Started DNS server: address=127.0.0.1:30018 network=udp writer.go:29: 2021-01-29T19:32:06.656Z [INFO] TestAgent_unloadServices/normal: Started HTTP server: address=127.0.0.1:30019 network=tcp writer.go:29: 2021-01-29T19:32:06.657Z [INFO] TestAgent_unloadServices/normal: started state syncer === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.672Z [WARN] TestAgent_unloadServices/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:06.673Z [INFO] TestAgent_unloadServices/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:30029 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:06.674Z [DEBUG] TestAgent_unloadServices/service_manager.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:06.675Z [DEBUG] TestAgent_unloadServices/service_manager.server.raft: vote granted: from=58f611f4-7146-d369-0dd7-85cae64d1ef6 term=2 tally=1 writer.go:29: 2021-01-29T19:32:06.676Z [INFO] TestAgent_unloadServices/service_manager.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:06.676Z [INFO] TestAgent_unloadServices/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:30029 [Leader]" writer.go:29: 2021-01-29T19:32:06.677Z [INFO] TestAgent_unloadServices/service_manager.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:06.678Z [INFO] TestAgent_unloadServices/service_manager.server: New leader elected: payload=Node-58f611f4-7146-d369-0dd7-85cae64d1ef6 writer.go:29: 2021-01-29T19:32:06.678Z [DEBUG] TestAgent_unloadServices/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30029 writer.go:29: 2021-01-29T19:32:06.693Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:06.697Z [INFO] TestAgent_unloadServices/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.698Z [WARN] TestAgent_unloadServices/normal.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.698Z [INFO] TestAgent_unloadServices/service_manager.leader: started routine: routine="CA root pruning" === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.699Z [INFO] TestAgent_unloadServices/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:30023 [Candidate]" term=2 === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.699Z [DEBUG] TestAgent_unloadServices/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-58f611f4-7146-d369-0dd7-85cae64d1ef6 === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.700Z [DEBUG] TestAgent_unloadServices/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:06.701Z [DEBUG] TestAgent_unloadServices/normal.server.raft: vote granted: from=836dfdff-e534-0bc1-76d5-f26de7416136 term=2 tally=1 === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.701Z [INFO] TestAgent_unloadServices/service_manager.server: member joined, marking health alive: member=Node-58f611f4-7146-d369-0dd7-85cae64d1ef6 === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.702Z [INFO] TestAgent_unloadServices/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:06.703Z [INFO] TestAgent_unloadServices/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:30023 [Leader]" writer.go:29: 2021-01-29T19:32:06.704Z [INFO] TestAgent_unloadServices/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:06.704Z [INFO] TestAgent_unloadServices/normal.server: New leader elected: payload=Node-836dfdff-e534-0bc1-76d5-f26de7416136 writer.go:29: 2021-01-29T19:32:06.705Z [DEBUG] TestAgent_unloadServices/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30023 writer.go:29: 2021-01-29T19:32:06.710Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:06.713Z [INFO] TestAgent_unloadServices/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:06.714Z [INFO] TestAgent_unloadServices/normal.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:06.715Z [DEBUG] TestAgent_unloadServices/normal.server: Skipping self join check for node since the cluster is too small: node=Node-836dfdff-e534-0bc1-76d5-f26de7416136 writer.go:29: 2021-01-29T19:32:06.716Z [INFO] TestAgent_unloadServices/normal.server: member joined, marking health alive: member=Node-836dfdff-e534-0bc1-76d5-f26de7416136 writer.go:29: 2021-01-29T19:32:06.909Z [DEBUG] TestAgent_unloadServices/normal: removed service: service=redis writer.go:29: 2021-01-29T19:32:06.912Z [INFO] TestAgent_unloadServices/normal: Requesting shutdown writer.go:29: 2021-01-29T19:32:06.913Z [INFO] TestAgent_unloadServices/normal.server: shutting down server writer.go:29: 2021-01-29T19:32:06.914Z [DEBUG] TestAgent_unloadServices/normal.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:06.915Z [WARN] TestAgent_unloadServices/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:06.916Z [ERROR] TestAgent_unloadServices/normal.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:06.916Z [DEBUG] TestAgent_unloadServices/normal.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.917Z [DEBUG] TestAgent_unloadServices/service_manager: removed service: service=redis writer.go:29: 2021-01-29T19:32:06.919Z [INFO] TestAgent_unloadServices/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:32:06.919Z [INFO] TestAgent_unloadServices/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:32:06.920Z [DEBUG] TestAgent_unloadServices/service_manager.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.925Z [WARN] TestAgent_unloadServices/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:06.926Z [INFO] TestAgent_unloadServices/normal.server.router.manager: shutting down === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.921Z [WARN] TestAgent_unloadServices/service_manager.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.926Z [INFO] TestAgent_unloadServices/normal: consul server down === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.922Z [ERROR] TestAgent_unloadServices/service_manager.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.928Z [INFO] TestAgent_unloadServices/normal: shutdown complete === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.924Z [DEBUG] TestAgent_unloadServices/service_manager.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.929Z [INFO] TestAgent_unloadServices/normal: Stopping server: protocol=DNS address=127.0.0.1:30018 network=tcp writer.go:29: 2021-01-29T19:32:06.930Z [INFO] TestAgent_unloadServices/normal: Stopping server: protocol=DNS address=127.0.0.1:30018 network=udp === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.930Z [WARN] TestAgent_unloadServices/service_manager.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:06.931Z [INFO] TestAgent_unloadServices/normal: Stopping server: protocol=HTTP address=127.0.0.1:30019 network=tcp === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:06.934Z [INFO] TestAgent_unloadServices/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:06.935Z [INFO] TestAgent_unloadServices/service_manager: consul server down writer.go:29: 2021-01-29T19:32:06.936Z [INFO] TestAgent_unloadServices/service_manager: shutdown complete writer.go:29: 2021-01-29T19:32:06.937Z [INFO] TestAgent_unloadServices/service_manager: Stopping server: protocol=DNS address=127.0.0.1:30024 network=tcp writer.go:29: 2021-01-29T19:32:06.938Z [INFO] TestAgent_unloadServices/service_manager: Stopping server: protocol=DNS address=127.0.0.1:30024 network=udp writer.go:29: 2021-01-29T19:32:06.940Z [INFO] TestAgent_unloadServices/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:30025 network=tcp === CONT TestAgent_unloadServices/normal writer.go:29: 2021-01-29T19:32:07.431Z [INFO] TestAgent_unloadServices/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:07.433Z [INFO] TestAgent_unloadServices/normal: Endpoints down === CONT TestAgent_unloadServices/service_manager writer.go:29: 2021-01-29T19:32:07.441Z [INFO] TestAgent_unloadServices/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:07.442Z [INFO] TestAgent_unloadServices/service_manager: Endpoints down --- PASS: TestAgent_unloadServices (0.00s) --- PASS: TestAgent_unloadServices/normal (0.83s) --- PASS: TestAgent_unloadServices/service_manager (0.84s) === RUN TestAgent_Service_MaintenanceMode === PAUSE TestAgent_Service_MaintenanceMode === RUN TestAgent_Service_Reap writer.go:29: 2021-01-29T19:32:07.457Z [WARN] TestAgent_Service_Reap: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:07.458Z [DEBUG] TestAgent_Service_Reap.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:07.460Z [DEBUG] TestAgent_Service_Reap.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:07.465Z [INFO] TestAgent_Service_Reap.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2941e7dc-c9f3-8731-6344-89c5d70b7e0e Address:127.0.0.1:30035}]" writer.go:29: 2021-01-29T19:32:07.467Z [INFO] TestAgent_Service_Reap.server.raft: entering follower state: follower="Node at 127.0.0.1:30035 [Follower]" leader= writer.go:29: 2021-01-29T19:32:07.467Z [INFO] TestAgent_Service_Reap.server.serf.wan: serf: EventMemberJoin: Node-2941e7dc-c9f3-8731-6344-89c5d70b7e0e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:07.473Z [INFO] TestAgent_Service_Reap.server.serf.lan: serf: EventMemberJoin: Node-2941e7dc-c9f3-8731-6344-89c5d70b7e0e 127.0.0.1 writer.go:29: 2021-01-29T19:32:07.492Z [INFO] TestAgent_Service_Reap.server: Handled event for server in area: event=member-join server=Node-2941e7dc-c9f3-8731-6344-89c5d70b7e0e.dc1 area=wan writer.go:29: 2021-01-29T19:32:07.492Z [INFO] TestAgent_Service_Reap.server: Adding LAN server: server="Node-2941e7dc-c9f3-8731-6344-89c5d70b7e0e (Addr: tcp/127.0.0.1:30035) (DC: dc1)" writer.go:29: 2021-01-29T19:32:07.493Z [INFO] TestAgent_Service_Reap: Started DNS server: address=127.0.0.1:30030 network=udp writer.go:29: 2021-01-29T19:32:07.496Z [INFO] TestAgent_Service_Reap: Started DNS server: address=127.0.0.1:30030 network=tcp writer.go:29: 2021-01-29T19:32:07.498Z [INFO] TestAgent_Service_Reap: Started HTTP server: address=127.0.0.1:30031 network=tcp writer.go:29: 2021-01-29T19:32:07.499Z [INFO] TestAgent_Service_Reap: started state syncer writer.go:29: 2021-01-29T19:32:07.523Z [WARN] TestAgent_Service_Reap.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:07.524Z [INFO] TestAgent_Service_Reap.server.raft: entering candidate state: node="Node at 127.0.0.1:30035 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:07.526Z [DEBUG] TestAgent_Service_Reap.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:07.527Z [DEBUG] TestAgent_Service_Reap.server.raft: vote granted: from=2941e7dc-c9f3-8731-6344-89c5d70b7e0e term=2 tally=1 writer.go:29: 2021-01-29T19:32:07.527Z [INFO] TestAgent_Service_Reap.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:07.528Z [INFO] TestAgent_Service_Reap.server.raft: entering leader state: leader="Node at 127.0.0.1:30035 [Leader]" writer.go:29: 2021-01-29T19:32:07.529Z [INFO] TestAgent_Service_Reap.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:07.530Z [INFO] TestAgent_Service_Reap.server: New leader elected: payload=Node-2941e7dc-c9f3-8731-6344-89c5d70b7e0e writer.go:29: 2021-01-29T19:32:07.531Z [DEBUG] TestAgent_Service_Reap.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30035 writer.go:29: 2021-01-29T19:32:07.537Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:07.544Z [INFO] TestAgent_Service_Reap.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:07.547Z [INFO] TestAgent_Service_Reap.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:07.551Z [DEBUG] TestAgent_Service_Reap.server: Skipping self join check for node since the cluster is too small: node=Node-2941e7dc-c9f3-8731-6344-89c5d70b7e0e writer.go:29: 2021-01-29T19:32:07.554Z [INFO] TestAgent_Service_Reap.server: member joined, marking health alive: member=Node-2941e7dc-c9f3-8731-6344-89c5d70b7e0e writer.go:29: 2021-01-29T19:32:07.646Z [DEBUG] TestAgent_Service_Reap: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:07.649Z [INFO] TestAgent_Service_Reap: Synced node info writer.go:29: 2021-01-29T19:32:07.650Z [DEBUG] TestAgent_Service_Reap: Node info in sync writer.go:29: 2021-01-29T19:32:07.780Z [WARN] TestAgent_Service_Reap: Check missed TTL, is now critical: check=service:redis writer.go:29: 2021-01-29T19:32:07.858Z [DEBUG] TestAgent_Service_Reap: Check status updated: check=service:redis status=passing writer.go:29: 2021-01-29T19:32:07.885Z [WARN] TestAgent_Service_Reap: Check missed TTL, is now critical: check=service:redis writer.go:29: 2021-01-29T19:32:08.096Z [DEBUG] TestAgent_Service_Reap: removed check: check=service:redis writer.go:29: 2021-01-29T19:32:08.097Z [DEBUG] TestAgent_Service_Reap: removed service: service=redis writer.go:29: 2021-01-29T19:32:08.098Z [INFO] TestAgent_Service_Reap: Check for service has been critical for too long; deregistered service: service=redis check=service:redis writer.go:29: 2021-01-29T19:32:08.365Z [INFO] TestAgent_Service_Reap: Requesting shutdown writer.go:29: 2021-01-29T19:32:08.367Z [INFO] TestAgent_Service_Reap.server: shutting down server writer.go:29: 2021-01-29T19:32:08.368Z [DEBUG] TestAgent_Service_Reap.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:08.370Z [WARN] TestAgent_Service_Reap.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:08.370Z [DEBUG] TestAgent_Service_Reap.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:08.373Z [WARN] TestAgent_Service_Reap.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:08.375Z [INFO] TestAgent_Service_Reap.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:08.375Z [INFO] TestAgent_Service_Reap: consul server down writer.go:29: 2021-01-29T19:32:08.377Z [INFO] TestAgent_Service_Reap: shutdown complete writer.go:29: 2021-01-29T19:32:08.377Z [INFO] TestAgent_Service_Reap: Stopping server: protocol=DNS address=127.0.0.1:30030 network=tcp writer.go:29: 2021-01-29T19:32:08.378Z [INFO] TestAgent_Service_Reap: Stopping server: protocol=DNS address=127.0.0.1:30030 network=udp writer.go:29: 2021-01-29T19:32:08.379Z [INFO] TestAgent_Service_Reap: Stopping server: protocol=HTTP address=127.0.0.1:30031 network=tcp writer.go:29: 2021-01-29T19:32:08.881Z [INFO] TestAgent_Service_Reap: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:08.883Z [INFO] TestAgent_Service_Reap: Endpoints down --- PASS: TestAgent_Service_Reap (1.44s) === RUN TestAgent_Service_NoReap writer.go:29: 2021-01-29T19:32:08.914Z [WARN] TestAgent_Service_NoReap: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:08.916Z [DEBUG] TestAgent_Service_NoReap.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:08.918Z [DEBUG] TestAgent_Service_NoReap.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:08.925Z [INFO] TestAgent_Service_NoReap.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a0350bfb-beb9-a305-6b4a-d4fecf47ff1c Address:127.0.0.1:30041}]" writer.go:29: 2021-01-29T19:32:08.933Z [INFO] TestAgent_Service_NoReap.server.raft: entering follower state: follower="Node at 127.0.0.1:30041 [Follower]" leader= writer.go:29: 2021-01-29T19:32:08.936Z [INFO] TestAgent_Service_NoReap.server.serf.wan: serf: EventMemberJoin: Node-a0350bfb-beb9-a305-6b4a-d4fecf47ff1c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:08.944Z [INFO] TestAgent_Service_NoReap.server.serf.lan: serf: EventMemberJoin: Node-a0350bfb-beb9-a305-6b4a-d4fecf47ff1c 127.0.0.1 writer.go:29: 2021-01-29T19:32:08.965Z [INFO] TestAgent_Service_NoReap.server: Adding LAN server: server="Node-a0350bfb-beb9-a305-6b4a-d4fecf47ff1c (Addr: tcp/127.0.0.1:30041) (DC: dc1)" writer.go:29: 2021-01-29T19:32:08.966Z [INFO] TestAgent_Service_NoReap.server: Handled event for server in area: event=member-join server=Node-a0350bfb-beb9-a305-6b4a-d4fecf47ff1c.dc1 area=wan writer.go:29: 2021-01-29T19:32:08.968Z [INFO] TestAgent_Service_NoReap: Started DNS server: address=127.0.0.1:30036 network=udp writer.go:29: 2021-01-29T19:32:08.969Z [INFO] TestAgent_Service_NoReap: Started DNS server: address=127.0.0.1:30036 network=tcp writer.go:29: 2021-01-29T19:32:08.972Z [INFO] TestAgent_Service_NoReap: Started HTTP server: address=127.0.0.1:30037 network=tcp writer.go:29: 2021-01-29T19:32:08.973Z [INFO] TestAgent_Service_NoReap: started state syncer writer.go:29: 2021-01-29T19:32:08.975Z [WARN] TestAgent_Service_NoReap.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:08.976Z [INFO] TestAgent_Service_NoReap.server.raft: entering candidate state: node="Node at 127.0.0.1:30041 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:08.994Z [DEBUG] TestAgent_Service_NoReap.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:08.995Z [DEBUG] TestAgent_Service_NoReap.server.raft: vote granted: from=a0350bfb-beb9-a305-6b4a-d4fecf47ff1c term=2 tally=1 writer.go:29: 2021-01-29T19:32:08.996Z [INFO] TestAgent_Service_NoReap.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:08.997Z [INFO] TestAgent_Service_NoReap.server.raft: entering leader state: leader="Node at 127.0.0.1:30041 [Leader]" writer.go:29: 2021-01-29T19:32:08.998Z [INFO] TestAgent_Service_NoReap.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:09.006Z [INFO] TestAgent_Service_NoReap.server: New leader elected: payload=Node-a0350bfb-beb9-a305-6b4a-d4fecf47ff1c writer.go:29: 2021-01-29T19:32:09.006Z [DEBUG] TestAgent_Service_NoReap.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30041 writer.go:29: 2021-01-29T19:32:09.012Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:09.018Z [INFO] TestAgent_Service_NoReap.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:09.019Z [INFO] TestAgent_Service_NoReap.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:09.020Z [DEBUG] TestAgent_Service_NoReap.server: Skipping self join check for node since the cluster is too small: node=Node-a0350bfb-beb9-a305-6b4a-d4fecf47ff1c writer.go:29: 2021-01-29T19:32:09.021Z [INFO] TestAgent_Service_NoReap.server: member joined, marking health alive: member=Node-a0350bfb-beb9-a305-6b4a-d4fecf47ff1c writer.go:29: 2021-01-29T19:32:09.188Z [DEBUG] TestAgent_Service_NoReap: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:09.190Z [INFO] TestAgent_Service_NoReap: Synced node info writer.go:29: 2021-01-29T19:32:09.292Z [WARN] TestAgent_Service_NoReap: Check missed TTL, is now critical: check=service:redis writer.go:29: 2021-01-29T19:32:09.671Z [INFO] TestAgent_Service_NoReap: Requesting shutdown writer.go:29: 2021-01-29T19:32:09.673Z [INFO] TestAgent_Service_NoReap.server: shutting down server writer.go:29: 2021-01-29T19:32:09.674Z [DEBUG] TestAgent_Service_NoReap.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:09.674Z [WARN] TestAgent_Service_NoReap.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:09.674Z [DEBUG] TestAgent_Service_NoReap.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:09.677Z [WARN] TestAgent_Service_NoReap.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:09.679Z [INFO] TestAgent_Service_NoReap.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:09.679Z [INFO] TestAgent_Service_NoReap: consul server down writer.go:29: 2021-01-29T19:32:09.680Z [INFO] TestAgent_Service_NoReap: shutdown complete writer.go:29: 2021-01-29T19:32:09.681Z [INFO] TestAgent_Service_NoReap: Stopping server: protocol=DNS address=127.0.0.1:30036 network=tcp writer.go:29: 2021-01-29T19:32:09.682Z [INFO] TestAgent_Service_NoReap: Stopping server: protocol=DNS address=127.0.0.1:30036 network=udp writer.go:29: 2021-01-29T19:32:09.683Z [INFO] TestAgent_Service_NoReap: Stopping server: protocol=HTTP address=127.0.0.1:30037 network=tcp writer.go:29: 2021-01-29T19:32:10.184Z [INFO] TestAgent_Service_NoReap: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:10.186Z [INFO] TestAgent_Service_NoReap: Endpoints down --- PASS: TestAgent_Service_NoReap (1.30s) === RUN TestAgent_AddService_restoresSnapshot === RUN TestAgent_AddService_restoresSnapshot/normal === PAUSE TestAgent_AddService_restoresSnapshot/normal === RUN TestAgent_AddService_restoresSnapshot/service_manager === PAUSE TestAgent_AddService_restoresSnapshot/service_manager === CONT TestAgent_AddService_restoresSnapshot/normal === CONT TestAgent_AddService_restoresSnapshot/service_manager === CONT TestAgent_AddService_restoresSnapshot/normal writer.go:29: 2021-01-29T19:32:10.199Z [WARN] TestAgent_AddService_restoresSnapshot/normal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:10.201Z [DEBUG] TestAgent_AddService_restoresSnapshot/normal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:10.210Z [DEBUG] TestAgent_AddService_restoresSnapshot/normal.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:10.230Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:eee1dd60-6f56-32e0-f9b0-23ffbf29ddcf Address:127.0.0.1:30047}]" writer.go:29: 2021-01-29T19:32:10.241Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server.raft: entering follower state: follower="Node at 127.0.0.1:30047 [Follower]" leader= writer.go:29: 2021-01-29T19:32:10.252Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server.serf.wan: serf: EventMemberJoin: Node-eee1dd60-6f56-32e0-f9b0-23ffbf29ddcf.dc1 127.0.0.1 === CONT TestAgent_AddService_restoresSnapshot/service_manager writer.go:29: 2021-01-29T19:32:10.261Z [WARN] TestAgent_AddService_restoresSnapshot/service_manager: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:10.262Z [DEBUG] TestAgent_AddService_restoresSnapshot/service_manager.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:10.263Z [DEBUG] TestAgent_AddService_restoresSnapshot/service_manager.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:10.266Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3d3c140a-9fd2-8328-5f46-79e2966f85a7 Address:127.0.0.1:30053}]" writer.go:29: 2021-01-29T19:32:10.268Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server.raft: entering follower state: follower="Node at 127.0.0.1:30053 [Follower]" leader= writer.go:29: 2021-01-29T19:32:10.269Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server.serf.wan: serf: EventMemberJoin: Node-3d3c140a-9fd2-8328-5f46-79e2966f85a7.dc1 127.0.0.1 === CONT TestAgent_AddService_restoresSnapshot/normal writer.go:29: 2021-01-29T19:32:10.266Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server.serf.lan: serf: EventMemberJoin: Node-eee1dd60-6f56-32e0-f9b0-23ffbf29ddcf 127.0.0.1 === CONT TestAgent_AddService_restoresSnapshot/service_manager writer.go:29: 2021-01-29T19:32:10.272Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server.serf.lan: serf: EventMemberJoin: Node-3d3c140a-9fd2-8328-5f46-79e2966f85a7 127.0.0.1 writer.go:29: 2021-01-29T19:32:10.274Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server: Handled event for server in area: event=member-join server=Node-3d3c140a-9fd2-8328-5f46-79e2966f85a7.dc1 area=wan writer.go:29: 2021-01-29T19:32:10.275Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: Started DNS server: address=127.0.0.1:30048 network=udp writer.go:29: 2021-01-29T19:32:10.274Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server: Adding LAN server: server="Node-3d3c140a-9fd2-8328-5f46-79e2966f85a7 (Addr: tcp/127.0.0.1:30053) (DC: dc1)" writer.go:29: 2021-01-29T19:32:10.276Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: Started DNS server: address=127.0.0.1:30048 network=tcp writer.go:29: 2021-01-29T19:32:10.278Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: Started HTTP server: address=127.0.0.1:30049 network=tcp === CONT TestAgent_AddService_restoresSnapshot/normal writer.go:29: 2021-01-29T19:32:10.279Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server: Adding LAN server: server="Node-eee1dd60-6f56-32e0-f9b0-23ffbf29ddcf (Addr: tcp/127.0.0.1:30047) (DC: dc1)" === CONT TestAgent_AddService_restoresSnapshot/service_manager writer.go:29: 2021-01-29T19:32:10.279Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: started state syncer === CONT TestAgent_AddService_restoresSnapshot/normal writer.go:29: 2021-01-29T19:32:10.280Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server: Handled event for server in area: event=member-join server=Node-eee1dd60-6f56-32e0-f9b0-23ffbf29ddcf.dc1 area=wan writer.go:29: 2021-01-29T19:32:10.280Z [WARN] TestAgent_AddService_restoresSnapshot/normal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:10.282Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server.raft: entering candidate state: node="Node at 127.0.0.1:30047 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:10.283Z [DEBUG] TestAgent_AddService_restoresSnapshot/normal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:10.285Z [DEBUG] TestAgent_AddService_restoresSnapshot/normal.server.raft: vote granted: from=eee1dd60-6f56-32e0-f9b0-23ffbf29ddcf term=2 tally=1 writer.go:29: 2021-01-29T19:32:10.285Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:10.284Z [INFO] TestAgent_AddService_restoresSnapshot/normal: Started DNS server: address=127.0.0.1:30042 network=tcp writer.go:29: 2021-01-29T19:32:10.287Z [INFO] TestAgent_AddService_restoresSnapshot/normal: Started DNS server: address=127.0.0.1:30042 network=udp writer.go:29: 2021-01-29T19:32:10.286Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server.raft: entering leader state: leader="Node at 127.0.0.1:30047 [Leader]" writer.go:29: 2021-01-29T19:32:10.289Z [INFO] TestAgent_AddService_restoresSnapshot/normal: Started HTTP server: address=127.0.0.1:30043 network=tcp writer.go:29: 2021-01-29T19:32:10.289Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:10.290Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server: New leader elected: payload=Node-eee1dd60-6f56-32e0-f9b0-23ffbf29ddcf writer.go:29: 2021-01-29T19:32:10.291Z [DEBUG] TestAgent_AddService_restoresSnapshot/normal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30047 writer.go:29: 2021-01-29T19:32:10.291Z [INFO] TestAgent_AddService_restoresSnapshot/normal: started state syncer writer.go:29: 2021-01-29T19:32:10.294Z [INFO] TestAgent_AddService_restoresSnapshot/normal: Synced node info writer.go:29: 2021-01-29T19:32:10.297Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:10.301Z [INFO] TestAgent_AddService_restoresSnapshot/normal: Requesting shutdown writer.go:29: 2021-01-29T19:32:10.301Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:10.304Z [INFO] TestAgent_AddService_restoresSnapshot/normal.leader: started routine: routine="CA root pruning" === CONT TestAgent_AddService_restoresSnapshot/service_manager writer.go:29: 2021-01-29T19:32:10.307Z [WARN] TestAgent_AddService_restoresSnapshot/service_manager.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:10.308Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server.raft: entering candidate state: node="Node at 127.0.0.1:30053 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:10.310Z [DEBUG] TestAgent_AddService_restoresSnapshot/service_manager.server.raft: votes: needed=1 === CONT TestAgent_AddService_restoresSnapshot/normal writer.go:29: 2021-01-29T19:32:10.302Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server: shutting down server === CONT TestAgent_AddService_restoresSnapshot/service_manager writer.go:29: 2021-01-29T19:32:10.310Z [DEBUG] TestAgent_AddService_restoresSnapshot/service_manager.server.raft: vote granted: from=3d3c140a-9fd2-8328-5f46-79e2966f85a7 term=2 tally=1 writer.go:29: 2021-01-29T19:32:10.311Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server.raft: election won: tally=1 === CONT TestAgent_AddService_restoresSnapshot/normal writer.go:29: 2021-01-29T19:32:10.311Z [DEBUG] TestAgent_AddService_restoresSnapshot/normal.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_AddService_restoresSnapshot/service_manager writer.go:29: 2021-01-29T19:32:10.312Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server.raft: entering leader state: leader="Node at 127.0.0.1:30053 [Leader]" === CONT TestAgent_AddService_restoresSnapshot/normal writer.go:29: 2021-01-29T19:32:10.312Z [WARN] TestAgent_AddService_restoresSnapshot/normal.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:10.306Z [DEBUG] TestAgent_AddService_restoresSnapshot/normal.server: Skipping self join check for node since the cluster is too small: node=Node-eee1dd60-6f56-32e0-f9b0-23ffbf29ddcf === CONT TestAgent_AddService_restoresSnapshot/service_manager writer.go:29: 2021-01-29T19:32:10.313Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server: cluster leadership acquired === CONT TestAgent_AddService_restoresSnapshot/normal writer.go:29: 2021-01-29T19:32:10.313Z [DEBUG] TestAgent_AddService_restoresSnapshot/normal.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:10.316Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server: member joined, marking health alive: member=Node-eee1dd60-6f56-32e0-f9b0-23ffbf29ddcf writer.go:29: 2021-01-29T19:32:10.318Z [WARN] TestAgent_AddService_restoresSnapshot/normal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:10.322Z [INFO] TestAgent_AddService_restoresSnapshot/normal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:10.322Z [INFO] TestAgent_AddService_restoresSnapshot/normal: consul server down === CONT TestAgent_AddService_restoresSnapshot/service_manager writer.go:29: 2021-01-29T19:32:10.313Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server: New leader elected: payload=Node-3d3c140a-9fd2-8328-5f46-79e2966f85a7 === CONT TestAgent_AddService_restoresSnapshot/normal writer.go:29: 2021-01-29T19:32:10.324Z [INFO] TestAgent_AddService_restoresSnapshot/normal: shutdown complete writer.go:29: 2021-01-29T19:32:10.324Z [INFO] TestAgent_AddService_restoresSnapshot/normal: Stopping server: protocol=DNS address=127.0.0.1:30042 network=tcp writer.go:29: 2021-01-29T19:32:10.325Z [INFO] TestAgent_AddService_restoresSnapshot/normal: Stopping server: protocol=DNS address=127.0.0.1:30042 network=udp writer.go:29: 2021-01-29T19:32:10.326Z [INFO] TestAgent_AddService_restoresSnapshot/normal: Stopping server: protocol=HTTP address=127.0.0.1:30043 network=tcp === CONT TestAgent_AddService_restoresSnapshot/service_manager writer.go:29: 2021-01-29T19:32:10.319Z [DEBUG] TestAgent_AddService_restoresSnapshot/service_manager.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30053 writer.go:29: 2021-01-29T19:32:10.341Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:10.343Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: Synced node info writer.go:29: 2021-01-29T19:32:10.350Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:10.352Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:10.354Z [DEBUG] TestAgent_AddService_restoresSnapshot/service_manager.server: Skipping self join check for node since the cluster is too small: node=Node-3d3c140a-9fd2-8328-5f46-79e2966f85a7 writer.go:29: 2021-01-29T19:32:10.357Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server: member joined, marking health alive: member=Node-3d3c140a-9fd2-8328-5f46-79e2966f85a7 writer.go:29: 2021-01-29T19:32:10.637Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: Requesting shutdown writer.go:29: 2021-01-29T19:32:10.639Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server: shutting down server writer.go:29: 2021-01-29T19:32:10.640Z [DEBUG] TestAgent_AddService_restoresSnapshot/service_manager.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:10.642Z [WARN] TestAgent_AddService_restoresSnapshot/service_manager.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:10.642Z [DEBUG] TestAgent_AddService_restoresSnapshot/service_manager.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:10.645Z [WARN] TestAgent_AddService_restoresSnapshot/service_manager.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:10.646Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:10.648Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: consul server down writer.go:29: 2021-01-29T19:32:10.649Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: shutdown complete writer.go:29: 2021-01-29T19:32:10.649Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: Stopping server: protocol=DNS address=127.0.0.1:30048 network=tcp writer.go:29: 2021-01-29T19:32:10.650Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: Stopping server: protocol=DNS address=127.0.0.1:30048 network=udp writer.go:29: 2021-01-29T19:32:10.651Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: Stopping server: protocol=HTTP address=127.0.0.1:30049 network=tcp === CONT TestAgent_AddService_restoresSnapshot/normal writer.go:29: 2021-01-29T19:32:10.827Z [INFO] TestAgent_AddService_restoresSnapshot/normal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:10.828Z [INFO] TestAgent_AddService_restoresSnapshot/normal: Endpoints down === CONT TestAgent_AddService_restoresSnapshot/service_manager writer.go:29: 2021-01-29T19:32:11.152Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:11.154Z [INFO] TestAgent_AddService_restoresSnapshot/service_manager: Endpoints down --- PASS: TestAgent_AddService_restoresSnapshot (0.00s) --- PASS: TestAgent_AddService_restoresSnapshot/normal (0.64s) --- PASS: TestAgent_AddService_restoresSnapshot/service_manager (0.97s) === RUN TestAgent_AddCheck_restoresSnapshot === PAUSE TestAgent_AddCheck_restoresSnapshot === RUN TestAgent_NodeMaintenanceMode === PAUSE TestAgent_NodeMaintenanceMode === RUN TestAgent_checkStateSnapshot === PAUSE TestAgent_checkStateSnapshot === RUN TestAgent_loadChecks_checkFails === PAUSE TestAgent_loadChecks_checkFails === RUN TestAgent_persistCheckState === PAUSE TestAgent_persistCheckState === RUN TestAgent_loadCheckState === PAUSE TestAgent_loadCheckState === RUN TestAgent_purgeCheckState === PAUSE TestAgent_purgeCheckState === RUN TestAgent_GetCoordinate === PAUSE TestAgent_GetCoordinate === RUN TestAgent_reloadWatches === PAUSE TestAgent_reloadWatches === RUN TestAgent_reloadWatchesHTTPS === PAUSE TestAgent_reloadWatchesHTTPS === RUN TestAgent_loadTokens === PAUSE TestAgent_loadTokens === RUN TestAgent_ReloadConfigOutgoingRPCConfig === PAUSE TestAgent_ReloadConfigOutgoingRPCConfig === RUN TestAgent_ReloadConfigAndKeepChecksStatus agent_test.go:3443: DM-skipped --- SKIP: TestAgent_ReloadConfigAndKeepChecksStatus (0.01s) === RUN TestAgent_ReloadConfigIncomingRPCConfig === PAUSE TestAgent_ReloadConfigIncomingRPCConfig === RUN TestAgent_ReloadConfigTLSConfigFailure === PAUSE TestAgent_ReloadConfigTLSConfigFailure === RUN TestAgent_consulConfig_AutoEncryptAllowTLS === PAUSE TestAgent_consulConfig_AutoEncryptAllowTLS === RUN TestAgent_consulConfig_RaftTrailingLogs === PAUSE TestAgent_consulConfig_RaftTrailingLogs === RUN TestAgent_grpcInjectAddr === RUN TestAgent_grpcInjectAddr/localhost_web_svc === RUN TestAgent_grpcInjectAddr/localhost_no_svc === RUN TestAgent_grpcInjectAddr/ipv4_web_svc === RUN TestAgent_grpcInjectAddr/ipv4_no_svc === RUN TestAgent_grpcInjectAddr/ipv6_no_svc === RUN TestAgent_grpcInjectAddr/ipv6_web_svc === RUN TestAgent_grpcInjectAddr/zone_ipv6_web_svc === RUN TestAgent_grpcInjectAddr/ipv6_literal_web_svc === RUN TestAgent_grpcInjectAddr/ipv6_injected_into_ipv6_url === RUN TestAgent_grpcInjectAddr/ipv6_injected_into_ipv6_url_with_svc === RUN TestAgent_grpcInjectAddr/ipv6_injected_into_ipv6_url_with_special --- PASS: TestAgent_grpcInjectAddr (0.02s) --- PASS: TestAgent_grpcInjectAddr/localhost_web_svc (0.00s) --- PASS: TestAgent_grpcInjectAddr/localhost_no_svc (0.00s) --- PASS: TestAgent_grpcInjectAddr/ipv4_web_svc (0.00s) --- PASS: TestAgent_grpcInjectAddr/ipv4_no_svc (0.00s) --- PASS: TestAgent_grpcInjectAddr/ipv6_no_svc (0.00s) --- PASS: TestAgent_grpcInjectAddr/ipv6_web_svc (0.00s) --- PASS: TestAgent_grpcInjectAddr/zone_ipv6_web_svc (0.00s) --- PASS: TestAgent_grpcInjectAddr/ipv6_literal_web_svc (0.00s) --- PASS: TestAgent_grpcInjectAddr/ipv6_injected_into_ipv6_url (0.00s) --- PASS: TestAgent_grpcInjectAddr/ipv6_injected_into_ipv6_url_with_svc (0.00s) --- PASS: TestAgent_grpcInjectAddr/ipv6_injected_into_ipv6_url_with_special (0.00s) === RUN TestAgent_httpInjectAddr === RUN TestAgent_httpInjectAddr/localhost_health === RUN TestAgent_httpInjectAddr/https_localhost_health === RUN TestAgent_httpInjectAddr/https_ipv4_health === RUN TestAgent_httpInjectAddr/https_ipv4_without_path === RUN TestAgent_httpInjectAddr/https_ipv6_health === RUN TestAgent_httpInjectAddr/https_ipv6_with_zone === RUN TestAgent_httpInjectAddr/https_ipv6_literal === RUN TestAgent_httpInjectAddr/https_ipv6_without_path === RUN TestAgent_httpInjectAddr/ipv6_injected_into_ipv6_url === RUN TestAgent_httpInjectAddr/ipv6_with_brackets_injected_into_ipv6_url === RUN TestAgent_httpInjectAddr/short_domain_health === RUN TestAgent_httpInjectAddr/nested_url_in_query --- PASS: TestAgent_httpInjectAddr (0.02s) --- PASS: TestAgent_httpInjectAddr/localhost_health (0.00s) --- PASS: TestAgent_httpInjectAddr/https_localhost_health (0.00s) --- PASS: TestAgent_httpInjectAddr/https_ipv4_health (0.00s) --- PASS: TestAgent_httpInjectAddr/https_ipv4_without_path (0.00s) --- PASS: TestAgent_httpInjectAddr/https_ipv6_health (0.00s) --- PASS: TestAgent_httpInjectAddr/https_ipv6_with_zone (0.00s) --- PASS: TestAgent_httpInjectAddr/https_ipv6_literal (0.00s) --- PASS: TestAgent_httpInjectAddr/https_ipv6_without_path (0.00s) --- PASS: TestAgent_httpInjectAddr/ipv6_injected_into_ipv6_url (0.00s) --- PASS: TestAgent_httpInjectAddr/ipv6_with_brackets_injected_into_ipv6_url (0.00s) --- PASS: TestAgent_httpInjectAddr/short_domain_health (0.00s) --- PASS: TestAgent_httpInjectAddr/nested_url_in_query (0.00s) === RUN TestDefaultIfEmpty --- PASS: TestDefaultIfEmpty (0.00s) === RUN TestConfigSourceFromName === RUN TestConfigSourceFromName/local === RUN TestConfigSourceFromName/remote === RUN TestConfigSourceFromName/#00 === RUN TestConfigSourceFromName/LOCAL === RUN TestConfigSourceFromName/REMOTE === RUN TestConfigSourceFromName/garbage === RUN TestConfigSourceFromName/_ --- PASS: TestConfigSourceFromName (0.02s) --- PASS: TestConfigSourceFromName/local (0.00s) --- PASS: TestConfigSourceFromName/remote (0.00s) --- PASS: TestConfigSourceFromName/#00 (0.00s) --- PASS: TestConfigSourceFromName/LOCAL (0.00s) --- PASS: TestConfigSourceFromName/REMOTE (0.00s) --- PASS: TestConfigSourceFromName/garbage (0.00s) --- PASS: TestConfigSourceFromName/_ (0.00s) === RUN TestAgent_RerouteExistingHTTPChecks === PAUSE TestAgent_RerouteExistingHTTPChecks === RUN TestAgent_RerouteNewHTTPChecks === PAUSE TestAgent_RerouteNewHTTPChecks === RUN TestAgentCache_serviceInConfigFile_initialFetchErrors_Issue6521 agent_test.go:4062: DM-skipped --- SKIP: TestAgentCache_serviceInConfigFile_initialFetchErrors_Issue6521 (0.00s) === RUN TestBlacklist === PAUSE TestBlacklist === RUN TestCatalogRegister_Service_InvalidAddress === PAUSE TestCatalogRegister_Service_InvalidAddress === RUN TestCatalogDeregister === PAUSE TestCatalogDeregister === RUN TestCatalogDatacenters === PAUSE TestCatalogDatacenters === RUN TestCatalogNodes === PAUSE TestCatalogNodes === RUN TestCatalogNodes_MetaFilter === PAUSE TestCatalogNodes_MetaFilter === RUN TestCatalogNodes_Filter === PAUSE TestCatalogNodes_Filter === RUN TestCatalogNodes_WanTranslation catalog_endpoint_test.go:194: DM-skipped --- SKIP: TestCatalogNodes_WanTranslation (0.00s) === RUN TestCatalogNodes_Blocking === PAUSE TestCatalogNodes_Blocking === RUN TestCatalogNodes_DistanceSort === PAUSE TestCatalogNodes_DistanceSort === RUN TestCatalogServices === PAUSE TestCatalogServices === RUN TestCatalogServices_NodeMetaFilter === PAUSE TestCatalogServices_NodeMetaFilter === RUN TestCatalogRegister_checkRegistration === PAUSE TestCatalogRegister_checkRegistration === RUN TestCatalogServiceNodes === PAUSE TestCatalogServiceNodes === RUN TestCatalogServiceNodes_NodeMetaFilter === PAUSE TestCatalogServiceNodes_NodeMetaFilter === RUN TestCatalogServiceNodes_Filter === PAUSE TestCatalogServiceNodes_Filter === RUN TestCatalogServiceNodes_WanTranslation catalog_endpoint_test.go:813: DM-skipped --- SKIP: TestCatalogServiceNodes_WanTranslation (0.00s) === RUN TestCatalogServiceNodes_DistanceSort === PAUSE TestCatalogServiceNodes_DistanceSort === RUN TestCatalogServiceNodes_ConnectProxy === PAUSE TestCatalogServiceNodes_ConnectProxy === RUN TestCatalogConnectServiceNodes_good === PAUSE TestCatalogConnectServiceNodes_good === RUN TestCatalogConnectServiceNodes_Filter catalog_endpoint_test.go:1043: DM-skipped --- SKIP: TestCatalogConnectServiceNodes_Filter (0.00s) === RUN TestCatalogNodeServices === PAUSE TestCatalogNodeServices === RUN TestCatalogNodeServiceList === PAUSE TestCatalogNodeServiceList === RUN TestCatalogNodeServices_Filter === PAUSE TestCatalogNodeServices_Filter === RUN TestCatalogNodeServices_ConnectProxy === PAUSE TestCatalogNodeServices_ConnectProxy === RUN TestCatalogNodeServices_WanTranslation catalog_endpoint_test.go:1240: DM-skipped --- SKIP: TestCatalogNodeServices_WanTranslation (0.00s) === RUN TestConfig_Get === PAUSE TestConfig_Get === RUN TestConfig_Delete === PAUSE TestConfig_Delete === RUN TestConfig_Apply === PAUSE TestConfig_Apply === RUN TestConfig_Apply_ProxyDefaultsMeshGateway === PAUSE TestConfig_Apply_ProxyDefaultsMeshGateway === RUN TestConfig_Apply_CAS === PAUSE TestConfig_Apply_CAS === RUN TestConfig_Apply_Decoding === PAUSE TestConfig_Apply_Decoding === RUN TestConfig_Apply_ProxyDefaultsExpose === PAUSE TestConfig_Apply_ProxyDefaultsExpose === RUN TestConnectCARoots_empty === PAUSE TestConnectCARoots_empty === RUN TestConnectCARoots_list === PAUSE TestConnectCARoots_list === RUN TestConnectCAConfig === PAUSE TestConnectCAConfig === RUN TestCoordinate_Disabled_Response === PAUSE TestCoordinate_Disabled_Response === RUN TestCoordinate_Datacenters coordinate_endpoint_test.go:54: DM-skipped --- SKIP: TestCoordinate_Datacenters (0.00s) === RUN TestCoordinate_Nodes coordinate_endpoint_test.go:81: DM-skipped --- SKIP: TestCoordinate_Nodes (0.00s) === RUN TestCoordinate_Node === PAUSE TestCoordinate_Node === RUN TestCoordinate_Update === PAUSE TestCoordinate_Update === RUN TestCoordinate_Update_ACLDeny === PAUSE TestCoordinate_Update_ACLDeny === RUN TestDiscoveryChainRead === PAUSE TestDiscoveryChainRead === RUN TestRecursorAddr === PAUSE TestRecursorAddr === RUN TestEncodeKVasRFC1464 --- PASS: TestEncodeKVasRFC1464 (0.00s) === RUN TestDNS_Over_TCP === PAUSE TestDNS_Over_TCP === RUN TestDNS_EmptyAltDomain === PAUSE TestDNS_EmptyAltDomain === RUN TestDNS_NodeLookup dns_test.go:207: DM-skipped --- SKIP: TestDNS_NodeLookup (0.00s) === RUN TestDNS_CaseInsensitiveNodeLookup === PAUSE TestDNS_CaseInsensitiveNodeLookup === RUN TestDNS_NodeLookup_PeriodName === PAUSE TestDNS_NodeLookup_PeriodName === RUN TestDNS_NodeLookup_AAAA === PAUSE TestDNS_NodeLookup_AAAA === RUN TestDNSCycleRecursorCheck === PAUSE TestDNSCycleRecursorCheck === RUN TestDNSCycleRecursorCheckAllFail dns_test.go:452: DM-skipped --- SKIP: TestDNSCycleRecursorCheckAllFail (0.00s) === RUN TestDNS_NodeLookup_CNAME === PAUSE TestDNS_NodeLookup_CNAME === RUN TestDNS_NodeLookup_TXT writer.go:29: 2021-01-29T19:32:11.381Z [WARN] TestDNS_NodeLookup_TXT: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:11.383Z [DEBUG] TestDNS_NodeLookup_TXT.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:11.385Z [DEBUG] TestDNS_NodeLookup_TXT.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:11.391Z [INFO] TestDNS_NodeLookup_TXT.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6db876a3-3d95-dfc6-7300-54e0e1656d07 Address:127.0.0.1:30059}]" writer.go:29: 2021-01-29T19:32:11.393Z [INFO] TestDNS_NodeLookup_TXT.server.raft: entering follower state: follower="Node at 127.0.0.1:30059 [Follower]" leader= writer.go:29: 2021-01-29T19:32:11.396Z [INFO] TestDNS_NodeLookup_TXT.server.serf.wan: serf: EventMemberJoin: Node-6db876a3-3d95-dfc6-7300-54e0e1656d07.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:11.416Z [INFO] TestDNS_NodeLookup_TXT.server.serf.lan: serf: EventMemberJoin: Node-6db876a3-3d95-dfc6-7300-54e0e1656d07 127.0.0.1 writer.go:29: 2021-01-29T19:32:11.424Z [INFO] TestDNS_NodeLookup_TXT.server: Handled event for server in area: event=member-join server=Node-6db876a3-3d95-dfc6-7300-54e0e1656d07.dc1 area=wan writer.go:29: 2021-01-29T19:32:11.424Z [INFO] TestDNS_NodeLookup_TXT.server: Adding LAN server: server="Node-6db876a3-3d95-dfc6-7300-54e0e1656d07 (Addr: tcp/127.0.0.1:30059) (DC: dc1)" writer.go:29: 2021-01-29T19:32:11.430Z [INFO] TestDNS_NodeLookup_TXT: Started DNS server: address=127.0.0.1:30054 network=tcp writer.go:29: 2021-01-29T19:32:11.433Z [INFO] TestDNS_NodeLookup_TXT: Started DNS server: address=127.0.0.1:30054 network=udp writer.go:29: 2021-01-29T19:32:11.438Z [INFO] TestDNS_NodeLookup_TXT: Started HTTP server: address=127.0.0.1:30055 network=tcp writer.go:29: 2021-01-29T19:32:11.439Z [INFO] TestDNS_NodeLookup_TXT: started state syncer writer.go:29: 2021-01-29T19:32:11.441Z [WARN] TestDNS_NodeLookup_TXT.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:11.442Z [INFO] TestDNS_NodeLookup_TXT.server.raft: entering candidate state: node="Node at 127.0.0.1:30059 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:11.444Z [DEBUG] TestDNS_NodeLookup_TXT.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:11.446Z [DEBUG] TestDNS_NodeLookup_TXT.server.raft: vote granted: from=6db876a3-3d95-dfc6-7300-54e0e1656d07 term=2 tally=1 writer.go:29: 2021-01-29T19:32:11.447Z [INFO] TestDNS_NodeLookup_TXT.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:11.448Z [INFO] TestDNS_NodeLookup_TXT.server.raft: entering leader state: leader="Node at 127.0.0.1:30059 [Leader]" writer.go:29: 2021-01-29T19:32:11.449Z [INFO] TestDNS_NodeLookup_TXT.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:11.450Z [INFO] TestDNS_NodeLookup_TXT.server: New leader elected: payload=Node-6db876a3-3d95-dfc6-7300-54e0e1656d07 writer.go:29: 2021-01-29T19:32:11.461Z [DEBUG] TestDNS_NodeLookup_TXT.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30059 writer.go:29: 2021-01-29T19:32:11.483Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:11.487Z [INFO] TestDNS_NodeLookup_TXT.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:11.488Z [INFO] TestDNS_NodeLookup_TXT.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:11.490Z [DEBUG] TestDNS_NodeLookup_TXT.server: Skipping self join check for node since the cluster is too small: node=Node-6db876a3-3d95-dfc6-7300-54e0e1656d07 writer.go:29: 2021-01-29T19:32:11.491Z [INFO] TestDNS_NodeLookup_TXT.server: member joined, marking health alive: member=Node-6db876a3-3d95-dfc6-7300-54e0e1656d07 writer.go:29: 2021-01-29T19:32:11.518Z [DEBUG] TestDNS_NodeLookup_TXT.dns: request served from client: name=google.node.consul. type=TXT class=IN latency=208.644µs client=127.0.0.1:34146 client_network=udp writer.go:29: 2021-01-29T19:32:11.518Z [INFO] TestDNS_NodeLookup_TXT: Requesting shutdown writer.go:29: 2021-01-29T19:32:11.522Z [INFO] TestDNS_NodeLookup_TXT.server: shutting down server writer.go:29: 2021-01-29T19:32:11.523Z [DEBUG] TestDNS_NodeLookup_TXT.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:11.525Z [WARN] TestDNS_NodeLookup_TXT.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:11.527Z [ERROR] TestDNS_NodeLookup_TXT.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:11.528Z [DEBUG] TestDNS_NodeLookup_TXT.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:11.530Z [WARN] TestDNS_NodeLookup_TXT.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:11.533Z [INFO] TestDNS_NodeLookup_TXT.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:11.535Z [INFO] TestDNS_NodeLookup_TXT: consul server down writer.go:29: 2021-01-29T19:32:11.537Z [INFO] TestDNS_NodeLookup_TXT: shutdown complete writer.go:29: 2021-01-29T19:32:11.539Z [INFO] TestDNS_NodeLookup_TXT: Stopping server: protocol=DNS address=127.0.0.1:30054 network=tcp writer.go:29: 2021-01-29T19:32:11.542Z [INFO] TestDNS_NodeLookup_TXT: Stopping server: protocol=DNS address=127.0.0.1:30054 network=udp writer.go:29: 2021-01-29T19:32:11.544Z [INFO] TestDNS_NodeLookup_TXT: Stopping server: protocol=HTTP address=127.0.0.1:30055 network=tcp writer.go:29: 2021-01-29T19:32:12.046Z [INFO] TestDNS_NodeLookup_TXT: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:12.047Z [INFO] TestDNS_NodeLookup_TXT: Endpoints down --- PASS: TestDNS_NodeLookup_TXT (0.68s) === RUN TestDNS_NodeLookup_TXT_DontSuppress writer.go:29: 2021-01-29T19:32:12.058Z [WARN] TestDNS_NodeLookup_TXT_DontSuppress: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:12.059Z [DEBUG] TestDNS_NodeLookup_TXT_DontSuppress.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:12.060Z [DEBUG] TestDNS_NodeLookup_TXT_DontSuppress.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:12.065Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ce1473c4-53fa-ec08-c229-e6731212e7a5 Address:127.0.0.1:30065}]" writer.go:29: 2021-01-29T19:32:12.069Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server.serf.wan: serf: EventMemberJoin: Node-ce1473c4-53fa-ec08-c229-e6731212e7a5.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:12.070Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server.raft: entering follower state: follower="Node at 127.0.0.1:30065 [Follower]" leader= writer.go:29: 2021-01-29T19:32:12.076Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server.serf.lan: serf: EventMemberJoin: Node-ce1473c4-53fa-ec08-c229-e6731212e7a5 127.0.0.1 writer.go:29: 2021-01-29T19:32:12.082Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server: Adding LAN server: server="Node-ce1473c4-53fa-ec08-c229-e6731212e7a5 (Addr: tcp/127.0.0.1:30065) (DC: dc1)" writer.go:29: 2021-01-29T19:32:12.084Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server: Handled event for server in area: event=member-join server=Node-ce1473c4-53fa-ec08-c229-e6731212e7a5.dc1 area=wan writer.go:29: 2021-01-29T19:32:12.087Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: Started DNS server: address=127.0.0.1:30060 network=udp writer.go:29: 2021-01-29T19:32:12.089Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: Started DNS server: address=127.0.0.1:30060 network=tcp writer.go:29: 2021-01-29T19:32:12.112Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: Started HTTP server: address=127.0.0.1:30061 network=tcp writer.go:29: 2021-01-29T19:32:12.113Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: started state syncer writer.go:29: 2021-01-29T19:32:12.145Z [WARN] TestDNS_NodeLookup_TXT_DontSuppress.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:12.146Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server.raft: entering candidate state: node="Node at 127.0.0.1:30065 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:12.149Z [DEBUG] TestDNS_NodeLookup_TXT_DontSuppress.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:12.151Z [DEBUG] TestDNS_NodeLookup_TXT_DontSuppress.server.raft: vote granted: from=ce1473c4-53fa-ec08-c229-e6731212e7a5 term=2 tally=1 writer.go:29: 2021-01-29T19:32:12.152Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:12.154Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server.raft: entering leader state: leader="Node at 127.0.0.1:30065 [Leader]" writer.go:29: 2021-01-29T19:32:12.155Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:12.156Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server: New leader elected: payload=Node-ce1473c4-53fa-ec08-c229-e6731212e7a5 writer.go:29: 2021-01-29T19:32:12.157Z [DEBUG] TestDNS_NodeLookup_TXT_DontSuppress.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30065 writer.go:29: 2021-01-29T19:32:12.164Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:12.170Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:12.171Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:12.172Z [DEBUG] TestDNS_NodeLookup_TXT_DontSuppress.server: Skipping self join check for node since the cluster is too small: node=Node-ce1473c4-53fa-ec08-c229-e6731212e7a5 writer.go:29: 2021-01-29T19:32:12.173Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server: member joined, marking health alive: member=Node-ce1473c4-53fa-ec08-c229-e6731212e7a5 writer.go:29: 2021-01-29T19:32:12.239Z [DEBUG] TestDNS_NodeLookup_TXT_DontSuppress.dns: request served from client: name=google.node.consul. type=TXT class=IN latency=143.687µs client=127.0.0.1:41244 client_network=udp writer.go:29: 2021-01-29T19:32:12.239Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: Requesting shutdown writer.go:29: 2021-01-29T19:32:12.243Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server: shutting down server writer.go:29: 2021-01-29T19:32:12.245Z [DEBUG] TestDNS_NodeLookup_TXT_DontSuppress.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:12.246Z [WARN] TestDNS_NodeLookup_TXT_DontSuppress.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:12.249Z [ERROR] TestDNS_NodeLookup_TXT_DontSuppress.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:12.251Z [DEBUG] TestDNS_NodeLookup_TXT_DontSuppress.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:12.259Z [WARN] TestDNS_NodeLookup_TXT_DontSuppress.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:12.264Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:12.266Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: consul server down writer.go:29: 2021-01-29T19:32:12.268Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: shutdown complete writer.go:29: 2021-01-29T19:32:12.270Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: Stopping server: protocol=DNS address=127.0.0.1:30060 network=tcp writer.go:29: 2021-01-29T19:32:12.272Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: Stopping server: protocol=DNS address=127.0.0.1:30060 network=udp writer.go:29: 2021-01-29T19:32:12.273Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: Stopping server: protocol=HTTP address=127.0.0.1:30061 network=tcp writer.go:29: 2021-01-29T19:32:12.775Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:12.777Z [INFO] TestDNS_NodeLookup_TXT_DontSuppress: Endpoints down --- PASS: TestDNS_NodeLookup_TXT_DontSuppress (0.73s) === RUN TestDNS_NodeLookup_ANY writer.go:29: 2021-01-29T19:32:12.865Z [WARN] TestDNS_NodeLookup_ANY: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:12.867Z [DEBUG] TestDNS_NodeLookup_ANY.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:12.870Z [DEBUG] TestDNS_NodeLookup_ANY.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:12.900Z [INFO] TestDNS_NodeLookup_ANY.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:263c853c-3490-f7d0-16e8-763a25490004 Address:127.0.0.1:30071}]" writer.go:29: 2021-01-29T19:32:12.904Z [INFO] TestDNS_NodeLookup_ANY.server.serf.wan: serf: EventMemberJoin: Node-263c853c-3490-f7d0-16e8-763a25490004.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:12.914Z [INFO] TestDNS_NodeLookup_ANY.server.raft: entering follower state: follower="Node at 127.0.0.1:30071 [Follower]" leader= writer.go:29: 2021-01-29T19:32:12.942Z [INFO] TestDNS_NodeLookup_ANY.server.serf.lan: serf: EventMemberJoin: Node-263c853c-3490-f7d0-16e8-763a25490004 127.0.0.1 writer.go:29: 2021-01-29T19:32:12.947Z [INFO] TestDNS_NodeLookup_ANY: Started DNS server: address=127.0.0.1:30066 network=udp writer.go:29: 2021-01-29T19:32:12.950Z [INFO] TestDNS_NodeLookup_ANY.server: Adding LAN server: server="Node-263c853c-3490-f7d0-16e8-763a25490004 (Addr: tcp/127.0.0.1:30071) (DC: dc1)" writer.go:29: 2021-01-29T19:32:12.951Z [INFO] TestDNS_NodeLookup_ANY.server: Handled event for server in area: event=member-join server=Node-263c853c-3490-f7d0-16e8-763a25490004.dc1 area=wan writer.go:29: 2021-01-29T19:32:12.954Z [INFO] TestDNS_NodeLookup_ANY: Started DNS server: address=127.0.0.1:30066 network=tcp writer.go:29: 2021-01-29T19:32:12.969Z [INFO] TestDNS_NodeLookup_ANY: Started HTTP server: address=127.0.0.1:30067 network=tcp writer.go:29: 2021-01-29T19:32:12.970Z [INFO] TestDNS_NodeLookup_ANY: started state syncer writer.go:29: 2021-01-29T19:32:12.981Z [WARN] TestDNS_NodeLookup_ANY.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:12.983Z [INFO] TestDNS_NodeLookup_ANY.server.raft: entering candidate state: node="Node at 127.0.0.1:30071 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:12.989Z [DEBUG] TestDNS_NodeLookup_ANY.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:12.998Z [DEBUG] TestDNS_NodeLookup_ANY.server.raft: vote granted: from=263c853c-3490-f7d0-16e8-763a25490004 term=2 tally=1 writer.go:29: 2021-01-29T19:32:12.999Z [INFO] TestDNS_NodeLookup_ANY.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:13.001Z [INFO] TestDNS_NodeLookup_ANY.server.raft: entering leader state: leader="Node at 127.0.0.1:30071 [Leader]" writer.go:29: 2021-01-29T19:32:13.002Z [INFO] TestDNS_NodeLookup_ANY.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:13.003Z [INFO] TestDNS_NodeLookup_ANY.server: New leader elected: payload=Node-263c853c-3490-f7d0-16e8-763a25490004 writer.go:29: 2021-01-29T19:32:13.004Z [DEBUG] TestDNS_NodeLookup_ANY.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30071 writer.go:29: 2021-01-29T19:32:13.017Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:13.025Z [INFO] TestDNS_NodeLookup_ANY.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:13.027Z [INFO] TestDNS_NodeLookup_ANY.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:13.029Z [DEBUG] TestDNS_NodeLookup_ANY.server: Skipping self join check for node since the cluster is too small: node=Node-263c853c-3490-f7d0-16e8-763a25490004 writer.go:29: 2021-01-29T19:32:13.031Z [INFO] TestDNS_NodeLookup_ANY.server: member joined, marking health alive: member=Node-263c853c-3490-f7d0-16e8-763a25490004 writer.go:29: 2021-01-29T19:32:13.044Z [DEBUG] TestDNS_NodeLookup_ANY: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:13.047Z [INFO] TestDNS_NodeLookup_ANY: Synced node info writer.go:29: 2021-01-29T19:32:13.105Z [DEBUG] TestDNS_NodeLookup_ANY.dns: request served from client: name=bar.node.consul. type=ANY class=IN latency=172.45µs client=127.0.0.1:49123 client_network=udp writer.go:29: 2021-01-29T19:32:13.105Z [INFO] TestDNS_NodeLookup_ANY: Requesting shutdown writer.go:29: 2021-01-29T19:32:13.109Z [INFO] TestDNS_NodeLookup_ANY.server: shutting down server writer.go:29: 2021-01-29T19:32:13.110Z [DEBUG] TestDNS_NodeLookup_ANY.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:13.112Z [WARN] TestDNS_NodeLookup_ANY.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:13.112Z [DEBUG] TestDNS_NodeLookup_ANY.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:13.116Z [WARN] TestDNS_NodeLookup_ANY.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:13.118Z [INFO] TestDNS_NodeLookup_ANY.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:13.119Z [INFO] TestDNS_NodeLookup_ANY: consul server down writer.go:29: 2021-01-29T19:32:13.121Z [INFO] TestDNS_NodeLookup_ANY: shutdown complete writer.go:29: 2021-01-29T19:32:13.123Z [INFO] TestDNS_NodeLookup_ANY: Stopping server: protocol=DNS address=127.0.0.1:30066 network=tcp writer.go:29: 2021-01-29T19:32:13.124Z [INFO] TestDNS_NodeLookup_ANY: Stopping server: protocol=DNS address=127.0.0.1:30066 network=udp writer.go:29: 2021-01-29T19:32:13.126Z [INFO] TestDNS_NodeLookup_ANY: Stopping server: protocol=HTTP address=127.0.0.1:30067 network=tcp writer.go:29: 2021-01-29T19:32:13.628Z [INFO] TestDNS_NodeLookup_ANY: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:13.631Z [INFO] TestDNS_NodeLookup_ANY: Endpoints down --- PASS: TestDNS_NodeLookup_ANY (0.85s) === RUN TestDNS_NodeLookup_ANY_DontSuppressTXT writer.go:29: 2021-01-29T19:32:13.645Z [WARN] TestDNS_NodeLookup_ANY_DontSuppressTXT: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:13.646Z [DEBUG] TestDNS_NodeLookup_ANY_DontSuppressTXT.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:13.648Z [DEBUG] TestDNS_NodeLookup_ANY_DontSuppressTXT.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:13.652Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0cf1c9d8-b2b3-332e-fca9-e75ca9aa6285 Address:127.0.0.1:30077}]" writer.go:29: 2021-01-29T19:32:13.654Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.raft: entering follower state: follower="Node at 127.0.0.1:30077 [Follower]" leader= writer.go:29: 2021-01-29T19:32:13.657Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.serf.wan: serf: EventMemberJoin: Node-0cf1c9d8-b2b3-332e-fca9-e75ca9aa6285.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:13.659Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.serf.lan: serf: EventMemberJoin: Node-0cf1c9d8-b2b3-332e-fca9-e75ca9aa6285 127.0.0.1 writer.go:29: 2021-01-29T19:32:13.662Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server: Adding LAN server: server="Node-0cf1c9d8-b2b3-332e-fca9-e75ca9aa6285 (Addr: tcp/127.0.0.1:30077) (DC: dc1)" writer.go:29: 2021-01-29T19:32:13.663Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server: Handled event for server in area: event=member-join server=Node-0cf1c9d8-b2b3-332e-fca9-e75ca9aa6285.dc1 area=wan writer.go:29: 2021-01-29T19:32:13.665Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: Started DNS server: address=127.0.0.1:30072 network=tcp writer.go:29: 2021-01-29T19:32:13.666Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: Started DNS server: address=127.0.0.1:30072 network=udp writer.go:29: 2021-01-29T19:32:13.669Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: Started HTTP server: address=127.0.0.1:30073 network=tcp writer.go:29: 2021-01-29T19:32:13.669Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: started state syncer writer.go:29: 2021-01-29T19:32:13.696Z [WARN] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:13.697Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.raft: entering candidate state: node="Node at 127.0.0.1:30077 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:13.698Z [DEBUG] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:13.699Z [DEBUG] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.raft: vote granted: from=0cf1c9d8-b2b3-332e-fca9-e75ca9aa6285 term=2 tally=1 writer.go:29: 2021-01-29T19:32:13.700Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:13.701Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.raft: entering leader state: leader="Node at 127.0.0.1:30077 [Leader]" writer.go:29: 2021-01-29T19:32:13.702Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:13.702Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server: New leader elected: payload=Node-0cf1c9d8-b2b3-332e-fca9-e75ca9aa6285 writer.go:29: 2021-01-29T19:32:13.703Z [DEBUG] TestDNS_NodeLookup_ANY_DontSuppressTXT.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30077 writer.go:29: 2021-01-29T19:32:13.706Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: Synced node info writer.go:29: 2021-01-29T19:32:13.710Z [DEBUG] TestDNS_NodeLookup_ANY_DontSuppressTXT: Node info in sync writer.go:29: 2021-01-29T19:32:13.710Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:13.725Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:13.727Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:13.728Z [DEBUG] TestDNS_NodeLookup_ANY_DontSuppressTXT.server: Skipping self join check for node since the cluster is too small: node=Node-0cf1c9d8-b2b3-332e-fca9-e75ca9aa6285 writer.go:29: 2021-01-29T19:32:13.730Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server: member joined, marking health alive: member=Node-0cf1c9d8-b2b3-332e-fca9-e75ca9aa6285 writer.go:29: 2021-01-29T19:32:14.099Z [DEBUG] TestDNS_NodeLookup_ANY_DontSuppressTXT.dns: request served from client: name=bar.node.consul. type=ANY class=IN latency=108.138µs client=127.0.0.1:43148 client_network=udp writer.go:29: 2021-01-29T19:32:14.099Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: Requesting shutdown writer.go:29: 2021-01-29T19:32:14.101Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server: shutting down server writer.go:29: 2021-01-29T19:32:14.102Z [DEBUG] TestDNS_NodeLookup_ANY_DontSuppressTXT.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:14.103Z [WARN] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:14.103Z [DEBUG] TestDNS_NodeLookup_ANY_DontSuppressTXT.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:14.105Z [WARN] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:14.112Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:14.113Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: consul server down writer.go:29: 2021-01-29T19:32:14.118Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: shutdown complete writer.go:29: 2021-01-29T19:32:14.119Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: Stopping server: protocol=DNS address=127.0.0.1:30072 network=tcp writer.go:29: 2021-01-29T19:32:14.120Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: Stopping server: protocol=DNS address=127.0.0.1:30072 network=udp writer.go:29: 2021-01-29T19:32:14.121Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: Stopping server: protocol=HTTP address=127.0.0.1:30073 network=tcp writer.go:29: 2021-01-29T19:32:14.622Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:14.624Z [INFO] TestDNS_NodeLookup_ANY_DontSuppressTXT: Endpoints down --- PASS: TestDNS_NodeLookup_ANY_DontSuppressTXT (0.99s) === RUN TestDNS_NodeLookup_A_SuppressTXT writer.go:29: 2021-01-29T19:32:14.642Z [WARN] TestDNS_NodeLookup_A_SuppressTXT: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:14.644Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:14.648Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:14.656Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fab9a43f-b44d-6fa0-6a60-8ce7a2d4fe37 Address:127.0.0.1:30083}]" writer.go:29: 2021-01-29T19:32:14.657Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server.raft: entering follower state: follower="Node at 127.0.0.1:30083 [Follower]" leader= writer.go:29: 2021-01-29T19:32:14.660Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server.serf.wan: serf: EventMemberJoin: Node-fab9a43f-b44d-6fa0-6a60-8ce7a2d4fe37.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:14.663Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server.serf.lan: serf: EventMemberJoin: Node-fab9a43f-b44d-6fa0-6a60-8ce7a2d4fe37 127.0.0.1 writer.go:29: 2021-01-29T19:32:14.666Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server: Adding LAN server: server="Node-fab9a43f-b44d-6fa0-6a60-8ce7a2d4fe37 (Addr: tcp/127.0.0.1:30083) (DC: dc1)" writer.go:29: 2021-01-29T19:32:14.667Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server: Handled event for server in area: event=member-join server=Node-fab9a43f-b44d-6fa0-6a60-8ce7a2d4fe37.dc1 area=wan writer.go:29: 2021-01-29T19:32:14.670Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: Started DNS server: address=127.0.0.1:30078 network=tcp writer.go:29: 2021-01-29T19:32:14.672Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: Started DNS server: address=127.0.0.1:30078 network=udp writer.go:29: 2021-01-29T19:32:14.673Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: Started HTTP server: address=127.0.0.1:30079 network=tcp writer.go:29: 2021-01-29T19:32:14.674Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: started state syncer writer.go:29: 2021-01-29T19:32:14.712Z [WARN] TestDNS_NodeLookup_A_SuppressTXT.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:14.714Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server.raft: entering candidate state: node="Node at 127.0.0.1:30083 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:14.716Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:14.717Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT.server.raft: vote granted: from=fab9a43f-b44d-6fa0-6a60-8ce7a2d4fe37 term=2 tally=1 writer.go:29: 2021-01-29T19:32:14.719Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:14.720Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server.raft: entering leader state: leader="Node at 127.0.0.1:30083 [Leader]" writer.go:29: 2021-01-29T19:32:14.722Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:14.722Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server: New leader elected: payload=Node-fab9a43f-b44d-6fa0-6a60-8ce7a2d4fe37 writer.go:29: 2021-01-29T19:32:14.726Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30083 writer.go:29: 2021-01-29T19:32:14.731Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:14.743Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:14.745Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:14.750Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT.server: Skipping self join check for node since the cluster is too small: node=Node-fab9a43f-b44d-6fa0-6a60-8ce7a2d4fe37 writer.go:29: 2021-01-29T19:32:14.751Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server: member joined, marking health alive: member=Node-fab9a43f-b44d-6fa0-6a60-8ce7a2d4fe37 writer.go:29: 2021-01-29T19:32:14.935Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:14.948Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: Synced node info writer.go:29: 2021-01-29T19:32:14.950Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT: Node info in sync writer.go:29: 2021-01-29T19:32:15.101Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT.dns: request served from client: name=bar.node.consul. type=A class=IN latency=76.034µs client=127.0.0.1:46184 client_network=udp writer.go:29: 2021-01-29T19:32:15.103Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: Requesting shutdown writer.go:29: 2021-01-29T19:32:15.104Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server: shutting down server writer.go:29: 2021-01-29T19:32:15.105Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:15.105Z [WARN] TestDNS_NodeLookup_A_SuppressTXT.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:15.107Z [DEBUG] TestDNS_NodeLookup_A_SuppressTXT.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:15.110Z [WARN] TestDNS_NodeLookup_A_SuppressTXT.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:15.112Z [INFO] TestDNS_NodeLookup_A_SuppressTXT.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:15.113Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: consul server down writer.go:29: 2021-01-29T19:32:15.114Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: shutdown complete writer.go:29: 2021-01-29T19:32:15.115Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: Stopping server: protocol=DNS address=127.0.0.1:30078 network=tcp writer.go:29: 2021-01-29T19:32:15.116Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: Stopping server: protocol=DNS address=127.0.0.1:30078 network=udp writer.go:29: 2021-01-29T19:32:15.117Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: Stopping server: protocol=HTTP address=127.0.0.1:30079 network=tcp writer.go:29: 2021-01-29T19:32:15.618Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:15.619Z [INFO] TestDNS_NodeLookup_A_SuppressTXT: Endpoints down --- PASS: TestDNS_NodeLookup_A_SuppressTXT (0.99s) === RUN TestDNS_EDNS0 === PAUSE TestDNS_EDNS0 === RUN TestDNS_EDNS0_ECS dns_test.go:791: DM-skipped --- SKIP: TestDNS_EDNS0_ECS (0.00s) === RUN TestDNS_ReverseLookup === PAUSE TestDNS_ReverseLookup === RUN TestDNS_ReverseLookup_CustomDomain === PAUSE TestDNS_ReverseLookup_CustomDomain === RUN TestDNS_ReverseLookup_IPV6 === PAUSE TestDNS_ReverseLookup_IPV6 === RUN TestDNS_ServiceReverseLookup dns_test.go:1006: DM-skipped --- SKIP: TestDNS_ServiceReverseLookup (0.00s) === RUN TestDNS_ServiceReverseLookup_IPV6 === PAUSE TestDNS_ServiceReverseLookup_IPV6 === RUN TestDNS_ServiceReverseLookup_CustomDomain === PAUSE TestDNS_ServiceReverseLookup_CustomDomain === RUN TestDNS_SOA_Settings === PAUSE TestDNS_SOA_Settings === RUN TestDNS_ServiceReverseLookupNodeAddress === PAUSE TestDNS_ServiceReverseLookupNodeAddress === RUN TestDNS_ServiceLookupNoMultiCNAME dns_test.go:1234: DM-skipped --- SKIP: TestDNS_ServiceLookupNoMultiCNAME (0.00s) === RUN TestDNS_ServiceLookupPreferNoCNAME === PAUSE TestDNS_ServiceLookupPreferNoCNAME === RUN TestDNS_ServiceLookupMultiAddrNoCNAME === PAUSE TestDNS_ServiceLookupMultiAddrNoCNAME === RUN TestDNS_ServiceLookup === PAUSE TestDNS_ServiceLookup === RUN TestDNS_ServiceLookupWithInternalServiceAddress === PAUSE TestDNS_ServiceLookupWithInternalServiceAddress === RUN TestDNS_ConnectServiceLookup === PAUSE TestDNS_ConnectServiceLookup === RUN TestDNS_ExternalServiceLookup === PAUSE TestDNS_ExternalServiceLookup === RUN TestDNS_InifiniteRecursion === PAUSE TestDNS_InifiniteRecursion === RUN TestDNS_ExternalServiceToConsulCNAMELookup === PAUSE TestDNS_ExternalServiceToConsulCNAMELookup === RUN TestDNS_NSRecords dns_test.go:1857: DM-skipped --- SKIP: TestDNS_NSRecords (0.01s) === RUN TestDNS_NSRecords_IPV6 === PAUSE TestDNS_NSRecords_IPV6 === RUN TestDNS_ExternalServiceToConsulCNAMENestedLookup === PAUSE TestDNS_ExternalServiceToConsulCNAMENestedLookup === RUN TestDNS_ServiceLookup_ServiceAddress_A === PAUSE TestDNS_ServiceLookup_ServiceAddress_A === RUN TestDNS_ServiceLookup_ServiceAddress_SRV === PAUSE TestDNS_ServiceLookup_ServiceAddress_SRV === RUN TestDNS_ServiceLookup_ServiceAddressIPV6 === PAUSE TestDNS_ServiceLookup_ServiceAddressIPV6 === RUN TestDNS_ServiceLookup_WanTranslation === PAUSE TestDNS_ServiceLookup_WanTranslation === RUN TestDNS_Lookup_TaggedIPAddresses === PAUSE TestDNS_Lookup_TaggedIPAddresses === RUN TestDNS_CaseInsensitiveServiceLookup === PAUSE TestDNS_CaseInsensitiveServiceLookup === RUN TestDNS_ServiceLookup_TagPeriod === PAUSE TestDNS_ServiceLookup_TagPeriod === RUN TestDNS_PreparedQueryNearIPEDNS === PAUSE TestDNS_PreparedQueryNearIPEDNS === RUN TestDNS_PreparedQueryNearIP === PAUSE TestDNS_PreparedQueryNearIP === RUN TestDNS_ServiceLookup_PreparedQueryNamePeriod === PAUSE TestDNS_ServiceLookup_PreparedQueryNamePeriod === RUN TestDNS_ServiceLookup_Dedup dns_test.go:3213: DM-skipped --- SKIP: TestDNS_ServiceLookup_Dedup (0.00s) === RUN TestDNS_ServiceLookup_Dedup_SRV === PAUSE TestDNS_ServiceLookup_Dedup_SRV === RUN TestDNS_Recurse === PAUSE TestDNS_Recurse === RUN TestDNS_Recurse_Truncation === PAUSE TestDNS_Recurse_Truncation === RUN TestDNS_RecursorTimeout === PAUSE TestDNS_RecursorTimeout === RUN TestDNS_ServiceLookup_FilterCritical === PAUSE TestDNS_ServiceLookup_FilterCritical === RUN TestDNS_ServiceLookup_OnlyFailing === PAUSE TestDNS_ServiceLookup_OnlyFailing === RUN TestDNS_ServiceLookup_OnlyPassing === PAUSE TestDNS_ServiceLookup_OnlyPassing === RUN TestDNS_ServiceLookup_Randomize === PAUSE TestDNS_ServiceLookup_Randomize === RUN TestBinarySearch === PAUSE TestBinarySearch === RUN TestDNS_TCP_and_UDP_Truncate dns_test.go:4108: DM-skipped --- SKIP: TestDNS_TCP_and_UDP_Truncate (0.00s) === RUN TestDNS_ServiceLookup_Truncate === PAUSE TestDNS_ServiceLookup_Truncate === RUN TestDNS_ServiceLookup_LargeResponses === PAUSE TestDNS_ServiceLookup_LargeResponses === RUN TestDNS_ServiceLookup_ARecordLimits dns_test.go:4559: DM-skipped --- SKIP: TestDNS_ServiceLookup_ARecordLimits (0.00s) === RUN TestDNS_ServiceLookup_AnswerLimits === PAUSE TestDNS_ServiceLookup_AnswerLimits === RUN TestDNS_ServiceLookup_CNAME dns_test.go:4704: DM-skipped --- SKIP: TestDNS_ServiceLookup_CNAME (0.00s) === RUN TestDNS_ServiceLookup_ServiceAddress_CNAME === PAUSE TestDNS_ServiceLookup_ServiceAddress_CNAME === RUN TestDNS_NodeLookup_TTL === PAUSE TestDNS_NodeLookup_TTL === RUN TestDNS_ServiceLookup_TTL === PAUSE TestDNS_ServiceLookup_TTL === RUN TestDNS_PreparedQuery_TTL === PAUSE TestDNS_PreparedQuery_TTL === RUN TestDNS_PreparedQuery_Failover dns_test.go:5224: DM-skipped --- SKIP: TestDNS_PreparedQuery_Failover (0.00s) === RUN TestDNS_ServiceLookup_SRV_RFC dns_test.go:5337: DM-skipped --- SKIP: TestDNS_ServiceLookup_SRV_RFC (0.00s) === RUN TestDNS_ServiceLookup_SRV_RFC_TCP_Default === PAUSE TestDNS_ServiceLookup_SRV_RFC_TCP_Default === RUN TestDNS_ServiceLookup_FilterACL === PAUSE TestDNS_ServiceLookup_FilterACL === RUN TestDNS_ServiceLookup_MetaTXT writer.go:29: 2021-01-29T19:32:15.816Z [WARN] TestDNS_ServiceLookup_MetaTXT: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:15.817Z [DEBUG] TestDNS_ServiceLookup_MetaTXT.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:15.818Z [DEBUG] TestDNS_ServiceLookup_MetaTXT.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:15.844Z [INFO] TestDNS_ServiceLookup_MetaTXT.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ee9da84a-6ba9-9d4b-d738-7fe50bed88a9 Address:127.0.0.1:30089}]" writer.go:29: 2021-01-29T19:32:15.847Z [INFO] TestDNS_ServiceLookup_MetaTXT.server.raft: entering follower state: follower="Node at 127.0.0.1:30089 [Follower]" leader= writer.go:29: 2021-01-29T19:32:15.850Z [INFO] TestDNS_ServiceLookup_MetaTXT.server.serf.wan: serf: EventMemberJoin: Node-ee9da84a-6ba9-9d4b-d738-7fe50bed88a9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:15.855Z [INFO] TestDNS_ServiceLookup_MetaTXT.server.serf.lan: serf: EventMemberJoin: Node-ee9da84a-6ba9-9d4b-d738-7fe50bed88a9 127.0.0.1 writer.go:29: 2021-01-29T19:32:15.859Z [INFO] TestDNS_ServiceLookup_MetaTXT.server: Adding LAN server: server="Node-ee9da84a-6ba9-9d4b-d738-7fe50bed88a9 (Addr: tcp/127.0.0.1:30089) (DC: dc1)" writer.go:29: 2021-01-29T19:32:15.859Z [INFO] TestDNS_ServiceLookup_MetaTXT.server: Handled event for server in area: event=member-join server=Node-ee9da84a-6ba9-9d4b-d738-7fe50bed88a9.dc1 area=wan writer.go:29: 2021-01-29T19:32:15.861Z [INFO] TestDNS_ServiceLookup_MetaTXT: Started DNS server: address=127.0.0.1:30084 network=udp writer.go:29: 2021-01-29T19:32:15.861Z [INFO] TestDNS_ServiceLookup_MetaTXT: Started DNS server: address=127.0.0.1:30084 network=tcp writer.go:29: 2021-01-29T19:32:15.863Z [INFO] TestDNS_ServiceLookup_MetaTXT: Started HTTP server: address=127.0.0.1:30085 network=tcp writer.go:29: 2021-01-29T19:32:15.864Z [INFO] TestDNS_ServiceLookup_MetaTXT: started state syncer writer.go:29: 2021-01-29T19:32:15.888Z [WARN] TestDNS_ServiceLookup_MetaTXT.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:15.890Z [INFO] TestDNS_ServiceLookup_MetaTXT.server.raft: entering candidate state: node="Node at 127.0.0.1:30089 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:15.894Z [DEBUG] TestDNS_ServiceLookup_MetaTXT.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:15.895Z [DEBUG] TestDNS_ServiceLookup_MetaTXT.server.raft: vote granted: from=ee9da84a-6ba9-9d4b-d738-7fe50bed88a9 term=2 tally=1 writer.go:29: 2021-01-29T19:32:15.897Z [INFO] TestDNS_ServiceLookup_MetaTXT.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:15.898Z [INFO] TestDNS_ServiceLookup_MetaTXT.server.raft: entering leader state: leader="Node at 127.0.0.1:30089 [Leader]" writer.go:29: 2021-01-29T19:32:15.900Z [INFO] TestDNS_ServiceLookup_MetaTXT.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:15.901Z [INFO] TestDNS_ServiceLookup_MetaTXT.server: New leader elected: payload=Node-ee9da84a-6ba9-9d4b-d738-7fe50bed88a9 writer.go:29: 2021-01-29T19:32:15.906Z [DEBUG] TestDNS_ServiceLookup_MetaTXT.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30089 writer.go:29: 2021-01-29T19:32:15.912Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:15.928Z [INFO] TestDNS_ServiceLookup_MetaTXT.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:15.930Z [INFO] TestDNS_ServiceLookup_MetaTXT.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:15.931Z [DEBUG] TestDNS_ServiceLookup_MetaTXT.server: Skipping self join check for node since the cluster is too small: node=Node-ee9da84a-6ba9-9d4b-d738-7fe50bed88a9 writer.go:29: 2021-01-29T19:32:15.933Z [INFO] TestDNS_ServiceLookup_MetaTXT.server: member joined, marking health alive: member=Node-ee9da84a-6ba9-9d4b-d738-7fe50bed88a9 writer.go:29: 2021-01-29T19:32:16.064Z [DEBUG] TestDNS_ServiceLookup_MetaTXT.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=160.399µs client=127.0.0.1:45952 client_network=udp writer.go:29: 2021-01-29T19:32:16.064Z [INFO] TestDNS_ServiceLookup_MetaTXT: Requesting shutdown writer.go:29: 2021-01-29T19:32:16.067Z [INFO] TestDNS_ServiceLookup_MetaTXT.server: shutting down server writer.go:29: 2021-01-29T19:32:16.069Z [DEBUG] TestDNS_ServiceLookup_MetaTXT.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:16.070Z [WARN] TestDNS_ServiceLookup_MetaTXT.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:16.072Z [ERROR] TestDNS_ServiceLookup_MetaTXT.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:16.072Z [DEBUG] TestDNS_ServiceLookup_MetaTXT.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:16.074Z [WARN] TestDNS_ServiceLookup_MetaTXT.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:16.076Z [INFO] TestDNS_ServiceLookup_MetaTXT.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:16.076Z [INFO] TestDNS_ServiceLookup_MetaTXT: consul server down writer.go:29: 2021-01-29T19:32:16.080Z [INFO] TestDNS_ServiceLookup_MetaTXT: shutdown complete writer.go:29: 2021-01-29T19:32:16.082Z [INFO] TestDNS_ServiceLookup_MetaTXT: Stopping server: protocol=DNS address=127.0.0.1:30084 network=tcp writer.go:29: 2021-01-29T19:32:16.084Z [INFO] TestDNS_ServiceLookup_MetaTXT: Stopping server: protocol=DNS address=127.0.0.1:30084 network=udp writer.go:29: 2021-01-29T19:32:16.085Z [INFO] TestDNS_ServiceLookup_MetaTXT: Stopping server: protocol=HTTP address=127.0.0.1:30085 network=tcp writer.go:29: 2021-01-29T19:32:16.587Z [INFO] TestDNS_ServiceLookup_MetaTXT: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:16.589Z [INFO] TestDNS_ServiceLookup_MetaTXT: Endpoints down --- PASS: TestDNS_ServiceLookup_MetaTXT (0.82s) === RUN TestDNS_ServiceLookup_SuppressTXT writer.go:29: 2021-01-29T19:32:16.605Z [WARN] TestDNS_ServiceLookup_SuppressTXT: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:16.607Z [DEBUG] TestDNS_ServiceLookup_SuppressTXT.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:16.609Z [DEBUG] TestDNS_ServiceLookup_SuppressTXT.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:16.615Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:18845157-df63-197f-9d9e-89084658a619 Address:127.0.0.1:30095}]" writer.go:29: 2021-01-29T19:32:16.616Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server.raft: entering follower state: follower="Node at 127.0.0.1:30095 [Follower]" leader= writer.go:29: 2021-01-29T19:32:16.617Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server.serf.wan: serf: EventMemberJoin: Node-18845157-df63-197f-9d9e-89084658a619.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:16.624Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server.serf.lan: serf: EventMemberJoin: Node-18845157-df63-197f-9d9e-89084658a619 127.0.0.1 writer.go:29: 2021-01-29T19:32:16.630Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server: Adding LAN server: server="Node-18845157-df63-197f-9d9e-89084658a619 (Addr: tcp/127.0.0.1:30095) (DC: dc1)" writer.go:29: 2021-01-29T19:32:16.630Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server: Handled event for server in area: event=member-join server=Node-18845157-df63-197f-9d9e-89084658a619.dc1 area=wan writer.go:29: 2021-01-29T19:32:16.636Z [INFO] TestDNS_ServiceLookup_SuppressTXT: Started DNS server: address=127.0.0.1:30090 network=tcp writer.go:29: 2021-01-29T19:32:16.640Z [INFO] TestDNS_ServiceLookup_SuppressTXT: Started DNS server: address=127.0.0.1:30090 network=udp writer.go:29: 2021-01-29T19:32:16.645Z [INFO] TestDNS_ServiceLookup_SuppressTXT: Started HTTP server: address=127.0.0.1:30091 network=tcp writer.go:29: 2021-01-29T19:32:16.647Z [INFO] TestDNS_ServiceLookup_SuppressTXT: started state syncer writer.go:29: 2021-01-29T19:32:16.672Z [WARN] TestDNS_ServiceLookup_SuppressTXT.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:16.674Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server.raft: entering candidate state: node="Node at 127.0.0.1:30095 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:16.677Z [DEBUG] TestDNS_ServiceLookup_SuppressTXT.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:16.680Z [DEBUG] TestDNS_ServiceLookup_SuppressTXT.server.raft: vote granted: from=18845157-df63-197f-9d9e-89084658a619 term=2 tally=1 writer.go:29: 2021-01-29T19:32:16.682Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:16.683Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server.raft: entering leader state: leader="Node at 127.0.0.1:30095 [Leader]" writer.go:29: 2021-01-29T19:32:16.685Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:16.686Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server: New leader elected: payload=Node-18845157-df63-197f-9d9e-89084658a619 writer.go:29: 2021-01-29T19:32:16.690Z [DEBUG] TestDNS_ServiceLookup_SuppressTXT.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30095 writer.go:29: 2021-01-29T19:32:16.697Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:16.702Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:16.704Z [INFO] TestDNS_ServiceLookup_SuppressTXT.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:16.706Z [DEBUG] TestDNS_ServiceLookup_SuppressTXT.server: Skipping self join check for node since the cluster is too small: node=Node-18845157-df63-197f-9d9e-89084658a619 writer.go:29: 2021-01-29T19:32:16.707Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server: member joined, marking health alive: member=Node-18845157-df63-197f-9d9e-89084658a619 writer.go:29: 2021-01-29T19:32:16.792Z [DEBUG] TestDNS_ServiceLookup_SuppressTXT.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=97.096µs client=127.0.0.1:60272 client_network=udp writer.go:29: 2021-01-29T19:32:16.792Z [INFO] TestDNS_ServiceLookup_SuppressTXT: Requesting shutdown writer.go:29: 2021-01-29T19:32:16.796Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server: shutting down server writer.go:29: 2021-01-29T19:32:16.797Z [DEBUG] TestDNS_ServiceLookup_SuppressTXT.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:16.799Z [WARN] TestDNS_ServiceLookup_SuppressTXT.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:16.801Z [ERROR] TestDNS_ServiceLookup_SuppressTXT.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:16.801Z [DEBUG] TestDNS_ServiceLookup_SuppressTXT.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:16.802Z [WARN] TestDNS_ServiceLookup_SuppressTXT.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:16.808Z [INFO] TestDNS_ServiceLookup_SuppressTXT.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:16.809Z [INFO] TestDNS_ServiceLookup_SuppressTXT: consul server down writer.go:29: 2021-01-29T19:32:16.811Z [INFO] TestDNS_ServiceLookup_SuppressTXT: shutdown complete writer.go:29: 2021-01-29T19:32:16.812Z [INFO] TestDNS_ServiceLookup_SuppressTXT: Stopping server: protocol=DNS address=127.0.0.1:30090 network=tcp writer.go:29: 2021-01-29T19:32:16.814Z [INFO] TestDNS_ServiceLookup_SuppressTXT: Stopping server: protocol=DNS address=127.0.0.1:30090 network=udp writer.go:29: 2021-01-29T19:32:16.816Z [INFO] TestDNS_ServiceLookup_SuppressTXT: Stopping server: protocol=HTTP address=127.0.0.1:30091 network=tcp writer.go:29: 2021-01-29T19:32:17.317Z [INFO] TestDNS_ServiceLookup_SuppressTXT: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:17.320Z [INFO] TestDNS_ServiceLookup_SuppressTXT: Endpoints down --- PASS: TestDNS_ServiceLookup_SuppressTXT (0.73s) === RUN TestDNS_AddressLookup === PAUSE TestDNS_AddressLookup === RUN TestDNS_AddressLookupIPV6 dns_test.go:5666: DM-skipped --- SKIP: TestDNS_AddressLookupIPV6 (0.00s) === RUN TestDNS_NonExistingLookup === PAUSE TestDNS_NonExistingLookup === RUN TestDNS_NonExistingLookupEmptyAorAAAA === PAUSE TestDNS_NonExistingLookupEmptyAorAAAA === RUN TestDNS_AltDomains_Service === PAUSE TestDNS_AltDomains_Service === RUN TestDNS_AltDomains_SOA === PAUSE TestDNS_AltDomains_SOA === RUN TestDNS_AltDomains_Overlap === PAUSE TestDNS_AltDomains_Overlap === RUN TestDNS_PreparedQuery_AllowStale === PAUSE TestDNS_PreparedQuery_AllowStale === RUN TestDNS_InvalidQueries === PAUSE TestDNS_InvalidQueries === RUN TestDNS_PreparedQuery_AgentSource === PAUSE TestDNS_PreparedQuery_AgentSource === RUN TestDNS_trimUDPResponse_NoTrim === PAUSE TestDNS_trimUDPResponse_NoTrim === RUN TestDNS_trimUDPResponse_TrimLimit === PAUSE TestDNS_trimUDPResponse_TrimLimit === RUN TestDNS_trimUDPResponse_TrimSize === PAUSE TestDNS_trimUDPResponse_TrimSize === RUN TestDNS_trimUDPResponse_TrimSizeEDNS === PAUSE TestDNS_trimUDPResponse_TrimSizeEDNS === RUN TestDNS_syncExtra === PAUSE TestDNS_syncExtra === RUN TestDNS_Compression_trimUDPResponse === PAUSE TestDNS_Compression_trimUDPResponse === RUN TestDNS_Compression_Query === PAUSE TestDNS_Compression_Query === RUN TestDNS_Compression_ReverseLookup === PAUSE TestDNS_Compression_ReverseLookup === RUN TestDNS_Compression_Recurse === PAUSE TestDNS_Compression_Recurse === RUN TestDNSInvalidRegex === RUN TestDNSInvalidRegex/Valid_Hostname === RUN TestDNSInvalidRegex/Valid_Hostname#01 === RUN TestDNSInvalidRegex/Invalid_Hostname_with_special_chars === RUN TestDNSInvalidRegex/Invalid_Hostname_with_special_chars_in_the_end === RUN TestDNSInvalidRegex/Whitespace === RUN TestDNSInvalidRegex/Only_special_chars --- PASS: TestDNSInvalidRegex (0.01s) --- PASS: TestDNSInvalidRegex/Valid_Hostname (0.00s) --- PASS: TestDNSInvalidRegex/Valid_Hostname#01 (0.00s) --- PASS: TestDNSInvalidRegex/Invalid_Hostname_with_special_chars (0.00s) --- PASS: TestDNSInvalidRegex/Invalid_Hostname_with_special_chars_in_the_end (0.00s) --- PASS: TestDNSInvalidRegex/Whitespace (0.00s) --- PASS: TestDNSInvalidRegex/Only_special_chars (0.00s) === RUN TestDNS_ConfigReload === PAUSE TestDNS_ConfigReload === RUN TestDNS_ReloadConfig_DuringQuery dns_test.go:6941: DM-skipped --- SKIP: TestDNS_ReloadConfig_DuringQuery (0.00s) === RUN TestEventFire === PAUSE TestEventFire === RUN TestEventFire_token === PAUSE TestEventFire_token === RUN TestEventList === PAUSE TestEventList === RUN TestEventList_Filter === PAUSE TestEventList_Filter === RUN TestEventList_ACLFilter === PAUSE TestEventList_ACLFilter === RUN TestEventList_Blocking === PAUSE TestEventList_Blocking === RUN TestEventList_EventBufOrder === PAUSE TestEventList_EventBufOrder === RUN TestUUIDToUint64 === PAUSE TestUUIDToUint64 === RUN TestHealthChecksInState health_endpoint_test.go:23: DM-skipped --- SKIP: TestHealthChecksInState (0.00s) === RUN TestHealthChecksInState_NodeMetaFilter === PAUSE TestHealthChecksInState_NodeMetaFilter === RUN TestHealthChecksInState_Filter === PAUSE TestHealthChecksInState_Filter === RUN TestHealthChecksInState_DistanceSort === PAUSE TestHealthChecksInState_DistanceSort === RUN TestHealthNodeChecks === PAUSE TestHealthNodeChecks === RUN TestHealthNodeChecks_Filtering === PAUSE TestHealthNodeChecks_Filtering === RUN TestHealthServiceChecks === PAUSE TestHealthServiceChecks === RUN TestHealthServiceChecks_NodeMetaFilter === PAUSE TestHealthServiceChecks_NodeMetaFilter === RUN TestHealthServiceChecks_Filtering === PAUSE TestHealthServiceChecks_Filtering === RUN TestHealthServiceChecks_DistanceSort === PAUSE TestHealthServiceChecks_DistanceSort === RUN TestHealthServiceNodes === PAUSE TestHealthServiceNodes === RUN TestHealthServiceNodes_NodeMetaFilter === PAUSE TestHealthServiceNodes_NodeMetaFilter === RUN TestHealthServiceNodes_Filter health_endpoint_test.go:741: DM-skipped --- SKIP: TestHealthServiceNodes_Filter (0.00s) === RUN TestHealthServiceNodes_DistanceSort === PAUSE TestHealthServiceNodes_DistanceSort === RUN TestHealthServiceNodes_PassingFilter health_endpoint_test.go:883: DM-skipped --- SKIP: TestHealthServiceNodes_PassingFilter (0.00s) === RUN TestHealthServiceNodes_CheckType === PAUSE TestHealthServiceNodes_CheckType === RUN TestHealthServiceNodes_WanTranslation === PAUSE TestHealthServiceNodes_WanTranslation === RUN TestHealthConnectServiceNodes === PAUSE TestHealthConnectServiceNodes === RUN TestHealthConnectServiceNodes_Filter === PAUSE TestHealthConnectServiceNodes_Filter === RUN TestHealthConnectServiceNodes_PassingFilter === PAUSE TestHealthConnectServiceNodes_PassingFilter === RUN TestFilterNonPassing === PAUSE TestFilterNonPassing === RUN TestDecodeACLPolicyWrite === RUN TestDecodeACLPolicyWrite/hashes_base64_encoded === RUN TestDecodeACLPolicyWrite/hashes_not-base64_encoded === RUN TestDecodeACLPolicyWrite/hashes_empty_string === RUN TestDecodeACLPolicyWrite/hashes_null === RUN TestDecodeACLPolicyWrite/hashes_numeric_value --- PASS: TestDecodeACLPolicyWrite (0.04s) --- PASS: TestDecodeACLPolicyWrite/hashes_base64_encoded (0.01s) --- PASS: TestDecodeACLPolicyWrite/hashes_not-base64_encoded (0.00s) --- PASS: TestDecodeACLPolicyWrite/hashes_empty_string (0.00s) --- PASS: TestDecodeACLPolicyWrite/hashes_null (0.00s) --- PASS: TestDecodeACLPolicyWrite/hashes_numeric_value (0.01s) === RUN TestDecodeACLToken === RUN TestDecodeACLToken/timestamps_correctly_RFC3339_formatted === RUN TestDecodeACLToken/timestamps_incorrectly_formatted_(RFC822) === RUN TestDecodeACLToken/timestamps_incorrectly_formatted_(RFC850) === RUN TestDecodeACLToken/timestamps_empty_string === RUN TestDecodeACLToken/timestamps_null === RUN TestDecodeACLToken/durations_correctly_formatted === RUN TestDecodeACLToken/durations_small,_correctly_formatted === RUN TestDecodeACLToken/durations_incorrectly_formatted === RUN TestDecodeACLToken/durations_empty_string === RUN TestDecodeACLToken/durations_string_without_quotes === RUN TestDecodeACLToken/durations_numeric === RUN TestDecodeACLToken/durations_negative === RUN TestDecodeACLToken/durations_numeric_and_negative === RUN TestDecodeACLToken/hashes_base64_encoded === RUN TestDecodeACLToken/hashes_not-base64_encoded === RUN TestDecodeACLToken/hashes_empty_string === RUN TestDecodeACLToken/hashes_null === RUN TestDecodeACLToken/hashes_numeric_value --- PASS: TestDecodeACLToken (0.04s) --- PASS: TestDecodeACLToken/timestamps_correctly_RFC3339_formatted (0.00s) --- PASS: TestDecodeACLToken/timestamps_incorrectly_formatted_(RFC822) (0.00s) --- PASS: TestDecodeACLToken/timestamps_incorrectly_formatted_(RFC850) (0.00s) --- PASS: TestDecodeACLToken/timestamps_empty_string (0.00s) --- PASS: TestDecodeACLToken/timestamps_null (0.00s) --- PASS: TestDecodeACLToken/durations_correctly_formatted (0.00s) --- PASS: TestDecodeACLToken/durations_small,_correctly_formatted (0.00s) --- PASS: TestDecodeACLToken/durations_incorrectly_formatted (0.00s) --- PASS: TestDecodeACLToken/durations_empty_string (0.00s) --- PASS: TestDecodeACLToken/durations_string_without_quotes (0.00s) --- PASS: TestDecodeACLToken/durations_numeric (0.00s) --- PASS: TestDecodeACLToken/durations_negative (0.00s) --- PASS: TestDecodeACLToken/durations_numeric_and_negative (0.00s) --- PASS: TestDecodeACLToken/hashes_base64_encoded (0.00s) --- PASS: TestDecodeACLToken/hashes_not-base64_encoded (0.00s) --- PASS: TestDecodeACLToken/hashes_empty_string (0.00s) --- PASS: TestDecodeACLToken/hashes_null (0.00s) --- PASS: TestDecodeACLToken/hashes_numeric_value (0.00s) === RUN TestDecodeACLRoleWrite === RUN TestDecodeACLRoleWrite/hashes_base64_encoded === RUN TestDecodeACLRoleWrite/hashes_not-base64_encoded === RUN TestDecodeACLRoleWrite/hashes_empty_string === RUN TestDecodeACLRoleWrite/hashes_null === RUN TestDecodeACLRoleWrite/hashes_numeric_value --- PASS: TestDecodeACLRoleWrite (0.02s) --- PASS: TestDecodeACLRoleWrite/hashes_base64_encoded (0.00s) --- PASS: TestDecodeACLRoleWrite/hashes_not-base64_encoded (0.00s) --- PASS: TestDecodeACLRoleWrite/hashes_empty_string (0.00s) --- PASS: TestDecodeACLRoleWrite/hashes_null (0.00s) --- PASS: TestDecodeACLRoleWrite/hashes_numeric_value (0.00s) === RUN TestDecodeAgentRegisterCheck === RUN TestDecodeAgentRegisterCheck/durations_correctly_formatted === RUN TestDecodeAgentRegisterCheck/durations_small,_correctly_formatted === RUN TestDecodeAgentRegisterCheck/durations_incorrectly_formatted === RUN TestDecodeAgentRegisterCheck/durations_empty_string === RUN TestDecodeAgentRegisterCheck/durations_string_without_quotes === RUN TestDecodeAgentRegisterCheck/durations_numeric === RUN TestDecodeAgentRegisterCheck/durations_negative === RUN TestDecodeAgentRegisterCheck/durations_numeric_and_negative === RUN TestDecodeAgentRegisterCheck/filled_in_map === RUN TestDecodeAgentRegisterCheck/empty_map === RUN TestDecodeAgentRegisterCheck/empty_map#01 === RUN TestDecodeAgentRegisterCheck/malformatted_map === RUN TestDecodeAgentRegisterCheck/not_a_map_(slice) === RUN TestDecodeAgentRegisterCheck/not_a_map_(int) === RUN TestDecodeAgentRegisterCheck/scriptArgs:_all_set === RUN TestDecodeAgentRegisterCheck/scriptArgs:_first_and_second_set === RUN TestDecodeAgentRegisterCheck/scriptArgs:_first_and_third_set === RUN TestDecodeAgentRegisterCheck/scriptArgs:_second_and_third_set === RUN TestDecodeAgentRegisterCheck/scriptArgs:_first_set === RUN TestDecodeAgentRegisterCheck/scriptArgs:_second_set === RUN TestDecodeAgentRegisterCheck/scriptArgs:_third_set === RUN TestDecodeAgentRegisterCheck/scriptArgs:_none_set === RUN TestDecodeAgentRegisterCheck/deregister:_both_set === RUN TestDecodeAgentRegisterCheck/deregister:_first_set === RUN TestDecodeAgentRegisterCheck/deregister:_second_set === RUN TestDecodeAgentRegisterCheck/deregister:_neither_set === RUN TestDecodeAgentRegisterCheck/dockerContainerID:_both_set === RUN TestDecodeAgentRegisterCheck/dockerContainerID:_first_set === RUN TestDecodeAgentRegisterCheck/dockerContainerID:_second_set === RUN TestDecodeAgentRegisterCheck/dockerContainerID:_neither_set === RUN TestDecodeAgentRegisterCheck/tlsSkipVerify:_both_set === RUN TestDecodeAgentRegisterCheck/tlsSkipVerify:_first_set === RUN TestDecodeAgentRegisterCheck/tlsSkipVerify:_second_set === RUN TestDecodeAgentRegisterCheck/tlsSkipVerify:_neither_set === RUN TestDecodeAgentRegisterCheck/serviceID:_both_set === RUN TestDecodeAgentRegisterCheck/serviceID:_first_set === RUN TestDecodeAgentRegisterCheck/serviceID:_second_set === RUN TestDecodeAgentRegisterCheck/serviceID:_neither_set --- PASS: TestDecodeAgentRegisterCheck (0.09s) --- PASS: TestDecodeAgentRegisterCheck/durations_correctly_formatted (0.00s) --- PASS: TestDecodeAgentRegisterCheck/durations_small,_correctly_formatted (0.00s) --- PASS: TestDecodeAgentRegisterCheck/durations_incorrectly_formatted (0.00s) --- PASS: TestDecodeAgentRegisterCheck/durations_empty_string (0.00s) --- PASS: TestDecodeAgentRegisterCheck/durations_string_without_quotes (0.00s) --- PASS: TestDecodeAgentRegisterCheck/durations_numeric (0.00s) --- PASS: TestDecodeAgentRegisterCheck/durations_negative (0.00s) --- PASS: TestDecodeAgentRegisterCheck/durations_numeric_and_negative (0.00s) --- PASS: TestDecodeAgentRegisterCheck/filled_in_map (0.00s) --- PASS: TestDecodeAgentRegisterCheck/empty_map (0.00s) --- PASS: TestDecodeAgentRegisterCheck/empty_map#01 (0.00s) --- PASS: TestDecodeAgentRegisterCheck/malformatted_map (0.00s) --- PASS: TestDecodeAgentRegisterCheck/not_a_map_(slice) (0.00s) --- PASS: TestDecodeAgentRegisterCheck/not_a_map_(int) (0.00s) --- PASS: TestDecodeAgentRegisterCheck/scriptArgs:_all_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/scriptArgs:_first_and_second_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/scriptArgs:_first_and_third_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/scriptArgs:_second_and_third_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/scriptArgs:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/scriptArgs:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/scriptArgs:_third_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/scriptArgs:_none_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/deregister:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/deregister:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/deregister:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/deregister:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/dockerContainerID:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/dockerContainerID:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/dockerContainerID:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/dockerContainerID:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/tlsSkipVerify:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/tlsSkipVerify:_first_set (0.01s) --- PASS: TestDecodeAgentRegisterCheck/tlsSkipVerify:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/tlsSkipVerify:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/serviceID:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/serviceID:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/serviceID:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterCheck/serviceID:_neither_set (0.00s) === RUN TestDecodeAgentRegisterService === RUN TestDecodeAgentRegisterService/translateEnableTagTCs:_both_set === RUN TestDecodeAgentRegisterService/translateEnableTagTCs:_first_set === RUN TestDecodeAgentRegisterService/translateEnableTagTCs:_second_set === RUN TestDecodeAgentRegisterService/translateEnableTagTCs:_neither_set === RUN TestDecodeAgentRegisterService/DestinationName:_both_set === RUN TestDecodeAgentRegisterService/DestinationName:_first_set === RUN TestDecodeAgentRegisterService/DestinationName:_second_set === RUN TestDecodeAgentRegisterService/DestinationName:_neither_set === RUN TestDecodeAgentRegisterService/DestinationType:_both_set === RUN TestDecodeAgentRegisterService/DestinationType:_first_set === RUN TestDecodeAgentRegisterService/DestinationType:_second_set === RUN TestDecodeAgentRegisterService/DestinationType:_neither_set === RUN TestDecodeAgentRegisterService/DestinationNamespace:_both_set === RUN TestDecodeAgentRegisterService/DestinationNamespace:_first_set === RUN TestDecodeAgentRegisterService/DestinationNamespace:_second_set === RUN TestDecodeAgentRegisterService/DestinationNamespace:_neither_set === RUN TestDecodeAgentRegisterService/LocalBindPort:_both_set === RUN TestDecodeAgentRegisterService/LocalBindPort:_first_set === RUN TestDecodeAgentRegisterService/LocalBindPort:_second_set === RUN TestDecodeAgentRegisterService/LocalBindPort:_neither_set === RUN TestDecodeAgentRegisterService/LocalBindAddress:_both_set === RUN TestDecodeAgentRegisterService/LocalBindAddress:_first_set === RUN TestDecodeAgentRegisterService/LocalBindAddress:_second_set === RUN TestDecodeAgentRegisterService/LocalBindAddress:_neither_set === RUN TestDecodeAgentRegisterService/DestinationServiceName:_both_set === RUN TestDecodeAgentRegisterService/DestinationServiceName:_first_set === RUN TestDecodeAgentRegisterService/DestinationServiceName:_second_set === RUN TestDecodeAgentRegisterService/DestinationServiceName:_neither_set === RUN TestDecodeAgentRegisterService/DestinationServiceID:_both_set === RUN TestDecodeAgentRegisterService/DestinationServiceID:_first_set === RUN TestDecodeAgentRegisterService/DestinationServiceID:_second_set === RUN TestDecodeAgentRegisterService/DestinationServiceID:_neither_set === RUN TestDecodeAgentRegisterService/LocalServicePort:_both_set === RUN TestDecodeAgentRegisterService/LocalServicePort:_first_set === RUN TestDecodeAgentRegisterService/LocalServicePort:_second_set === RUN TestDecodeAgentRegisterService/LocalServicePort:_neither_set === RUN TestDecodeAgentRegisterService/LocalServiceAddress:_both_set === RUN TestDecodeAgentRegisterService/LocalServiceAddress:_first_set === RUN TestDecodeAgentRegisterService/LocalServiceAddress:_second_set === RUN TestDecodeAgentRegisterService/LocalServiceAddress:_neither_set === RUN TestDecodeAgentRegisterService/SidecarService:_both_set === RUN TestDecodeAgentRegisterService/SidecarService:_first_set === RUN TestDecodeAgentRegisterService/SidecarService:_second_set === RUN TestDecodeAgentRegisterService/SidecarService:_neither_set === RUN TestDecodeAgentRegisterService/LocalPathPort:_both_set === RUN TestDecodeAgentRegisterService/LocalPathPort:_first_set === RUN TestDecodeAgentRegisterService/LocalPathPort:_second_set === RUN TestDecodeAgentRegisterService/LocalPathPort:_neither_set === RUN TestDecodeAgentRegisterService/ListenerPort:_both_set === RUN TestDecodeAgentRegisterService/ListenerPort:_first_set === RUN TestDecodeAgentRegisterService/ListenerPort:_second_set === RUN TestDecodeAgentRegisterService/ListenerPort:_neither_set === RUN TestDecodeAgentRegisterService/TaggedAddresses:_both_set === RUN TestDecodeAgentRegisterService/TaggedAddresses:_first_set === RUN TestDecodeAgentRegisterService/TaggedAddresses:_second_set === RUN TestDecodeAgentRegisterService/TaggedAddresses:_neither_set === RUN TestDecodeAgentRegisterService/durations_correctly_formatted === RUN TestDecodeAgentRegisterService/durations_small,_correctly_formatted === RUN TestDecodeAgentRegisterService/durations_incorrectly_formatted === RUN TestDecodeAgentRegisterService/durations_empty_string === RUN TestDecodeAgentRegisterService/durations_string_without_quotes === RUN TestDecodeAgentRegisterService/durations_numeric === RUN TestDecodeAgentRegisterService/durations_negative === RUN TestDecodeAgentRegisterService/durations_numeric_and_negative === RUN TestDecodeAgentRegisterService/filled_in_map === RUN TestDecodeAgentRegisterService/empty_map === RUN TestDecodeAgentRegisterService/empty_map#01 === RUN TestDecodeAgentRegisterService/malformatted_map === RUN TestDecodeAgentRegisterService/not_a_map_(slice) === RUN TestDecodeAgentRegisterService/not_a_map_(int) === RUN TestDecodeAgentRegisterService/scriptArgs:_all_set === RUN TestDecodeAgentRegisterService/scriptArgs:_first_and_second_set === RUN TestDecodeAgentRegisterService/scriptArgs:_first_and_third_set === RUN TestDecodeAgentRegisterService/scriptArgs:_second_and_third_set === RUN TestDecodeAgentRegisterService/scriptArgs:_first_set === RUN TestDecodeAgentRegisterService/scriptArgs:_second_set === RUN TestDecodeAgentRegisterService/scriptArgs:_third_set === RUN TestDecodeAgentRegisterService/scriptArgs:_none_set === RUN TestDecodeAgentRegisterService/deregister:_both_set === RUN TestDecodeAgentRegisterService/deregister:_first_set === RUN TestDecodeAgentRegisterService/deregister:_second_set === RUN TestDecodeAgentRegisterService/deregister:_neither_set === RUN TestDecodeAgentRegisterService/dockerContainerID:_both_set === RUN TestDecodeAgentRegisterService/dockerContainerID:_first_set === RUN TestDecodeAgentRegisterService/dockerContainerID:_second_set === RUN TestDecodeAgentRegisterService/dockerContainerID:_neither_set === RUN TestDecodeAgentRegisterService/tlsSkipVerify:_both_set === RUN TestDecodeAgentRegisterService/tlsSkipVerify:_first_set === RUN TestDecodeAgentRegisterService/tlsSkipVerify:_second_set === RUN TestDecodeAgentRegisterService/tlsSkipVerify:_neither_set === RUN TestDecodeAgentRegisterService/serviceID:_both_set === RUN TestDecodeAgentRegisterService/serviceID:_first_set === RUN TestDecodeAgentRegisterService/serviceID:_second_set === RUN TestDecodeAgentRegisterService/serviceID:_neither_set --- PASS: TestDecodeAgentRegisterService (0.42s) --- PASS: TestDecodeAgentRegisterService/translateEnableTagTCs:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/translateEnableTagTCs:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/translateEnableTagTCs:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/translateEnableTagTCs:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationName:_both_set (0.01s) --- PASS: TestDecodeAgentRegisterService/DestinationName:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationName:_second_set (0.01s) --- PASS: TestDecodeAgentRegisterService/DestinationName:_neither_set (0.01s) --- PASS: TestDecodeAgentRegisterService/DestinationType:_both_set (0.01s) --- PASS: TestDecodeAgentRegisterService/DestinationType:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationType:_second_set (0.01s) --- PASS: TestDecodeAgentRegisterService/DestinationType:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationNamespace:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationNamespace:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationNamespace:_second_set (0.02s) --- PASS: TestDecodeAgentRegisterService/DestinationNamespace:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalBindPort:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalBindPort:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalBindPort:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalBindPort:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalBindAddress:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalBindAddress:_first_set (0.01s) --- PASS: TestDecodeAgentRegisterService/LocalBindAddress:_second_set (0.01s) --- PASS: TestDecodeAgentRegisterService/LocalBindAddress:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationServiceName:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationServiceName:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationServiceName:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationServiceName:_neither_set (0.01s) --- PASS: TestDecodeAgentRegisterService/DestinationServiceID:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationServiceID:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationServiceID:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/DestinationServiceID:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalServicePort:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalServicePort:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalServicePort:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalServicePort:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalServiceAddress:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalServiceAddress:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalServiceAddress:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalServiceAddress:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/SidecarService:_both_set (0.01s) --- PASS: TestDecodeAgentRegisterService/SidecarService:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/SidecarService:_second_set (0.01s) --- PASS: TestDecodeAgentRegisterService/SidecarService:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalPathPort:_both_set (0.02s) --- PASS: TestDecodeAgentRegisterService/LocalPathPort:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/LocalPathPort:_second_set (0.01s) --- PASS: TestDecodeAgentRegisterService/LocalPathPort:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/ListenerPort:_both_set (0.01s) --- PASS: TestDecodeAgentRegisterService/ListenerPort:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/ListenerPort:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/ListenerPort:_neither_set (0.01s) --- PASS: TestDecodeAgentRegisterService/TaggedAddresses:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/TaggedAddresses:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/TaggedAddresses:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/TaggedAddresses:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/durations_correctly_formatted (0.00s) --- PASS: TestDecodeAgentRegisterService/durations_small,_correctly_formatted (0.00s) --- PASS: TestDecodeAgentRegisterService/durations_incorrectly_formatted (0.00s) --- PASS: TestDecodeAgentRegisterService/durations_empty_string (0.00s) --- PASS: TestDecodeAgentRegisterService/durations_string_without_quotes (0.00s) --- PASS: TestDecodeAgentRegisterService/durations_numeric (0.00s) --- PASS: TestDecodeAgentRegisterService/durations_negative (0.00s) --- PASS: TestDecodeAgentRegisterService/durations_numeric_and_negative (0.00s) --- PASS: TestDecodeAgentRegisterService/filled_in_map (0.00s) --- PASS: TestDecodeAgentRegisterService/empty_map (0.00s) --- PASS: TestDecodeAgentRegisterService/empty_map#01 (0.00s) --- PASS: TestDecodeAgentRegisterService/malformatted_map (0.00s) --- PASS: TestDecodeAgentRegisterService/not_a_map_(slice) (0.00s) --- PASS: TestDecodeAgentRegisterService/not_a_map_(int) (0.00s) --- PASS: TestDecodeAgentRegisterService/scriptArgs:_all_set (0.00s) --- PASS: TestDecodeAgentRegisterService/scriptArgs:_first_and_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/scriptArgs:_first_and_third_set (0.01s) --- PASS: TestDecodeAgentRegisterService/scriptArgs:_second_and_third_set (0.00s) --- PASS: TestDecodeAgentRegisterService/scriptArgs:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/scriptArgs:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/scriptArgs:_third_set (0.00s) --- PASS: TestDecodeAgentRegisterService/scriptArgs:_none_set (0.00s) --- PASS: TestDecodeAgentRegisterService/deregister:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/deregister:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/deregister:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/deregister:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/dockerContainerID:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/dockerContainerID:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/dockerContainerID:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/dockerContainerID:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/tlsSkipVerify:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/tlsSkipVerify:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/tlsSkipVerify:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/tlsSkipVerify:_neither_set (0.00s) --- PASS: TestDecodeAgentRegisterService/serviceID:_both_set (0.00s) --- PASS: TestDecodeAgentRegisterService/serviceID:_first_set (0.00s) --- PASS: TestDecodeAgentRegisterService/serviceID:_second_set (0.00s) --- PASS: TestDecodeAgentRegisterService/serviceID:_neither_set (0.00s) === RUN TestDecodeCatalogRegister === RUN TestDecodeCatalogRegister/durations_correctly_formatted === RUN TestDecodeCatalogRegister/durations_small,_correctly_formatted === RUN TestDecodeCatalogRegister/durations_incorrectly_formatted === RUN TestDecodeCatalogRegister/durations_empty_string === RUN TestDecodeCatalogRegister/durations_string_without_quotes === RUN TestDecodeCatalogRegister/durations_numeric === RUN TestDecodeCatalogRegister/durations_negative === RUN TestDecodeCatalogRegister/durations_numeric_and_negative --- PASS: TestDecodeCatalogRegister (0.04s) --- PASS: TestDecodeCatalogRegister/durations_correctly_formatted (0.01s) --- PASS: TestDecodeCatalogRegister/durations_small,_correctly_formatted (0.00s) --- PASS: TestDecodeCatalogRegister/durations_incorrectly_formatted (0.00s) --- PASS: TestDecodeCatalogRegister/durations_empty_string (0.00s) --- PASS: TestDecodeCatalogRegister/durations_string_without_quotes (0.00s) --- PASS: TestDecodeCatalogRegister/durations_numeric (0.00s) --- PASS: TestDecodeCatalogRegister/durations_negative (0.00s) --- PASS: TestDecodeCatalogRegister/durations_numeric_and_negative (0.00s) === RUN TestDecodeDiscoveryChainRead === RUN TestDecodeDiscoveryChainRead/durations_correctly_formatted === RUN TestDecodeDiscoveryChainRead/durations_small,_correctly_formatted === RUN TestDecodeDiscoveryChainRead/durations_incorrectly_formatted === RUN TestDecodeDiscoveryChainRead/durations_empty_string === RUN TestDecodeDiscoveryChainRead/durations_string_without_quotes === RUN TestDecodeDiscoveryChainRead/durations_numeric === RUN TestDecodeDiscoveryChainRead/durations_negative === RUN TestDecodeDiscoveryChainRead/durations_numeric_and_negative === RUN TestDecodeDiscoveryChainRead/positive_string_integer_(weakly_typed) === RUN TestDecodeDiscoveryChainRead/negative_string_integer_(weakly_typed) === RUN TestDecodeDiscoveryChainRead/positive_integer_for_string_field_(weakly_typed) === RUN TestDecodeDiscoveryChainRead/negative_integer_for_string_field_(weakly_typed) === RUN TestDecodeDiscoveryChainRead/bool_for_string_field_(weakly_typed) === RUN TestDecodeDiscoveryChainRead/float_for_string_field_(weakly_typed) === RUN TestDecodeDiscoveryChainRead/map_for_string_field_(weakly_typed) === RUN TestDecodeDiscoveryChainRead/slice_for_string_field_(weakly_typed) === RUN TestDecodeDiscoveryChainRead/OverrideMeshGateway:_both_set === RUN TestDecodeDiscoveryChainRead/OverrideMeshGateway:_first_set === RUN TestDecodeDiscoveryChainRead/OverrideMeshGateway:_second_set === RUN TestDecodeDiscoveryChainRead/OverrideMeshGateway:_neither_set === RUN TestDecodeDiscoveryChainRead/OverrideProtocol:_both_set === RUN TestDecodeDiscoveryChainRead/OverrideProtocol:_first_set === RUN TestDecodeDiscoveryChainRead/OverrideProtocol:_second_set === RUN TestDecodeDiscoveryChainRead/OverrideProtocol:_neither_set === RUN TestDecodeDiscoveryChainRead/OverrideConnectTimeout:_both_set === RUN TestDecodeDiscoveryChainRead/OverrideConnectTimeout:_first_set === RUN TestDecodeDiscoveryChainRead/OverrideConnectTimeout:_second_set === RUN TestDecodeDiscoveryChainRead/OverrideConnectTimeout:_neither_set --- PASS: TestDecodeDiscoveryChainRead (0.06s) --- PASS: TestDecodeDiscoveryChainRead/durations_correctly_formatted (0.00s) --- PASS: TestDecodeDiscoveryChainRead/durations_small,_correctly_formatted (0.00s) --- PASS: TestDecodeDiscoveryChainRead/durations_incorrectly_formatted (0.00s) --- PASS: TestDecodeDiscoveryChainRead/durations_empty_string (0.00s) --- PASS: TestDecodeDiscoveryChainRead/durations_string_without_quotes (0.00s) --- PASS: TestDecodeDiscoveryChainRead/durations_numeric (0.00s) --- PASS: TestDecodeDiscoveryChainRead/durations_negative (0.00s) --- PASS: TestDecodeDiscoveryChainRead/durations_numeric_and_negative (0.00s) --- PASS: TestDecodeDiscoveryChainRead/positive_string_integer_(weakly_typed) (0.00s) --- PASS: TestDecodeDiscoveryChainRead/negative_string_integer_(weakly_typed) (0.00s) --- PASS: TestDecodeDiscoveryChainRead/positive_integer_for_string_field_(weakly_typed) (0.00s) --- PASS: TestDecodeDiscoveryChainRead/negative_integer_for_string_field_(weakly_typed) (0.00s) --- PASS: TestDecodeDiscoveryChainRead/bool_for_string_field_(weakly_typed) (0.00s) --- PASS: TestDecodeDiscoveryChainRead/float_for_string_field_(weakly_typed) (0.00s) --- PASS: TestDecodeDiscoveryChainRead/map_for_string_field_(weakly_typed) (0.00s) --- PASS: TestDecodeDiscoveryChainRead/slice_for_string_field_(weakly_typed) (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideMeshGateway:_both_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideMeshGateway:_first_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideMeshGateway:_second_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideMeshGateway:_neither_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideProtocol:_both_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideProtocol:_first_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideProtocol:_second_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideProtocol:_neither_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideConnectTimeout:_both_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideConnectTimeout:_first_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideConnectTimeout:_second_set (0.00s) --- PASS: TestDecodeDiscoveryChainRead/OverrideConnectTimeout:_neither_set (0.00s) === RUN TestDecodeIntentionCreate === RUN TestDecodeIntentionCreate/hashes_base64_encoded === RUN TestDecodeIntentionCreate/hashes_not-base64_encoded === RUN TestDecodeIntentionCreate/hashes_empty_string === RUN TestDecodeIntentionCreate/hashes_null === RUN TestDecodeIntentionCreate/hashes_numeric_value === RUN TestDecodeIntentionCreate/timestamps_correctly_RFC3339_formatted === RUN TestDecodeIntentionCreate/timestamps_incorrectly_formatted_(RFC822) === RUN TestDecodeIntentionCreate/timestamps_incorrectly_formatted_(RFC850) === RUN TestDecodeIntentionCreate/timestamps_empty_string === RUN TestDecodeIntentionCreate/timestamps_null --- PASS: TestDecodeIntentionCreate (0.02s) --- PASS: TestDecodeIntentionCreate/hashes_base64_encoded (0.00s) --- PASS: TestDecodeIntentionCreate/hashes_not-base64_encoded (0.00s) --- PASS: TestDecodeIntentionCreate/hashes_empty_string (0.00s) --- PASS: TestDecodeIntentionCreate/hashes_null (0.00s) --- PASS: TestDecodeIntentionCreate/hashes_numeric_value (0.00s) --- PASS: TestDecodeIntentionCreate/timestamps_correctly_RFC3339_formatted (0.00s) --- PASS: TestDecodeIntentionCreate/timestamps_incorrectly_formatted_(RFC822) (0.00s) --- PASS: TestDecodeIntentionCreate/timestamps_incorrectly_formatted_(RFC850) (0.00s) --- PASS: TestDecodeIntentionCreate/timestamps_empty_string (0.00s) --- PASS: TestDecodeIntentionCreate/timestamps_null (0.00s) === RUN TestDecodeOperatorAutopilotConfiguration === RUN TestDecodeOperatorAutopilotConfiguration/durations_correctly_formatted === RUN TestDecodeOperatorAutopilotConfiguration/durations_small,_correctly_formatted === RUN TestDecodeOperatorAutopilotConfiguration/durations_incorrectly_formatted === RUN TestDecodeOperatorAutopilotConfiguration/durations_empty_string === RUN TestDecodeOperatorAutopilotConfiguration/durations_string_without_quotes === RUN TestDecodeOperatorAutopilotConfiguration/durations_numeric === RUN TestDecodeOperatorAutopilotConfiguration/durations_negative === RUN TestDecodeOperatorAutopilotConfiguration/durations_numeric_and_negative --- PASS: TestDecodeOperatorAutopilotConfiguration (0.01s) --- PASS: TestDecodeOperatorAutopilotConfiguration/durations_correctly_formatted (0.00s) --- PASS: TestDecodeOperatorAutopilotConfiguration/durations_small,_correctly_formatted (0.00s) --- PASS: TestDecodeOperatorAutopilotConfiguration/durations_incorrectly_formatted (0.00s) --- PASS: TestDecodeOperatorAutopilotConfiguration/durations_empty_string (0.00s) --- PASS: TestDecodeOperatorAutopilotConfiguration/durations_string_without_quotes (0.00s) --- PASS: TestDecodeOperatorAutopilotConfiguration/durations_numeric (0.00s) --- PASS: TestDecodeOperatorAutopilotConfiguration/durations_negative (0.00s) --- PASS: TestDecodeOperatorAutopilotConfiguration/durations_numeric_and_negative (0.00s) === RUN TestDecodeSessionCreate === RUN TestDecodeSessionCreate/durations_correctly_formatted === RUN TestDecodeSessionCreate/durations_small,_correctly_formatted === RUN TestDecodeSessionCreate/durations_incorrectly_formatted === RUN TestDecodeSessionCreate/durations_empty_string === RUN TestDecodeSessionCreate/durations_string_without_quotes === RUN TestDecodeSessionCreate/durations_numeric === RUN TestDecodeSessionCreate/duration_small,_numeric_(<_lockDelayMinThreshold) === RUN TestDecodeSessionCreate/duration_string,_no_unit === RUN TestDecodeSessionCreate/duration_small,_string,_already_duration === RUN TestDecodeSessionCreate/duration_small,_numeric,_negative === RUN TestDecodeSessionCreate/many_check_ids === RUN TestDecodeSessionCreate/one_check_ids === RUN TestDecodeSessionCreate/empty_check_id_slice === RUN TestDecodeSessionCreate/null_check_ids === RUN TestDecodeSessionCreate/empty_value_check_ids === RUN TestDecodeSessionCreate/malformatted_check_ids_(string) --- PASS: TestDecodeSessionCreate (0.03s) --- PASS: TestDecodeSessionCreate/durations_correctly_formatted (0.00s) --- PASS: TestDecodeSessionCreate/durations_small,_correctly_formatted (0.00s) --- PASS: TestDecodeSessionCreate/durations_incorrectly_formatted (0.00s) --- PASS: TestDecodeSessionCreate/durations_empty_string (0.00s) --- PASS: TestDecodeSessionCreate/durations_string_without_quotes (0.00s) --- PASS: TestDecodeSessionCreate/durations_numeric (0.00s) --- PASS: TestDecodeSessionCreate/duration_small,_numeric_(<_lockDelayMinThreshold) (0.00s) --- PASS: TestDecodeSessionCreate/duration_string,_no_unit (0.00s) --- PASS: TestDecodeSessionCreate/duration_small,_string,_already_duration (0.00s) --- PASS: TestDecodeSessionCreate/duration_small,_numeric,_negative (0.00s) --- PASS: TestDecodeSessionCreate/many_check_ids (0.00s) --- PASS: TestDecodeSessionCreate/one_check_ids (0.00s) --- PASS: TestDecodeSessionCreate/empty_check_id_slice (0.00s) --- PASS: TestDecodeSessionCreate/null_check_ids (0.00s) --- PASS: TestDecodeSessionCreate/empty_value_check_ids (0.00s) --- PASS: TestDecodeSessionCreate/malformatted_check_ids_(string) (0.00s) === RUN TestDecodeTxnConvertOps === RUN TestDecodeTxnConvertOps/durations_correctly_formatted === RUN TestDecodeTxnConvertOps/durations_small,_correctly_formatted === RUN TestDecodeTxnConvertOps/durations_incorrectly_formatted === RUN TestDecodeTxnConvertOps/durations_empty_string === RUN TestDecodeTxnConvertOps/durations_string_without_quotes === RUN TestDecodeTxnConvertOps/durations_numeric === RUN TestDecodeTxnConvertOps/durations_negative === RUN TestDecodeTxnConvertOps/durations_numeric_and_negative --- PASS: TestDecodeTxnConvertOps (0.02s) --- PASS: TestDecodeTxnConvertOps/durations_correctly_formatted (0.00s) --- PASS: TestDecodeTxnConvertOps/durations_small,_correctly_formatted (0.00s) --- PASS: TestDecodeTxnConvertOps/durations_incorrectly_formatted (0.00s) --- PASS: TestDecodeTxnConvertOps/durations_empty_string (0.00s) --- PASS: TestDecodeTxnConvertOps/durations_string_without_quotes (0.00s) --- PASS: TestDecodeTxnConvertOps/durations_numeric (0.00s) --- PASS: TestDecodeTxnConvertOps/durations_negative (0.00s) --- PASS: TestDecodeTxnConvertOps/durations_numeric_and_negative (0.00s) === RUN TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.293Z [WARN] TestHTTPAPI_MethodNotAllowed_OSS: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:18.295Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:18.298Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:18.347Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e53e7bb0-3022-af35-e38a-a8369bb74b4a Address:127.0.0.1:30101}]" writer.go:29: 2021-01-29T19:32:18.351Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.raft: entering follower state: follower="Node at 127.0.0.1:30101 [Follower]" leader= writer.go:29: 2021-01-29T19:32:18.355Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.serf.wan: serf: EventMemberJoin: Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:18.364Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.serf.lan: serf: EventMemberJoin: Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a 127.0.0.1 writer.go:29: 2021-01-29T19:32:18.369Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: Started DNS server: address=127.0.0.1:30096 network=udp writer.go:29: 2021-01-29T19:32:18.373Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: Handled event for server in area: event=member-join server=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a.dc1 area=wan writer.go:29: 2021-01-29T19:32:18.374Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: Adding LAN server: server="Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a (Addr: tcp/127.0.0.1:30101) (DC: dc1)" writer.go:29: 2021-01-29T19:32:18.377Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: Started DNS server: address=127.0.0.1:30096 network=tcp writer.go:29: 2021-01-29T19:32:18.392Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: Started HTTP server: address=127.0.0.1:30097 network=tcp writer.go:29: 2021-01-29T19:32:18.394Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: started state syncer writer.go:29: 2021-01-29T19:32:18.409Z [WARN] TestHTTPAPI_MethodNotAllowed_OSS.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:18.411Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.raft: entering candidate state: node="Node at 127.0.0.1:30101 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:18.423Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:18.426Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:18.429Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.server.raft: vote granted: from=e53e7bb0-3022-af35-e38a-a8369bb74b4a term=2 tally=1 writer.go:29: 2021-01-29T19:32:18.433Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:18.436Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.raft: entering leader state: leader="Node at 127.0.0.1:30101 [Leader]" writer.go:29: 2021-01-29T19:32:18.440Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:18.441Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: New leader elected: payload=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a writer.go:29: 2021-01-29T19:32:18.444Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: initializing acls writer.go:29: 2021-01-29T19:32:18.461Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:18.464Z [WARN] TestHTTPAPI_MethodNotAllowed_OSS.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:18.470Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:18.476Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:18.478Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:18.487Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:18.491Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.serf.lan: serf: EventMemberUpdate: Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a writer.go:29: 2021-01-29T19:32:18.494Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: Updating LAN server: server="Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a (Addr: tcp/127.0.0.1:30101) (DC: dc1)" writer.go:29: 2021-01-29T19:32:18.496Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.serf.wan: serf: EventMemberUpdate: Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a.dc1 writer.go:29: 2021-01-29T19:32:18.497Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: Handled event for server in area: event=member-update server=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a.dc1 area=wan writer.go:29: 2021-01-29T19:32:18.512Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:18.518Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:18.519Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:18.521Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.server: Skipping self join check for node since the cluster is too small: node=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a writer.go:29: 2021-01-29T19:32:18.523Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: member joined, marking health alive: member=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a writer.go:29: 2021-01-29T19:32:18.525Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.server: Skipping self join check for node since the cluster is too small: node=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a writer.go:29: 2021-01-29T19:32:18.561Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.acl: dropping node from result due to ACLs: node=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/query === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.575Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/query from=127.0.0.1:49968 latency=3.52407ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/query === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.582Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/query from=127.0.0.1:49970 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:18.586Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/query from=127.0.0.1:49970 latency=3.32179ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/query === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.597Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/query from=127.0.0.1:49972 latency=310.493µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/query === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.603Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/query from=127.0.0.1:49974 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:18.604Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/query from=127.0.0.1:49974 latency=1.102595ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/query === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.608Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/query from=127.0.0.1:49976 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:18.610Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/query from=127.0.0.1:49976 latency=1.140912ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/query === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.614Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/query from=127.0.0.1:49976 latency=1.878µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/query/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.618Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/query/ from=127.0.0.1:49976 error="failed prepared query lookup: index error: UUID must be 36 characters" writer.go:29: 2021-01-29T19:32:18.620Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/query/ from=127.0.0.1:49976 latency=2.227531ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/query/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.626Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/query/ from=127.0.0.1:49978 error="Prepared Query lookup failed: failed prepared query lookup: index error: UUID must be 36 characters" writer.go:29: 2021-01-29T19:32:18.630Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/query/ from=127.0.0.1:49978 latency=3.465327ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/query/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.637Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/query/ from=127.0.0.1:49980 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:18.639Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/query/ from=127.0.0.1:49980 latency=2.515915ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/query/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.649Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/query/ from=127.0.0.1:49982 error="Prepared Query lookup failed: failed prepared query lookup: index error: UUID must be 36 characters" writer.go:29: 2021-01-29T19:32:18.650Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/query/ from=127.0.0.1:49982 latency=1.715339ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/query/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.657Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/query/ from=127.0.0.1:49984 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:18.659Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/query/ from=127.0.0.1:49984 latency=2.13086ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/query/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.665Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/query/ from=127.0.0.1:49984 latency=111.687µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/query/xxx/execute === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.686Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/query/xxx/execute from=127.0.0.1:49986 latency=166.88µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/query/xxx/execute === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.692Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/query/xxx/execute from=127.0.0.1:49988 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:18.694Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/query/xxx/execute from=127.0.0.1:49988 latency=2.267203ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/query/xxx/execute === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.705Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/query/xxx/execute from=127.0.0.1:49990 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:18.712Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/query/xxx/execute from=127.0.0.1:49990 latency=7.192814ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/query/xxx/execute === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.720Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/query/xxx/execute from=127.0.0.1:49992 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:18.721Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/query/xxx/execute from=127.0.0.1:49992 latency=1.126198ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/query/xxx/execute === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.732Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/query/xxx/execute from=127.0.0.1:49994 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:18.734Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/query/xxx/execute from=127.0.0.1:49994 latency=1.467037ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/query/xxx/execute === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.737Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/query/xxx/execute from=127.0.0.1:49994 latency=117.856µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/query/xxx/explain === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.746Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/query/xxx/explain from=127.0.0.1:49996 latency=133.852µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/query/xxx/explain === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.753Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/query/xxx/explain from=127.0.0.1:49998 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:18.755Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/query/xxx/explain from=127.0.0.1:49998 latency=1.830144ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/query/xxx/explain === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.761Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/query/xxx/explain from=127.0.0.1:50000 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:18.763Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/query/xxx/explain from=127.0.0.1:50000 latency=2.055941ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/query/xxx/explain === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.764Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:18.770Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/query/xxx/explain from=127.0.0.1:50002 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:18.772Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/query/xxx/explain from=127.0.0.1:50002 latency=2.167715ms writer.go:29: 2021-01-29T19:32:18.775Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: Synced node info === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/query/xxx/explain === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.778Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS: Node info in sync writer.go:29: 2021-01-29T19:32:18.784Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/query/xxx/explain from=127.0.0.1:50004 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:18.788Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/query/xxx/explain from=127.0.0.1:50004 latency=4.273215ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/query/xxx/explain === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.794Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/query/xxx/explain from=127.0.0.1:50004 latency=70.552µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/update/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.800Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/check/update/ from=127.0.0.1:50006 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:18.802Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/check/update/ from=127.0.0.1:50006 latency=2.442055ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/update/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.812Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/check/update/ from=127.0.0.1:50008 latency=62.157µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/update/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.819Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/check/update/ from=127.0.0.1:50010 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:18.822Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/check/update/ from=127.0.0.1:50010 latency=2.464646ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/update/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.833Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/check/update/ from=127.0.0.1:50012 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:18.834Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/check/update/ from=127.0.0.1:50012 latency=1.860293ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/update/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.841Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/check/update/ from=127.0.0.1:50014 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:18.843Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/check/update/ from=127.0.0.1:50014 latency=1.785521ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/update/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.848Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/check/update/ from=127.0.0.1:50014 latency=2.124µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.854Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/catalog/datacenters from=127.0.0.1:50014 latency=2.101124ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.862Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/catalog/datacenters from=127.0.0.1:50016 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:18.863Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/catalog/datacenters from=127.0.0.1:50016 latency=1.112874ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.868Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/catalog/datacenters from=127.0.0.1:50018 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:18.873Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/catalog/datacenters from=127.0.0.1:50018 latency=4.424012ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.885Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/catalog/datacenters from=127.0.0.1:50020 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:18.887Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/catalog/datacenters from=127.0.0.1:50020 latency=1.986294ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.894Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/catalog/datacenters from=127.0.0.1:50022 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:18.896Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/catalog/datacenters from=127.0.0.1:50022 latency=1.75502ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.903Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/catalog/datacenters from=127.0.0.1:50022 latency=1.463µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/logout === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.907Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/logout from=127.0.0.1:50022 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:18.908Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/logout from=127.0.0.1:50022 latency=1.17979ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/logout === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.913Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/logout from=127.0.0.1:50024 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:18.914Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/logout from=127.0.0.1:50024 latency=1.148197ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/logout === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.920Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/logout from=127.0.0.1:50026 error="ACL not found" writer.go:29: 2021-01-29T19:32:18.923Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/logout from=127.0.0.1:50026 latency=2.551926ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/logout === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.942Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/logout from=127.0.0.1:50028 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:18.943Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/logout from=127.0.0.1:50028 latency=1.142494ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/logout === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.948Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/logout from=127.0.0.1:50030 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:18.950Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/logout from=127.0.0.1:50030 latency=1.8258ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/logout === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.954Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/logout from=127.0.0.1:50030 latency=2.648µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/role === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.958Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/role from=127.0.0.1:50030 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:18.960Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/role from=127.0.0.1:50030 latency=2.01825ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/role === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.968Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/role from=127.0.0.1:50032 error="Bad request: Role decoding failed: EOF" writer.go:29: 2021-01-29T19:32:18.970Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:50032 latency=1.81319ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/role === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.978Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/role from=127.0.0.1:50034 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:18.980Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/role from=127.0.0.1:50034 latency=2.130524ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/role === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.987Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/role from=127.0.0.1:50036 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:18.989Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/role from=127.0.0.1:50036 latency=1.373493ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/role === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:18.996Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/role from=127.0.0.1:50038 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:18.998Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/role from=127.0.0.1:50038 latency=1.079649ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/role === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.001Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/role from=127.0.0.1:50038 latency=2.892µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/maintenance === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.006Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/maintenance from=127.0.0.1:50038 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.008Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/maintenance from=127.0.0.1:50038 latency=1.924562ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/maintenance === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.015Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/maintenance from=127.0.0.1:50040 latency=256.057µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/maintenance === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.026Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/maintenance from=127.0.0.1:50042 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.027Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/maintenance from=127.0.0.1:50042 latency=1.849126ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/maintenance === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.034Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/maintenance from=127.0.0.1:50044 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.051Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/maintenance from=127.0.0.1:50044 latency=16.336145ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/maintenance === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.057Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/maintenance from=127.0.0.1:50046 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.059Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/maintenance from=127.0.0.1:50046 latency=1.979038ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/maintenance === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.062Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/maintenance from=127.0.0.1:50046 latency=1.981µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.066Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/session/node/ from=127.0.0.1:50046 latency=102.477µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.075Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/session/node/ from=127.0.0.1:50048 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:19.076Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/session/node/ from=127.0.0.1:50048 latency=1.900622ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.088Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/session/node/ from=127.0.0.1:50050 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.090Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/session/node/ from=127.0.0.1:50050 latency=1.874011ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.097Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/session/node/ from=127.0.0.1:50052 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.099Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/session/node/ from=127.0.0.1:50052 latency=1.824041ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.105Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/session/node/ from=127.0.0.1:50054 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.107Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/session/node/ from=127.0.0.1:50054 latency=1.846748ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.111Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/session/node/ from=127.0.0.1:50054 latency=1.557µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/policy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.116Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/policy/ from=127.0.0.1:50054 error="Bad request: Missing policy ID" writer.go:29: 2021-01-29T19:32:19.118Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/policy/ from=127.0.0.1:50054 latency=2.303799ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/policy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.125Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/policy/ from=127.0.0.1:50056 error="Bad request: Policy decoding failed: EOF" writer.go:29: 2021-01-29T19:32:19.127Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/policy/ from=127.0.0.1:50056 latency=1.263611ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/policy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.132Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/policy/ from=127.0.0.1:50058 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.133Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/policy/ from=127.0.0.1:50058 latency=1.211095ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/policy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.141Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/policy/ from=127.0.0.1:50060 error="Bad request: Missing policy ID" writer.go:29: 2021-01-29T19:32:19.143Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/policy/ from=127.0.0.1:50060 latency=1.879625ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/policy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.150Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/policy/ from=127.0.0.1:50062 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.152Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/policy/ from=127.0.0.1:50062 latency=2.260076ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/policy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.156Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/policy/ from=127.0.0.1:50062 latency=1.738µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/role/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.159Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/role/name/ from=127.0.0.1:50062 error="Bad request: Missing role Name" writer.go:29: 2021-01-29T19:32:19.160Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/role/name/ from=127.0.0.1:50062 latency=1.119702ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/role/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.166Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/role/name/ from=127.0.0.1:50064 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:19.168Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/role/name/ from=127.0.0.1:50064 latency=2.114513ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/role/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.174Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/role/name/ from=127.0.0.1:50066 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.189Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/role/name/ from=127.0.0.1:50066 latency=15.668145ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/role/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.205Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/role/name/ from=127.0.0.1:50068 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.206Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/role/name/ from=127.0.0.1:50068 latency=1.204376ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/role/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.212Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/role/name/ from=127.0.0.1:50070 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.213Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/role/name/ from=127.0.0.1:50070 latency=1.538475ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/role/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.218Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/role/name/ from=127.0.0.1:50070 latency=3.646µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/event/fire/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.223Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/event/fire/ from=127.0.0.1:50070 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.226Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/event/fire/ from=127.0.0.1:50070 latency=2.547943ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/event/fire/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.235Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/event/fire/ from=127.0.0.1:50072 latency=79.567µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/event/fire/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.248Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/event/fire/ from=127.0.0.1:50074 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.250Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/event/fire/ from=127.0.0.1:50074 latency=1.989474ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/event/fire/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.257Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/event/fire/ from=127.0.0.1:50076 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.259Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/event/fire/ from=127.0.0.1:50076 latency=1.922254ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/event/fire/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.285Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/event/fire/ from=127.0.0.1:50078 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.288Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/event/fire/ from=127.0.0.1:50078 latency=2.449901ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/event/fire/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.293Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/event/fire/ from=127.0.0.1:50078 latency=2.056µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/internal/acl/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.298Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/internal/acl/authorize from=127.0.0.1:50078 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.301Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/internal/acl/authorize from=127.0.0.1:50078 latency=2.466105ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/internal/acl/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.309Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/internal/acl/authorize from=127.0.0.1:50080 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:19.311Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/internal/acl/authorize from=127.0.0.1:50080 latency=2.189978ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/internal/acl/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.319Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/internal/acl/authorize from=127.0.0.1:50082 error="Bad request: Failed to decode request body: EOF" writer.go:29: 2021-01-29T19:32:19.321Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/internal/acl/authorize from=127.0.0.1:50082 latency=1.727385ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/internal/acl/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.340Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/internal/acl/authorize from=127.0.0.1:50084 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.341Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/internal/acl/authorize from=127.0.0.1:50084 latency=1.106478ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/internal/acl/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.345Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/internal/acl/authorize from=127.0.0.1:50086 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.347Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/internal/acl/authorize from=127.0.0.1:50086 latency=1.28719ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/internal/acl/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.351Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/internal/acl/authorize from=127.0.0.1:50086 latency=1.769µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/auth-method/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.353Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/auth-method/ from=127.0.0.1:50086 error="Bad request: Missing auth method name" writer.go:29: 2021-01-29T19:32:19.354Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/auth-method/ from=127.0.0.1:50086 latency=1.23238ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/auth-method/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.359Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/auth-method/ from=127.0.0.1:50088 error="Bad request: AuthMethod decoding failed: EOF" writer.go:29: 2021-01-29T19:32:19.362Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/auth-method/ from=127.0.0.1:50088 latency=2.501818ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/auth-method/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.374Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/auth-method/ from=127.0.0.1:50090 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.377Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/auth-method/ from=127.0.0.1:50090 latency=2.83001ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/auth-method/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.388Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/auth-method/ from=127.0.0.1:50092 error="Bad request: Missing auth method name" writer.go:29: 2021-01-29T19:32:19.390Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/auth-method/ from=127.0.0.1:50092 latency=2.05787ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/auth-method/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.398Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/auth-method/ from=127.0.0.1:50094 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.400Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/auth-method/ from=127.0.0.1:50094 latency=2.260519ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/auth-method/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.408Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/auth-method/ from=127.0.0.1:50094 latency=2.147µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/auth-method === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.413Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/auth-method from=127.0.0.1:50094 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.415Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/auth-method from=127.0.0.1:50094 latency=2.260009ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/auth-method === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.425Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/auth-method from=127.0.0.1:50096 error="Bad request: AuthMethod decoding failed: EOF" writer.go:29: 2021-01-29T19:32:19.427Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:50096 latency=2.050626ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/auth-method === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.435Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/auth-method from=127.0.0.1:50098 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.438Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/auth-method from=127.0.0.1:50098 latency=2.968681ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/auth-method === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.446Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/auth-method from=127.0.0.1:50100 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.448Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/auth-method from=127.0.0.1:50100 latency=2.40026ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/auth-method === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.456Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/auth-method from=127.0.0.1:50102 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.458Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/auth-method from=127.0.0.1:50102 latency=1.994713ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/auth-method === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.464Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/auth-method from=127.0.0.1:50102 latency=2.61µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/rules/translate === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.468Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/rules/translate from=127.0.0.1:50102 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.471Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/rules/translate from=127.0.0.1:50102 latency=2.201401ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/rules/translate === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.479Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/rules/translate from=127.0.0.1:50104 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:19.482Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/rules/translate from=127.0.0.1:50104 latency=2.494613ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/rules/translate === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.491Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/rules/translate from=127.0.0.1:50106 error="Permission denied" writer.go:29: 2021-01-29T19:32:19.493Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/rules/translate from=127.0.0.1:50106 latency=2.019423ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/rules/translate === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.500Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/rules/translate from=127.0.0.1:50108 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.502Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/rules/translate from=127.0.0.1:50108 latency=1.921914ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/rules/translate === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.510Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/rules/translate from=127.0.0.1:50110 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.511Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/rules/translate from=127.0.0.1:50110 latency=1.11498ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/rules/translate === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.518Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/rules/translate from=127.0.0.1:50110 latency=2.353µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.521Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/check/register from=127.0.0.1:50110 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.524Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/check/register from=127.0.0.1:50110 latency=2.310363ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.530Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/check/register from=127.0.0.1:50112 latency=93.688µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.539Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/check/register from=127.0.0.1:50114 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.543Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/check/register from=127.0.0.1:50114 latency=3.821352ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.555Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/check/register from=127.0.0.1:50116 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.557Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/check/register from=127.0.0.1:50116 latency=1.672293ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.562Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/check/register from=127.0.0.1:50118 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.563Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/check/register from=127.0.0.1:50118 latency=1.029489ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.566Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/check/register from=127.0.0.1:50118 latency=1.172µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/bootstrap === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.569Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/bootstrap from=127.0.0.1:50118 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.570Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/bootstrap from=127.0.0.1:50118 latency=1.23039ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/bootstrap === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.574Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/bootstrap from=127.0.0.1:50120 latency=145.157µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/bootstrap === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.580Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/bootstrap from=127.0.0.1:50122 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.582Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/bootstrap from=127.0.0.1:50122 latency=1.819298ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/bootstrap === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.586Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/bootstrap from=127.0.0.1:50124 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.587Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/bootstrap from=127.0.0.1:50124 latency=1.046366ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/bootstrap === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.592Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/bootstrap from=127.0.0.1:50126 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.594Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/bootstrap from=127.0.0.1:50126 latency=1.168809ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/bootstrap === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.596Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/bootstrap from=127.0.0.1:50126 latency=1.228µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/roles === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.599Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/roles from=127.0.0.1:50126 error="Permission denied" writer.go:29: 2021-01-29T19:32:19.600Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/roles from=127.0.0.1:50126 latency=1.185044ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/roles === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.605Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/roles from=127.0.0.1:50128 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:19.606Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/roles from=127.0.0.1:50128 latency=1.222205ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/roles === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.611Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/roles from=127.0.0.1:50130 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.613Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/roles from=127.0.0.1:50130 latency=2.183381ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/roles === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.619Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/roles from=127.0.0.1:50132 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.620Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/roles from=127.0.0.1:50132 latency=1.144005ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/roles === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.625Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/roles from=127.0.0.1:50134 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.627Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/roles from=127.0.0.1:50134 latency=1.262959ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/roles === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.629Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/roles from=127.0.0.1:50134 latency=1.523µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/event/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.633Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/event/list from=127.0.0.1:50134 latency=177.931µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/event/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.637Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/event/list from=127.0.0.1:50136 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:19.639Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/event/list from=127.0.0.1:50136 latency=2.000789ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/event/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.646Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/event/list from=127.0.0.1:50138 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.648Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/event/list from=127.0.0.1:50138 latency=1.468337ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/event/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.652Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/event/list from=127.0.0.1:50140 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.653Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/event/list from=127.0.0.1:50140 latency=1.156075ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/event/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.658Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/event/list from=127.0.0.1:50142 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.659Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/event/list from=127.0.0.1:50142 latency=1.112925ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/event/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.662Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/event/list from=127.0.0.1:50142 latency=1.397µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/internal/ui/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.667Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.acl: dropping node from result due to ACLs: node=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a writer.go:29: 2021-01-29T19:32:19.670Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/internal/ui/services from=127.0.0.1:50142 latency=3.302634ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/internal/ui/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.675Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/internal/ui/services from=127.0.0.1:50144 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:19.677Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/internal/ui/services from=127.0.0.1:50144 latency=1.156892ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/internal/ui/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.682Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/internal/ui/services from=127.0.0.1:50146 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.684Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/internal/ui/services from=127.0.0.1:50146 latency=1.704375ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/internal/ui/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.690Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/internal/ui/services from=127.0.0.1:50148 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.692Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/internal/ui/services from=127.0.0.1:50148 latency=1.23183ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/internal/ui/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.696Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/internal/ui/services from=127.0.0.1:50150 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.697Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/internal/ui/services from=127.0.0.1:50150 latency=1.057754ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/internal/ui/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.704Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/internal/ui/services from=127.0.0.1:50150 latency=2.131µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/policy === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.707Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/policy from=127.0.0.1:50150 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.710Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/policy from=127.0.0.1:50150 latency=2.433438ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/policy === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.721Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/policy from=127.0.0.1:50152 error="Bad request: Policy decoding failed: EOF" writer.go:29: 2021-01-29T19:32:19.724Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:50152 latency=3.033024ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/policy === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.731Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/policy from=127.0.0.1:50154 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.733Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/policy from=127.0.0.1:50154 latency=2.579151ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/policy === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.740Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/policy from=127.0.0.1:50156 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.741Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/policy from=127.0.0.1:50156 latency=1.157836ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/policy === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.745Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/policy from=127.0.0.1:50158 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.746Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/policy from=127.0.0.1:50158 latency=1.068806ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/policy === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.749Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/policy from=127.0.0.1:50158 latency=863ns === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/service/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.751Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/service/register from=127.0.0.1:50158 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.752Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/service/register from=127.0.0.1:50158 latency=1.149832ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/service/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.757Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:50160 latency=111.003µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/service/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.761Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/service/register from=127.0.0.1:50162 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.763Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/service/register from=127.0.0.1:50162 latency=1.239062ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/service/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.768Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/service/register from=127.0.0.1:50164 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.769Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/service/register from=127.0.0.1:50164 latency=1.606697ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/service/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.774Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/service/register from=127.0.0.1:50166 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.775Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/service/register from=127.0.0.1:50166 latency=1.037416ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/service/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.777Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/service/register from=127.0.0.1:50166 latency=1.036µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/config === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.780Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/config from=127.0.0.1:50166 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.781Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/config from=127.0.0.1:50166 latency=1.187037ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/config === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.786Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/config from=127.0.0.1:50168 error="Bad request: Request decoding failed: EOF" writer.go:29: 2021-01-29T19:32:19.787Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/config from=127.0.0.1:50168 latency=1.243079ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/config === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.791Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/config from=127.0.0.1:50170 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.792Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/config from=127.0.0.1:50170 latency=1.151624ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/config === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.797Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/config from=127.0.0.1:50172 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.798Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/config from=127.0.0.1:50172 latency=1.070325ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/config === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.802Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/config from=127.0.0.1:50174 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.803Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/config from=127.0.0.1:50174 latency=1.045698ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/config === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.806Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/config from=127.0.0.1:50174 latency=1.261µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.809Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/catalog/connect/ from=127.0.0.1:50174 latency=96.617µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.813Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/catalog/connect/ from=127.0.0.1:50176 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:19.814Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/catalog/connect/ from=127.0.0.1:50176 latency=1.202883ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.819Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/catalog/connect/ from=127.0.0.1:50178 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.820Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/catalog/connect/ from=127.0.0.1:50178 latency=1.121913ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.825Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/catalog/connect/ from=127.0.0.1:50180 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.826Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/catalog/connect/ from=127.0.0.1:50180 latency=1.158544ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.831Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/catalog/connect/ from=127.0.0.1:50182 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.833Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/catalog/connect/ from=127.0.0.1:50182 latency=1.087277ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.836Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/catalog/connect/ from=127.0.0.1:50182 latency=2.969µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/binding-rule === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.841Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/binding-rule from=127.0.0.1:50182 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.842Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/binding-rule from=127.0.0.1:50182 latency=1.123915ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/binding-rule === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.846Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:50184 error="Bad request: BindingRule decoding failed: EOF" writer.go:29: 2021-01-29T19:32:19.847Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:50184 latency=1.172993ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/binding-rule === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.852Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/binding-rule from=127.0.0.1:50186 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.853Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/binding-rule from=127.0.0.1:50186 latency=1.343084ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/binding-rule === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.858Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/binding-rule from=127.0.0.1:50188 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.859Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/binding-rule from=127.0.0.1:50188 latency=1.231214ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/binding-rule === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.865Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/binding-rule from=127.0.0.1:50190 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.866Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/binding-rule from=127.0.0.1:50190 latency=1.172199ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/binding-rule === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.869Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/binding-rule from=127.0.0.1:50190 latency=1.125µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/members === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.871Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS: dropping node from result due to ACLs: node=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a accessorID= writer.go:29: 2021-01-29T19:32:19.874Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/members from=127.0.0.1:50190 latency=2.230692ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/members === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.881Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/members from=127.0.0.1:50192 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:19.882Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/members from=127.0.0.1:50192 latency=1.897703ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/members === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.890Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/members from=127.0.0.1:50194 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.892Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/members from=127.0.0.1:50194 latency=2.505909ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/members === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.900Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/members from=127.0.0.1:50196 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.901Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/members from=127.0.0.1:50196 latency=1.200063ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/members === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.906Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/members from=127.0.0.1:50198 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.907Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/members from=127.0.0.1:50198 latency=1.415486ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/members === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.912Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/members from=127.0.0.1:50198 latency=2.408µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/operator/keyring === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.916Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/operator/keyring from=127.0.0.1:50198 error="Reading keyring denied by ACLs" writer.go:29: 2021-01-29T19:32:19.918Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/operator/keyring from=127.0.0.1:50198 latency=2.227348ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/operator/keyring === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.926Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/operator/keyring from=127.0.0.1:50200 error="Bad request: Request decode failed: EOF" writer.go:29: 2021-01-29T19:32:19.928Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/operator/keyring from=127.0.0.1:50200 latency=2.077031ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/operator/keyring === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.934Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/operator/keyring from=127.0.0.1:50202 error="Bad request: Request decode failed: EOF" writer.go:29: 2021-01-29T19:32:19.935Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/operator/keyring from=127.0.0.1:50202 latency=1.557841ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/operator/keyring === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.941Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/operator/keyring from=127.0.0.1:50204 error="Bad request: Request decode failed: EOF" writer.go:29: 2021-01-29T19:32:19.943Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/operator/keyring from=127.0.0.1:50204 latency=2.213235ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/operator/keyring === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.951Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/operator/keyring from=127.0.0.1:50206 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.952Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/operator/keyring from=127.0.0.1:50206 latency=1.082684ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/operator/keyring === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.954Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/operator/keyring from=127.0.0.1:50206 latency=2.173µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/txn === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.957Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/txn from=127.0.0.1:50206 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:19.959Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/txn from=127.0.0.1:50206 latency=1.378234ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/txn === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.965Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/txn from=127.0.0.1:50208 latency=104.763µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/txn === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.971Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/txn from=127.0.0.1:50210 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:19.974Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/txn from=127.0.0.1:50210 latency=2.894826ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/txn === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.983Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/txn from=127.0.0.1:50212 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:19.985Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/txn from=127.0.0.1:50212 latency=1.983088ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/txn === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.992Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/txn from=127.0.0.1:50214 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:19.993Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/txn from=127.0.0.1:50214 latency=1.134913ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/txn === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:19.998Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/txn from=127.0.0.1:50214 latency=1.902µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/checks === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.001Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/checks from=127.0.0.1:50214 latency=219.693µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/checks === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.010Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/checks from=127.0.0.1:50216 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.013Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/checks from=127.0.0.1:50216 latency=2.90371ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/checks === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.026Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/checks from=127.0.0.1:50218 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.028Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/checks from=127.0.0.1:50218 latency=2.65588ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/checks === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.034Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/checks from=127.0.0.1:50220 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.036Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/checks from=127.0.0.1:50220 latency=1.222027ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/checks === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.040Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/checks from=127.0.0.1:50222 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.043Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/checks from=127.0.0.1:50222 latency=2.362063ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/checks === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.048Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/checks from=127.0.0.1:50222 latency=2.523µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/kv/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.053Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/kv/ from=127.0.0.1:50222 latency=68.741µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/kv/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.056Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:20.057Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS: Node info in sync writer.go:29: 2021-01-29T19:32:20.058Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/kv/ from=127.0.0.1:50224 latency=34.128µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/kv/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.064Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/kv/ from=127.0.0.1:50226 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.066Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/kv/ from=127.0.0.1:50226 latency=2.355727ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/kv/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.082Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/kv/ from=127.0.0.1:50228 latency=64.244µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/kv/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.097Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/kv/ from=127.0.0.1:50230 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.099Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/kv/ from=127.0.0.1:50230 latency=1.494493ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/kv/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.102Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/kv/ from=127.0.0.1:50230 latency=2.937µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/host === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.109Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/host from=127.0.0.1:50230 error="Permission denied" writer.go:29: 2021-01-29T19:32:20.111Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/host from=127.0.0.1:50230 latency=2.343804ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/host === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.122Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/host from=127.0.0.1:50232 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.128Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/host from=127.0.0.1:50232 latency=5.468839ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/host === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.141Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/host from=127.0.0.1:50234 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.143Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/host from=127.0.0.1:50234 latency=2.358981ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/host === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.150Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/host from=127.0.0.1:50236 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.151Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/host from=127.0.0.1:50236 latency=1.199154ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/host === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.156Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/host from=127.0.0.1:50238 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.157Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/host from=127.0.0.1:50238 latency=1.43612ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/host === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.161Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/host from=127.0.0.1:50238 latency=1.98µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/warn/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.164Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/check/warn/ from=127.0.0.1:50238 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:20.166Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/check/warn/ from=127.0.0.1:50238 latency=1.497707ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/warn/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.172Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/check/warn/ from=127.0.0.1:50240 error="Unknown check """ writer.go:29: 2021-01-29T19:32:20.174Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/check/warn/ from=127.0.0.1:50240 latency=1.775554ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/warn/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.179Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/check/warn/ from=127.0.0.1:50242 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.181Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/check/warn/ from=127.0.0.1:50242 latency=1.503538ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/warn/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.186Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/check/warn/ from=127.0.0.1:50244 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.187Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/check/warn/ from=127.0.0.1:50244 latency=1.350021ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/warn/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.194Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/check/warn/ from=127.0.0.1:50246 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.196Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/check/warn/ from=127.0.0.1:50246 latency=2.164313ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/warn/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.201Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/check/warn/ from=127.0.0.1:50246 latency=4.408µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.204Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/catalog/service/ from=127.0.0.1:50246 latency=72.486µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.209Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/catalog/service/ from=127.0.0.1:50248 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.211Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/catalog/service/ from=127.0.0.1:50248 latency=1.736728ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.217Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/catalog/service/ from=127.0.0.1:50250 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.218Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/catalog/service/ from=127.0.0.1:50250 latency=1.323535ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.223Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/catalog/service/ from=127.0.0.1:50252 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.224Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/catalog/service/ from=127.0.0.1:50252 latency=1.131251ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.228Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/catalog/service/ from=127.0.0.1:50254 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.229Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/catalog/service/ from=127.0.0.1:50254 latency=1.132357ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.234Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/catalog/service/ from=127.0.0.1:50254 latency=1.964µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/coordinate/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.237Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/coordinate/update from=127.0.0.1:50254 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:20.238Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/coordinate/update from=127.0.0.1:50254 latency=1.408112ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/coordinate/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.242Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/coordinate/update from=127.0.0.1:50256 latency=67.58µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/coordinate/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.248Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/coordinate/update from=127.0.0.1:50258 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.250Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/coordinate/update from=127.0.0.1:50258 latency=2.491151ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/coordinate/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.256Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/coordinate/update from=127.0.0.1:50260 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.257Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/coordinate/update from=127.0.0.1:50260 latency=1.281177ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/coordinate/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.262Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/coordinate/update from=127.0.0.1:50262 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.263Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/coordinate/update from=127.0.0.1:50262 latency=1.822778ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/coordinate/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.269Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/coordinate/update from=127.0.0.1:50262 latency=1.86µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/auth-methods === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.273Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/auth-methods from=127.0.0.1:50262 error="Permission denied" writer.go:29: 2021-01-29T19:32:20.274Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/auth-methods from=127.0.0.1:50262 latency=1.322819ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/auth-methods === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.283Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/auth-methods from=127.0.0.1:50264 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.287Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/auth-methods from=127.0.0.1:50264 latency=4.050265ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/auth-methods === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.294Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/auth-methods from=127.0.0.1:50266 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.295Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/auth-methods from=127.0.0.1:50266 latency=1.156763ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/auth-methods === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.300Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/auth-methods from=127.0.0.1:50268 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.301Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/auth-methods from=127.0.0.1:50268 latency=1.16297ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/auth-methods === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.305Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/auth-methods from=127.0.0.1:50270 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.306Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/auth-methods from=127.0.0.1:50270 latency=1.12878ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/auth-methods === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.309Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/auth-methods from=127.0.0.1:50270 latency=1.746µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/connect/ca/leaf/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.319Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50270 error="URI must be either service or agent" writer.go:29: 2021-01-29T19:32:20.321Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50270 latency=8.994283ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/connect/ca/leaf/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.329Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50272 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.332Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50272 latency=2.274135ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/connect/ca/leaf/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.340Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50274 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.342Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50274 latency=2.000735ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/connect/ca/leaf/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.350Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50276 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.352Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50276 latency=1.823803ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/connect/ca/leaf/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.358Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50278 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.360Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50278 latency=1.932117ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/connect/ca/leaf/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.374Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/connect/ca/leaf/ from=127.0.0.1:50278 latency=1.767µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/coordinate/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.382Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/coordinate/nodes from=127.0.0.1:50278 latency=2.808373ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/coordinate/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.390Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/coordinate/nodes from=127.0.0.1:50280 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.392Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/coordinate/nodes from=127.0.0.1:50280 latency=2.106752ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/coordinate/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.399Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/coordinate/nodes from=127.0.0.1:50282 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.402Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/coordinate/nodes from=127.0.0.1:50282 latency=2.720435ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/coordinate/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.410Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/coordinate/nodes from=127.0.0.1:50284 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.412Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/coordinate/nodes from=127.0.0.1:50284 latency=1.909505ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/coordinate/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.420Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/coordinate/nodes from=127.0.0.1:50286 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.422Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/coordinate/nodes from=127.0.0.1:50286 latency=2.408275ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/coordinate/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.429Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/coordinate/nodes from=127.0.0.1:50286 latency=2.117µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/clone/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.434Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/clone/ from=127.0.0.1:50286 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:20.436Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/clone/ from=127.0.0.1:50286 latency=1.113515ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/clone/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.440Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/clone/ from=127.0.0.1:50288 latency=95.604µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/clone/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.447Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/clone/ from=127.0.0.1:50290 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.449Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/clone/ from=127.0.0.1:50290 latency=2.489968ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/clone/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.456Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/clone/ from=127.0.0.1:50292 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.458Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/clone/ from=127.0.0.1:50292 latency=1.313222ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/clone/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.463Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/clone/ from=127.0.0.1:50294 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.464Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/clone/ from=127.0.0.1:50294 latency=1.062734ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/clone/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.467Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/clone/ from=127.0.0.1:50294 latency=1.218µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.470Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/catalog/node/ from=127.0.0.1:50294 latency=111.006µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.474Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/catalog/node/ from=127.0.0.1:50296 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.475Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/catalog/node/ from=127.0.0.1:50296 latency=1.050733ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.480Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/catalog/node/ from=127.0.0.1:50298 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.481Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/catalog/node/ from=127.0.0.1:50298 latency=1.284514ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.501Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:20.507Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/catalog/node/ from=127.0.0.1:50300 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.513Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/catalog/node/ from=127.0.0.1:50300 latency=6.699461ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.519Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/catalog/node/ from=127.0.0.1:50302 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.520Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/catalog/node/ from=127.0.0.1:50302 latency=1.325523ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.524Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/catalog/node/ from=127.0.0.1:50302 latency=1.579µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.534Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/connect/ca/roots from=127.0.0.1:50302 latency=3.725286ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.540Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/connect/ca/roots from=127.0.0.1:50304 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.542Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/connect/ca/roots from=127.0.0.1:50304 latency=1.226749ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.546Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/connect/ca/roots from=127.0.0.1:50306 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.547Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/connect/ca/roots from=127.0.0.1:50306 latency=1.122516ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.552Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/connect/ca/roots from=127.0.0.1:50308 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.554Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/connect/ca/roots from=127.0.0.1:50308 latency=1.941284ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.563Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/connect/ca/roots from=127.0.0.1:50310 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.564Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/connect/ca/roots from=127.0.0.1:50310 latency=1.070045ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.567Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/connect/ca/roots from=127.0.0.1:50310 latency=1.878µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/health/checks/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.571Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/health/checks/ from=127.0.0.1:50310 latency=84.844µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/health/checks/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.576Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/health/checks/ from=127.0.0.1:50312 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.578Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/health/checks/ from=127.0.0.1:50312 latency=1.89961ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/health/checks/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.594Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/health/checks/ from=127.0.0.1:50314 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.596Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/health/checks/ from=127.0.0.1:50314 latency=1.84383ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/health/checks/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.600Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/health/checks/ from=127.0.0.1:50316 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.601Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/health/checks/ from=127.0.0.1:50316 latency=1.160553ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/health/checks/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.612Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/health/checks/ from=127.0.0.1:50318 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.614Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/health/checks/ from=127.0.0.1:50318 latency=2.170253ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/health/checks/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.619Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/health/checks/ from=127.0.0.1:50318 latency=1.303µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/discovery-chain/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.623Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/discovery-chain/ from=127.0.0.1:50318 error="Bad request: Missing chain name" writer.go:29: 2021-01-29T19:32:20.625Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/discovery-chain/ from=127.0.0.1:50318 latency=1.934784ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/discovery-chain/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.631Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/discovery-chain/ from=127.0.0.1:50320 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.634Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/discovery-chain/ from=127.0.0.1:50320 latency=2.573874ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/discovery-chain/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.640Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/discovery-chain/ from=127.0.0.1:50322 error="Bad request: Missing chain name" writer.go:29: 2021-01-29T19:32:20.641Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/discovery-chain/ from=127.0.0.1:50322 latency=1.191294ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/discovery-chain/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.654Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/discovery-chain/ from=127.0.0.1:50324 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.655Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/discovery-chain/ from=127.0.0.1:50324 latency=1.085519ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/discovery-chain/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.663Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/discovery-chain/ from=127.0.0.1:50326 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.664Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/discovery-chain/ from=127.0.0.1:50326 latency=1.047171ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/discovery-chain/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.666Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/discovery-chain/ from=127.0.0.1:50326 latency=1.11µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/health/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.669Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/health/node/ from=127.0.0.1:50326 latency=107.466µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/health/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.673Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/health/node/ from=127.0.0.1:50328 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.675Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/health/node/ from=127.0.0.1:50328 latency=2.067538ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/health/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.682Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/health/node/ from=127.0.0.1:50330 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.684Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/health/node/ from=127.0.0.1:50330 latency=1.824148ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/health/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.688Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/health/node/ from=127.0.0.1:50332 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.689Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/health/node/ from=127.0.0.1:50332 latency=1.425021ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/health/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.695Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/health/node/ from=127.0.0.1:50334 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.696Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/health/node/ from=127.0.0.1:50334 latency=1.389128ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/health/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.699Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/health/node/ from=127.0.0.1:50334 latency=1.061µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/status/leader === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.703Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/status/leader from=127.0.0.1:50334 latency=131.027µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/status/leader === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.708Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/status/leader from=127.0.0.1:50336 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.710Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/status/leader from=127.0.0.1:50336 latency=1.877158ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/status/leader === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.714Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/status/leader from=127.0.0.1:50338 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.716Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/status/leader from=127.0.0.1:50338 latency=1.272811ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/status/leader === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.720Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/status/leader from=127.0.0.1:50340 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.722Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/status/leader from=127.0.0.1:50340 latency=2.078383ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/status/leader === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.727Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/status/leader from=127.0.0.1:50342 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.729Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/status/leader from=127.0.0.1:50342 latency=1.865614ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/status/leader === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.734Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/status/leader from=127.0.0.1:50342 latency=2.017µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/intentions/match === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.738Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/connect/intentions/match from=127.0.0.1:50342 error="required query parameter 'by' not set" writer.go:29: 2021-01-29T19:32:20.741Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/connect/intentions/match from=127.0.0.1:50342 latency=2.460376ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/intentions/match === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.750Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/connect/intentions/match from=127.0.0.1:50344 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.752Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/connect/intentions/match from=127.0.0.1:50344 latency=2.035442ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/intentions/match === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.759Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/connect/intentions/match from=127.0.0.1:50346 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.762Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/connect/intentions/match from=127.0.0.1:50346 latency=2.017715ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/intentions/match === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.778Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/connect/intentions/match from=127.0.0.1:50348 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.782Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/connect/intentions/match from=127.0.0.1:50348 latency=3.579971ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/intentions/match === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.792Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/connect/intentions/match from=127.0.0.1:50350 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.793Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/connect/intentions/match from=127.0.0.1:50350 latency=1.163011ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/intentions/match === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.796Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/connect/intentions/match from=127.0.0.1:50350 latency=1.785µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/renew/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.798Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/session/renew/ from=127.0.0.1:50350 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:20.801Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/session/renew/ from=127.0.0.1:50350 latency=2.418884ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/renew/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.822Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/session/renew/ from=127.0.0.1:50354 latency=102.735µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/renew/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.826Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/session/renew/ from=127.0.0.1:50358 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.828Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/session/renew/ from=127.0.0.1:50358 latency=1.224979ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/renew/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.831Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/session/renew/ from=127.0.0.1:50362 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.832Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/session/renew/ from=127.0.0.1:50362 latency=1.04756ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/renew/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.836Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/session/renew/ from=127.0.0.1:50364 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.837Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/session/renew/ from=127.0.0.1:50364 latency=1.042013ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/renew/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.842Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/session/renew/ from=127.0.0.1:50364 latency=1.797µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.851Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/create from=127.0.0.1:50364 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:20.855Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/create from=127.0.0.1:50364 latency=3.464988ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.863Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/create from=127.0.0.1:50366 error="Permission denied" writer.go:29: 2021-01-29T19:32:20.864Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/create from=127.0.0.1:50366 latency=1.2785ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.869Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/create from=127.0.0.1:50368 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.870Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/create from=127.0.0.1:50368 latency=1.079436ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.875Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/create from=127.0.0.1:50370 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:20.878Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/create from=127.0.0.1:50370 latency=3.009001ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.885Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/create from=127.0.0.1:50372 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.887Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/create from=127.0.0.1:50372 latency=1.748713ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.891Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/create from=127.0.0.1:50372 latency=2.175µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/config/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.901Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/config/ from=127.0.0.1:50372 latency=623.427µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/config/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.914Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/config/ from=127.0.0.1:50380 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.917Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/config/ from=127.0.0.1:50380 latency=2.871869ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/config/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.928Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/config/ from=127.0.0.1:50382 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:20.933Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/config/ from=127.0.0.1:50382 latency=4.92165ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/config/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.950Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/config/ from=127.0.0.1:50388 latency=52.301µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/config/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.957Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/config/ from=127.0.0.1:50392 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:20.958Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/config/ from=127.0.0.1:50392 latency=1.128805ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/config/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.968Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/config/ from=127.0.0.1:50392 latency=1.222µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/intentions === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.970Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:50392 latency=214.463µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/intentions === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.977Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/connect/intentions from=127.0.0.1:50396 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:20.979Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/connect/intentions from=127.0.0.1:50396 latency=1.785562ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/intentions === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:20.991Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/connect/intentions from=127.0.0.1:50400 error="Failed to decode request body: EOF" writer.go:29: 2021-01-29T19:32:20.993Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:50400 latency=1.974749ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/intentions === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.003Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/connect/intentions from=127.0.0.1:50402 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.005Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/connect/intentions from=127.0.0.1:50402 latency=1.229913ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/intentions === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.017Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/connect/intentions from=127.0.0.1:50406 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.018Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/connect/intentions from=127.0.0.1:50406 latency=1.01285ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/intentions === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.021Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/connect/intentions from=127.0.0.1:50406 latency=1.107µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/operator/raft/peer === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.024Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/operator/raft/peer from=127.0.0.1:50406 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:21.025Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/operator/raft/peer from=127.0.0.1:50406 latency=1.233647ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/operator/raft/peer === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.030Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/operator/raft/peer from=127.0.0.1:50408 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.032Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/operator/raft/peer from=127.0.0.1:50408 latency=1.090374ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/operator/raft/peer === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.037Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/operator/raft/peer from=127.0.0.1:50410 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.038Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/operator/raft/peer from=127.0.0.1:50410 latency=1.951131ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/operator/raft/peer === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.043Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/operator/raft/peer from=127.0.0.1:50412 error="Bad request: Must specify either ?id with the server's ID or ?address with IP:port of peer to remove" writer.go:29: 2021-01-29T19:32:21.044Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/operator/raft/peer from=127.0.0.1:50412 latency=1.051888ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/operator/raft/peer === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.049Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/operator/raft/peer from=127.0.0.1:50416 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.050Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/operator/raft/peer from=127.0.0.1:50416 latency=1.105684ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/operator/raft/peer === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.053Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/operator/raft/peer from=127.0.0.1:50416 latency=1.752µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.055Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/check/deregister/ from=127.0.0.1:50416 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:21.056Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/check/deregister/ from=127.0.0.1:50416 latency=1.252008ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.061Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/check/deregister/ from=127.0.0.1:50418 error="Unknown check """ writer.go:29: 2021-01-29T19:32:21.063Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/check/deregister/ from=127.0.0.1:50418 latency=2.041106ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.069Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/check/deregister/ from=127.0.0.1:50420 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.070Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/check/deregister/ from=127.0.0.1:50420 latency=1.347062ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.074Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/check/deregister/ from=127.0.0.1:50422 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.076Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/check/deregister/ from=127.0.0.1:50422 latency=1.152336ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.080Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/check/deregister/ from=127.0.0.1:50424 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.081Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/check/deregister/ from=127.0.0.1:50424 latency=1.019147ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.086Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/check/deregister/ from=127.0.0.1:50424 latency=972ns === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/connect/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.090Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/connect/authorize from=127.0.0.1:50424 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:21.092Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/connect/authorize from=127.0.0.1:50424 latency=2.384779ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/connect/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.097Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/connect/authorize from=127.0.0.1:50430 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.099Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/connect/authorize from=127.0.0.1:50430 latency=1.136725ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/connect/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.103Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/connect/authorize from=127.0.0.1:50432 error="Bad request: Request decode failed: EOF" writer.go:29: 2021-01-29T19:32:21.104Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/connect/authorize from=127.0.0.1:50432 latency=1.096453ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/connect/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.108Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/connect/authorize from=127.0.0.1:50434 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.109Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/connect/authorize from=127.0.0.1:50434 latency=1.060569ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/connect/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.114Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/connect/authorize from=127.0.0.1:50436 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.116Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/connect/authorize from=127.0.0.1:50436 latency=1.68074ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/connect/authorize === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.121Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/connect/authorize from=127.0.0.1:50436 latency=2.071µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.127Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/session/destroy/ from=127.0.0.1:50436 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:21.129Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/session/destroy/ from=127.0.0.1:50436 latency=1.850871ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.134Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/session/destroy/ from=127.0.0.1:50438 latency=54.762µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.139Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/session/destroy/ from=127.0.0.1:50440 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.141Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/session/destroy/ from=127.0.0.1:50440 latency=1.980598ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.148Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/session/destroy/ from=127.0.0.1:50442 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.150Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/session/destroy/ from=127.0.0.1:50442 latency=1.980338ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.157Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/session/destroy/ from=127.0.0.1:50444 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.159Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/session/destroy/ from=127.0.0.1:50444 latency=1.740777ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.163Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/session/destroy/ from=127.0.0.1:50444 latency=3.215µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/policies === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.167Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/policies from=127.0.0.1:50444 error="Permission denied" writer.go:29: 2021-01-29T19:32:21.169Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/policies from=127.0.0.1:50444 latency=2.897715ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/policies === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.186Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/policies from=127.0.0.1:50446 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.187Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/policies from=127.0.0.1:50446 latency=1.214596ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/policies === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.191Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/policies from=127.0.0.1:50448 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.192Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/policies from=127.0.0.1:50448 latency=1.113035ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/policies === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.197Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/policies from=127.0.0.1:50450 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.198Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/policies from=127.0.0.1:50450 latency=1.161322ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/policies === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.203Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/policies from=127.0.0.1:50452 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.204Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/policies from=127.0.0.1:50452 latency=1.214699ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/policies === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.207Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/policies from=127.0.0.1:50452 latency=2.118µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/deregister === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.210Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/catalog/deregister from=127.0.0.1:50452 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:21.212Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/catalog/deregister from=127.0.0.1:50452 latency=1.37515ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/deregister === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.217Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/catalog/deregister from=127.0.0.1:50454 latency=77.825µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/deregister === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.222Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/catalog/deregister from=127.0.0.1:50456 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.224Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/catalog/deregister from=127.0.0.1:50456 latency=2.021889ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/deregister === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.230Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/catalog/deregister from=127.0.0.1:50458 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.233Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/catalog/deregister from=127.0.0.1:50458 latency=2.603149ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/deregister === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.240Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/catalog/deregister from=127.0.0.1:50460 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.242Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/catalog/deregister from=127.0.0.1:50460 latency=2.10618ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/deregister === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.246Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/catalog/deregister from=127.0.0.1:50460 latency=2.414µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/snapshot === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.250Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/snapshot from=127.0.0.1:50460 error="Permission denied" writer.go:29: 2021-01-29T19:32:21.251Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/snapshot from=127.0.0.1:50460 latency=1.137671ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/snapshot === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.256Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/snapshot from=127.0.0.1:50462 error="Permission denied" writer.go:29: 2021-01-29T19:32:21.258Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/snapshot from=127.0.0.1:50462 latency=1.910983ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/snapshot === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.264Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/snapshot from=127.0.0.1:50464 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.265Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/snapshot from=127.0.0.1:50464 latency=1.099037ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/snapshot === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.272Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/snapshot from=127.0.0.1:50466 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.275Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/snapshot from=127.0.0.1:50466 latency=2.613185ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/snapshot === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.282Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/snapshot from=127.0.0.1:50468 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.284Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/snapshot from=127.0.0.1:50468 latency=2.013088ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/snapshot === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.287Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/snapshot from=127.0.0.1:50468 latency=1.622µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/join/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.290Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/join/ from=127.0.0.1:50468 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:21.292Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/join/ from=127.0.0.1:50468 latency=1.120902ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/join/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.296Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/join/ from=127.0.0.1:50470 error="Permission denied" writer.go:29: 2021-01-29T19:32:21.297Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/join/ from=127.0.0.1:50470 latency=1.173684ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/join/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.303Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/join/ from=127.0.0.1:50472 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.305Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/join/ from=127.0.0.1:50472 latency=2.077001ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/join/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.311Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/join/ from=127.0.0.1:50474 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.313Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/join/ from=127.0.0.1:50474 latency=1.783311ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/join/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.319Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/join/ from=127.0.0.1:50476 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.320Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/join/ from=127.0.0.1:50476 latency=1.044031ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/join/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.326Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/join/ from=127.0.0.1:50476 latency=2.498µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/service/maintenance/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.333Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/service/maintenance/ from=127.0.0.1:50476 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:21.334Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/service/maintenance/ from=127.0.0.1:50476 latency=1.383704ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/service/maintenance/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.339Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/service/maintenance/ from=127.0.0.1:50478 latency=109.814µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/service/maintenance/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.346Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/service/maintenance/ from=127.0.0.1:50480 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.348Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/service/maintenance/ from=127.0.0.1:50480 latency=2.000623ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/service/maintenance/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.353Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/service/maintenance/ from=127.0.0.1:50482 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.355Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/service/maintenance/ from=127.0.0.1:50482 latency=1.805624ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/service/maintenance/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.363Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/service/maintenance/ from=127.0.0.1:50484 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.365Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/service/maintenance/ from=127.0.0.1:50484 latency=1.845525ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/service/maintenance/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.376Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/service/maintenance/ from=127.0.0.1:50484 latency=5.449µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/operator/raft/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.386Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/operator/raft/configuration from=127.0.0.1:50484 error="Permission denied" writer.go:29: 2021-01-29T19:32:21.388Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/operator/raft/configuration from=127.0.0.1:50484 latency=5.790196ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/operator/raft/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.402Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/operator/raft/configuration from=127.0.0.1:50486 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.405Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/operator/raft/configuration from=127.0.0.1:50486 latency=2.507935ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/operator/raft/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.423Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/operator/raft/configuration from=127.0.0.1:50488 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.425Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/operator/raft/configuration from=127.0.0.1:50488 latency=2.11758ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/operator/raft/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.433Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/operator/raft/configuration from=127.0.0.1:50490 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.435Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/operator/raft/configuration from=127.0.0.1:50490 latency=1.761193ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/operator/raft/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.440Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/operator/raft/configuration from=127.0.0.1:50492 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.442Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/operator/raft/configuration from=127.0.0.1:50492 latency=1.829336ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/operator/raft/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.446Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/operator/raft/configuration from=127.0.0.1:50492 latency=1.89µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.450Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/session/info/ from=127.0.0.1:50492 latency=68.868µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.456Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/session/info/ from=127.0.0.1:50494 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.459Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/session/info/ from=127.0.0.1:50494 latency=3.38227ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.465Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/session/info/ from=127.0.0.1:50496 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.467Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/session/info/ from=127.0.0.1:50496 latency=2.343717ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.474Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/session/info/ from=127.0.0.1:50498 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.475Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/session/info/ from=127.0.0.1:50498 latency=1.090816ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.481Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/session/info/ from=127.0.0.1:50500 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.484Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/session/info/ from=127.0.0.1:50500 latency=2.17145ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.488Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/session/info/ from=127.0.0.1:50500 latency=2.25µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/replication === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.493Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/replication from=127.0.0.1:50500 latency=256.432µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/replication === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.507Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/replication from=127.0.0.1:50502 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.509Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/replication from=127.0.0.1:50502 latency=1.946961ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/replication === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.514Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/replication from=127.0.0.1:50504 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.515Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/replication from=127.0.0.1:50504 latency=1.192211ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/replication === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.521Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/replication from=127.0.0.1:50506 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.522Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/replication from=127.0.0.1:50506 latency=1.195238ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/replication === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.528Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/replication from=127.0.0.1:50508 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.529Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/replication from=127.0.0.1:50508 latency=1.005563ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/replication === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.532Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/replication from=127.0.0.1:50508 latency=1.288µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.536Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/token/ from=127.0.0.1:50508 error="Bad request: Missing token ID" writer.go:29: 2021-01-29T19:32:21.537Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/token/ from=127.0.0.1:50508 latency=1.225395ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.542Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/token/ from=127.0.0.1:50510 error="Bad request: Token decoding failed: EOF" writer.go:29: 2021-01-29T19:32:21.544Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/token/ from=127.0.0.1:50510 latency=1.950206ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.551Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/token/ from=127.0.0.1:50512 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.552Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/token/ from=127.0.0.1:50512 latency=1.820724ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.559Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/token/ from=127.0.0.1:50514 error="Bad request: Missing token ID" writer.go:29: 2021-01-29T19:32:21.561Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/token/ from=127.0.0.1:50514 latency=1.819874ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.567Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/token/ from=127.0.0.1:50516 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.568Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/token/ from=127.0.0.1:50516 latency=1.557516ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.572Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/token/ from=127.0.0.1:50516 latency=2.015µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/metrics === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.576Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/metrics from=127.0.0.1:50516 error="Permission denied" writer.go:29: 2021-01-29T19:32:21.577Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/metrics from=127.0.0.1:50516 latency=1.846212ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/metrics === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.583Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/metrics from=127.0.0.1:50518 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.585Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/metrics from=127.0.0.1:50518 latency=1.812848ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/metrics === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.591Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/metrics from=127.0.0.1:50520 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.593Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/metrics from=127.0.0.1:50520 latency=1.419813ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/metrics === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.598Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/metrics from=127.0.0.1:50522 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.600Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/metrics from=127.0.0.1:50522 latency=1.47762ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/metrics === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.605Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/metrics from=127.0.0.1:50524 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.607Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/metrics from=127.0.0.1:50524 latency=1.60067ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/metrics === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.612Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/metrics from=127.0.0.1:50524 latency=1.816µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.615Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:50524 latency=122.899µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.621Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/services from=127.0.0.1:50526 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.624Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/services from=127.0.0.1:50526 latency=2.832623ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.630Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/services from=127.0.0.1:50528 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.631Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/services from=127.0.0.1:50528 latency=1.310025ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.637Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/services from=127.0.0.1:50530 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.638Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/services from=127.0.0.1:50530 latency=1.366487ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.643Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/services from=127.0.0.1:50532 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.645Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/services from=127.0.0.1:50532 latency=1.369323ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.648Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/services from=127.0.0.1:50532 latency=2.229µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.652Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/service/ from=127.0.0.1:50532 latency=171.97µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.667Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/service/ from=127.0.0.1:50534 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.669Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/service/ from=127.0.0.1:50534 latency=1.800661ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.674Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/service/ from=127.0.0.1:50536 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.676Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/service/ from=127.0.0.1:50536 latency=1.439159ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.681Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/service/ from=127.0.0.1:50538 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.683Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/service/ from=127.0.0.1:50538 latency=1.287278ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.690Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/service/ from=127.0.0.1:50540 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.691Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/service/ from=127.0.0.1:50540 latency=1.55788ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.702Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/service/ from=127.0.0.1:50540 latency=2.801µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/role/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.709Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/role/ from=127.0.0.1:50540 error="Bad request: Missing role ID" writer.go:29: 2021-01-29T19:32:21.710Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/role/ from=127.0.0.1:50540 latency=1.451063ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/role/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.716Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/role/ from=127.0.0.1:50542 error="Bad request: Role decoding failed: EOF" writer.go:29: 2021-01-29T19:32:21.717Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/role/ from=127.0.0.1:50542 latency=1.163841ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/role/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.722Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/role/ from=127.0.0.1:50544 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.724Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/role/ from=127.0.0.1:50544 latency=1.107521ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/role/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.732Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/role/ from=127.0.0.1:50546 error="Bad request: Missing role ID" writer.go:29: 2021-01-29T19:32:21.733Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/role/ from=127.0.0.1:50546 latency=1.386237ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/role/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.738Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/role/ from=127.0.0.1:50548 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.739Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/role/ from=127.0.0.1:50548 latency=1.770917ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/role/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.749Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/role/ from=127.0.0.1:50548 latency=2.128µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/service/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.758Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/service/deregister/ from=127.0.0.1:50548 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:21.760Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/service/deregister/ from=127.0.0.1:50548 latency=1.837466ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/service/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.766Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/service/deregister/ from=127.0.0.1:50550 error="Unknown service {"" {}}" writer.go:29: 2021-01-29T19:32:21.768Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/service/deregister/ from=127.0.0.1:50550 latency=1.990937ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/service/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.785Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/service/deregister/ from=127.0.0.1:50552 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.787Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/service/deregister/ from=127.0.0.1:50552 latency=1.988063ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/service/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.794Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/service/deregister/ from=127.0.0.1:50554 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.796Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/service/deregister/ from=127.0.0.1:50554 latency=1.772625ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/service/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.803Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/service/deregister/ from=127.0.0.1:50556 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.805Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/service/deregister/ from=127.0.0.1:50556 latency=1.880235ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/service/deregister/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.812Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/service/deregister/ from=127.0.0.1:50556 latency=3.983µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/coordinate/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.817Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/coordinate/datacenters from=127.0.0.1:50556 latency=199.939µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/coordinate/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.823Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/coordinate/datacenters from=127.0.0.1:50558 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.825Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/coordinate/datacenters from=127.0.0.1:50558 latency=1.980887ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/coordinate/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.833Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/coordinate/datacenters from=127.0.0.1:50560 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.834Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/coordinate/datacenters from=127.0.0.1:50560 latency=1.751054ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/coordinate/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.848Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/coordinate/datacenters from=127.0.0.1:50562 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.850Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/coordinate/datacenters from=127.0.0.1:50562 latency=1.517165ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/coordinate/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.861Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/coordinate/datacenters from=127.0.0.1:50564 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.863Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/coordinate/datacenters from=127.0.0.1:50564 latency=2.239547ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/coordinate/datacenters === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.868Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/coordinate/datacenters from=127.0.0.1:50564 latency=2.413µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/operator/autopilot/health === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.874Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/operator/autopilot/health from=127.0.0.1:50564 error="Permission denied" writer.go:29: 2021-01-29T19:32:21.876Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/operator/autopilot/health from=127.0.0.1:50564 latency=1.515585ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/operator/autopilot/health === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.882Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/operator/autopilot/health from=127.0.0.1:50566 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.884Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/operator/autopilot/health from=127.0.0.1:50566 latency=2.017011ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/operator/autopilot/health === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.893Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/operator/autopilot/health from=127.0.0.1:50568 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.896Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/operator/autopilot/health from=127.0.0.1:50568 latency=3.01172ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/operator/autopilot/health === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.913Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/operator/autopilot/health from=127.0.0.1:50570 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.915Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/operator/autopilot/health from=127.0.0.1:50570 latency=1.663296ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/operator/autopilot/health === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.921Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/operator/autopilot/health from=127.0.0.1:50572 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.922Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/operator/autopilot/health from=127.0.0.1:50572 latency=1.212255ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/operator/autopilot/health === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.926Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/operator/autopilot/health from=127.0.0.1:50572 latency=3.07µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/node-services/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.931Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/catalog/node-services/ from=127.0.0.1:50572 latency=132.592µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/node-services/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.938Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/catalog/node-services/ from=127.0.0.1:50574 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:21.940Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/catalog/node-services/ from=127.0.0.1:50574 latency=2.045885ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/node-services/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.947Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/catalog/node-services/ from=127.0.0.1:50576 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:21.949Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/catalog/node-services/ from=127.0.0.1:50576 latency=2.21097ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/node-services/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.966Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/catalog/node-services/ from=127.0.0.1:50578 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:21.968Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/catalog/node-services/ from=127.0.0.1:50578 latency=1.165346ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/node-services/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.980Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/catalog/node-services/ from=127.0.0.1:50580 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:21.981Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/catalog/node-services/ from=127.0.0.1:50580 latency=1.078186ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/node-services/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.984Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/catalog/node-services/ from=127.0.0.1:50580 latency=2.127µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/intentions/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.986Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/connect/intentions/ from=127.0.0.1:50580 error="Bad request: failed intention lookup: index error: UUID must be 36 characters" writer.go:29: 2021-01-29T19:32:21.988Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/connect/intentions/ from=127.0.0.1:50580 latency=1.200424ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/intentions/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:21.992Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/connect/intentions/ from=127.0.0.1:50582 error="Bad request: Request decode failed: EOF" writer.go:29: 2021-01-29T19:32:21.994Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/connect/intentions/ from=127.0.0.1:50582 latency=1.794675ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/intentions/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.000Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/connect/intentions/ from=127.0.0.1:50584 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.003Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/connect/intentions/ from=127.0.0.1:50584 latency=2.265706ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/intentions/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.020Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/connect/intentions/ from=127.0.0.1:50586 error="Intention lookup failed: failed intention lookup: index error: UUID must be 36 characters" writer.go:29: 2021-01-29T19:32:22.021Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/connect/intentions/ from=127.0.0.1:50586 latency=1.981071ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/intentions/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.041Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/connect/intentions/ from=127.0.0.1:50588 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.043Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/connect/intentions/ from=127.0.0.1:50588 latency=1.835761ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/intentions/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.050Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/connect/intentions/ from=127.0.0.1:50588 latency=2.647µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/rules/translate/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.056Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/rules/translate/ from=127.0.0.1:50588 error="Bad request: Missing token ID" writer.go:29: 2021-01-29T19:32:22.061Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/rules/translate/ from=127.0.0.1:50588 latency=4.914088ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/rules/translate/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.077Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/rules/translate/ from=127.0.0.1:50590 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:22.079Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/rules/translate/ from=127.0.0.1:50590 latency=2.001249ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/rules/translate/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.085Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/rules/translate/ from=127.0.0.1:50592 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.087Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/rules/translate/ from=127.0.0.1:50592 latency=1.943831ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/rules/translate/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.108Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/rules/translate/ from=127.0.0.1:50594 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.116Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/rules/translate/ from=127.0.0.1:50594 latency=7.499295ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/rules/translate/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.134Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/rules/translate/ from=127.0.0.1:50596 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.135Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/rules/translate/ from=127.0.0.1:50596 latency=1.209917ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/rules/translate/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.139Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/rules/translate/ from=127.0.0.1:50596 latency=2.24µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/tokens === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.142Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/tokens from=127.0.0.1:50596 error="Permission denied" writer.go:29: 2021-01-29T19:32:22.145Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/tokens from=127.0.0.1:50596 latency=2.17715ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/tokens === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.153Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/tokens from=127.0.0.1:50598 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:22.160Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/tokens from=127.0.0.1:50598 latency=7.85359ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/tokens === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.174Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/tokens from=127.0.0.1:50600 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.176Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/tokens from=127.0.0.1:50600 latency=2.095249ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/tokens === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.184Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/tokens from=127.0.0.1:50602 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.186Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/tokens from=127.0.0.1:50602 latency=1.875563ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/tokens === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.195Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/tokens from=127.0.0.1:50604 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.197Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/tokens from=127.0.0.1:50604 latency=1.788168ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/tokens === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.203Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/tokens from=127.0.0.1:50604 latency=1.929µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.210Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/update from=127.0.0.1:50604 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:22.212Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/update from=127.0.0.1:50604 latency=1.755619ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.220Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/update from=127.0.0.1:50606 latency=60.339µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.228Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/update from=127.0.0.1:50608 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.230Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/update from=127.0.0.1:50608 latency=1.906873ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.239Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/update from=127.0.0.1:50610 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.242Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/update from=127.0.0.1:50610 latency=2.092971ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.253Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/update from=127.0.0.1:50612 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.255Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/update from=127.0.0.1:50612 latency=1.994256ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/update === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.264Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/update from=127.0.0.1:50612 latency=3.402µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/binding-rules === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.269Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/binding-rules from=127.0.0.1:50612 error="Permission denied" writer.go:29: 2021-01-29T19:32:22.273Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:50612 latency=4.389253ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/binding-rules === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.295Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/binding-rules from=127.0.0.1:50614 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:22.297Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/binding-rules from=127.0.0.1:50614 latency=1.648054ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/binding-rules === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.305Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/binding-rules from=127.0.0.1:50616 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.308Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/binding-rules from=127.0.0.1:50616 latency=3.030714ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/binding-rules === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.317Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/binding-rules from=127.0.0.1:50618 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.322Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/binding-rules from=127.0.0.1:50618 latency=5.04755ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/binding-rules === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.335Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/binding-rules from=127.0.0.1:50620 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.336Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/binding-rules from=127.0.0.1:50620 latency=1.803748ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/binding-rules === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.350Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/binding-rules from=127.0.0.1:50620 latency=2.924µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/health/service/id/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.360Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/health/service/id/ from=127.0.0.1:50620 error="Bad request: Missing serviceID" writer.go:29: 2021-01-29T19:32:22.362Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/health/service/id/ from=127.0.0.1:50620 latency=2.013802ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/health/service/id/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.370Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/health/service/id/ from=127.0.0.1:50622 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:22.372Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/health/service/id/ from=127.0.0.1:50622 latency=2.150842ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/health/service/id/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.380Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/health/service/id/ from=127.0.0.1:50624 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.382Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/health/service/id/ from=127.0.0.1:50624 latency=2.020632ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/health/service/id/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.397Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/health/service/id/ from=127.0.0.1:50626 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.399Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/health/service/id/ from=127.0.0.1:50626 latency=1.950822ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/health/service/id/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.407Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/health/service/id/ from=127.0.0.1:50628 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.409Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/health/service/id/ from=127.0.0.1:50628 latency=1.849312ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/health/service/id/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.413Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/health/service/id/ from=127.0.0.1:50628 latency=1.657µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/login === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.417Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/login from=127.0.0.1:50628 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:22.419Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/login from=127.0.0.1:50628 latency=1.728073ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/login === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.436Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/login from=127.0.0.1:50630 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:22.438Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/login from=127.0.0.1:50630 latency=1.97259ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/login === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.452Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/login from=127.0.0.1:50632 error="Bad request: Failed to decode request body:: EOF" writer.go:29: 2021-01-29T19:32:22.454Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/login from=127.0.0.1:50632 latency=1.915308ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/login === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.462Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/login from=127.0.0.1:50634 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.465Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/login from=127.0.0.1:50634 latency=2.799901ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/login === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.476Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/login from=127.0.0.1:50636 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.489Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/login from=127.0.0.1:50636 latency=13.300674ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/login === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.500Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/login from=127.0.0.1:50636 latency=1.559µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.505Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/token/ from=127.0.0.1:50636 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:22.506Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/token/ from=127.0.0.1:50636 latency=1.149558ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.511Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/token/ from=127.0.0.1:50638 error="Permission denied" writer.go:29: 2021-01-29T19:32:22.513Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/token/ from=127.0.0.1:50638 latency=2.013283ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.539Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/token/ from=127.0.0.1:50640 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.541Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/token/ from=127.0.0.1:50640 latency=1.924803ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.547Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/token/ from=127.0.0.1:50642 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.549Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/token/ from=127.0.0.1:50642 latency=2.045738ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.555Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/token/ from=127.0.0.1:50644 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.556Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/token/ from=127.0.0.1:50644 latency=1.186944ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/token/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.564Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/token/ from=127.0.0.1:50644 latency=2.104µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.570Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/session/create from=127.0.0.1:50644 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:22.571Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/session/create from=127.0.0.1:50644 latency=1.122431ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.576Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/session/create from=127.0.0.1:50646 error="Permission denied" writer.go:29: 2021-01-29T19:32:22.579Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:50646 latency=2.105996ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.598Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/session/create from=127.0.0.1:50648 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.600Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/session/create from=127.0.0.1:50648 latency=2.045302ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.607Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/session/create from=127.0.0.1:50650 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.609Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/session/create from=127.0.0.1:50650 latency=2.221749ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.617Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/session/create from=127.0.0.1:50652 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.619Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/session/create from=127.0.0.1:50652 latency=1.85074ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/create === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.622Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/session/create from=127.0.0.1:50652 latency=2.416µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/health/state/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.626Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/health/state/ from=127.0.0.1:50652 latency=108.555µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/health/state/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.631Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/health/state/ from=127.0.0.1:50654 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:22.634Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/health/state/ from=127.0.0.1:50654 latency=2.514943ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/health/state/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.640Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/health/state/ from=127.0.0.1:50656 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.642Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/health/state/ from=127.0.0.1:50656 latency=2.201292ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/health/state/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.650Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/health/state/ from=127.0.0.1:50658 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.652Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/health/state/ from=127.0.0.1:50658 latency=2.383634ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/health/state/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.660Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/health/state/ from=127.0.0.1:50660 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.663Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/health/state/ from=127.0.0.1:50660 latency=2.357143ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/health/state/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.698Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/health/state/ from=127.0.0.1:50660 latency=2.924µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/token/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.704Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/token/self from=127.0.0.1:50660 error="ACL not found" writer.go:29: 2021-01-29T19:32:22.705Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:50660 latency=1.617064ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/token/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.712Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/token/self from=127.0.0.1:50662 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:22.714Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/token/self from=127.0.0.1:50662 latency=2.065636ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/token/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.722Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/token/self from=127.0.0.1:50664 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.724Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/token/self from=127.0.0.1:50664 latency=2.239709ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/token/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.740Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/token/self from=127.0.0.1:50666 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.742Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/token/self from=127.0.0.1:50666 latency=1.990423ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/token/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.748Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/token/self from=127.0.0.1:50668 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.750Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/token/self from=127.0.0.1:50668 latency=1.794143ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/token/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.756Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/token/self from=127.0.0.1:50668 latency=2.301µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/ca/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.761Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/connect/ca/configuration from=127.0.0.1:50668 error="Permission denied" writer.go:29: 2021-01-29T19:32:22.763Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/connect/ca/configuration from=127.0.0.1:50668 latency=2.257737ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/ca/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.784Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/connect/ca/configuration from=127.0.0.1:50670 error="Bad request: Request decode failed: EOF" writer.go:29: 2021-01-29T19:32:22.787Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/connect/ca/configuration from=127.0.0.1:50670 latency=2.502942ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/ca/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.795Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/connect/ca/configuration from=127.0.0.1:50672 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.798Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/connect/ca/configuration from=127.0.0.1:50672 latency=2.749688ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/ca/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.810Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/connect/ca/configuration from=127.0.0.1:50674 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.812Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/connect/ca/configuration from=127.0.0.1:50674 latency=2.284427ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/ca/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.821Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/connect/ca/configuration from=127.0.0.1:50676 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.823Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/connect/ca/configuration from=127.0.0.1:50676 latency=1.938461ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/ca/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.833Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/connect/ca/configuration from=127.0.0.1:50676 latency=1.708µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/intentions/check === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.839Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/connect/intentions/check from=127.0.0.1:50676 error="required query parameter 'source' not set" writer.go:29: 2021-01-29T19:32:22.841Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/connect/intentions/check from=127.0.0.1:50676 latency=2.272617ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/intentions/check === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.850Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/connect/intentions/check from=127.0.0.1:50678 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:22.852Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/connect/intentions/check from=127.0.0.1:50678 latency=2.143961ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/intentions/check === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.860Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/connect/intentions/check from=127.0.0.1:50680 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.863Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/connect/intentions/check from=127.0.0.1:50680 latency=2.382713ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/intentions/check === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.870Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/connect/intentions/check from=127.0.0.1:50682 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.872Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/connect/intentions/check from=127.0.0.1:50682 latency=1.772183ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/intentions/check === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.880Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/connect/intentions/check from=127.0.0.1:50684 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.882Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/connect/intentions/check from=127.0.0.1:50684 latency=2.412321ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/intentions/check === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.891Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/connect/intentions/check from=127.0.0.1:50684 latency=2.378µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.895Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/destroy/ from=127.0.0.1:50684 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:22.898Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/destroy/ from=127.0.0.1:50684 latency=2.522982ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.908Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/destroy/ from=127.0.0.1:50686 latency=171.925µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.915Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/destroy/ from=127.0.0.1:50688 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.916Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/destroy/ from=127.0.0.1:50688 latency=1.150218ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.922Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/destroy/ from=127.0.0.1:50690 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.923Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/destroy/ from=127.0.0.1:50690 latency=1.140443ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.931Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/destroy/ from=127.0.0.1:50692 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.934Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/destroy/ from=127.0.0.1:50692 latency=3.300655ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/destroy/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.939Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/destroy/ from=127.0.0.1:50692 latency=2.152µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.944Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/info/ from=127.0.0.1:50692 latency=66.664µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.950Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/info/ from=127.0.0.1:50694 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:22.952Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/info/ from=127.0.0.1:50694 latency=2.264813ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.964Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/info/ from=127.0.0.1:50696 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:22.966Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/info/ from=127.0.0.1:50696 latency=2.336294ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.973Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/info/ from=127.0.0.1:50698 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:22.976Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/info/ from=127.0.0.1:50698 latency=2.67625ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.983Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/info/ from=127.0.0.1:50700 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:22.985Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/info/ from=127.0.0.1:50700 latency=2.152446ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/info/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:22.991Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/info/ from=127.0.0.1:50700 latency=2.653µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.002Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/list from=127.0.0.1:50700 error="Permission denied" writer.go:29: 2021-01-29T19:32:23.004Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/list from=127.0.0.1:50700 latency=2.16882ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.013Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/list from=127.0.0.1:50702 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.015Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/list from=127.0.0.1:50702 latency=2.179658ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.032Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/list from=127.0.0.1:50704 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.037Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/list from=127.0.0.1:50704 latency=5.64504ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.045Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/list from=127.0.0.1:50706 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.047Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/list from=127.0.0.1:50706 latency=1.875585ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.054Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/list from=127.0.0.1:50708 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.056Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/list from=127.0.0.1:50708 latency=1.870642ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.061Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/list from=127.0.0.1:50708 latency=2.775µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/coordinate/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.067Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/coordinate/node/ from=127.0.0.1:50708 error="Permission denied" writer.go:29: 2021-01-29T19:32:23.069Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/coordinate/node/ from=127.0.0.1:50708 latency=2.140283ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/coordinate/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.075Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/coordinate/node/ from=127.0.0.1:50710 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.076Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/coordinate/node/ from=127.0.0.1:50710 latency=1.192415ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/coordinate/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.081Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/coordinate/node/ from=127.0.0.1:50712 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.083Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/coordinate/node/ from=127.0.0.1:50712 latency=1.359483ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/coordinate/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.088Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/coordinate/node/ from=127.0.0.1:50714 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.089Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/coordinate/node/ from=127.0.0.1:50714 latency=1.25815ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/coordinate/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.094Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/coordinate/node/ from=127.0.0.1:50716 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.095Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/coordinate/node/ from=127.0.0.1:50716 latency=1.33685ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/coordinate/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.099Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/coordinate/node/ from=127.0.0.1:50716 latency=2.677µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/health/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.104Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/health/connect/ from=127.0.0.1:50716 latency=70.469µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/health/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.120Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/health/connect/ from=127.0.0.1:50718 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.123Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/health/connect/ from=127.0.0.1:50718 latency=2.798794ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/health/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.128Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/health/connect/ from=127.0.0.1:50720 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.130Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/health/connect/ from=127.0.0.1:50720 latency=1.323626ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/health/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.134Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/health/connect/ from=127.0.0.1:50722 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.135Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/health/connect/ from=127.0.0.1:50722 latency=1.10851ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/health/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.140Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/health/connect/ from=127.0.0.1:50724 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.141Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/health/connect/ from=127.0.0.1:50724 latency=1.041672ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/health/connect/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.143Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/health/connect/ from=127.0.0.1:50724 latency=1.123µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/operator/autopilot/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.146Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/operator/autopilot/configuration from=127.0.0.1:50724 error="Permission denied" writer.go:29: 2021-01-29T19:32:23.147Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/operator/autopilot/configuration from=127.0.0.1:50724 latency=1.295503ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/operator/autopilot/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.154Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/operator/autopilot/configuration from=127.0.0.1:50726 error="Bad request: Error parsing autopilot config: EOF" writer.go:29: 2021-01-29T19:32:23.155Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/operator/autopilot/configuration from=127.0.0.1:50726 latency=1.132799ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/operator/autopilot/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.161Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/operator/autopilot/configuration from=127.0.0.1:50728 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.162Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/operator/autopilot/configuration from=127.0.0.1:50728 latency=1.597746ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/operator/autopilot/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.168Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/operator/autopilot/configuration from=127.0.0.1:50730 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.171Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/operator/autopilot/configuration from=127.0.0.1:50730 latency=2.926481ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/operator/autopilot/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.184Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/operator/autopilot/configuration from=127.0.0.1:50732 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.185Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/operator/autopilot/configuration from=127.0.0.1:50732 latency=1.069324ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/operator/autopilot/configuration === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.187Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/operator/autopilot/configuration from=127.0.0.1:50732 latency=3.126µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/force-leave/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.190Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/force-leave/ from=127.0.0.1:50732 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:23.191Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/force-leave/ from=127.0.0.1:50732 latency=1.101855ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/force-leave/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.198Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/force-leave/ from=127.0.0.1:50734 error="Permission denied" writer.go:29: 2021-01-29T19:32:23.200Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/force-leave/ from=127.0.0.1:50734 latency=1.941956ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/force-leave/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.206Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/force-leave/ from=127.0.0.1:50736 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.207Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/force-leave/ from=127.0.0.1:50736 latency=1.283336ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/force-leave/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.212Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/force-leave/ from=127.0.0.1:50738 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.213Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/force-leave/ from=127.0.0.1:50738 latency=1.287843ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/force-leave/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.218Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/force-leave/ from=127.0.0.1:50740 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.219Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/force-leave/ from=127.0.0.1:50740 latency=1.2339ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/force-leave/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.224Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/force-leave/ from=127.0.0.1:50740 latency=1.675µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/pass/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.226Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/check/pass/ from=127.0.0.1:50740 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:23.229Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/check/pass/ from=127.0.0.1:50740 latency=2.555616ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/pass/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.236Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/check/pass/ from=127.0.0.1:50742 error="Unknown check """ writer.go:29: 2021-01-29T19:32:23.238Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/check/pass/ from=127.0.0.1:50742 latency=2.631703ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/pass/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.250Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/check/pass/ from=127.0.0.1:50744 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.252Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/check/pass/ from=127.0.0.1:50744 latency=2.134247ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/pass/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.262Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/check/pass/ from=127.0.0.1:50746 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.266Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/check/pass/ from=127.0.0.1:50746 latency=3.869743ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/pass/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.282Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/check/pass/ from=127.0.0.1:50748 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.283Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/check/pass/ from=127.0.0.1:50748 latency=1.081908ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/pass/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.288Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/check/pass/ from=127.0.0.1:50748 latency=2.148µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/fail/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.292Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/check/fail/ from=127.0.0.1:50748 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:23.293Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/check/fail/ from=127.0.0.1:50748 latency=1.182038ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/fail/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.299Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/check/fail/ from=127.0.0.1:50750 error="Unknown check """ writer.go:29: 2021-01-29T19:32:23.301Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/check/fail/ from=127.0.0.1:50750 latency=2.000385ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/fail/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.308Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/check/fail/ from=127.0.0.1:50752 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.310Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/check/fail/ from=127.0.0.1:50752 latency=1.886732ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/fail/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.316Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/check/fail/ from=127.0.0.1:50754 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.318Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/check/fail/ from=127.0.0.1:50754 latency=2.236848ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/fail/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.326Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/check/fail/ from=127.0.0.1:50756 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.328Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/check/fail/ from=127.0.0.1:50756 latency=1.281756ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/fail/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.331Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/check/fail/ from=127.0.0.1:50756 latency=1.065µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.334Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/self from=127.0.0.1:50756 error="Permission denied" writer.go:29: 2021-01-29T19:32:23.337Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:50756 latency=2.560266ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.345Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/self from=127.0.0.1:50758 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.346Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/self from=127.0.0.1:50758 latency=1.131159ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.352Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/self from=127.0.0.1:50760 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.353Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/self from=127.0.0.1:50760 latency=1.352304ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.362Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/self from=127.0.0.1:50762 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.363Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/self from=127.0.0.1:50762 latency=1.169902ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.368Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/self from=127.0.0.1:50764 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.369Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/self from=127.0.0.1:50764 latency=1.081103ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/self === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.372Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/self from=127.0.0.1:50764 latency=2.058µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/leave === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.377Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/leave from=127.0.0.1:50764 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:23.379Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/leave from=127.0.0.1:50764 latency=1.839623ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/leave === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.386Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/leave from=127.0.0.1:50766 error="Permission denied" writer.go:29: 2021-01-29T19:32:23.387Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/leave from=127.0.0.1:50766 latency=1.293706ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/leave === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.393Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/leave from=127.0.0.1:50768 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.395Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/leave from=127.0.0.1:50768 latency=2.503635ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/leave === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.403Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/leave from=127.0.0.1:50770 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.405Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/leave from=127.0.0.1:50770 latency=1.226957ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/leave === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.416Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/leave from=127.0.0.1:50772 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.417Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/leave from=127.0.0.1:50772 latency=1.307455ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/leave === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.422Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/leave from=127.0.0.1:50772 latency=2.383µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/internal/ui/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.426Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/internal/ui/node/ from=127.0.0.1:50772 latency=103.491µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/internal/ui/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.433Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/internal/ui/node/ from=127.0.0.1:50774 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.434Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/internal/ui/node/ from=127.0.0.1:50774 latency=1.335096ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/internal/ui/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.441Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/internal/ui/node/ from=127.0.0.1:50776 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.442Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/internal/ui/node/ from=127.0.0.1:50776 latency=1.236691ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/internal/ui/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.447Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/internal/ui/node/ from=127.0.0.1:50778 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.448Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/internal/ui/node/ from=127.0.0.1:50778 latency=1.195531ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/internal/ui/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.453Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/internal/ui/node/ from=127.0.0.1:50780 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.456Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/internal/ui/node/ from=127.0.0.1:50780 latency=2.383762ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/internal/ui/node/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.469Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/internal/ui/node/ from=127.0.0.1:50780 latency=3.954µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.473Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/session/list from=127.0.0.1:50780 latency=290.869µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.478Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/session/list from=127.0.0.1:50782 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.479Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/session/list from=127.0.0.1:50782 latency=1.551594ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.487Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/session/list from=127.0.0.1:50784 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.488Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/session/list from=127.0.0.1:50784 latency=1.236175ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.494Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/session/list from=127.0.0.1:50786 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.495Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/session/list from=127.0.0.1:50786 latency=1.224792ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.500Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/session/list from=127.0.0.1:50788 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.501Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/session/list from=127.0.0.1:50788 latency=1.092468ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/list === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.504Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/session/list from=127.0.0.1:50788 latency=2.49µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.509Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/connect/ca/roots from=127.0.0.1:50788 latency=149.203µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.514Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/connect/ca/roots from=127.0.0.1:50790 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.516Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/connect/ca/roots from=127.0.0.1:50790 latency=1.272981ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.521Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/connect/ca/roots from=127.0.0.1:50792 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.522Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/connect/ca/roots from=127.0.0.1:50792 latency=1.304982ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.528Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/connect/ca/roots from=127.0.0.1:50794 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.529Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/connect/ca/roots from=127.0.0.1:50794 latency=1.249353ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.548Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/connect/ca/roots from=127.0.0.1:50796 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.549Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/connect/ca/roots from=127.0.0.1:50796 latency=1.316503ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/connect/ca/roots === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.553Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/connect/ca/roots from=127.0.0.1:50796 latency=1.338µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.556Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.acl: dropping service from result due to ACLs: service=consul writer.go:29: 2021-01-29T19:32:23.559Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/catalog/services from=127.0.0.1:50796 latency=3.240404ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.565Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/catalog/services from=127.0.0.1:50798 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.566Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/catalog/services from=127.0.0.1:50798 latency=1.238666ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.571Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/catalog/services from=127.0.0.1:50800 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.573Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/catalog/services from=127.0.0.1:50800 latency=1.582693ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.585Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/catalog/services from=127.0.0.1:50802 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.590Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/catalog/services from=127.0.0.1:50802 latency=5.474834ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.601Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/catalog/services from=127.0.0.1:50804 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.603Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/catalog/services from=127.0.0.1:50804 latency=1.993708ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/services === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.607Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/catalog/services from=127.0.0.1:50804 latency=1.858µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/internal/ui/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.614Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.acl: dropping node from result due to ACLs: node=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a writer.go:29: 2021-01-29T19:32:23.621Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/internal/ui/nodes from=127.0.0.1:50804 latency=8.890092ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/internal/ui/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.627Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/internal/ui/nodes from=127.0.0.1:50806 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.629Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/internal/ui/nodes from=127.0.0.1:50806 latency=1.909183ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/internal/ui/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.637Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/internal/ui/nodes from=127.0.0.1:50808 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.639Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/internal/ui/nodes from=127.0.0.1:50808 latency=2.170636ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/internal/ui/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.658Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/internal/ui/nodes from=127.0.0.1:50810 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.659Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/internal/ui/nodes from=127.0.0.1:50810 latency=1.259176ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/internal/ui/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.664Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/internal/ui/nodes from=127.0.0.1:50812 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.665Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/internal/ui/nodes from=127.0.0.1:50812 latency=1.16181ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/internal/ui/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.673Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/internal/ui/nodes from=127.0.0.1:50812 latency=1.598µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/health/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.675Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/health/service/ from=127.0.0.1:50812 latency=88.753µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/health/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.683Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/health/service/ from=127.0.0.1:50814 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.685Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/health/service/ from=127.0.0.1:50814 latency=2.722182ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/health/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.693Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/health/service/ from=127.0.0.1:50816 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.695Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/health/service/ from=127.0.0.1:50816 latency=1.636124ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/health/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.701Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/health/service/ from=127.0.0.1:50818 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.702Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/health/service/ from=127.0.0.1:50818 latency=1.146864ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/health/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.706Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/health/service/ from=127.0.0.1:50820 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.708Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/health/service/ from=127.0.0.1:50820 latency=1.070059ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/health/service/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.710Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/health/service/ from=127.0.0.1:50820 latency=1.354µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.714Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.acl: dropping node from result due to ACLs: node=Node-e53e7bb0-3022-af35-e38a-a8369bb74b4a writer.go:29: 2021-01-29T19:32:23.717Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/catalog/nodes from=127.0.0.1:50820 latency=3.149029ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.723Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/catalog/nodes from=127.0.0.1:50822 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.724Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/catalog/nodes from=127.0.0.1:50822 latency=1.216142ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.729Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/catalog/nodes from=127.0.0.1:50824 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.730Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/catalog/nodes from=127.0.0.1:50824 latency=1.229929ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.737Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/catalog/nodes from=127.0.0.1:50826 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.738Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/catalog/nodes from=127.0.0.1:50826 latency=1.159506ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.743Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/catalog/nodes from=127.0.0.1:50828 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.745Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/catalog/nodes from=127.0.0.1:50828 latency=1.792178ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/nodes === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.751Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/catalog/nodes from=127.0.0.1:50828 latency=4.023µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.757Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/catalog/register from=127.0.0.1:50828 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:23.759Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/catalog/register from=127.0.0.1:50828 latency=2.126788ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.768Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/catalog/register from=127.0.0.1:50830 latency=108.294µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.773Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/catalog/register from=127.0.0.1:50832 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.776Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/catalog/register from=127.0.0.1:50832 latency=2.386121ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.783Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/catalog/register from=127.0.0.1:50834 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.786Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/catalog/register from=127.0.0.1:50834 latency=2.460126ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.793Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/catalog/register from=127.0.0.1:50836 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.795Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/catalog/register from=127.0.0.1:50836 latency=1.829446ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/register === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.799Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/catalog/register from=127.0.0.1:50836 latency=2.089µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/policy/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.805Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/policy/name/ from=127.0.0.1:50836 error="Bad request: Missing policy Name" writer.go:29: 2021-01-29T19:32:23.806Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/policy/name/ from=127.0.0.1:50836 latency=1.163037ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/policy/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.812Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/policy/name/ from=127.0.0.1:50838 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.814Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/policy/name/ from=127.0.0.1:50838 latency=2.250109ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/policy/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.822Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/policy/name/ from=127.0.0.1:50840 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.825Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/policy/name/ from=127.0.0.1:50840 latency=2.070945ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/policy/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.830Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/policy/name/ from=127.0.0.1:50842 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.833Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/policy/name/ from=127.0.0.1:50842 latency=2.236094ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/policy/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.839Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/policy/name/ from=127.0.0.1:50844 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.842Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/policy/name/ from=127.0.0.1:50844 latency=2.605352ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/policy/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.846Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/policy/name/ from=127.0.0.1:50844 latency=5.273µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/binding-rule/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.852Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/binding-rule/ from=127.0.0.1:50844 error="Bad request: Missing binding rule ID" writer.go:29: 2021-01-29T19:32:23.854Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/binding-rule/ from=127.0.0.1:50844 latency=1.748374ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/binding-rule/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.859Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/binding-rule/ from=127.0.0.1:50846 error="Bad request: BindingRule decoding failed: EOF" writer.go:29: 2021-01-29T19:32:23.861Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/binding-rule/ from=127.0.0.1:50846 latency=1.869875ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/binding-rule/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.879Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/binding-rule/ from=127.0.0.1:50848 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.881Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/binding-rule/ from=127.0.0.1:50848 latency=2.088062ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/binding-rule/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.888Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/binding-rule/ from=127.0.0.1:50850 error="Bad request: Missing binding rule ID" writer.go:29: 2021-01-29T19:32:23.890Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/binding-rule/ from=127.0.0.1:50850 latency=1.985891ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/binding-rule/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.897Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/binding-rule/ from=127.0.0.1:50852 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.898Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/binding-rule/ from=127.0.0.1:50852 latency=1.076186ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/binding-rule/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.901Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/binding-rule/ from=127.0.0.1:50852 latency=1.827µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/health/service/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.905Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/agent/health/service/name/ from=127.0.0.1:50852 error="Bad request: Missing service Name" writer.go:29: 2021-01-29T19:32:23.906Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/agent/health/service/name/ from=127.0.0.1:50852 latency=1.243408ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/health/service/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.916Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/agent/health/service/name/ from=127.0.0.1:50854 error="method PUT not allowed" writer.go:29: 2021-01-29T19:32:23.922Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/agent/health/service/name/ from=127.0.0.1:50854 latency=6.806346ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/health/service/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.936Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/agent/health/service/name/ from=127.0.0.1:50856 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:23.939Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/agent/health/service/name/ from=127.0.0.1:50856 latency=3.4049ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/health/service/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.949Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/agent/health/service/name/ from=127.0.0.1:50858 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:23.951Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/agent/health/service/name/ from=127.0.0.1:50858 latency=1.969197ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/health/service/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.975Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/agent/health/service/name/ from=127.0.0.1:50860 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:23.978Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/agent/health/service/name/ from=127.0.0.1:50860 latency=2.684374ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/health/service/name/ === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.984Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/agent/health/service/name/ from=127.0.0.1:50860 latency=9.487µs === RUN TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/token === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:23.990Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=GET url=/v1/acl/token from=127.0.0.1:50860 error="method GET not allowed" writer.go:29: 2021-01-29T19:32:23.993Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=GET url=/v1/acl/token from=127.0.0.1:50860 latency=2.494424ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/token === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:24.006Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=PUT url=/v1/acl/token from=127.0.0.1:50862 error="Bad request: Token decoding failed: EOF" writer.go:29: 2021-01-29T19:32:24.011Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:50862 latency=4.871435ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/token === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:24.023Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=POST url=/v1/acl/token from=127.0.0.1:50864 error="method POST not allowed" writer.go:29: 2021-01-29T19:32:24.025Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=POST url=/v1/acl/token from=127.0.0.1:50864 latency=2.232083ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/token === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:24.033Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=DELETE url=/v1/acl/token from=127.0.0.1:50866 error="method DELETE not allowed" writer.go:29: 2021-01-29T19:32:24.035Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=DELETE url=/v1/acl/token from=127.0.0.1:50866 latency=2.129384ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/token === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:24.044Z [ERROR] TestHTTPAPI_MethodNotAllowed_OSS.http: Request error: method=HEAD url=/v1/acl/token from=127.0.0.1:50868 error="method HEAD not allowed" writer.go:29: 2021-01-29T19:32:24.046Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=HEAD url=/v1/acl/token from=127.0.0.1:50868 latency=2.083874ms === RUN TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/token === CONT TestHTTPAPI_MethodNotAllowed_OSS writer.go:29: 2021-01-29T19:32:24.052Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.http: Request finished: method=OPTIONS url=/v1/acl/token from=127.0.0.1:50868 latency=3.197µs writer.go:29: 2021-01-29T19:32:24.055Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: Requesting shutdown writer.go:29: 2021-01-29T19:32:24.060Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server: shutting down server writer.go:29: 2021-01-29T19:32:24.061Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:24.063Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:24.065Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:24.066Z [WARN] TestHTTPAPI_MethodNotAllowed_OSS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:24.067Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:24.067Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:24.067Z [DEBUG] TestHTTPAPI_MethodNotAllowed_OSS.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:24.071Z [WARN] TestHTTPAPI_MethodNotAllowed_OSS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:24.082Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:24.082Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: consul server down writer.go:29: 2021-01-29T19:32:24.087Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: shutdown complete writer.go:29: 2021-01-29T19:32:24.089Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: Stopping server: protocol=DNS address=127.0.0.1:30096 network=tcp writer.go:29: 2021-01-29T19:32:24.102Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: Stopping server: protocol=DNS address=127.0.0.1:30096 network=udp writer.go:29: 2021-01-29T19:32:24.103Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: Stopping server: protocol=HTTP address=127.0.0.1:30097 network=tcp writer.go:29: 2021-01-29T19:32:24.605Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:24.609Z [INFO] TestHTTPAPI_MethodNotAllowed_OSS: Endpoints down --- PASS: TestHTTPAPI_MethodNotAllowed_OSS (6.35s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/query (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/query (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/query (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/query (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/query (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/query (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/query/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/query/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/query/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/query/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/query/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/query/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/query/xxx/execute (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/query/xxx/execute (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/query/xxx/execute (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/query/xxx/execute (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/query/xxx/execute (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/query/xxx/execute (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/query/xxx/explain (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/query/xxx/explain (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/query/xxx/explain (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/query/xxx/explain (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/query/xxx/explain (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/query/xxx/explain (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/update/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/update/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/update/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/update/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/update/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/update/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/datacenters (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/datacenters (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/datacenters (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/datacenters (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/datacenters (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/datacenters (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/logout (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/logout (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/logout (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/logout (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/logout (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/logout (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/role (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/role (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/role (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/role (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/role (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/role (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/maintenance (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/maintenance (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/maintenance (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/maintenance (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/maintenance (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/maintenance (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/node/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/node/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/policy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/policy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/policy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/policy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/policy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/policy/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/role/name/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/role/name/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/role/name/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/role/name/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/role/name/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/role/name/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/event/fire/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/event/fire/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/event/fire/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/event/fire/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/event/fire/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/event/fire/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/internal/acl/authorize (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/internal/acl/authorize (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/internal/acl/authorize (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/internal/acl/authorize (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/internal/acl/authorize (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/internal/acl/authorize (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/auth-method/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/auth-method/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/auth-method/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/auth-method/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/auth-method/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/auth-method/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/auth-method (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/auth-method (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/auth-method (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/auth-method (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/auth-method (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/auth-method (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/rules/translate (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/rules/translate (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/rules/translate (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/rules/translate (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/rules/translate (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/rules/translate (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/register (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/register (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/bootstrap (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/bootstrap (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/bootstrap (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/bootstrap (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/bootstrap (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/bootstrap (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/roles (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/roles (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/roles (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/roles (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/roles (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/roles (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/event/list (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/event/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/event/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/event/list (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/event/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/event/list (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/internal/ui/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/internal/ui/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/internal/ui/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/internal/ui/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/internal/ui/services (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/internal/ui/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/policy (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/policy (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/policy (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/policy (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/policy (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/policy (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/service/register (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/service/register (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/service/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/service/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/service/register (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/service/register (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/config (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/config (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/config (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/config (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/config (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/config (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/connect/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/connect/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/connect/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/connect/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/connect/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/connect/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/binding-rule (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/binding-rule (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/binding-rule (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/binding-rule (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/binding-rule (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/binding-rule (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/members (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/members (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/members (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/members (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/members (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/members (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/operator/keyring (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/operator/keyring (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/operator/keyring (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/operator/keyring (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/operator/keyring (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/operator/keyring (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/txn (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/txn (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/txn (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/txn (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/txn (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/txn (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/checks (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/checks (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/checks (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/checks (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/checks (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/checks (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/kv/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/kv/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/kv/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/kv/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/kv/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/kv/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/host (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/host (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/host (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/host (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/host (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/host (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/warn/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/warn/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/warn/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/warn/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/warn/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/warn/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/service/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/service/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/service/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/coordinate/update (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/coordinate/update (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/coordinate/update (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/coordinate/update (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/coordinate/update (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/coordinate/update (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/auth-methods (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/auth-methods (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/auth-methods (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/auth-methods (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/auth-methods (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/auth-methods (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/connect/ca/leaf/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/connect/ca/leaf/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/connect/ca/leaf/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/connect/ca/leaf/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/connect/ca/leaf/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/connect/ca/leaf/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/coordinate/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/coordinate/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/coordinate/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/coordinate/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/coordinate/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/coordinate/nodes (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/clone/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/clone/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/clone/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/clone/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/clone/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/clone/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/node/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/node/ (0.03s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/node/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/ca/roots (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/ca/roots (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/ca/roots (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/ca/roots (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/ca/roots (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/ca/roots (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/health/checks/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/health/checks/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/health/checks/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/health/checks/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/health/checks/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/health/checks/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/discovery-chain/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/discovery-chain/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/discovery-chain/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/discovery-chain/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/discovery-chain/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/discovery-chain/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/health/node/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/health/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/health/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/health/node/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/health/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/health/node/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/status/leader (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/status/leader (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/status/leader (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/status/leader (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/status/leader (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/status/leader (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/intentions/match (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/intentions/match (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/intentions/match (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/intentions/match (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/intentions/match (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/intentions/match (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/renew/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/renew/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/renew/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/renew/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/renew/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/renew/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/create (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/create (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/create (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/create (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/create (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/create (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/config/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/config/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/config/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/config/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/config/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/config/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/intentions (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/intentions (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/intentions (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/intentions (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/intentions (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/intentions (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/operator/raft/peer (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/operator/raft/peer (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/operator/raft/peer (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/operator/raft/peer (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/operator/raft/peer (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/operator/raft/peer (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/deregister/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/deregister/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/deregister/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/deregister/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/deregister/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/deregister/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/connect/authorize (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/connect/authorize (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/connect/authorize (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/connect/authorize (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/connect/authorize (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/connect/authorize (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/destroy/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/destroy/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/destroy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/destroy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/destroy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/destroy/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/policies (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/policies (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/policies (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/policies (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/policies (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/policies (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/deregister (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/deregister (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/deregister (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/deregister (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/deregister (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/deregister (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/snapshot (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/snapshot (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/snapshot (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/snapshot (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/snapshot (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/snapshot (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/join/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/join/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/join/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/join/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/join/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/join/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/service/maintenance/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/service/maintenance/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/service/maintenance/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/service/maintenance/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/service/maintenance/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/service/maintenance/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/operator/raft/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/operator/raft/configuration (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/operator/raft/configuration (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/operator/raft/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/operator/raft/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/operator/raft/configuration (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/info/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/info/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/info/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/info/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/info/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/info/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/replication (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/replication (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/replication (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/replication (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/replication (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/replication (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/token/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/token/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/token/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/token/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/token/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/token/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/metrics (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/metrics (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/metrics (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/metrics (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/metrics (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/metrics (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/services (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/services (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/role/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/role/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/role/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/role/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/role/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/role/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/service/deregister/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/service/deregister/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/service/deregister/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/service/deregister/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/service/deregister/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/service/deregister/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/coordinate/datacenters (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/coordinate/datacenters (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/coordinate/datacenters (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/coordinate/datacenters (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/coordinate/datacenters (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/coordinate/datacenters (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/operator/autopilot/health (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/operator/autopilot/health (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/operator/autopilot/health (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/operator/autopilot/health (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/operator/autopilot/health (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/operator/autopilot/health (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/node-services/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/node-services/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/node-services/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/node-services/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/node-services/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/node-services/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/intentions/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/intentions/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/intentions/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/intentions/ (0.03s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/intentions/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/intentions/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/rules/translate/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/rules/translate/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/rules/translate/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/rules/translate/ (0.03s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/rules/translate/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/rules/translate/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/tokens (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/tokens (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/tokens (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/tokens (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/tokens (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/tokens (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/update (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/update (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/update (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/update (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/update (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/update (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/binding-rules (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/binding-rules (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/binding-rules (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/binding-rules (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/binding-rules (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/binding-rules (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/health/service/id/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/health/service/id/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/health/service/id/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/health/service/id/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/health/service/id/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/health/service/id/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/login (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/login (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/login (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/login (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/login (0.03s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/login (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/token/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/token/ (0.03s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/token/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/token/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/token/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/token/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/create (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/create (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/create (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/create (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/create (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/create (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/health/state/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/health/state/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/health/state/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/health/state/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/health/state/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/health/state/ (0.03s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/token/self (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/token/self (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/token/self (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/token/self (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/token/self (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/token/self (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/ca/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/ca/configuration (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/ca/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/ca/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/ca/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/ca/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/connect/intentions/check (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/connect/intentions/check (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/connect/intentions/check (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/connect/intentions/check (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/connect/intentions/check (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/connect/intentions/check (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/destroy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/destroy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/destroy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/destroy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/destroy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/destroy/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/info/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/info/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/info/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/info/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/info/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/info/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/list (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/list (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/coordinate/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/coordinate/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/coordinate/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/coordinate/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/coordinate/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/coordinate/node/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/health/connect/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/health/connect/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/health/connect/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/health/connect/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/health/connect/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/health/connect/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/operator/autopilot/configuration (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/operator/autopilot/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/operator/autopilot/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/operator/autopilot/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/operator/autopilot/configuration (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/operator/autopilot/configuration (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/force-leave/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/force-leave/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/force-leave/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/force-leave/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/force-leave/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/force-leave/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/pass/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/pass/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/pass/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/pass/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/pass/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/pass/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/check/fail/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/check/fail/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/check/fail/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/check/fail/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/check/fail/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/check/fail/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/self (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/self (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/self (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/self (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/self (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/self (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/leave (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/leave (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/leave (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/leave (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/leave (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/leave (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/internal/ui/node/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/internal/ui/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/internal/ui/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/internal/ui/node/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/internal/ui/node/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/internal/ui/node/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/session/list (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/session/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/session/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/session/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/session/list (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/session/list (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/connect/ca/roots (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/connect/ca/roots (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/connect/ca/roots (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/connect/ca/roots (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/connect/ca/roots (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/connect/ca/roots (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/services (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/services (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/services (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/internal/ui/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/internal/ui/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/internal/ui/nodes (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/internal/ui/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/internal/ui/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/internal/ui/nodes (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/health/service/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/health/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/health/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/health/service/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/health/service/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/health/service/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/nodes (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/catalog/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/catalog/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/catalog/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/catalog/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/catalog/register (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/catalog/register (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/policy/name/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/policy/name/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/policy/name/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/policy/name/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/policy/name/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/policy/name/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/binding-rule/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/binding-rule/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/binding-rule/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/binding-rule/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/binding-rule/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/binding-rule/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/agent/health/service/name/ (0.00s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/agent/health/service/name/ (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/agent/health/service/name/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/agent/health/service/name/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/agent/health/service/name/ (0.03s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/agent/health/service/name/ (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/GET_/v1/acl/token (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/PUT_/v1/acl/token (0.02s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/POST_/v1/acl/token (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/DELETE_/v1/acl/token (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/HEAD_/v1/acl/token (0.01s) --- PASS: TestHTTPAPI_MethodNotAllowed_OSS/OPTIONS_/v1/acl/token (0.00s) === RUN TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:24.731Z [WARN] TestHTTPAPI_OptionMethod_OSS: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:32:24.732Z [WARN] TestHTTPAPI_OptionMethod_OSS: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:24.735Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:24.740Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:24.754Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8d441da2-2c43-88a4-63be-4420bff2dfa1 Address:127.0.0.1:30107}]" writer.go:29: 2021-01-29T19:32:24.774Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.raft: entering follower state: follower="Node at 127.0.0.1:30107 [Follower]" leader= writer.go:29: 2021-01-29T19:32:24.776Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.serf.wan: serf: EventMemberJoin: Node-8d441da2-2c43-88a4-63be-4420bff2dfa1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:24.792Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.serf.lan: serf: EventMemberJoin: Node-8d441da2-2c43-88a4-63be-4420bff2dfa1 127.0.0.1 writer.go:29: 2021-01-29T19:32:24.796Z [INFO] TestHTTPAPI_OptionMethod_OSS: Started DNS server: address=127.0.0.1:30102 network=udp writer.go:29: 2021-01-29T19:32:24.797Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: Handled event for server in area: event=member-join server=Node-8d441da2-2c43-88a4-63be-4420bff2dfa1.dc1 area=wan writer.go:29: 2021-01-29T19:32:24.805Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: Adding LAN server: server="Node-8d441da2-2c43-88a4-63be-4420bff2dfa1 (Addr: tcp/127.0.0.1:30107) (DC: dc1)" writer.go:29: 2021-01-29T19:32:24.807Z [INFO] TestHTTPAPI_OptionMethod_OSS: Started DNS server: address=127.0.0.1:30102 network=tcp writer.go:29: 2021-01-29T19:32:24.820Z [INFO] TestHTTPAPI_OptionMethod_OSS: Started HTTP server: address=127.0.0.1:30103 network=tcp writer.go:29: 2021-01-29T19:32:24.825Z [INFO] TestHTTPAPI_OptionMethod_OSS: started state syncer writer.go:29: 2021-01-29T19:32:24.824Z [WARN] TestHTTPAPI_OptionMethod_OSS.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:24.834Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.raft: entering candidate state: node="Node at 127.0.0.1:30107 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:24.840Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:24.843Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.server.raft: vote granted: from=8d441da2-2c43-88a4-63be-4420bff2dfa1 term=2 tally=1 writer.go:29: 2021-01-29T19:32:24.846Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:24.849Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.raft: entering leader state: leader="Node at 127.0.0.1:30107 [Leader]" writer.go:29: 2021-01-29T19:32:24.851Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.server: Cannot upgrade to new ACLs: leaderMode=2 mode=2 found=true leader=127.0.0.1:30107 writer.go:29: 2021-01-29T19:32:24.855Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:24.857Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: New leader elected: payload=Node-8d441da2-2c43-88a4-63be-4420bff2dfa1 writer.go:29: 2021-01-29T19:32:24.859Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: initializing acls writer.go:29: 2021-01-29T19:32:24.862Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:24.867Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:24.868Z [INFO] TestHTTPAPI_OptionMethod_OSS.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:24.869Z [INFO] TestHTTPAPI_OptionMethod_OSS.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:24.871Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.serf.lan: serf: EventMemberUpdate: Node-8d441da2-2c43-88a4-63be-4420bff2dfa1 writer.go:29: 2021-01-29T19:32:24.874Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.serf.wan: serf: EventMemberUpdate: Node-8d441da2-2c43-88a4-63be-4420bff2dfa1.dc1 writer.go:29: 2021-01-29T19:32:24.878Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: Updating LAN server: server="Node-8d441da2-2c43-88a4-63be-4420bff2dfa1 (Addr: tcp/127.0.0.1:30107) (DC: dc1)" writer.go:29: 2021-01-29T19:32:24.882Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: Handled event for server in area: event=member-update server=Node-8d441da2-2c43-88a4-63be-4420bff2dfa1.dc1 area=wan writer.go:29: 2021-01-29T19:32:24.899Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:24.904Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:24.905Z [INFO] TestHTTPAPI_OptionMethod_OSS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:24.906Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.server: Skipping self join check for node since the cluster is too small: node=Node-8d441da2-2c43-88a4-63be-4420bff2dfa1 writer.go:29: 2021-01-29T19:32:24.907Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: member joined, marking health alive: member=Node-8d441da2-2c43-88a4-63be-4420bff2dfa1 writer.go:29: 2021-01-29T19:32:24.906Z [INFO] TestHTTPAPI_OptionMethod_OSS: Synced node info writer.go:29: 2021-01-29T19:32:24.911Z [DEBUG] TestHTTPAPI_OptionMethod_OSS: Node info in sync writer.go:29: 2021-01-29T19:32:24.910Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.server: Skipping self join check for node since the cluster is too small: node=Node-8d441da2-2c43-88a4-63be-4420bff2dfa1 === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/query === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.267Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/query from= latency=3.103µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/query/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.272Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/query/ from= latency=2.492776ms === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/query/xxx/execute === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.275Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/query/xxx/execute from= latency=979.172µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/query/xxx/explain === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.278Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/query/xxx/explain from= latency=913.234µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/register === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.280Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/check/register from= latency=1.362µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/bootstrap === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.283Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/bootstrap from= latency=2.692µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/roles === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.285Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/roles from= latency=1.235µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/auth-method === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.289Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/auth-method from= latency=2.979µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/rules/translate === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.291Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/rules/translate from= latency=1.215µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/policy === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.295Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/policy from= latency=1.225µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/service/register === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.298Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/service/register from= latency=1.531µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/config === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.301Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/config from= latency=2.249µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/event/list === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.303Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/event/list from= latency=1.126µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/internal/ui/services === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.306Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/internal/ui/services from= latency=702ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/binding-rule === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.309Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/binding-rule from= latency=1.308µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/connect/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.311Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/catalog/connect/ from= latency=787ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/txn === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.315Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/txn from= latency=73.72µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/checks === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.333Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/checks from= latency=1.553µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/members === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.336Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/members from= latency=802ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/operator/keyring === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.338Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/operator/keyring from= latency=1.311µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/host === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.341Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/host from= latency=2.398µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/warn/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.347Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/check/warn/ from= latency=2.033µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/service/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.353Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/catalog/service/ from= latency=2.477µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/kv/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.358Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/kv/ from= latency=1.902µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/auth-methods === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.365Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/auth-methods from= latency=2.099µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/connect/ca/leaf/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.377Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/connect/ca/leaf/ from= latency=1.781µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/coordinate/nodes === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.382Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/coordinate/nodes from= latency=2.544µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/coordinate/update === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.389Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/coordinate/update from= latency=7.81µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/clone/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.398Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/clone/ from= latency=1.424µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/node/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.403Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/catalog/node/ from= latency=2.019µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/ca/roots === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.408Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/connect/ca/roots from= latency=1.793µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/health/checks/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.415Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/health/checks/ from= latency=983ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/discovery-chain/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.419Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/discovery-chain/ from= latency=2.057µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/health/node/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.422Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/health/node/ from= latency=1.943µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/status/leader === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.425Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/status/leader from= latency=756ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/create === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.428Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/create from= latency=1.208µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/config/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.431Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/config/ from= latency=1.429µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/intentions === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.439Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/connect/intentions from= latency=2.789µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/intentions/match === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.451Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/connect/intentions/match from= latency=3.136µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/renew/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.454Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/session/renew/ from= latency=838ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/deregister/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.457Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/check/deregister/ from= latency=977ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/connect/authorize === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.460Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/connect/authorize from= latency=1.158µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/operator/raft/peer === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.473Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/operator/raft/peer from= latency=2.035µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/destroy/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.479Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/session/destroy/ from= latency=1.518µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/policies === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.484Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/policies from= latency=1.229µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/deregister === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.487Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/catalog/deregister from= latency=778ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/snapshot === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.490Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/snapshot from= latency=1.135µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/operator/raft/configuration === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.497Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/operator/raft/configuration from= latency=926ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/info/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.502Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/session/info/ from= latency=1.981µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/replication === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.505Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/replication from= latency=5.041µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/join/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.507Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/join/ from= latency=682ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/service/maintenance/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.510Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/service/maintenance/ from= latency=689ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/token/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.513Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/token/ from= latency=1.218µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/metrics === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.515Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/metrics from= latency=1.1µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/services === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.518Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/services from= latency=1.641µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/service/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.521Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/service/ from= latency=1.249µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/role/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.524Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/role/ from= latency=975ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/service/deregister/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.526Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/service/deregister/ from= latency=779ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/coordinate/datacenters === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.529Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/coordinate/datacenters from= latency=1.365µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/operator/autopilot/health === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.531Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/operator/autopilot/health from= latency=932ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/intentions/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.536Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/connect/intentions/ from= latency=1.869µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/rules/translate/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.541Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/rules/translate/ from= latency=1.846µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/tokens === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.545Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/tokens from= latency=2.254µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/node-services/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.549Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/catalog/node-services/ from= latency=1.117µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/update === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.551Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/update from= latency=810ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/binding-rules === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.554Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/binding-rules from= latency=742ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/health/service/id/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.556Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/health/service/id/ from= latency=727ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/login === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.559Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/login from= latency=3.717µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/token/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.602Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/token/ from= latency=1.748µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/create === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.604Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/session/create from= latency=3.009µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/health/state/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.607Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/health/state/ from= latency=715ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/intentions/check === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.609Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/connect/intentions/check from= latency=759ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/destroy/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.612Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/destroy/ from= latency=984ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/info/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.614Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/info/ from= latency=927ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/list === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.617Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/list from= latency=2.93µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/token/self === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.619Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/token/self from= latency=991ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/ca/configuration === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.622Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/connect/ca/configuration from= latency=1.392µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/operator/autopilot/configuration === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.625Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/operator/autopilot/configuration from= latency=1.37µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/force-leave/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.629Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/force-leave/ from= latency=1.862µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/pass/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.636Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/check/pass/ from= latency=1.574µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/fail/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.639Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/check/fail/ from= latency=1.653µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/coordinate/node/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.641Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/coordinate/node/ from= latency=1.013µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/health/connect/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.644Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/health/connect/ from= latency=1.016µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/self === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.646Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/self from= latency=733ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/leave === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.649Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/leave from= latency=932ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/internal/ui/node/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.651Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/internal/ui/node/ from= latency=1.264µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/connect/ca/roots === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.654Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/connect/ca/roots from= latency=1.396µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/services === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.656Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/catalog/services from= latency=648ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/internal/ui/nodes === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.659Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/internal/ui/nodes from= latency=726ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/list === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.670Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/session/list from= latency=2.728µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/nodes === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.675Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/catalog/nodes from= latency=1.166µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/health/service/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.676Z [DEBUG] TestHTTPAPI_OptionMethod_OSS: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:25.679Z [DEBUG] TestHTTPAPI_OptionMethod_OSS: Node info in sync writer.go:29: 2021-01-29T19:32:25.679Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/health/service/ from= latency=1.851µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/policy/name/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.684Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/policy/name/ from= latency=939ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/binding-rule/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.690Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/binding-rule/ from= latency=1.44µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/health/service/name/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.693Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/health/service/name/ from= latency=973ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/register === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.696Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/catalog/register from= latency=785ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/token === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.699Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/token from= latency=5.609µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/logout === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.703Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/logout from= latency=1.641µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/role === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.707Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/role from= latency=2.073µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/maintenance === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.711Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/maintenance from= latency=3.884µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/update/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.715Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/agent/check/update/ from= latency=1.022µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/datacenters === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.717Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/catalog/datacenters from= latency=844ns === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/node/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.720Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/session/node/ from= latency=1.69µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/policy/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.724Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/policy/ from= latency=2.188µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/role/name/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.727Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/role/name/ from= latency=1.26µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/auth-method/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.731Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/acl/auth-method/ from= latency=2.548µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/event/fire/ === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.735Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/event/fire/ from= latency=1.687µs === RUN TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/internal/acl/authorize === CONT TestHTTPAPI_OptionMethod_OSS writer.go:29: 2021-01-29T19:32:25.738Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.http: Request finished: method=OPTIONS url=http://127.0.0.1:30103/v1/internal/acl/authorize from= latency=764ns writer.go:29: 2021-01-29T19:32:25.739Z [INFO] TestHTTPAPI_OptionMethod_OSS: Requesting shutdown writer.go:29: 2021-01-29T19:32:25.741Z [INFO] TestHTTPAPI_OptionMethod_OSS.server: shutting down server writer.go:29: 2021-01-29T19:32:25.742Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:25.743Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:25.744Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:25.745Z [WARN] TestHTTPAPI_OptionMethod_OSS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:25.743Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:25.744Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:25.745Z [DEBUG] TestHTTPAPI_OptionMethod_OSS.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:25.747Z [WARN] TestHTTPAPI_OptionMethod_OSS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:25.752Z [INFO] TestHTTPAPI_OptionMethod_OSS.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:25.753Z [INFO] TestHTTPAPI_OptionMethod_OSS: consul server down writer.go:29: 2021-01-29T19:32:25.754Z [INFO] TestHTTPAPI_OptionMethod_OSS: shutdown complete writer.go:29: 2021-01-29T19:32:25.755Z [INFO] TestHTTPAPI_OptionMethod_OSS: Stopping server: protocol=DNS address=127.0.0.1:30102 network=tcp writer.go:29: 2021-01-29T19:32:25.756Z [INFO] TestHTTPAPI_OptionMethod_OSS: Stopping server: protocol=DNS address=127.0.0.1:30102 network=udp writer.go:29: 2021-01-29T19:32:25.757Z [INFO] TestHTTPAPI_OptionMethod_OSS: Stopping server: protocol=HTTP address=127.0.0.1:30103 network=tcp writer.go:29: 2021-01-29T19:32:26.258Z [INFO] TestHTTPAPI_OptionMethod_OSS: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:26.259Z [INFO] TestHTTPAPI_OptionMethod_OSS: Endpoints down --- PASS: TestHTTPAPI_OptionMethod_OSS (1.64s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/query (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/query/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/query/xxx/execute (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/query/xxx/explain (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/register (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/bootstrap (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/roles (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/auth-method (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/rules/translate (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/policy (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/service/register (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/config (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/event/list (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/internal/ui/services (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/binding-rule (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/connect/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/txn (0.02s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/checks (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/members (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/operator/keyring (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/host (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/warn/ (0.01s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/service/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/kv/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/auth-methods (0.01s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/connect/ca/leaf/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/coordinate/nodes (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/coordinate/update (0.01s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/clone/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/node/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/ca/roots (0.01s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/health/checks/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/discovery-chain/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/health/node/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/status/leader (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/create (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/config/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/intentions (0.01s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/intentions/match (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/renew/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/deregister/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/connect/authorize (0.01s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/operator/raft/peer (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/destroy/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/policies (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/deregister (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/snapshot (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/operator/raft/configuration (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/info/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/replication (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/join/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/service/maintenance/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/token/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/metrics (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/services (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/service/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/role/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/service/deregister/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/coordinate/datacenters (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/operator/autopilot/health (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/intentions/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/rules/translate/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/tokens (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/node-services/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/update (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/binding-rules (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/health/service/id/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/login (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/token/ (0.04s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/create (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/health/state/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/intentions/check (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/destroy/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/info/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/list (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/token/self (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/connect/ca/configuration (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/operator/autopilot/configuration (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/force-leave/ (0.01s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/pass/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/fail/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/coordinate/node/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/health/connect/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/self (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/leave (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/internal/ui/node/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/connect/ca/roots (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/services (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/internal/ui/nodes (0.01s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/list (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/nodes (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/health/service/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/policy/name/ (0.01s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/binding-rule/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/health/service/name/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/register (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/token (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/logout (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/role (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/maintenance (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/agent/check/update/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/catalog/datacenters (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/session/node/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/policy/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/role/name/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/acl/auth-method/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/event/fire/ (0.00s) --- PASS: TestHTTPAPI_OptionMethod_OSS/OPTIONS_/v1/internal/acl/authorize (0.00s) === RUN TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.272Z [WARN] TestHTTPAPI_AllowedNets_OSS: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:32:26.273Z [WARN] TestHTTPAPI_AllowedNets_OSS: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:26.274Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:26.275Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:26.282Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:46e5d9b2-820f-64aa-c610-058264dce0d1 Address:127.0.0.1:30113}]" writer.go:29: 2021-01-29T19:32:26.284Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.raft: entering follower state: follower="Node at 127.0.0.1:30113 [Follower]" leader= writer.go:29: 2021-01-29T19:32:26.284Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.serf.wan: serf: EventMemberJoin: Node-46e5d9b2-820f-64aa-c610-058264dce0d1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:26.288Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.serf.lan: serf: EventMemberJoin: Node-46e5d9b2-820f-64aa-c610-058264dce0d1 127.0.0.1 writer.go:29: 2021-01-29T19:32:26.290Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: Adding LAN server: server="Node-46e5d9b2-820f-64aa-c610-058264dce0d1 (Addr: tcp/127.0.0.1:30113) (DC: dc1)" writer.go:29: 2021-01-29T19:32:26.291Z [INFO] TestHTTPAPI_AllowedNets_OSS: Started DNS server: address=127.0.0.1:30108 network=udp writer.go:29: 2021-01-29T19:32:26.292Z [INFO] TestHTTPAPI_AllowedNets_OSS: Started DNS server: address=127.0.0.1:30108 network=tcp writer.go:29: 2021-01-29T19:32:26.294Z [INFO] TestHTTPAPI_AllowedNets_OSS: Started HTTP server: address=127.0.0.1:30109 network=tcp writer.go:29: 2021-01-29T19:32:26.295Z [INFO] TestHTTPAPI_AllowedNets_OSS: started state syncer writer.go:29: 2021-01-29T19:32:26.291Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: Handled event for server in area: event=member-join server=Node-46e5d9b2-820f-64aa-c610-058264dce0d1.dc1 area=wan writer.go:29: 2021-01-29T19:32:26.341Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:26.354Z [WARN] TestHTTPAPI_AllowedNets_OSS.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:26.356Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.raft: entering candidate state: node="Node at 127.0.0.1:30113 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:26.358Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:26.359Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.server.raft: vote granted: from=46e5d9b2-820f-64aa-c610-058264dce0d1 term=2 tally=1 writer.go:29: 2021-01-29T19:32:26.360Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:26.361Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.raft: entering leader state: leader="Node at 127.0.0.1:30113 [Leader]" writer.go:29: 2021-01-29T19:32:26.362Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:26.363Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: New leader elected: payload=Node-46e5d9b2-820f-64aa-c610-058264dce0d1 writer.go:29: 2021-01-29T19:32:26.365Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: initializing acls writer.go:29: 2021-01-29T19:32:26.368Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:26.371Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:26.372Z [INFO] TestHTTPAPI_AllowedNets_OSS.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:26.374Z [INFO] TestHTTPAPI_AllowedNets_OSS.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:26.375Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.serf.lan: serf: EventMemberUpdate: Node-46e5d9b2-820f-64aa-c610-058264dce0d1 writer.go:29: 2021-01-29T19:32:26.384Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: Updating LAN server: server="Node-46e5d9b2-820f-64aa-c610-058264dce0d1 (Addr: tcp/127.0.0.1:30113) (DC: dc1)" writer.go:29: 2021-01-29T19:32:26.385Z [INFO] TestHTTPAPI_AllowedNets_OSS: Synced node info writer.go:29: 2021-01-29T19:32:26.395Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.serf.wan: serf: EventMemberUpdate: Node-46e5d9b2-820f-64aa-c610-058264dce0d1.dc1 writer.go:29: 2021-01-29T19:32:26.403Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: Handled event for server in area: event=member-update server=Node-46e5d9b2-820f-64aa-c610-058264dce0d1.dc1 area=wan writer.go:29: 2021-01-29T19:32:26.408Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:26.412Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:26.413Z [INFO] TestHTTPAPI_AllowedNets_OSS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:26.414Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.server: Skipping self join check for node since the cluster is too small: node=Node-46e5d9b2-820f-64aa-c610-058264dce0d1 writer.go:29: 2021-01-29T19:32:26.418Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: member joined, marking health alive: member=Node-46e5d9b2-820f-64aa-c610-058264dce0d1 writer.go:29: 2021-01-29T19:32:26.421Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.server: Skipping self join check for node since the cluster is too small: node=Node-46e5d9b2-820f-64aa-c610-058264dce0d1 === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/operator/autopilot/configuration === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.490Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/operator/autopilot/configuration from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.493Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/operator/autopilot/configuration from=192.168.1.2:5555 latency=3.017392ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/force-leave/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.499Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/force-leave/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.502Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/force-leave/ from=192.168.1.2:5555 latency=2.918021ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/pass/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.507Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/check/pass/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.508Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/check/pass/ from=192.168.1.2:5555 latency=1.568369ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/fail/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.510Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/check/fail/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.512Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/check/fail/ from=192.168.1.2:5555 latency=1.591695ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/leave === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.514Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/leave from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.516Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/leave from=192.168.1.2:5555 latency=1.488423ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/binding-rule/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.518Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/binding-rule/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.520Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/binding-rule/ from=192.168.1.2:5555 latency=1.642983ms === RUN TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/acl/binding-rule/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.522Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=DELETE url=http://127.0.0.1:30109/v1/acl/binding-rule/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.530Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=DELETE url=http://127.0.0.1:30109/v1/acl/binding-rule/ from=192.168.1.2:5555 latency=7.154112ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/catalog/register === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.532Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/catalog/register from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.534Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/catalog/register from=192.168.1.2:5555 latency=1.700793ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/token === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.536Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/token from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.538Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/token from=192.168.1.2:5555 latency=1.653324ms === RUN TestHTTPAPI_AllowedNets_OSS/POST_/v1/acl/logout === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.541Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=POST url=http://127.0.0.1:30109/v1/acl/logout from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.544Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=POST url=http://127.0.0.1:30109/v1/acl/logout from=192.168.1.2:5555 latency=2.76252ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/role === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.552Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/role from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.555Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/role from=192.168.1.2:5555 latency=2.630876ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/maintenance === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.558Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/maintenance from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.561Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/maintenance from=192.168.1.2:5555 latency=2.953607ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/update/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.569Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/check/update/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.575Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/check/update/ from=192.168.1.2:5555 latency=5.778168ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/policy/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.587Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/policy/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.590Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/policy/ from=192.168.1.2:5555 latency=3.577623ms === RUN TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/acl/policy/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.597Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=DELETE url=http://127.0.0.1:30109/v1/acl/policy/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.600Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=DELETE url=http://127.0.0.1:30109/v1/acl/policy/ from=192.168.1.2:5555 latency=3.182519ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/auth-method/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.605Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/auth-method/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.609Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/auth-method/ from=192.168.1.2:5555 latency=3.665362ms === RUN TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/acl/auth-method/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.613Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=DELETE url=http://127.0.0.1:30109/v1/acl/auth-method/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.615Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=DELETE url=http://127.0.0.1:30109/v1/acl/auth-method/ from=192.168.1.2:5555 latency=1.46139ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/event/fire/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.617Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/event/fire/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.619Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/event/fire/ from=192.168.1.2:5555 latency=1.649806ms === RUN TestHTTPAPI_AllowedNets_OSS/POST_/v1/internal/acl/authorize === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.633Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=POST url=http://127.0.0.1:30109/v1/internal/acl/authorize from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.635Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=POST url=http://127.0.0.1:30109/v1/internal/acl/authorize from=192.168.1.2:5555 latency=2.57024ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/register === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.640Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/check/register from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.645Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/check/register from=192.168.1.2:5555 latency=5.452201ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/bootstrap === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.653Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/bootstrap from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.655Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/bootstrap from=192.168.1.2:5555 latency=2.537992ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/auth-method === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.659Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/auth-method from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.662Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/auth-method from=192.168.1.2:5555 latency=3.278875ms === RUN TestHTTPAPI_AllowedNets_OSS/POST_/v1/acl/rules/translate === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.668Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=POST url=http://127.0.0.1:30109/v1/acl/rules/translate from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.672Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=POST url=http://127.0.0.1:30109/v1/acl/rules/translate from=192.168.1.2:5555 latency=4.049505ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/policy === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.677Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/policy from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.680Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/policy from=192.168.1.2:5555 latency=2.935234ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/service/register === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.685Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/service/register from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.688Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/service/register from=192.168.1.2:5555 latency=3.023314ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/config === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.692Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/config from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.693Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/config from=192.168.1.2:5555 latency=1.598546ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/binding-rule === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.695Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/binding-rule from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.697Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/binding-rule from=192.168.1.2:5555 latency=1.598026ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/txn === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.702Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/txn from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.704Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/txn from=192.168.1.2:5555 latency=1.903377ms === RUN TestHTTPAPI_AllowedNets_OSS/POST_/v1/operator/keyring === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.710Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=POST url=http://127.0.0.1:30109/v1/operator/keyring from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.713Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=POST url=http://127.0.0.1:30109/v1/operator/keyring from=192.168.1.2:5555 latency=3.806032ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/operator/keyring === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.718Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/operator/keyring from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.720Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/operator/keyring from=192.168.1.2:5555 latency=2.823885ms === RUN TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/operator/keyring === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.723Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=DELETE url=http://127.0.0.1:30109/v1/operator/keyring from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.725Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=DELETE url=http://127.0.0.1:30109/v1/operator/keyring from=192.168.1.2:5555 latency=1.544649ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/warn/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.727Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/check/warn/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.729Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/check/warn/ from=192.168.1.2:5555 latency=1.546871ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/kv/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.731Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/kv/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.734Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/kv/ from=192.168.1.2:5555 latency=3.155848ms === RUN TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/kv/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.738Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=DELETE url=http://127.0.0.1:30109/v1/kv/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.739Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=DELETE url=http://127.0.0.1:30109/v1/kv/ from=192.168.1.2:5555 latency=1.422ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/coordinate/update === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.741Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/coordinate/update from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.743Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/coordinate/update from=192.168.1.2:5555 latency=1.43924ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/clone/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.745Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/clone/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.747Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/clone/ from=192.168.1.2:5555 latency=1.890452ms === RUN TestHTTPAPI_AllowedNets_OSS/POST_/v1/discovery-chain/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.751Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=POST url=http://127.0.0.1:30109/v1/discovery-chain/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.755Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=POST url=http://127.0.0.1:30109/v1/discovery-chain/ from=192.168.1.2:5555 latency=3.560586ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/create === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.762Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/create from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.764Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/create from=192.168.1.2:5555 latency=2.74675ms === RUN TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/config/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.769Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=DELETE url=http://127.0.0.1:30109/v1/config/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.773Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=DELETE url=http://127.0.0.1:30109/v1/config/ from=192.168.1.2:5555 latency=3.2542ms === RUN TestHTTPAPI_AllowedNets_OSS/POST_/v1/connect/intentions === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.780Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=POST url=http://127.0.0.1:30109/v1/connect/intentions from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.787Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=POST url=http://127.0.0.1:30109/v1/connect/intentions from=192.168.1.2:5555 latency=6.355587ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/session/renew/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.797Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/session/renew/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.800Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/session/renew/ from=192.168.1.2:5555 latency=3.04068ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/deregister/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.814Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/check/deregister/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.817Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/check/deregister/ from=192.168.1.2:5555 latency=2.494372ms === RUN TestHTTPAPI_AllowedNets_OSS/POST_/v1/agent/connect/authorize === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.821Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=POST url=http://127.0.0.1:30109/v1/agent/connect/authorize from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.833Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=POST url=http://127.0.0.1:30109/v1/agent/connect/authorize from=192.168.1.2:5555 latency=11.369834ms === RUN TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/operator/raft/peer === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.837Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=DELETE url=http://127.0.0.1:30109/v1/operator/raft/peer from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.839Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=DELETE url=http://127.0.0.1:30109/v1/operator/raft/peer from=192.168.1.2:5555 latency=1.689963ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/session/destroy/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.841Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/session/destroy/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.843Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/session/destroy/ from=192.168.1.2:5555 latency=1.588795ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/catalog/deregister === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.846Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/catalog/deregister from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.851Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/catalog/deregister from=192.168.1.2:5555 latency=4.777918ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/snapshot === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.862Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/snapshot from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.864Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/snapshot from=192.168.1.2:5555 latency=1.471341ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/join/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.866Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/join/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.868Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/join/ from=192.168.1.2:5555 latency=2.186348ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/service/maintenance/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.873Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/service/maintenance/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.875Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/service/maintenance/ from=192.168.1.2:5555 latency=2.340539ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/token/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.879Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/token/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.881Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/token/ from=192.168.1.2:5555 latency=2.307613ms === RUN TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/acl/token/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.884Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=DELETE url=http://127.0.0.1:30109/v1/acl/token/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.887Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=DELETE url=http://127.0.0.1:30109/v1/acl/token/ from=192.168.1.2:5555 latency=2.577059ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/role/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.913Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/role/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.915Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/role/ from=192.168.1.2:5555 latency=2.594051ms === RUN TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/acl/role/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.919Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=DELETE url=http://127.0.0.1:30109/v1/acl/role/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.923Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=DELETE url=http://127.0.0.1:30109/v1/acl/role/ from=192.168.1.2:5555 latency=4.015652ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/service/deregister/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.946Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/service/deregister/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.948Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/service/deregister/ from=192.168.1.2:5555 latency=2.589491ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/connect/intentions/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.953Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/connect/intentions/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.955Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/connect/intentions/ from=192.168.1.2:5555 latency=2.613328ms === RUN TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/connect/intentions/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.960Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=DELETE url=http://127.0.0.1:30109/v1/connect/intentions/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.964Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=DELETE url=http://127.0.0.1:30109/v1/connect/intentions/ from=192.168.1.2:5555 latency=3.187516ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/update === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.969Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/update from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.978Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/update from=192.168.1.2:5555 latency=8.966844ms === RUN TestHTTPAPI_AllowedNets_OSS/POST_/v1/acl/login === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.986Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=POST url=http://127.0.0.1:30109/v1/acl/login from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.989Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=POST url=http://127.0.0.1:30109/v1/acl/login from=192.168.1.2:5555 latency=2.206477ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/token/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.992Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/agent/token/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:26.995Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/agent/token/ from=192.168.1.2:5555 latency=2.201405ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/session/create === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:26.999Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/session/create from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:27.001Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/session/create from=192.168.1.2:5555 latency=2.712954ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/destroy/ === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:27.004Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/acl/destroy/ from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:27.008Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/acl/destroy/ from=192.168.1.2:5555 latency=3.663382ms === RUN TestHTTPAPI_AllowedNets_OSS/PUT_/v1/connect/ca/configuration === CONT TestHTTPAPI_AllowedNets_OSS writer.go:29: 2021-01-29T19:32:27.014Z [ERROR] TestHTTPAPI_AllowedNets_OSS.http: Request error: method=PUT url=http://127.0.0.1:30109/v1/connect/ca/configuration from=192.168.1.2:5555 error="Access is restricted" writer.go:29: 2021-01-29T19:32:27.021Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.http: Request finished: method=PUT url=http://127.0.0.1:30109/v1/connect/ca/configuration from=192.168.1.2:5555 latency=6.393467ms writer.go:29: 2021-01-29T19:32:27.027Z [INFO] TestHTTPAPI_AllowedNets_OSS: Requesting shutdown writer.go:29: 2021-01-29T19:32:27.030Z [INFO] TestHTTPAPI_AllowedNets_OSS.server: shutting down server writer.go:29: 2021-01-29T19:32:27.031Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:27.032Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:27.033Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:27.033Z [WARN] TestHTTPAPI_AllowedNets_OSS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:27.035Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:27.036Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:27.037Z [DEBUG] TestHTTPAPI_AllowedNets_OSS.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:27.046Z [WARN] TestHTTPAPI_AllowedNets_OSS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:27.048Z [INFO] TestHTTPAPI_AllowedNets_OSS.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:27.048Z [INFO] TestHTTPAPI_AllowedNets_OSS: consul server down writer.go:29: 2021-01-29T19:32:27.050Z [INFO] TestHTTPAPI_AllowedNets_OSS: shutdown complete writer.go:29: 2021-01-29T19:32:27.050Z [INFO] TestHTTPAPI_AllowedNets_OSS: Stopping server: protocol=DNS address=127.0.0.1:30108 network=tcp writer.go:29: 2021-01-29T19:32:27.051Z [INFO] TestHTTPAPI_AllowedNets_OSS: Stopping server: protocol=DNS address=127.0.0.1:30108 network=udp writer.go:29: 2021-01-29T19:32:27.052Z [INFO] TestHTTPAPI_AllowedNets_OSS: Stopping server: protocol=HTTP address=127.0.0.1:30109 network=tcp writer.go:29: 2021-01-29T19:32:27.553Z [INFO] TestHTTPAPI_AllowedNets_OSS: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:27.557Z [INFO] TestHTTPAPI_AllowedNets_OSS: Endpoints down --- PASS: TestHTTPAPI_AllowedNets_OSS (1.30s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/operator/autopilot/configuration (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/force-leave/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/pass/ (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/fail/ (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/leave (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/binding-rule/ (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/acl/binding-rule/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/catalog/register (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/token (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/POST_/v1/acl/logout (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/role (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/maintenance (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/update/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/policy/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/acl/policy/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/auth-method/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/acl/auth-method/ (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/event/fire/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/POST_/v1/internal/acl/authorize (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/register (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/bootstrap (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/auth-method (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/POST_/v1/acl/rules/translate (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/policy (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/service/register (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/config (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/binding-rule (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/txn (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/POST_/v1/operator/keyring (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/operator/keyring (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/operator/keyring (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/warn/ (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/kv/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/kv/ (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/coordinate/update (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/clone/ (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/POST_/v1/discovery-chain/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/create (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/config/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/POST_/v1/connect/intentions (0.02s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/session/renew/ (0.02s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/check/deregister/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/POST_/v1/agent/connect/authorize (0.02s) --- PASS: TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/operator/raft/peer (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/session/destroy/ (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/catalog/deregister (0.02s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/snapshot (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/join/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/service/maintenance/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/token/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/acl/token/ (0.03s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/role/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/acl/role/ (0.03s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/service/deregister/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/connect/intentions/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/DELETE_/v1/connect/intentions/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/update (0.02s) --- PASS: TestHTTPAPI_AllowedNets_OSS/POST_/v1/acl/login (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/agent/token/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/session/create (0.00s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/acl/destroy/ (0.01s) --- PASS: TestHTTPAPI_AllowedNets_OSS/PUT_/v1/connect/ca/configuration (0.01s) === RUN TestHTTPServer_UnixSocket === PAUSE TestHTTPServer_UnixSocket === RUN TestHTTPServer_UnixSocket_FileExists === PAUSE TestHTTPServer_UnixSocket_FileExists === RUN TestHTTPServer_H2 http_test.go:132: DM-skipped --- SKIP: TestHTTPServer_H2 (0.00s) === RUN TestSetIndex === PAUSE TestSetIndex === RUN TestSetKnownLeader === PAUSE TestSetKnownLeader === RUN TestSetLastContact === PAUSE TestSetLastContact === RUN TestSetMeta === PAUSE TestSetMeta === RUN TestHTTPAPI_BlockEndpoints === PAUSE TestHTTPAPI_BlockEndpoints === RUN TestHTTPAPI_Ban_Nonprintable_Characters http_test.go:324: DM-skipped --- SKIP: TestHTTPAPI_Ban_Nonprintable_Characters (0.00s) === RUN TestHTTPAPI_Allow_Nonprintable_Characters_With_Flag http_test.go:344: DM-skipped --- SKIP: TestHTTPAPI_Allow_Nonprintable_Characters_With_Flag (0.00s) === RUN TestHTTPAPI_TranslateAddrHeader === PAUSE TestHTTPAPI_TranslateAddrHeader === RUN TestHTTPAPIResponseHeaders === PAUSE TestHTTPAPIResponseHeaders === RUN TestUIResponseHeaders === PAUSE TestUIResponseHeaders === RUN TestContentTypeIsJSON === PAUSE TestContentTypeIsJSON === RUN TestHTTP_wrap_obfuscateLog === PAUSE TestHTTP_wrap_obfuscateLog === RUN TestPrettyPrint === PAUSE TestPrettyPrint === RUN TestPrettyPrintBare === PAUSE TestPrettyPrintBare === RUN TestParseSource === PAUSE TestParseSource === RUN TestParseCacheControl === RUN TestParseCacheControl/empty_header === RUN TestParseCacheControl/simple_max-age === RUN TestParseCacheControl/zero_max-age === RUN TestParseCacheControl/must-revalidate === RUN TestParseCacheControl/mixes_age,_must-revalidate === RUN TestParseCacheControl/quoted_max-age === RUN TestParseCacheControl/mixed_case_max-age === RUN TestParseCacheControl/simple_stale-if-error === RUN TestParseCacheControl/combined_with_space === RUN TestParseCacheControl/combined_no_space === RUN TestParseCacheControl/unsupported_directive === RUN TestParseCacheControl/mixed_unsupported_directive === RUN TestParseCacheControl/garbage_value === RUN TestParseCacheControl/garbage_value_with_quotes --- PASS: TestParseCacheControl (0.05s) --- PASS: TestParseCacheControl/empty_header (0.00s) --- PASS: TestParseCacheControl/simple_max-age (0.00s) --- PASS: TestParseCacheControl/zero_max-age (0.00s) --- PASS: TestParseCacheControl/must-revalidate (0.00s) --- PASS: TestParseCacheControl/mixes_age,_must-revalidate (0.01s) --- PASS: TestParseCacheControl/quoted_max-age (0.00s) --- PASS: TestParseCacheControl/mixed_case_max-age (0.00s) --- PASS: TestParseCacheControl/simple_stale-if-error (0.00s) --- PASS: TestParseCacheControl/combined_with_space (0.00s) --- PASS: TestParseCacheControl/combined_no_space (0.00s) --- PASS: TestParseCacheControl/unsupported_directive (0.00s) --- PASS: TestParseCacheControl/mixed_unsupported_directive (0.00s) --- PASS: TestParseCacheControl/garbage_value (0.00s) --- PASS: TestParseCacheControl/garbage_value_with_quotes (0.00s) === RUN TestParseWait === PAUSE TestParseWait === RUN TestPProfHandlers_EnableDebug === PAUSE TestPProfHandlers_EnableDebug === RUN TestPProfHandlers_DisableDebugNoACLs === PAUSE TestPProfHandlers_DisableDebugNoACLs === RUN TestPProfHandlers_ACLs === PAUSE TestPProfHandlers_ACLs === RUN TestParseWait_InvalidTime === PAUSE TestParseWait_InvalidTime === RUN TestParseWait_InvalidIndex === PAUSE TestParseWait_InvalidIndex === RUN TestParseConsistency === PAUSE TestParseConsistency === RUN TestParseConsistencyAndMaxStale writer.go:29: 2021-01-29T19:32:27.688Z [WARN] TestParseConsistencyAndMaxStale: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:27.702Z [DEBUG] TestParseConsistencyAndMaxStale.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:27.703Z [DEBUG] TestParseConsistencyAndMaxStale.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:27.708Z [INFO] TestParseConsistencyAndMaxStale.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cea13c51-c279-c9c4-3d60-bd9510531c7c Address:127.0.0.1:30119}]" writer.go:29: 2021-01-29T19:32:27.709Z [INFO] TestParseConsistencyAndMaxStale.server.raft: entering follower state: follower="Node at 127.0.0.1:30119 [Follower]" leader= writer.go:29: 2021-01-29T19:32:27.712Z [INFO] TestParseConsistencyAndMaxStale.server.serf.wan: serf: EventMemberJoin: Node-cea13c51-c279-c9c4-3d60-bd9510531c7c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:27.719Z [INFO] TestParseConsistencyAndMaxStale.server.serf.lan: serf: EventMemberJoin: Node-cea13c51-c279-c9c4-3d60-bd9510531c7c 127.0.0.1 writer.go:29: 2021-01-29T19:32:27.724Z [INFO] TestParseConsistencyAndMaxStale.server: Handled event for server in area: event=member-join server=Node-cea13c51-c279-c9c4-3d60-bd9510531c7c.dc1 area=wan writer.go:29: 2021-01-29T19:32:27.725Z [INFO] TestParseConsistencyAndMaxStale.server: Adding LAN server: server="Node-cea13c51-c279-c9c4-3d60-bd9510531c7c (Addr: tcp/127.0.0.1:30119) (DC: dc1)" writer.go:29: 2021-01-29T19:32:27.726Z [INFO] TestParseConsistencyAndMaxStale: Started DNS server: address=127.0.0.1:30114 network=udp writer.go:29: 2021-01-29T19:32:27.728Z [INFO] TestParseConsistencyAndMaxStale: Started DNS server: address=127.0.0.1:30114 network=tcp writer.go:29: 2021-01-29T19:32:27.730Z [INFO] TestParseConsistencyAndMaxStale: Started HTTP server: address=127.0.0.1:30115 network=tcp writer.go:29: 2021-01-29T19:32:27.732Z [INFO] TestParseConsistencyAndMaxStale: started state syncer writer.go:29: 2021-01-29T19:32:27.762Z [WARN] TestParseConsistencyAndMaxStale.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:27.764Z [INFO] TestParseConsistencyAndMaxStale.server.raft: entering candidate state: node="Node at 127.0.0.1:30119 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:27.766Z [DEBUG] TestParseConsistencyAndMaxStale.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:27.768Z [DEBUG] TestParseConsistencyAndMaxStale.server.raft: vote granted: from=cea13c51-c279-c9c4-3d60-bd9510531c7c term=2 tally=1 writer.go:29: 2021-01-29T19:32:27.770Z [INFO] TestParseConsistencyAndMaxStale.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:27.771Z [INFO] TestParseConsistencyAndMaxStale.server.raft: entering leader state: leader="Node at 127.0.0.1:30119 [Leader]" writer.go:29: 2021-01-29T19:32:27.773Z [INFO] TestParseConsistencyAndMaxStale.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:27.774Z [INFO] TestParseConsistencyAndMaxStale.server: New leader elected: payload=Node-cea13c51-c279-c9c4-3d60-bd9510531c7c writer.go:29: 2021-01-29T19:32:27.775Z [DEBUG] TestParseConsistencyAndMaxStale.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30119 writer.go:29: 2021-01-29T19:32:27.781Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:27.785Z [INFO] TestParseConsistencyAndMaxStale.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:27.786Z [INFO] TestParseConsistencyAndMaxStale.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:27.786Z [DEBUG] TestParseConsistencyAndMaxStale.server: Skipping self join check for node since the cluster is too small: node=Node-cea13c51-c279-c9c4-3d60-bd9510531c7c writer.go:29: 2021-01-29T19:32:27.787Z [INFO] TestParseConsistencyAndMaxStale.server: member joined, marking health alive: member=Node-cea13c51-c279-c9c4-3d60-bd9510531c7c writer.go:29: 2021-01-29T19:32:27.824Z [INFO] TestParseConsistencyAndMaxStale: Requesting shutdown writer.go:29: 2021-01-29T19:32:27.825Z [INFO] TestParseConsistencyAndMaxStale.server: shutting down server writer.go:29: 2021-01-29T19:32:27.826Z [DEBUG] TestParseConsistencyAndMaxStale.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:27.827Z [WARN] TestParseConsistencyAndMaxStale.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:27.826Z [ERROR] TestParseConsistencyAndMaxStale.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:27.827Z [DEBUG] TestParseConsistencyAndMaxStale.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:27.829Z [WARN] TestParseConsistencyAndMaxStale.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:27.832Z [INFO] TestParseConsistencyAndMaxStale.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:27.834Z [INFO] TestParseConsistencyAndMaxStale: consul server down writer.go:29: 2021-01-29T19:32:27.835Z [INFO] TestParseConsistencyAndMaxStale: shutdown complete writer.go:29: 2021-01-29T19:32:27.835Z [INFO] TestParseConsistencyAndMaxStale: Stopping server: protocol=DNS address=127.0.0.1:30114 network=tcp writer.go:29: 2021-01-29T19:32:27.836Z [INFO] TestParseConsistencyAndMaxStale: Stopping server: protocol=DNS address=127.0.0.1:30114 network=udp writer.go:29: 2021-01-29T19:32:27.837Z [INFO] TestParseConsistencyAndMaxStale: Stopping server: protocol=HTTP address=127.0.0.1:30115 network=tcp writer.go:29: 2021-01-29T19:32:28.338Z [INFO] TestParseConsistencyAndMaxStale: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:28.340Z [INFO] TestParseConsistencyAndMaxStale: Endpoints down --- PASS: TestParseConsistencyAndMaxStale (0.69s) === RUN TestParseConsistency_Invalid === PAUSE TestParseConsistency_Invalid === RUN TestACLResolution === PAUSE TestACLResolution === RUN TestEnableWebUI === PAUSE TestEnableWebUI === RUN TestAllowedNets http_test.go:1166: DM-skipped --- SKIP: TestAllowedNets (0.00s) === RUN TestHTTPServer_HandshakeTimeout === PAUSE TestHTTPServer_HandshakeTimeout === RUN TestRPC_HTTPSMaxConnsPerClient === PAUSE TestRPC_HTTPSMaxConnsPerClient === RUN TestIntentionsList_empty === PAUSE TestIntentionsList_empty === RUN TestIntentionsList_values === PAUSE TestIntentionsList_values === RUN TestIntentionsMatch_basic === PAUSE TestIntentionsMatch_basic === RUN TestIntentionsMatch_noBy === PAUSE TestIntentionsMatch_noBy === RUN TestIntentionsMatch_byInvalid === PAUSE TestIntentionsMatch_byInvalid === RUN TestIntentionsMatch_noName === PAUSE TestIntentionsMatch_noName === RUN TestIntentionsCheck_basic === PAUSE TestIntentionsCheck_basic === RUN TestIntentionsCheck_noSource === PAUSE TestIntentionsCheck_noSource === RUN TestIntentionsCheck_noDestination === PAUSE TestIntentionsCheck_noDestination === RUN TestIntentionsCreate_good === PAUSE TestIntentionsCreate_good === RUN TestIntentionsCreate_noBody === PAUSE TestIntentionsCreate_noBody === RUN TestIntentionsSpecificGet_good === PAUSE TestIntentionsSpecificGet_good === RUN TestIntentionsSpecificGet_invalidId === PAUSE TestIntentionsSpecificGet_invalidId === RUN TestIntentionsSpecificUpdate_good === PAUSE TestIntentionsSpecificUpdate_good === RUN TestIntentionsSpecificDelete_good === PAUSE TestIntentionsSpecificDelete_good === RUN TestParseIntentionMatchEntry === RUN TestParseIntentionMatchEntry/foo === RUN TestParseIntentionMatchEntry/foo/bar === RUN TestParseIntentionMatchEntry/foo/bar/baz --- PASS: TestParseIntentionMatchEntry (0.00s) --- PASS: TestParseIntentionMatchEntry/foo (0.00s) --- PASS: TestParseIntentionMatchEntry/foo/bar (0.00s) --- PASS: TestParseIntentionMatchEntry/foo/bar/baz (0.00s) === RUN TestAgent_LoadKeyrings === PAUSE TestAgent_LoadKeyrings === RUN TestAgent_InmemKeyrings === PAUSE TestAgent_InmemKeyrings === RUN TestAgent_InitKeyring === PAUSE TestAgent_InitKeyring === RUN TestAgentKeyring_ACL === PAUSE TestAgentKeyring_ACL === RUN TestValidateLocalOnly --- PASS: TestValidateLocalOnly (0.00s) === RUN TestAgent_KeyringIsMissingKey --- PASS: TestAgent_KeyringIsMissingKey (0.00s) === RUN TestKVSEndpoint_PUT_GET_DELETE === PAUSE TestKVSEndpoint_PUT_GET_DELETE === RUN TestKVSEndpoint_Recurse === PAUSE TestKVSEndpoint_Recurse === RUN TestKVSEndpoint_DELETE_CAS === PAUSE TestKVSEndpoint_DELETE_CAS === RUN TestKVSEndpoint_CAS === PAUSE TestKVSEndpoint_CAS === RUN TestKVSEndpoint_ListKeys kvs_endpoint_test.go:294: DM-skipped --- SKIP: TestKVSEndpoint_ListKeys (0.00s) === RUN TestKVSEndpoint_AcquireRelease === PAUSE TestKVSEndpoint_AcquireRelease === RUN TestKVSEndpoint_GET_Raw kvs_endpoint_test.go:403: DM-skipped --- SKIP: TestKVSEndpoint_GET_Raw (0.00s) === RUN TestKVSEndpoint_PUT_ConflictingFlags === PAUSE TestKVSEndpoint_PUT_ConflictingFlags === RUN TestKVSEndpoint_DELETE_ConflictingFlags === PAUSE TestKVSEndpoint_DELETE_ConflictingFlags === RUN TestNotifyGroup --- PASS: TestNotifyGroup (0.00s) === RUN TestNotifyGroup_Clear --- PASS: TestNotifyGroup_Clear (0.00s) === RUN TestOperator_RaftConfiguration === PAUSE TestOperator_RaftConfiguration === RUN TestOperator_RaftPeer === PAUSE TestOperator_RaftPeer === RUN TestOperator_KeyringInstall === PAUSE TestOperator_KeyringInstall === RUN TestOperator_KeyringList === PAUSE TestOperator_KeyringList === RUN TestOperator_KeyringRemove === PAUSE TestOperator_KeyringRemove === RUN TestOperator_KeyringUse === PAUSE TestOperator_KeyringUse === RUN TestOperator_Keyring_InvalidRelayFactor === PAUSE TestOperator_Keyring_InvalidRelayFactor === RUN TestOperator_Keyring_LocalOnly === PAUSE TestOperator_Keyring_LocalOnly === RUN TestOperator_AutopilotGetConfiguration === PAUSE TestOperator_AutopilotGetConfiguration === RUN TestOperator_AutopilotSetConfiguration operator_endpoint_test.go:350: DM-skipped --- SKIP: TestOperator_AutopilotSetConfiguration (0.00s) === RUN TestOperator_AutopilotCASConfiguration === PAUSE TestOperator_AutopilotCASConfiguration === RUN TestOperator_ServerHealth === PAUSE TestOperator_ServerHealth === RUN TestOperator_ServerHealth_Unhealthy === PAUSE TestOperator_ServerHealth_Unhealthy === RUN TestPreparedQuery_Create === PAUSE TestPreparedQuery_Create === RUN TestPreparedQuery_List === PAUSE TestPreparedQuery_List === RUN TestPreparedQuery_Execute === PAUSE TestPreparedQuery_Execute === RUN TestPreparedQuery_ExecuteCached === PAUSE TestPreparedQuery_ExecuteCached === RUN TestPreparedQuery_Explain === PAUSE TestPreparedQuery_Explain === RUN TestPreparedQuery_Get === PAUSE TestPreparedQuery_Get === RUN TestPreparedQuery_Update === PAUSE TestPreparedQuery_Update === RUN TestPreparedQuery_Delete === PAUSE TestPreparedQuery_Delete === RUN TestPreparedQuery_parseLimit === PAUSE TestPreparedQuery_parseLimit === RUN TestPreparedQuery_Integration prepared_query_endpoint_test.go:994: DM-skipped --- SKIP: TestPreparedQuery_Integration (0.00s) === RUN TestRexecWriter remote_exec_test.go:28: DM-skipped --- SKIP: TestRexecWriter (0.00s) === RUN TestRemoteExecGetSpec === PAUSE TestRemoteExecGetSpec === RUN TestRemoteExecGetSpec_ACLToken === PAUSE TestRemoteExecGetSpec_ACLToken === RUN TestRemoteExecGetSpec_ACLAgentToken === PAUSE TestRemoteExecGetSpec_ACLAgentToken === RUN TestRemoteExecGetSpec_ACLDeny === PAUSE TestRemoteExecGetSpec_ACLDeny === RUN TestRemoteExecWrites === PAUSE TestRemoteExecWrites === RUN TestRemoteExecWrites_ACLToken === PAUSE TestRemoteExecWrites_ACLToken === RUN TestRemoteExecWrites_ACLAgentToken === PAUSE TestRemoteExecWrites_ACLAgentToken === RUN TestRemoteExecWrites_ACLDeny === PAUSE TestRemoteExecWrites_ACLDeny === RUN TestHandleRemoteExec === PAUSE TestHandleRemoteExec === RUN TestHandleRemoteExecFailed === PAUSE TestHandleRemoteExecFailed === RUN TestAgent_ServiceHTTPChecksNotification service_checks_test.go:19: DM-Disabled --- SKIP: TestAgent_ServiceHTTPChecksNotification (0.00s) === RUN TestServiceManager_RegisterService writer.go:29: 2021-01-29T19:32:28.491Z [WARN] TestServiceManager_RegisterService: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:28.493Z [DEBUG] TestServiceManager_RegisterService.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:28.495Z [DEBUG] TestServiceManager_RegisterService.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:28.500Z [INFO] TestServiceManager_RegisterService.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e86bf4eb-b141-a766-df73-61488a10b017 Address:127.0.0.1:30125}]" writer.go:29: 2021-01-29T19:32:28.503Z [INFO] TestServiceManager_RegisterService.server.raft: entering follower state: follower="Node at 127.0.0.1:30125 [Follower]" leader= writer.go:29: 2021-01-29T19:32:28.504Z [INFO] TestServiceManager_RegisterService.server.serf.wan: serf: EventMemberJoin: Node-e86bf4eb-b141-a766-df73-61488a10b017.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:28.507Z [INFO] TestServiceManager_RegisterService.server.serf.lan: serf: EventMemberJoin: Node-e86bf4eb-b141-a766-df73-61488a10b017 127.0.0.1 writer.go:29: 2021-01-29T19:32:28.509Z [INFO] TestServiceManager_RegisterService.server: Handled event for server in area: event=member-join server=Node-e86bf4eb-b141-a766-df73-61488a10b017.dc1 area=wan writer.go:29: 2021-01-29T19:32:28.509Z [INFO] TestServiceManager_RegisterService.server: Adding LAN server: server="Node-e86bf4eb-b141-a766-df73-61488a10b017 (Addr: tcp/127.0.0.1:30125) (DC: dc1)" writer.go:29: 2021-01-29T19:32:28.509Z [INFO] TestServiceManager_RegisterService: Started DNS server: address=127.0.0.1:30120 network=udp writer.go:29: 2021-01-29T19:32:28.511Z [INFO] TestServiceManager_RegisterService: Started DNS server: address=127.0.0.1:30120 network=tcp writer.go:29: 2021-01-29T19:32:28.513Z [INFO] TestServiceManager_RegisterService: Started HTTP server: address=127.0.0.1:30121 network=tcp writer.go:29: 2021-01-29T19:32:28.514Z [INFO] TestServiceManager_RegisterService: started state syncer writer.go:29: 2021-01-29T19:32:28.570Z [WARN] TestServiceManager_RegisterService.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:28.572Z [INFO] TestServiceManager_RegisterService.server.raft: entering candidate state: node="Node at 127.0.0.1:30125 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:28.575Z [DEBUG] TestServiceManager_RegisterService.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:28.577Z [DEBUG] TestServiceManager_RegisterService.server.raft: vote granted: from=e86bf4eb-b141-a766-df73-61488a10b017 term=2 tally=1 writer.go:29: 2021-01-29T19:32:28.578Z [INFO] TestServiceManager_RegisterService.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:28.580Z [INFO] TestServiceManager_RegisterService.server.raft: entering leader state: leader="Node at 127.0.0.1:30125 [Leader]" writer.go:29: 2021-01-29T19:32:28.582Z [INFO] TestServiceManager_RegisterService.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:28.582Z [INFO] TestServiceManager_RegisterService.server: New leader elected: payload=Node-e86bf4eb-b141-a766-df73-61488a10b017 writer.go:29: 2021-01-29T19:32:28.583Z [DEBUG] TestServiceManager_RegisterService.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30125 writer.go:29: 2021-01-29T19:32:28.587Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:28.595Z [INFO] TestServiceManager_RegisterService.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:28.596Z [INFO] TestServiceManager_RegisterService.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:28.597Z [DEBUG] TestServiceManager_RegisterService.server: Skipping self join check for node since the cluster is too small: node=Node-e86bf4eb-b141-a766-df73-61488a10b017 writer.go:29: 2021-01-29T19:32:28.598Z [INFO] TestServiceManager_RegisterService.server: member joined, marking health alive: member=Node-e86bf4eb-b141-a766-df73-61488a10b017 writer.go:29: 2021-01-29T19:32:28.809Z [DEBUG] TestServiceManager_RegisterService: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:28.813Z [INFO] TestServiceManager_RegisterService: Synced node info writer.go:29: 2021-01-29T19:32:28.953Z [INFO] TestServiceManager_RegisterService: Requesting shutdown writer.go:29: 2021-01-29T19:32:28.954Z [INFO] TestServiceManager_RegisterService.server: shutting down server writer.go:29: 2021-01-29T19:32:28.955Z [DEBUG] TestServiceManager_RegisterService.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:28.956Z [WARN] TestServiceManager_RegisterService.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:28.957Z [DEBUG] TestServiceManager_RegisterService.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:28.959Z [WARN] TestServiceManager_RegisterService.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:28.961Z [INFO] TestServiceManager_RegisterService.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:28.961Z [INFO] TestServiceManager_RegisterService: consul server down writer.go:29: 2021-01-29T19:32:28.963Z [INFO] TestServiceManager_RegisterService: shutdown complete writer.go:29: 2021-01-29T19:32:28.963Z [INFO] TestServiceManager_RegisterService: Stopping server: protocol=DNS address=127.0.0.1:30120 network=tcp writer.go:29: 2021-01-29T19:32:28.964Z [INFO] TestServiceManager_RegisterService: Stopping server: protocol=DNS address=127.0.0.1:30120 network=udp writer.go:29: 2021-01-29T19:32:28.965Z [INFO] TestServiceManager_RegisterService: Stopping server: protocol=HTTP address=127.0.0.1:30121 network=tcp writer.go:29: 2021-01-29T19:32:29.466Z [INFO] TestServiceManager_RegisterService: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:29.468Z [INFO] TestServiceManager_RegisterService: Endpoints down --- PASS: TestServiceManager_RegisterService (0.99s) === RUN TestServiceManager_RegisterSidecar writer.go:29: 2021-01-29T19:32:29.481Z [WARN] TestServiceManager_RegisterSidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:29.483Z [DEBUG] TestServiceManager_RegisterSidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:29.485Z [DEBUG] TestServiceManager_RegisterSidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:29.503Z [INFO] TestServiceManager_RegisterSidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7294398b-70fe-bc63-e168-3aa0390a9293 Address:127.0.0.1:30131}]" writer.go:29: 2021-01-29T19:32:29.505Z [INFO] TestServiceManager_RegisterSidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:30131 [Follower]" leader= writer.go:29: 2021-01-29T19:32:29.505Z [INFO] TestServiceManager_RegisterSidecar.server.serf.wan: serf: EventMemberJoin: Node-7294398b-70fe-bc63-e168-3aa0390a9293.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:29.509Z [INFO] TestServiceManager_RegisterSidecar.server.serf.lan: serf: EventMemberJoin: Node-7294398b-70fe-bc63-e168-3aa0390a9293 127.0.0.1 writer.go:29: 2021-01-29T19:32:29.511Z [INFO] TestServiceManager_RegisterSidecar.server: Handled event for server in area: event=member-join server=Node-7294398b-70fe-bc63-e168-3aa0390a9293.dc1 area=wan writer.go:29: 2021-01-29T19:32:29.512Z [INFO] TestServiceManager_RegisterSidecar.server: Adding LAN server: server="Node-7294398b-70fe-bc63-e168-3aa0390a9293 (Addr: tcp/127.0.0.1:30131) (DC: dc1)" writer.go:29: 2021-01-29T19:32:29.512Z [INFO] TestServiceManager_RegisterSidecar: Started DNS server: address=127.0.0.1:30126 network=udp writer.go:29: 2021-01-29T19:32:29.515Z [INFO] TestServiceManager_RegisterSidecar: Started DNS server: address=127.0.0.1:30126 network=tcp writer.go:29: 2021-01-29T19:32:29.517Z [INFO] TestServiceManager_RegisterSidecar: Started HTTP server: address=127.0.0.1:30127 network=tcp writer.go:29: 2021-01-29T19:32:29.518Z [INFO] TestServiceManager_RegisterSidecar: started state syncer writer.go:29: 2021-01-29T19:32:29.574Z [WARN] TestServiceManager_RegisterSidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:29.576Z [INFO] TestServiceManager_RegisterSidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:30131 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:29.579Z [DEBUG] TestServiceManager_RegisterSidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:29.580Z [DEBUG] TestServiceManager_RegisterSidecar.server.raft: vote granted: from=7294398b-70fe-bc63-e168-3aa0390a9293 term=2 tally=1 writer.go:29: 2021-01-29T19:32:29.582Z [INFO] TestServiceManager_RegisterSidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:29.583Z [INFO] TestServiceManager_RegisterSidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:30131 [Leader]" writer.go:29: 2021-01-29T19:32:29.584Z [INFO] TestServiceManager_RegisterSidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:29.585Z [INFO] TestServiceManager_RegisterSidecar.server: New leader elected: payload=Node-7294398b-70fe-bc63-e168-3aa0390a9293 writer.go:29: 2021-01-29T19:32:29.586Z [DEBUG] TestServiceManager_RegisterSidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30131 writer.go:29: 2021-01-29T19:32:29.594Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:29.617Z [INFO] TestServiceManager_RegisterSidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:29.619Z [INFO] TestServiceManager_RegisterSidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:29.621Z [DEBUG] TestServiceManager_RegisterSidecar.server: Skipping self join check for node since the cluster is too small: node=Node-7294398b-70fe-bc63-e168-3aa0390a9293 writer.go:29: 2021-01-29T19:32:29.623Z [INFO] TestServiceManager_RegisterSidecar.server: member joined, marking health alive: member=Node-7294398b-70fe-bc63-e168-3aa0390a9293 writer.go:29: 2021-01-29T19:32:29.836Z [DEBUG] TestServiceManager_RegisterSidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:29.838Z [INFO] TestServiceManager_RegisterSidecar: Synced node info writer.go:29: 2021-01-29T19:32:29.881Z [DEBUG] TestServiceManager_RegisterSidecar: added local registration for service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:32:29.883Z [INFO] TestServiceManager_RegisterSidecar: Requesting shutdown writer.go:29: 2021-01-29T19:32:29.885Z [INFO] TestServiceManager_RegisterSidecar.server: shutting down server writer.go:29: 2021-01-29T19:32:29.886Z [DEBUG] TestServiceManager_RegisterSidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:29.887Z [WARN] TestServiceManager_RegisterSidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:29.888Z [DEBUG] TestServiceManager_RegisterSidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:29.892Z [WARN] TestServiceManager_RegisterSidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:29.897Z [INFO] TestServiceManager_RegisterSidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:29.897Z [INFO] TestServiceManager_RegisterSidecar: consul server down writer.go:29: 2021-01-29T19:32:29.899Z [INFO] TestServiceManager_RegisterSidecar: shutdown complete writer.go:29: 2021-01-29T19:32:29.900Z [INFO] TestServiceManager_RegisterSidecar: Stopping server: protocol=DNS address=127.0.0.1:30126 network=tcp writer.go:29: 2021-01-29T19:32:29.901Z [INFO] TestServiceManager_RegisterSidecar: Stopping server: protocol=DNS address=127.0.0.1:30126 network=udp writer.go:29: 2021-01-29T19:32:29.902Z [INFO] TestServiceManager_RegisterSidecar: Stopping server: protocol=HTTP address=127.0.0.1:30127 network=tcp writer.go:29: 2021-01-29T19:32:30.403Z [INFO] TestServiceManager_RegisterSidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:30.404Z [INFO] TestServiceManager_RegisterSidecar: Endpoints down --- PASS: TestServiceManager_RegisterSidecar (0.94s) === RUN TestServiceManager_RegisterMeshGateway writer.go:29: 2021-01-29T19:32:30.426Z [WARN] TestServiceManager_RegisterMeshGateway: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:30.428Z [DEBUG] TestServiceManager_RegisterMeshGateway.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:30.431Z [DEBUG] TestServiceManager_RegisterMeshGateway.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:30.438Z [INFO] TestServiceManager_RegisterMeshGateway.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:72bf3ff6-a603-8e04-d611-4ea8970a1999 Address:127.0.0.1:30137}]" writer.go:29: 2021-01-29T19:32:30.439Z [INFO] TestServiceManager_RegisterMeshGateway.server.raft: entering follower state: follower="Node at 127.0.0.1:30137 [Follower]" leader= writer.go:29: 2021-01-29T19:32:30.442Z [INFO] TestServiceManager_RegisterMeshGateway.server.serf.wan: serf: EventMemberJoin: Node-72bf3ff6-a603-8e04-d611-4ea8970a1999.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:30.450Z [INFO] TestServiceManager_RegisterMeshGateway.server.serf.lan: serf: EventMemberJoin: Node-72bf3ff6-a603-8e04-d611-4ea8970a1999 127.0.0.1 writer.go:29: 2021-01-29T19:32:30.454Z [INFO] TestServiceManager_RegisterMeshGateway.server: Adding LAN server: server="Node-72bf3ff6-a603-8e04-d611-4ea8970a1999 (Addr: tcp/127.0.0.1:30137) (DC: dc1)" writer.go:29: 2021-01-29T19:32:30.456Z [INFO] TestServiceManager_RegisterMeshGateway.server: Handled event for server in area: event=member-join server=Node-72bf3ff6-a603-8e04-d611-4ea8970a1999.dc1 area=wan writer.go:29: 2021-01-29T19:32:30.458Z [INFO] TestServiceManager_RegisterMeshGateway: Started DNS server: address=127.0.0.1:30132 network=tcp writer.go:29: 2021-01-29T19:32:30.460Z [INFO] TestServiceManager_RegisterMeshGateway: Started DNS server: address=127.0.0.1:30132 network=udp writer.go:29: 2021-01-29T19:32:30.461Z [INFO] TestServiceManager_RegisterMeshGateway: Started HTTP server: address=127.0.0.1:30133 network=tcp writer.go:29: 2021-01-29T19:32:30.463Z [INFO] TestServiceManager_RegisterMeshGateway: started state syncer writer.go:29: 2021-01-29T19:32:30.501Z [WARN] TestServiceManager_RegisterMeshGateway.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:30.502Z [INFO] TestServiceManager_RegisterMeshGateway.server.raft: entering candidate state: node="Node at 127.0.0.1:30137 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:30.515Z [DEBUG] TestServiceManager_RegisterMeshGateway.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:30.517Z [DEBUG] TestServiceManager_RegisterMeshGateway.server.raft: vote granted: from=72bf3ff6-a603-8e04-d611-4ea8970a1999 term=2 tally=1 writer.go:29: 2021-01-29T19:32:30.519Z [INFO] TestServiceManager_RegisterMeshGateway.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:30.520Z [INFO] TestServiceManager_RegisterMeshGateway.server.raft: entering leader state: leader="Node at 127.0.0.1:30137 [Leader]" writer.go:29: 2021-01-29T19:32:30.523Z [INFO] TestServiceManager_RegisterMeshGateway.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:30.523Z [INFO] TestServiceManager_RegisterMeshGateway.server: New leader elected: payload=Node-72bf3ff6-a603-8e04-d611-4ea8970a1999 writer.go:29: 2021-01-29T19:32:30.525Z [DEBUG] TestServiceManager_RegisterMeshGateway.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30137 writer.go:29: 2021-01-29T19:32:30.530Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:30.536Z [INFO] TestServiceManager_RegisterMeshGateway.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:30.536Z [INFO] TestServiceManager_RegisterMeshGateway.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:30.537Z [DEBUG] TestServiceManager_RegisterMeshGateway.server: Skipping self join check for node since the cluster is too small: node=Node-72bf3ff6-a603-8e04-d611-4ea8970a1999 writer.go:29: 2021-01-29T19:32:30.538Z [INFO] TestServiceManager_RegisterMeshGateway.server: member joined, marking health alive: member=Node-72bf3ff6-a603-8e04-d611-4ea8970a1999 writer.go:29: 2021-01-29T19:32:30.642Z [DEBUG] TestServiceManager_RegisterMeshGateway: added local registration for service: service=mesh-gateway writer.go:29: 2021-01-29T19:32:30.645Z [INFO] TestServiceManager_RegisterMeshGateway: Requesting shutdown writer.go:29: 2021-01-29T19:32:30.647Z [INFO] TestServiceManager_RegisterMeshGateway.server: shutting down server writer.go:29: 2021-01-29T19:32:30.647Z [DEBUG] TestServiceManager_RegisterMeshGateway.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:30.648Z [WARN] TestServiceManager_RegisterMeshGateway.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:30.650Z [ERROR] TestServiceManager_RegisterMeshGateway.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:30.651Z [DEBUG] TestServiceManager_RegisterMeshGateway.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:30.654Z [WARN] TestServiceManager_RegisterMeshGateway.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:30.657Z [INFO] TestServiceManager_RegisterMeshGateway.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:30.659Z [INFO] TestServiceManager_RegisterMeshGateway: consul server down writer.go:29: 2021-01-29T19:32:30.660Z [INFO] TestServiceManager_RegisterMeshGateway: shutdown complete writer.go:29: 2021-01-29T19:32:30.661Z [INFO] TestServiceManager_RegisterMeshGateway: Stopping server: protocol=DNS address=127.0.0.1:30132 network=tcp writer.go:29: 2021-01-29T19:32:30.663Z [INFO] TestServiceManager_RegisterMeshGateway: Stopping server: protocol=DNS address=127.0.0.1:30132 network=udp writer.go:29: 2021-01-29T19:32:30.664Z [INFO] TestServiceManager_RegisterMeshGateway: Stopping server: protocol=HTTP address=127.0.0.1:30133 network=tcp writer.go:29: 2021-01-29T19:32:31.166Z [INFO] TestServiceManager_RegisterMeshGateway: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:31.167Z [INFO] TestServiceManager_RegisterMeshGateway: Endpoints down --- PASS: TestServiceManager_RegisterMeshGateway (0.76s) === RUN TestServiceManager_PersistService_API service_manager_test.go:209: DM-disabled --- SKIP: TestServiceManager_PersistService_API (0.00s) === RUN TestServiceManager_PersistService_ConfigFiles === PAUSE TestServiceManager_PersistService_ConfigFiles === RUN TestServiceManager_Disabled writer.go:29: 2021-01-29T19:32:31.180Z [WARN] TestServiceManager_Disabled: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:31.181Z [DEBUG] TestServiceManager_Disabled.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:31.183Z [DEBUG] TestServiceManager_Disabled.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:31.187Z [INFO] TestServiceManager_Disabled.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a007c1c2-294b-bde3-8494-c82a3cd58eae Address:127.0.0.1:30143}]" writer.go:29: 2021-01-29T19:32:31.190Z [INFO] TestServiceManager_Disabled.server.raft: entering follower state: follower="Node at 127.0.0.1:30143 [Follower]" leader= writer.go:29: 2021-01-29T19:32:31.192Z [INFO] TestServiceManager_Disabled.server.serf.wan: serf: EventMemberJoin: Node-a007c1c2-294b-bde3-8494-c82a3cd58eae.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:31.195Z [INFO] TestServiceManager_Disabled.server.serf.lan: serf: EventMemberJoin: Node-a007c1c2-294b-bde3-8494-c82a3cd58eae 127.0.0.1 writer.go:29: 2021-01-29T19:32:31.198Z [INFO] TestServiceManager_Disabled.server: Adding LAN server: server="Node-a007c1c2-294b-bde3-8494-c82a3cd58eae (Addr: tcp/127.0.0.1:30143) (DC: dc1)" writer.go:29: 2021-01-29T19:32:31.199Z [INFO] TestServiceManager_Disabled.server: Handled event for server in area: event=member-join server=Node-a007c1c2-294b-bde3-8494-c82a3cd58eae.dc1 area=wan writer.go:29: 2021-01-29T19:32:31.202Z [INFO] TestServiceManager_Disabled: Started DNS server: address=127.0.0.1:30138 network=tcp writer.go:29: 2021-01-29T19:32:31.203Z [INFO] TestServiceManager_Disabled: Started DNS server: address=127.0.0.1:30138 network=udp writer.go:29: 2021-01-29T19:32:31.204Z [INFO] TestServiceManager_Disabled: Started HTTP server: address=127.0.0.1:30139 network=tcp writer.go:29: 2021-01-29T19:32:31.205Z [INFO] TestServiceManager_Disabled: started state syncer writer.go:29: 2021-01-29T19:32:31.231Z [WARN] TestServiceManager_Disabled.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:31.232Z [INFO] TestServiceManager_Disabled.server.raft: entering candidate state: node="Node at 127.0.0.1:30143 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:31.234Z [DEBUG] TestServiceManager_Disabled.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:31.235Z [DEBUG] TestServiceManager_Disabled.server.raft: vote granted: from=a007c1c2-294b-bde3-8494-c82a3cd58eae term=2 tally=1 writer.go:29: 2021-01-29T19:32:31.236Z [INFO] TestServiceManager_Disabled.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:31.237Z [INFO] TestServiceManager_Disabled.server.raft: entering leader state: leader="Node at 127.0.0.1:30143 [Leader]" writer.go:29: 2021-01-29T19:32:31.238Z [INFO] TestServiceManager_Disabled.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:31.239Z [INFO] TestServiceManager_Disabled.server: New leader elected: payload=Node-a007c1c2-294b-bde3-8494-c82a3cd58eae writer.go:29: 2021-01-29T19:32:31.239Z [DEBUG] TestServiceManager_Disabled.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30143 writer.go:29: 2021-01-29T19:32:31.244Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:31.272Z [INFO] TestServiceManager_Disabled.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:31.273Z [INFO] TestServiceManager_Disabled.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:31.274Z [DEBUG] TestServiceManager_Disabled.server: Skipping self join check for node since the cluster is too small: node=Node-a007c1c2-294b-bde3-8494-c82a3cd58eae writer.go:29: 2021-01-29T19:32:31.274Z [INFO] TestServiceManager_Disabled.server: member joined, marking health alive: member=Node-a007c1c2-294b-bde3-8494-c82a3cd58eae writer.go:29: 2021-01-29T19:32:31.306Z [INFO] TestServiceManager_Disabled: Requesting shutdown writer.go:29: 2021-01-29T19:32:31.307Z [INFO] TestServiceManager_Disabled.server: shutting down server writer.go:29: 2021-01-29T19:32:31.308Z [DEBUG] TestServiceManager_Disabled.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:31.310Z [WARN] TestServiceManager_Disabled.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:31.311Z [ERROR] TestServiceManager_Disabled.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:31.311Z [DEBUG] TestServiceManager_Disabled.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:31.313Z [WARN] TestServiceManager_Disabled.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:31.322Z [INFO] TestServiceManager_Disabled.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:31.322Z [INFO] TestServiceManager_Disabled: consul server down writer.go:29: 2021-01-29T19:32:31.323Z [INFO] TestServiceManager_Disabled: shutdown complete writer.go:29: 2021-01-29T19:32:31.324Z [INFO] TestServiceManager_Disabled: Stopping server: protocol=DNS address=127.0.0.1:30138 network=tcp writer.go:29: 2021-01-29T19:32:31.325Z [INFO] TestServiceManager_Disabled: Stopping server: protocol=DNS address=127.0.0.1:30138 network=udp writer.go:29: 2021-01-29T19:32:31.326Z [INFO] TestServiceManager_Disabled: Stopping server: protocol=HTTP address=127.0.0.1:30139 network=tcp writer.go:29: 2021-01-29T19:32:31.827Z [INFO] TestServiceManager_Disabled: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:31.829Z [INFO] TestServiceManager_Disabled: Endpoints down --- PASS: TestServiceManager_Disabled (0.66s) === RUN TestSessionCreate === PAUSE TestSessionCreate === RUN TestSessionCreate_NodeChecks === PAUSE TestSessionCreate_NodeChecks === RUN TestSessionCreate_Delete === PAUSE TestSessionCreate_Delete === RUN TestSessionCreate_DefaultCheck === PAUSE TestSessionCreate_DefaultCheck === RUN TestSessionCreate_NoCheck === PAUSE TestSessionCreate_NoCheck === RUN TestSessionDestroy === PAUSE TestSessionDestroy === RUN TestSessionCustomTTL === PAUSE TestSessionCustomTTL === RUN TestSessionTTLRenew session_endpoint_test.go:496: DM-skipped --- SKIP: TestSessionTTLRenew (0.00s) === RUN TestSessionGet === PAUSE TestSessionGet === RUN TestSessionList === RUN TestSessionList/#00 writer.go:29: 2021-01-29T19:32:31.874Z [WARN] TestSessionList/#00: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:31.876Z [DEBUG] TestSessionList/#00.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:31.878Z [DEBUG] TestSessionList/#00.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:31.883Z [INFO] TestSessionList/#00.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c35d605c-28b2-4eaf-b3c7-de49991172bf Address:127.0.0.1:30149}]" writer.go:29: 2021-01-29T19:32:31.887Z [INFO] TestSessionList/#00.server.raft: entering follower state: follower="Node at 127.0.0.1:30149 [Follower]" leader= writer.go:29: 2021-01-29T19:32:31.887Z [INFO] TestSessionList/#00.server.serf.wan: serf: EventMemberJoin: Node-c35d605c-28b2-4eaf-b3c7-de49991172bf.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:31.892Z [INFO] TestSessionList/#00.server.serf.lan: serf: EventMemberJoin: Node-c35d605c-28b2-4eaf-b3c7-de49991172bf 127.0.0.1 writer.go:29: 2021-01-29T19:32:31.894Z [INFO] TestSessionList/#00.server: Handled event for server in area: event=member-join server=Node-c35d605c-28b2-4eaf-b3c7-de49991172bf.dc1 area=wan writer.go:29: 2021-01-29T19:32:31.895Z [INFO] TestSessionList/#00.server: Adding LAN server: server="Node-c35d605c-28b2-4eaf-b3c7-de49991172bf (Addr: tcp/127.0.0.1:30149) (DC: dc1)" writer.go:29: 2021-01-29T19:32:31.895Z [INFO] TestSessionList/#00: Started DNS server: address=127.0.0.1:30144 network=udp writer.go:29: 2021-01-29T19:32:31.898Z [INFO] TestSessionList/#00: Started DNS server: address=127.0.0.1:30144 network=tcp writer.go:29: 2021-01-29T19:32:31.899Z [INFO] TestSessionList/#00: Started HTTP server: address=127.0.0.1:30145 network=tcp writer.go:29: 2021-01-29T19:32:31.902Z [INFO] TestSessionList/#00: started state syncer writer.go:29: 2021-01-29T19:32:31.937Z [WARN] TestSessionList/#00.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:31.939Z [INFO] TestSessionList/#00.server.raft: entering candidate state: node="Node at 127.0.0.1:30149 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:31.942Z [DEBUG] TestSessionList/#00.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:31.944Z [DEBUG] TestSessionList/#00.server.raft: vote granted: from=c35d605c-28b2-4eaf-b3c7-de49991172bf term=2 tally=1 writer.go:29: 2021-01-29T19:32:31.946Z [INFO] TestSessionList/#00.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:31.948Z [INFO] TestSessionList/#00.server.raft: entering leader state: leader="Node at 127.0.0.1:30149 [Leader]" writer.go:29: 2021-01-29T19:32:31.949Z [INFO] TestSessionList/#00.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:31.950Z [INFO] TestSessionList/#00.server: New leader elected: payload=Node-c35d605c-28b2-4eaf-b3c7-de49991172bf writer.go:29: 2021-01-29T19:32:31.951Z [DEBUG] TestSessionList/#00.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30149 writer.go:29: 2021-01-29T19:32:31.965Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:31.971Z [INFO] TestSessionList/#00.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:31.972Z [INFO] TestSessionList/#00.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:31.973Z [DEBUG] TestSessionList/#00.server: Skipping self join check for node since the cluster is too small: node=Node-c35d605c-28b2-4eaf-b3c7-de49991172bf writer.go:29: 2021-01-29T19:32:31.974Z [INFO] TestSessionList/#00.server: member joined, marking health alive: member=Node-c35d605c-28b2-4eaf-b3c7-de49991172bf writer.go:29: 2021-01-29T19:32:32.211Z [INFO] TestSessionList/#00: Requesting shutdown writer.go:29: 2021-01-29T19:32:32.212Z [INFO] TestSessionList/#00.server: shutting down server writer.go:29: 2021-01-29T19:32:32.214Z [DEBUG] TestSessionList/#00.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:32.216Z [WARN] TestSessionList/#00.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:32.218Z [ERROR] TestSessionList/#00.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:32.220Z [DEBUG] TestSessionList/#00.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:32.222Z [WARN] TestSessionList/#00.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:32.226Z [INFO] TestSessionList/#00.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:32.226Z [INFO] TestSessionList/#00: consul server down writer.go:29: 2021-01-29T19:32:32.229Z [INFO] TestSessionList/#00: shutdown complete writer.go:29: 2021-01-29T19:32:32.230Z [INFO] TestSessionList/#00: Stopping server: protocol=DNS address=127.0.0.1:30144 network=tcp writer.go:29: 2021-01-29T19:32:32.231Z [INFO] TestSessionList/#00: Stopping server: protocol=DNS address=127.0.0.1:30144 network=udp writer.go:29: 2021-01-29T19:32:32.233Z [INFO] TestSessionList/#00: Stopping server: protocol=HTTP address=127.0.0.1:30145 network=tcp writer.go:29: 2021-01-29T19:32:32.734Z [INFO] TestSessionList/#00: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:32.738Z [INFO] TestSessionList/#00: Endpoints down === RUN TestSessionList/#01 writer.go:29: 2021-01-29T19:32:32.749Z [WARN] TestSessionList/#01: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:32.751Z [DEBUG] TestSessionList/#01.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:32.764Z [DEBUG] TestSessionList/#01.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:32.767Z [INFO] TestSessionList/#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2cc265f4-f81a-b711-c21d-78812eeec5bb Address:127.0.0.1:30155}]" writer.go:29: 2021-01-29T19:32:32.769Z [INFO] TestSessionList/#01.server.raft: entering follower state: follower="Node at 127.0.0.1:30155 [Follower]" leader= writer.go:29: 2021-01-29T19:32:32.769Z [INFO] TestSessionList/#01.server.serf.wan: serf: EventMemberJoin: Node-2cc265f4-f81a-b711-c21d-78812eeec5bb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:32.775Z [INFO] TestSessionList/#01.server.serf.lan: serf: EventMemberJoin: Node-2cc265f4-f81a-b711-c21d-78812eeec5bb 127.0.0.1 writer.go:29: 2021-01-29T19:32:32.778Z [INFO] TestSessionList/#01.server: Adding LAN server: server="Node-2cc265f4-f81a-b711-c21d-78812eeec5bb (Addr: tcp/127.0.0.1:30155) (DC: dc1)" writer.go:29: 2021-01-29T19:32:32.778Z [INFO] TestSessionList/#01.server: Handled event for server in area: event=member-join server=Node-2cc265f4-f81a-b711-c21d-78812eeec5bb.dc1 area=wan writer.go:29: 2021-01-29T19:32:32.779Z [INFO] TestSessionList/#01: Started DNS server: address=127.0.0.1:30150 network=udp writer.go:29: 2021-01-29T19:32:32.782Z [INFO] TestSessionList/#01: Started DNS server: address=127.0.0.1:30150 network=tcp writer.go:29: 2021-01-29T19:32:32.783Z [INFO] TestSessionList/#01: Started HTTP server: address=127.0.0.1:30151 network=tcp writer.go:29: 2021-01-29T19:32:32.785Z [INFO] TestSessionList/#01: started state syncer writer.go:29: 2021-01-29T19:32:32.837Z [WARN] TestSessionList/#01.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:32.839Z [INFO] TestSessionList/#01.server.raft: entering candidate state: node="Node at 127.0.0.1:30155 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:32.841Z [DEBUG] TestSessionList/#01.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:32.846Z [DEBUG] TestSessionList/#01.server.raft: vote granted: from=2cc265f4-f81a-b711-c21d-78812eeec5bb term=2 tally=1 writer.go:29: 2021-01-29T19:32:32.847Z [INFO] TestSessionList/#01.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:32.849Z [INFO] TestSessionList/#01.server.raft: entering leader state: leader="Node at 127.0.0.1:30155 [Leader]" writer.go:29: 2021-01-29T19:32:32.853Z [INFO] TestSessionList/#01.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:32.854Z [INFO] TestSessionList/#01.server: New leader elected: payload=Node-2cc265f4-f81a-b711-c21d-78812eeec5bb writer.go:29: 2021-01-29T19:32:32.859Z [DEBUG] TestSessionList/#01.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30155 writer.go:29: 2021-01-29T19:32:32.885Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:32.890Z [INFO] TestSessionList/#01.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:32.892Z [INFO] TestSessionList/#01.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:32.894Z [DEBUG] TestSessionList/#01.server: Skipping self join check for node since the cluster is too small: node=Node-2cc265f4-f81a-b711-c21d-78812eeec5bb writer.go:29: 2021-01-29T19:32:32.896Z [INFO] TestSessionList/#01.server: member joined, marking health alive: member=Node-2cc265f4-f81a-b711-c21d-78812eeec5bb writer.go:29: 2021-01-29T19:32:32.917Z [DEBUG] TestSessionList/#01: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:32.920Z [INFO] TestSessionList/#01: Synced node info writer.go:29: 2021-01-29T19:32:32.922Z [DEBUG] TestSessionList/#01: Node info in sync writer.go:29: 2021-01-29T19:32:33.056Z [INFO] TestSessionList/#01: Requesting shutdown writer.go:29: 2021-01-29T19:32:33.058Z [INFO] TestSessionList/#01.server: shutting down server writer.go:29: 2021-01-29T19:32:33.059Z [DEBUG] TestSessionList/#01.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:33.061Z [WARN] TestSessionList/#01.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:33.062Z [DEBUG] TestSessionList/#01.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:33.066Z [WARN] TestSessionList/#01.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:33.068Z [INFO] TestSessionList/#01.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:33.068Z [INFO] TestSessionList/#01: consul server down writer.go:29: 2021-01-29T19:32:33.073Z [INFO] TestSessionList/#01: shutdown complete writer.go:29: 2021-01-29T19:32:33.076Z [INFO] TestSessionList/#01: Stopping server: protocol=DNS address=127.0.0.1:30150 network=tcp writer.go:29: 2021-01-29T19:32:33.084Z [INFO] TestSessionList/#01: Stopping server: protocol=DNS address=127.0.0.1:30150 network=udp writer.go:29: 2021-01-29T19:32:33.085Z [INFO] TestSessionList/#01: Stopping server: protocol=HTTP address=127.0.0.1:30151 network=tcp writer.go:29: 2021-01-29T19:32:33.586Z [INFO] TestSessionList/#01: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:33.587Z [INFO] TestSessionList/#01: Endpoints down --- PASS: TestSessionList (1.73s) --- PASS: TestSessionList/#00 (0.88s) --- PASS: TestSessionList/#01 (0.85s) === RUN TestSessionsForNode session_endpoint_test.go:678: DM-skipped --- SKIP: TestSessionsForNode (0.00s) === RUN TestSessionDeleteDestroy === PAUSE TestSessionDeleteDestroy === RUN TestAgent_sidecarServiceFromNodeService === RUN TestAgent_sidecarServiceFromNodeService/no_sidecar writer.go:29: 2021-01-29T19:32:33.610Z [WARN] jones: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:33.611Z [DEBUG] jones.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:33.613Z [DEBUG] jones.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:33.620Z [INFO] jones.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d81a87e6-c1e4-f4b2-d0f5-c22a74b95fb8 Address:127.0.0.1:30161}]" writer.go:29: 2021-01-29T19:32:33.622Z [INFO] jones.server.raft: entering follower state: follower="Node at 127.0.0.1:30161 [Follower]" leader= writer.go:29: 2021-01-29T19:32:33.623Z [INFO] jones.server.serf.wan: serf: EventMemberJoin: Node-d81a87e6-c1e4-f4b2-d0f5-c22a74b95fb8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:33.626Z [INFO] jones.server.serf.lan: serf: EventMemberJoin: Node-d81a87e6-c1e4-f4b2-d0f5-c22a74b95fb8 127.0.0.1 writer.go:29: 2021-01-29T19:32:33.628Z [INFO] jones.server: Handled event for server in area: event=member-join server=Node-d81a87e6-c1e4-f4b2-d0f5-c22a74b95fb8.dc1 area=wan writer.go:29: 2021-01-29T19:32:33.628Z [INFO] jones.server: Adding LAN server: server="Node-d81a87e6-c1e4-f4b2-d0f5-c22a74b95fb8 (Addr: tcp/127.0.0.1:30161) (DC: dc1)" writer.go:29: 2021-01-29T19:32:33.629Z [INFO] jones: Started DNS server: address=127.0.0.1:30156 network=udp writer.go:29: 2021-01-29T19:32:33.631Z [INFO] jones: Started DNS server: address=127.0.0.1:30156 network=tcp writer.go:29: 2021-01-29T19:32:33.633Z [INFO] jones: Started HTTP server: address=127.0.0.1:30157 network=tcp writer.go:29: 2021-01-29T19:32:33.634Z [INFO] jones: started state syncer writer.go:29: 2021-01-29T19:32:33.674Z [WARN] jones.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:33.675Z [INFO] jones.server.raft: entering candidate state: node="Node at 127.0.0.1:30161 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:33.677Z [DEBUG] jones.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:33.677Z [DEBUG] jones.server.raft: vote granted: from=d81a87e6-c1e4-f4b2-d0f5-c22a74b95fb8 term=2 tally=1 writer.go:29: 2021-01-29T19:32:33.678Z [INFO] jones.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:33.679Z [INFO] jones.server.raft: entering leader state: leader="Node at 127.0.0.1:30161 [Leader]" writer.go:29: 2021-01-29T19:32:33.680Z [INFO] jones.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:33.680Z [INFO] jones.server: New leader elected: payload=Node-d81a87e6-c1e4-f4b2-d0f5-c22a74b95fb8 writer.go:29: 2021-01-29T19:32:33.681Z [DEBUG] jones.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30161 writer.go:29: 2021-01-29T19:32:33.689Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:33.704Z [INFO] jones.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:33.706Z [INFO] jones.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:33.708Z [DEBUG] jones.server: Skipping self join check for node since the cluster is too small: node=Node-d81a87e6-c1e4-f4b2-d0f5-c22a74b95fb8 writer.go:29: 2021-01-29T19:32:33.710Z [INFO] jones.server: member joined, marking health alive: member=Node-d81a87e6-c1e4-f4b2-d0f5-c22a74b95fb8 writer.go:29: 2021-01-29T19:32:33.959Z [INFO] jones: Requesting shutdown writer.go:29: 2021-01-29T19:32:33.960Z [INFO] jones.server: shutting down server writer.go:29: 2021-01-29T19:32:33.961Z [DEBUG] jones.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:33.962Z [WARN] jones.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:33.961Z [ERROR] jones.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:33.962Z [DEBUG] jones.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:33.964Z [WARN] jones.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:33.969Z [INFO] jones.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:33.969Z [INFO] jones: consul server down writer.go:29: 2021-01-29T19:32:33.971Z [INFO] jones: shutdown complete writer.go:29: 2021-01-29T19:32:33.971Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30156 network=tcp writer.go:29: 2021-01-29T19:32:33.972Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30156 network=udp writer.go:29: 2021-01-29T19:32:33.974Z [INFO] jones: Stopping server: protocol=HTTP address=127.0.0.1:30157 network=tcp writer.go:29: 2021-01-29T19:32:34.475Z [INFO] jones: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:34.477Z [INFO] jones: Endpoints down === RUN TestAgent_sidecarServiceFromNodeService/all_the_defaults writer.go:29: 2021-01-29T19:32:34.516Z [WARN] jones: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:34.520Z [DEBUG] jones.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:34.525Z [DEBUG] jones.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:34.533Z [INFO] jones.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:63bb7d54-8423-876a-4ca1-075775b2be1f Address:127.0.0.1:30167}]" writer.go:29: 2021-01-29T19:32:34.536Z [INFO] jones.server.raft: entering follower state: follower="Node at 127.0.0.1:30167 [Follower]" leader= writer.go:29: 2021-01-29T19:32:34.539Z [INFO] jones.server.serf.wan: serf: EventMemberJoin: Node-63bb7d54-8423-876a-4ca1-075775b2be1f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:34.548Z [INFO] jones.server.serf.lan: serf: EventMemberJoin: Node-63bb7d54-8423-876a-4ca1-075775b2be1f 127.0.0.1 writer.go:29: 2021-01-29T19:32:34.564Z [INFO] jones.server: Adding LAN server: server="Node-63bb7d54-8423-876a-4ca1-075775b2be1f (Addr: tcp/127.0.0.1:30167) (DC: dc1)" writer.go:29: 2021-01-29T19:32:34.566Z [INFO] jones.server: Handled event for server in area: event=member-join server=Node-63bb7d54-8423-876a-4ca1-075775b2be1f.dc1 area=wan writer.go:29: 2021-01-29T19:32:34.569Z [INFO] jones: Started DNS server: address=127.0.0.1:30162 network=tcp writer.go:29: 2021-01-29T19:32:34.571Z [INFO] jones: Started DNS server: address=127.0.0.1:30162 network=udp writer.go:29: 2021-01-29T19:32:34.575Z [INFO] jones: Started HTTP server: address=127.0.0.1:30163 network=tcp writer.go:29: 2021-01-29T19:32:34.579Z [INFO] jones: started state syncer writer.go:29: 2021-01-29T19:32:34.598Z [WARN] jones.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:34.600Z [INFO] jones.server.raft: entering candidate state: node="Node at 127.0.0.1:30167 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:34.603Z [DEBUG] jones.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:34.604Z [DEBUG] jones.server.raft: vote granted: from=63bb7d54-8423-876a-4ca1-075775b2be1f term=2 tally=1 writer.go:29: 2021-01-29T19:32:34.606Z [INFO] jones.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:34.608Z [INFO] jones.server.raft: entering leader state: leader="Node at 127.0.0.1:30167 [Leader]" writer.go:29: 2021-01-29T19:32:34.610Z [INFO] jones.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:34.610Z [INFO] jones.server: New leader elected: payload=Node-63bb7d54-8423-876a-4ca1-075775b2be1f writer.go:29: 2021-01-29T19:32:34.622Z [DEBUG] jones.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30167 writer.go:29: 2021-01-29T19:32:34.633Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:34.638Z [INFO] jones.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:34.639Z [INFO] jones.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:34.640Z [DEBUG] jones.server: Skipping self join check for node since the cluster is too small: node=Node-63bb7d54-8423-876a-4ca1-075775b2be1f writer.go:29: 2021-01-29T19:32:34.641Z [INFO] jones.server: member joined, marking health alive: member=Node-63bb7d54-8423-876a-4ca1-075775b2be1f writer.go:29: 2021-01-29T19:32:34.841Z [DEBUG] jones: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:34.844Z [INFO] jones: Synced node info writer.go:29: 2021-01-29T19:32:34.882Z [INFO] jones: Requesting shutdown writer.go:29: 2021-01-29T19:32:34.883Z [INFO] jones.server: shutting down server writer.go:29: 2021-01-29T19:32:34.884Z [DEBUG] jones.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:34.885Z [WARN] jones.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:34.885Z [DEBUG] jones.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:34.887Z [WARN] jones.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:34.890Z [INFO] jones.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:34.891Z [INFO] jones: consul server down writer.go:29: 2021-01-29T19:32:34.893Z [INFO] jones: shutdown complete writer.go:29: 2021-01-29T19:32:34.894Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30162 network=tcp writer.go:29: 2021-01-29T19:32:34.896Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30162 network=udp writer.go:29: 2021-01-29T19:32:34.897Z [INFO] jones: Stopping server: protocol=HTTP address=127.0.0.1:30163 network=tcp writer.go:29: 2021-01-29T19:32:35.399Z [INFO] jones: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:35.400Z [INFO] jones: Endpoints down === RUN TestAgent_sidecarServiceFromNodeService/all_the_allowed_overrides writer.go:29: 2021-01-29T19:32:35.587Z [WARN] jones: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:35.594Z [DEBUG] jones.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:35.604Z [DEBUG] jones.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:35.634Z [INFO] jones.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:16196f2a-374e-2c4e-5707-3b2a296eb146 Address:127.0.0.1:30173}]" writer.go:29: 2021-01-29T19:32:35.640Z [INFO] jones.server.raft: entering follower state: follower="Node at 127.0.0.1:30173 [Follower]" leader= writer.go:29: 2021-01-29T19:32:35.661Z [INFO] jones.server.serf.wan: serf: EventMemberJoin: Node-16196f2a-374e-2c4e-5707-3b2a296eb146.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:35.680Z [INFO] jones.server.serf.lan: serf: EventMemberJoin: Node-16196f2a-374e-2c4e-5707-3b2a296eb146 127.0.0.1 writer.go:29: 2021-01-29T19:32:35.712Z [WARN] jones.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:35.738Z [INFO] jones.server.raft: entering candidate state: node="Node at 127.0.0.1:30173 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:35.758Z [INFO] jones.server: Adding LAN server: server="Node-16196f2a-374e-2c4e-5707-3b2a296eb146 (Addr: tcp/127.0.0.1:30173) (DC: dc1)" writer.go:29: 2021-01-29T19:32:35.761Z [INFO] jones: Started DNS server: address=127.0.0.1:30168 network=udp writer.go:29: 2021-01-29T19:32:35.762Z [INFO] jones: Started DNS server: address=127.0.0.1:30168 network=tcp writer.go:29: 2021-01-29T19:32:35.762Z [DEBUG] jones.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:35.765Z [DEBUG] jones.server.raft: vote granted: from=16196f2a-374e-2c4e-5707-3b2a296eb146 term=2 tally=1 writer.go:29: 2021-01-29T19:32:35.767Z [INFO] jones.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:35.766Z [INFO] jones: Started HTTP server: address=127.0.0.1:30169 network=tcp writer.go:29: 2021-01-29T19:32:35.766Z [INFO] jones.server: Handled event for server in area: event=member-join server=Node-16196f2a-374e-2c4e-5707-3b2a296eb146.dc1 area=wan writer.go:29: 2021-01-29T19:32:35.768Z [INFO] jones.server.raft: entering leader state: leader="Node at 127.0.0.1:30173 [Leader]" writer.go:29: 2021-01-29T19:32:35.770Z [INFO] jones: started state syncer writer.go:29: 2021-01-29T19:32:35.775Z [INFO] jones.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:35.776Z [INFO] jones: Synced node info writer.go:29: 2021-01-29T19:32:35.776Z [INFO] jones.server: New leader elected: payload=Node-16196f2a-374e-2c4e-5707-3b2a296eb146 writer.go:29: 2021-01-29T19:32:35.778Z [DEBUG] jones.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30173 writer.go:29: 2021-01-29T19:32:35.785Z [INFO] jones: Requesting shutdown writer.go:29: 2021-01-29T19:32:35.786Z [INFO] jones.server: shutting down server writer.go:29: 2021-01-29T19:32:35.787Z [WARN] jones.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:35.792Z [WARN] jones.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:35.794Z [INFO] jones.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:35.802Z [ERROR] jones.server: failed to establish leadership: error="leadership lost while committing log" writer.go:29: 2021-01-29T19:32:35.802Z [INFO] jones: consul server down writer.go:29: 2021-01-29T19:32:35.804Z [INFO] jones: shutdown complete writer.go:29: 2021-01-29T19:32:35.804Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30168 network=tcp writer.go:29: 2021-01-29T19:32:35.805Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30168 network=udp writer.go:29: 2021-01-29T19:32:35.806Z [INFO] jones: Stopping server: protocol=HTTP address=127.0.0.1:30169 network=tcp writer.go:29: 2021-01-29T19:32:35.807Z [ERROR] jones.server: failed to transfer leadership attempt, will retry: attempt=0 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:32:35.808Z [ERROR] jones.server: failed to transfer leadership attempt, will retry: attempt=1 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:32:36.307Z [INFO] jones: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:36.309Z [INFO] jones: Endpoints down === RUN TestAgent_sidecarServiceFromNodeService/no_auto_ports_available writer.go:29: 2021-01-29T19:32:36.329Z [WARN] jones: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:36.332Z [DEBUG] jones.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:36.335Z [DEBUG] jones.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:36.343Z [INFO] jones.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8e81e012-4c2d-476d-d425-f5ebdd0e7897 Address:127.0.0.1:30179}]" writer.go:29: 2021-01-29T19:32:36.345Z [INFO] jones.server.raft: entering follower state: follower="Node at 127.0.0.1:30179 [Follower]" leader= writer.go:29: 2021-01-29T19:32:36.348Z [INFO] jones.server.serf.wan: serf: EventMemberJoin: Node-8e81e012-4c2d-476d-d425-f5ebdd0e7897.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:36.363Z [INFO] jones.server.serf.lan: serf: EventMemberJoin: Node-8e81e012-4c2d-476d-d425-f5ebdd0e7897 127.0.0.1 writer.go:29: 2021-01-29T19:32:36.365Z [INFO] jones.server: Handled event for server in area: event=member-join server=Node-8e81e012-4c2d-476d-d425-f5ebdd0e7897.dc1 area=wan writer.go:29: 2021-01-29T19:32:36.365Z [INFO] jones.server: Adding LAN server: server="Node-8e81e012-4c2d-476d-d425-f5ebdd0e7897 (Addr: tcp/127.0.0.1:30179) (DC: dc1)" writer.go:29: 2021-01-29T19:32:36.369Z [INFO] jones: Started DNS server: address=127.0.0.1:30174 network=tcp writer.go:29: 2021-01-29T19:32:36.370Z [INFO] jones: Started DNS server: address=127.0.0.1:30174 network=udp writer.go:29: 2021-01-29T19:32:36.372Z [INFO] jones: Started HTTP server: address=127.0.0.1:30175 network=tcp writer.go:29: 2021-01-29T19:32:36.373Z [INFO] jones: started state syncer writer.go:29: 2021-01-29T19:32:36.408Z [WARN] jones.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:36.409Z [INFO] jones.server.raft: entering candidate state: node="Node at 127.0.0.1:30179 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:36.423Z [DEBUG] jones.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:36.424Z [DEBUG] jones.server.raft: vote granted: from=8e81e012-4c2d-476d-d425-f5ebdd0e7897 term=2 tally=1 writer.go:29: 2021-01-29T19:32:36.425Z [INFO] jones.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:36.425Z [INFO] jones.server.raft: entering leader state: leader="Node at 127.0.0.1:30179 [Leader]" writer.go:29: 2021-01-29T19:32:36.427Z [INFO] jones.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:36.428Z [INFO] jones.server: New leader elected: payload=Node-8e81e012-4c2d-476d-d425-f5ebdd0e7897 writer.go:29: 2021-01-29T19:32:36.429Z [DEBUG] jones.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30179 writer.go:29: 2021-01-29T19:32:36.445Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:36.455Z [INFO] jones.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:36.458Z [INFO] jones.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:36.460Z [DEBUG] jones.server: Skipping self join check for node since the cluster is too small: node=Node-8e81e012-4c2d-476d-d425-f5ebdd0e7897 writer.go:29: 2021-01-29T19:32:36.464Z [INFO] jones.server: member joined, marking health alive: member=Node-8e81e012-4c2d-476d-d425-f5ebdd0e7897 writer.go:29: 2021-01-29T19:32:36.597Z [DEBUG] jones: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:36.599Z [INFO] jones: Synced node info writer.go:29: 2021-01-29T19:32:36.805Z [INFO] jones: Requesting shutdown writer.go:29: 2021-01-29T19:32:36.811Z [ERROR] jones.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:32:36.811Z [INFO] jones.server: shutting down server writer.go:29: 2021-01-29T19:32:36.812Z [DEBUG] jones.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:36.813Z [WARN] jones.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:36.813Z [DEBUG] jones.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:36.817Z [WARN] jones.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:36.822Z [INFO] jones.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:36.823Z [INFO] jones: consul server down writer.go:29: 2021-01-29T19:32:36.824Z [INFO] jones: shutdown complete writer.go:29: 2021-01-29T19:32:36.825Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30174 network=tcp writer.go:29: 2021-01-29T19:32:36.826Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30174 network=udp writer.go:29: 2021-01-29T19:32:36.827Z [INFO] jones: Stopping server: protocol=HTTP address=127.0.0.1:30175 network=tcp writer.go:29: 2021-01-29T19:32:37.328Z [INFO] jones: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:37.330Z [INFO] jones: Endpoints down === RUN TestAgent_sidecarServiceFromNodeService/auto_ports_disabled writer.go:29: 2021-01-29T19:32:37.352Z [WARN] jones: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:37.354Z [DEBUG] jones.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:37.357Z [DEBUG] jones.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:37.364Z [INFO] jones.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e5d19edc-30c5-0196-c44b-7608b2d97cca Address:127.0.0.1:30185}]" writer.go:29: 2021-01-29T19:32:37.367Z [INFO] jones.server.raft: entering follower state: follower="Node at 127.0.0.1:30185 [Follower]" leader= writer.go:29: 2021-01-29T19:32:37.368Z [INFO] jones.server.serf.wan: serf: EventMemberJoin: Node-e5d19edc-30c5-0196-c44b-7608b2d97cca.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:37.371Z [INFO] jones.server.serf.lan: serf: EventMemberJoin: Node-e5d19edc-30c5-0196-c44b-7608b2d97cca 127.0.0.1 writer.go:29: 2021-01-29T19:32:37.374Z [INFO] jones.server: Handled event for server in area: event=member-join server=Node-e5d19edc-30c5-0196-c44b-7608b2d97cca.dc1 area=wan writer.go:29: 2021-01-29T19:32:37.374Z [INFO] jones.server: Adding LAN server: server="Node-e5d19edc-30c5-0196-c44b-7608b2d97cca (Addr: tcp/127.0.0.1:30185) (DC: dc1)" writer.go:29: 2021-01-29T19:32:37.376Z [INFO] jones: Started DNS server: address=127.0.0.1:30180 network=udp writer.go:29: 2021-01-29T19:32:37.378Z [INFO] jones: Started DNS server: address=127.0.0.1:30180 network=tcp writer.go:29: 2021-01-29T19:32:37.380Z [INFO] jones: Started HTTP server: address=127.0.0.1:30181 network=tcp writer.go:29: 2021-01-29T19:32:37.382Z [INFO] jones: started state syncer writer.go:29: 2021-01-29T19:32:37.414Z [WARN] jones.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:37.415Z [INFO] jones.server.raft: entering candidate state: node="Node at 127.0.0.1:30185 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:37.418Z [DEBUG] jones.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:37.420Z [DEBUG] jones.server.raft: vote granted: from=e5d19edc-30c5-0196-c44b-7608b2d97cca term=2 tally=1 writer.go:29: 2021-01-29T19:32:37.422Z [INFO] jones.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:37.423Z [INFO] jones.server.raft: entering leader state: leader="Node at 127.0.0.1:30185 [Leader]" writer.go:29: 2021-01-29T19:32:37.425Z [INFO] jones.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:37.427Z [INFO] jones.server: New leader elected: payload=Node-e5d19edc-30c5-0196-c44b-7608b2d97cca writer.go:29: 2021-01-29T19:32:37.431Z [DEBUG] jones.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30185 writer.go:29: 2021-01-29T19:32:37.437Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:37.442Z [INFO] jones.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:37.444Z [INFO] jones.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:37.445Z [DEBUG] jones.server: Skipping self join check for node since the cluster is too small: node=Node-e5d19edc-30c5-0196-c44b-7608b2d97cca writer.go:29: 2021-01-29T19:32:37.447Z [INFO] jones.server: member joined, marking health alive: member=Node-e5d19edc-30c5-0196-c44b-7608b2d97cca writer.go:29: 2021-01-29T19:32:37.543Z [INFO] jones: Requesting shutdown writer.go:29: 2021-01-29T19:32:37.544Z [INFO] jones.server: shutting down server writer.go:29: 2021-01-29T19:32:37.545Z [DEBUG] jones.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:37.546Z [WARN] jones.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:37.545Z [ERROR] jones.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:37.546Z [DEBUG] jones.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:37.548Z [WARN] jones.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:37.551Z [INFO] jones.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:37.551Z [INFO] jones: consul server down writer.go:29: 2021-01-29T19:32:37.554Z [INFO] jones: shutdown complete writer.go:29: 2021-01-29T19:32:37.556Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30180 network=tcp writer.go:29: 2021-01-29T19:32:37.557Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30180 network=udp writer.go:29: 2021-01-29T19:32:37.559Z [INFO] jones: Stopping server: protocol=HTTP address=127.0.0.1:30181 network=tcp writer.go:29: 2021-01-29T19:32:38.061Z [INFO] jones: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:38.063Z [INFO] jones: Endpoints down === RUN TestAgent_sidecarServiceFromNodeService/inherit_tags_and_meta writer.go:29: 2021-01-29T19:32:38.083Z [WARN] jones: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:38.084Z [DEBUG] jones.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:38.085Z [DEBUG] jones.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:38.091Z [INFO] jones.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ab7adc75-15b0-f29d-8549-68fb916fd075 Address:127.0.0.1:30191}]" writer.go:29: 2021-01-29T19:32:38.093Z [INFO] jones.server.raft: entering follower state: follower="Node at 127.0.0.1:30191 [Follower]" leader= writer.go:29: 2021-01-29T19:32:38.094Z [INFO] jones.server.serf.wan: serf: EventMemberJoin: Node-ab7adc75-15b0-f29d-8549-68fb916fd075.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:38.102Z [INFO] jones.server.serf.lan: serf: EventMemberJoin: Node-ab7adc75-15b0-f29d-8549-68fb916fd075 127.0.0.1 writer.go:29: 2021-01-29T19:32:38.105Z [INFO] jones.server: Adding LAN server: server="Node-ab7adc75-15b0-f29d-8549-68fb916fd075 (Addr: tcp/127.0.0.1:30191) (DC: dc1)" writer.go:29: 2021-01-29T19:32:38.107Z [INFO] jones: Started DNS server: address=127.0.0.1:30186 network=udp writer.go:29: 2021-01-29T19:32:38.108Z [INFO] jones: Started DNS server: address=127.0.0.1:30186 network=tcp writer.go:29: 2021-01-29T19:32:38.107Z [INFO] jones.server: Handled event for server in area: event=member-join server=Node-ab7adc75-15b0-f29d-8549-68fb916fd075.dc1 area=wan writer.go:29: 2021-01-29T19:32:38.111Z [INFO] jones: Started HTTP server: address=127.0.0.1:30187 network=tcp writer.go:29: 2021-01-29T19:32:38.112Z [INFO] jones: started state syncer writer.go:29: 2021-01-29T19:32:38.144Z [WARN] jones.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:38.145Z [INFO] jones.server.raft: entering candidate state: node="Node at 127.0.0.1:30191 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:38.148Z [DEBUG] jones.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:38.149Z [DEBUG] jones.server.raft: vote granted: from=ab7adc75-15b0-f29d-8549-68fb916fd075 term=2 tally=1 writer.go:29: 2021-01-29T19:32:38.150Z [INFO] jones.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:38.151Z [INFO] jones.server.raft: entering leader state: leader="Node at 127.0.0.1:30191 [Leader]" writer.go:29: 2021-01-29T19:32:38.152Z [INFO] jones.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:38.153Z [INFO] jones.server: New leader elected: payload=Node-ab7adc75-15b0-f29d-8549-68fb916fd075 writer.go:29: 2021-01-29T19:32:38.155Z [DEBUG] jones.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30191 writer.go:29: 2021-01-29T19:32:38.160Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:38.175Z [INFO] jones: Requesting shutdown writer.go:29: 2021-01-29T19:32:38.177Z [INFO] jones.server: shutting down server writer.go:29: 2021-01-29T19:32:38.178Z [WARN] jones.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:38.178Z [ERROR] jones.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:38.181Z [INFO] jones.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:38.182Z [WARN] jones.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:38.183Z [INFO] jones.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:38.185Z [DEBUG] jones.server: Skipping self join check for node since the cluster is too small: node=Node-ab7adc75-15b0-f29d-8549-68fb916fd075 writer.go:29: 2021-01-29T19:32:38.186Z [INFO] jones.server: member joined, marking health alive: member=Node-ab7adc75-15b0-f29d-8549-68fb916fd075 writer.go:29: 2021-01-29T19:32:38.188Z [DEBUG] jones.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:38.189Z [DEBUG] jones.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:38.189Z [INFO] jones.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:38.189Z [INFO] jones: consul server down writer.go:29: 2021-01-29T19:32:38.194Z [INFO] jones: shutdown complete writer.go:29: 2021-01-29T19:32:38.195Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30186 network=tcp writer.go:29: 2021-01-29T19:32:38.197Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30186 network=udp writer.go:29: 2021-01-29T19:32:38.198Z [INFO] jones: Stopping server: protocol=HTTP address=127.0.0.1:30187 network=tcp writer.go:29: 2021-01-29T19:32:38.700Z [INFO] jones: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:38.701Z [INFO] jones: Endpoints down === RUN TestAgent_sidecarServiceFromNodeService/invalid_check_type writer.go:29: 2021-01-29T19:32:38.713Z [WARN] jones: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:38.714Z [DEBUG] jones.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:38.716Z [DEBUG] jones.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:38.721Z [INFO] jones.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c68fc65e-27e1-911f-bdf3-6e8c4bba4107 Address:127.0.0.1:30197}]" writer.go:29: 2021-01-29T19:32:38.723Z [INFO] jones.server.raft: entering follower state: follower="Node at 127.0.0.1:30197 [Follower]" leader= writer.go:29: 2021-01-29T19:32:38.724Z [INFO] jones.server.serf.wan: serf: EventMemberJoin: Node-c68fc65e-27e1-911f-bdf3-6e8c4bba4107.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:38.734Z [INFO] jones.server.serf.lan: serf: EventMemberJoin: Node-c68fc65e-27e1-911f-bdf3-6e8c4bba4107 127.0.0.1 writer.go:29: 2021-01-29T19:32:38.740Z [INFO] jones.server: Adding LAN server: server="Node-c68fc65e-27e1-911f-bdf3-6e8c4bba4107 (Addr: tcp/127.0.0.1:30197) (DC: dc1)" writer.go:29: 2021-01-29T19:32:38.741Z [INFO] jones.server: Handled event for server in area: event=member-join server=Node-c68fc65e-27e1-911f-bdf3-6e8c4bba4107.dc1 area=wan writer.go:29: 2021-01-29T19:32:38.742Z [INFO] jones: Started DNS server: address=127.0.0.1:30192 network=udp writer.go:29: 2021-01-29T19:32:38.745Z [INFO] jones: Started DNS server: address=127.0.0.1:30192 network=tcp writer.go:29: 2021-01-29T19:32:38.749Z [INFO] jones: Started HTTP server: address=127.0.0.1:30193 network=tcp writer.go:29: 2021-01-29T19:32:38.751Z [INFO] jones: started state syncer writer.go:29: 2021-01-29T19:32:38.784Z [WARN] jones.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:38.791Z [INFO] jones.server.raft: entering candidate state: node="Node at 127.0.0.1:30197 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:38.794Z [DEBUG] jones.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:38.796Z [DEBUG] jones.server.raft: vote granted: from=c68fc65e-27e1-911f-bdf3-6e8c4bba4107 term=2 tally=1 writer.go:29: 2021-01-29T19:32:38.797Z [INFO] jones.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:38.799Z [INFO] jones.server.raft: entering leader state: leader="Node at 127.0.0.1:30197 [Leader]" writer.go:29: 2021-01-29T19:32:38.801Z [INFO] jones.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:38.801Z [INFO] jones.server: New leader elected: payload=Node-c68fc65e-27e1-911f-bdf3-6e8c4bba4107 writer.go:29: 2021-01-29T19:32:38.805Z [DEBUG] jones.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30197 writer.go:29: 2021-01-29T19:32:38.816Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:38.832Z [INFO] jones.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:38.842Z [INFO] jones.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:38.846Z [DEBUG] jones.server: Skipping self join check for node since the cluster is too small: node=Node-c68fc65e-27e1-911f-bdf3-6e8c4bba4107 writer.go:29: 2021-01-29T19:32:38.853Z [INFO] jones.server: member joined, marking health alive: member=Node-c68fc65e-27e1-911f-bdf3-6e8c4bba4107 writer.go:29: 2021-01-29T19:32:39.044Z [INFO] jones: Requesting shutdown writer.go:29: 2021-01-29T19:32:39.045Z [INFO] jones.server: shutting down server writer.go:29: 2021-01-29T19:32:39.046Z [DEBUG] jones.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:39.046Z [ERROR] jones.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:39.047Z [WARN] jones.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:39.047Z [DEBUG] jones.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:39.050Z [WARN] jones.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:39.054Z [INFO] jones.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:39.055Z [INFO] jones: consul server down writer.go:29: 2021-01-29T19:32:39.056Z [INFO] jones: shutdown complete writer.go:29: 2021-01-29T19:32:39.057Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30192 network=tcp writer.go:29: 2021-01-29T19:32:39.058Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30192 network=udp writer.go:29: 2021-01-29T19:32:39.059Z [INFO] jones: Stopping server: protocol=HTTP address=127.0.0.1:30193 network=tcp writer.go:29: 2021-01-29T19:32:39.560Z [INFO] jones: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:39.561Z [INFO] jones: Endpoints down === RUN TestAgent_sidecarServiceFromNodeService/invalid_meta writer.go:29: 2021-01-29T19:32:39.583Z [WARN] jones: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:39.585Z [DEBUG] jones.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:39.587Z [DEBUG] jones.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:39.593Z [INFO] jones.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2db899c4-c075-06d6-3a08-56fb8c9988ae Address:127.0.0.1:30203}]" writer.go:29: 2021-01-29T19:32:39.595Z [INFO] jones.server.raft: entering follower state: follower="Node at 127.0.0.1:30203 [Follower]" leader= writer.go:29: 2021-01-29T19:32:39.608Z [INFO] jones.server.serf.wan: serf: EventMemberJoin: Node-2db899c4-c075-06d6-3a08-56fb8c9988ae.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:39.615Z [INFO] jones.server.serf.lan: serf: EventMemberJoin: Node-2db899c4-c075-06d6-3a08-56fb8c9988ae 127.0.0.1 writer.go:29: 2021-01-29T19:32:39.617Z [INFO] jones: Started DNS server: address=127.0.0.1:30198 network=udp writer.go:29: 2021-01-29T19:32:39.618Z [INFO] jones.server: Adding LAN server: server="Node-2db899c4-c075-06d6-3a08-56fb8c9988ae (Addr: tcp/127.0.0.1:30203) (DC: dc1)" writer.go:29: 2021-01-29T19:32:39.619Z [INFO] jones.server: Handled event for server in area: event=member-join server=Node-2db899c4-c075-06d6-3a08-56fb8c9988ae.dc1 area=wan writer.go:29: 2021-01-29T19:32:39.620Z [INFO] jones: Started DNS server: address=127.0.0.1:30198 network=tcp writer.go:29: 2021-01-29T19:32:39.632Z [INFO] jones: Started HTTP server: address=127.0.0.1:30199 network=tcp writer.go:29: 2021-01-29T19:32:39.633Z [INFO] jones: started state syncer writer.go:29: 2021-01-29T19:32:39.636Z [WARN] jones.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:39.638Z [INFO] jones.server.raft: entering candidate state: node="Node at 127.0.0.1:30203 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:39.641Z [DEBUG] jones.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:39.642Z [DEBUG] jones.server.raft: vote granted: from=2db899c4-c075-06d6-3a08-56fb8c9988ae term=2 tally=1 writer.go:29: 2021-01-29T19:32:39.643Z [INFO] jones.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:39.643Z [INFO] jones.server.raft: entering leader state: leader="Node at 127.0.0.1:30203 [Leader]" writer.go:29: 2021-01-29T19:32:39.644Z [INFO] jones.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:39.645Z [INFO] jones.server: New leader elected: payload=Node-2db899c4-c075-06d6-3a08-56fb8c9988ae writer.go:29: 2021-01-29T19:32:39.653Z [DEBUG] jones.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30203 writer.go:29: 2021-01-29T19:32:39.660Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:39.688Z [INFO] jones.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:39.689Z [INFO] jones.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:39.691Z [DEBUG] jones.server: Skipping self join check for node since the cluster is too small: node=Node-2db899c4-c075-06d6-3a08-56fb8c9988ae writer.go:29: 2021-01-29T19:32:39.692Z [INFO] jones.server: member joined, marking health alive: member=Node-2db899c4-c075-06d6-3a08-56fb8c9988ae writer.go:29: 2021-01-29T19:32:39.783Z [DEBUG] jones: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:39.785Z [INFO] jones: Synced node info writer.go:29: 2021-01-29T19:32:39.786Z [DEBUG] jones: Node info in sync writer.go:29: 2021-01-29T19:32:39.847Z [INFO] jones: Requesting shutdown writer.go:29: 2021-01-29T19:32:39.848Z [INFO] jones.server: shutting down server writer.go:29: 2021-01-29T19:32:39.849Z [DEBUG] jones.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:39.850Z [WARN] jones.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:39.850Z [DEBUG] jones.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:39.852Z [WARN] jones.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:39.856Z [INFO] jones.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:39.856Z [INFO] jones: consul server down writer.go:29: 2021-01-29T19:32:39.858Z [INFO] jones: shutdown complete writer.go:29: 2021-01-29T19:32:39.859Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30198 network=tcp writer.go:29: 2021-01-29T19:32:39.860Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30198 network=udp writer.go:29: 2021-01-29T19:32:39.861Z [INFO] jones: Stopping server: protocol=HTTP address=127.0.0.1:30199 network=tcp writer.go:29: 2021-01-29T19:32:40.362Z [INFO] jones: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:40.364Z [INFO] jones: Endpoints down === RUN TestAgent_sidecarServiceFromNodeService/re-registering_same_sidecar_with_no_port_should_pick_same_one writer.go:29: 2021-01-29T19:32:40.413Z [WARN] jones: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:40.414Z [DEBUG] jones.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:40.415Z [DEBUG] jones.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:40.424Z [INFO] jones.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8e0d6ecf-a009-bf2f-754b-210fcb03a9b7 Address:127.0.0.1:30209}]" writer.go:29: 2021-01-29T19:32:40.429Z [INFO] jones.server.raft: entering follower state: follower="Node at 127.0.0.1:30209 [Follower]" leader= writer.go:29: 2021-01-29T19:32:40.435Z [INFO] jones.server.serf.wan: serf: EventMemberJoin: Node-8e0d6ecf-a009-bf2f-754b-210fcb03a9b7.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:40.452Z [INFO] jones.server.serf.lan: serf: EventMemberJoin: Node-8e0d6ecf-a009-bf2f-754b-210fcb03a9b7 127.0.0.1 writer.go:29: 2021-01-29T19:32:40.474Z [INFO] jones.server: Adding LAN server: server="Node-8e0d6ecf-a009-bf2f-754b-210fcb03a9b7 (Addr: tcp/127.0.0.1:30209) (DC: dc1)" writer.go:29: 2021-01-29T19:32:40.475Z [INFO] jones.server: Handled event for server in area: event=member-join server=Node-8e0d6ecf-a009-bf2f-754b-210fcb03a9b7.dc1 area=wan writer.go:29: 2021-01-29T19:32:40.479Z [INFO] jones: Started DNS server: address=127.0.0.1:30204 network=tcp writer.go:29: 2021-01-29T19:32:40.480Z [INFO] jones: Started DNS server: address=127.0.0.1:30204 network=udp writer.go:29: 2021-01-29T19:32:40.484Z [INFO] jones: Started HTTP server: address=127.0.0.1:30205 network=tcp writer.go:29: 2021-01-29T19:32:40.485Z [INFO] jones: started state syncer writer.go:29: 2021-01-29T19:32:40.502Z [WARN] jones.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:40.506Z [INFO] jones.server.raft: entering candidate state: node="Node at 127.0.0.1:30209 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:40.554Z [DEBUG] jones.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:40.555Z [DEBUG] jones.server.raft: vote granted: from=8e0d6ecf-a009-bf2f-754b-210fcb03a9b7 term=2 tally=1 writer.go:29: 2021-01-29T19:32:40.557Z [INFO] jones.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:40.559Z [INFO] jones.server.raft: entering leader state: leader="Node at 127.0.0.1:30209 [Leader]" writer.go:29: 2021-01-29T19:32:40.561Z [INFO] jones.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:40.563Z [INFO] jones.server: New leader elected: payload=Node-8e0d6ecf-a009-bf2f-754b-210fcb03a9b7 writer.go:29: 2021-01-29T19:32:40.564Z [DEBUG] jones.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30209 writer.go:29: 2021-01-29T19:32:40.584Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:40.599Z [INFO] jones.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:40.600Z [INFO] jones.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:40.601Z [DEBUG] jones.server: Skipping self join check for node since the cluster is too small: node=Node-8e0d6ecf-a009-bf2f-754b-210fcb03a9b7 writer.go:29: 2021-01-29T19:32:40.602Z [INFO] jones.server: member joined, marking health alive: member=Node-8e0d6ecf-a009-bf2f-754b-210fcb03a9b7 writer.go:29: 2021-01-29T19:32:40.710Z [INFO] jones: Requesting shutdown writer.go:29: 2021-01-29T19:32:40.711Z [INFO] jones.server: shutting down server writer.go:29: 2021-01-29T19:32:40.712Z [DEBUG] jones.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:40.712Z [ERROR] jones.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:40.713Z [WARN] jones.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:40.713Z [DEBUG] jones.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:40.715Z [WARN] jones.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:40.719Z [INFO] jones.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:40.720Z [INFO] jones: consul server down writer.go:29: 2021-01-29T19:32:40.723Z [INFO] jones: shutdown complete writer.go:29: 2021-01-29T19:32:40.724Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30204 network=tcp writer.go:29: 2021-01-29T19:32:40.725Z [INFO] jones: Stopping server: protocol=DNS address=127.0.0.1:30204 network=udp writer.go:29: 2021-01-29T19:32:40.727Z [INFO] jones: Stopping server: protocol=HTTP address=127.0.0.1:30205 network=tcp writer.go:29: 2021-01-29T19:32:41.228Z [INFO] jones: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:41.230Z [INFO] jones: Endpoints down --- PASS: TestAgent_sidecarServiceFromNodeService (7.63s) --- PASS: TestAgent_sidecarServiceFromNodeService/no_sidecar (0.88s) --- PASS: TestAgent_sidecarServiceFromNodeService/all_the_defaults (0.92s) --- PASS: TestAgent_sidecarServiceFromNodeService/all_the_allowed_overrides (0.91s) --- PASS: TestAgent_sidecarServiceFromNodeService/no_auto_ports_available (1.02s) --- PASS: TestAgent_sidecarServiceFromNodeService/auto_ports_disabled (0.73s) --- PASS: TestAgent_sidecarServiceFromNodeService/inherit_tags_and_meta (0.64s) --- PASS: TestAgent_sidecarServiceFromNodeService/invalid_check_type (0.86s) --- PASS: TestAgent_sidecarServiceFromNodeService/invalid_meta (0.80s) --- PASS: TestAgent_sidecarServiceFromNodeService/re-registering_same_sidecar_with_no_port_should_pick_same_one (0.86s) === RUN TestSnapshot snapshot_endpoint_test.go:16: DM-skipped --- SKIP: TestSnapshot (0.00s) === RUN TestSnapshot_Options === PAUSE TestSnapshot_Options === RUN TestStatusLeader status_endpoint_test.go:14: DM-skipped --- SKIP: TestStatusLeader (0.00s) === RUN TestStatusLeaderSecondary === PAUSE TestStatusLeaderSecondary === RUN TestStatusPeers === PAUSE TestStatusPeers === RUN TestStatusPeersSecondary === PAUSE TestStatusPeersSecondary === RUN TestDefaultConfig === RUN TestDefaultConfig/#00 === PAUSE TestDefaultConfig/#00 === RUN TestDefaultConfig/#01 === PAUSE TestDefaultConfig/#01 === RUN TestDefaultConfig/#02 === PAUSE TestDefaultConfig/#02 === RUN TestDefaultConfig/#03 === PAUSE TestDefaultConfig/#03 === RUN TestDefaultConfig/#04 === PAUSE TestDefaultConfig/#04 === RUN TestDefaultConfig/#05 === PAUSE TestDefaultConfig/#05 === RUN TestDefaultConfig/#06 === PAUSE TestDefaultConfig/#06 === RUN TestDefaultConfig/#07 === PAUSE TestDefaultConfig/#07 === RUN TestDefaultConfig/#08 === PAUSE TestDefaultConfig/#08 === RUN TestDefaultConfig/#09 === PAUSE TestDefaultConfig/#09 === RUN TestDefaultConfig/#10 === PAUSE TestDefaultConfig/#10 === RUN TestDefaultConfig/#11 === PAUSE TestDefaultConfig/#11 === RUN TestDefaultConfig/#12 === PAUSE TestDefaultConfig/#12 === RUN TestDefaultConfig/#13 === PAUSE TestDefaultConfig/#13 === RUN TestDefaultConfig/#14 === PAUSE TestDefaultConfig/#14 === RUN TestDefaultConfig/#15 === PAUSE TestDefaultConfig/#15 === RUN TestDefaultConfig/#16 === PAUSE TestDefaultConfig/#16 === RUN TestDefaultConfig/#17 === PAUSE TestDefaultConfig/#17 === RUN TestDefaultConfig/#18 === PAUSE TestDefaultConfig/#18 === RUN TestDefaultConfig/#19 === PAUSE TestDefaultConfig/#19 === RUN TestDefaultConfig/#20 === PAUSE TestDefaultConfig/#20 === RUN TestDefaultConfig/#21 === PAUSE TestDefaultConfig/#21 === RUN TestDefaultConfig/#22 === PAUSE TestDefaultConfig/#22 === RUN TestDefaultConfig/#23 === PAUSE TestDefaultConfig/#23 === RUN TestDefaultConfig/#24 === PAUSE TestDefaultConfig/#24 === RUN TestDefaultConfig/#25 === PAUSE TestDefaultConfig/#25 === RUN TestDefaultConfig/#26 === PAUSE TestDefaultConfig/#26 === RUN TestDefaultConfig/#27 === PAUSE TestDefaultConfig/#27 === RUN TestDefaultConfig/#28 === PAUSE TestDefaultConfig/#28 === RUN TestDefaultConfig/#29 === PAUSE TestDefaultConfig/#29 === RUN TestDefaultConfig/#30 === PAUSE TestDefaultConfig/#30 === RUN TestDefaultConfig/#31 === PAUSE TestDefaultConfig/#31 === RUN TestDefaultConfig/#32 === PAUSE TestDefaultConfig/#32 === RUN TestDefaultConfig/#33 === PAUSE TestDefaultConfig/#33 === RUN TestDefaultConfig/#34 === PAUSE TestDefaultConfig/#34 === RUN TestDefaultConfig/#35 === PAUSE TestDefaultConfig/#35 === RUN TestDefaultConfig/#36 === PAUSE TestDefaultConfig/#36 === RUN TestDefaultConfig/#37 === PAUSE TestDefaultConfig/#37 === RUN TestDefaultConfig/#38 === PAUSE TestDefaultConfig/#38 === RUN TestDefaultConfig/#39 === PAUSE TestDefaultConfig/#39 === RUN TestDefaultConfig/#40 === PAUSE TestDefaultConfig/#40 === RUN TestDefaultConfig/#41 === PAUSE TestDefaultConfig/#41 === RUN TestDefaultConfig/#42 === PAUSE TestDefaultConfig/#42 === RUN TestDefaultConfig/#43 === PAUSE TestDefaultConfig/#43 === RUN TestDefaultConfig/#44 === PAUSE TestDefaultConfig/#44 === RUN TestDefaultConfig/#45 === PAUSE TestDefaultConfig/#45 === RUN TestDefaultConfig/#46 === PAUSE TestDefaultConfig/#46 === RUN TestDefaultConfig/#47 === PAUSE TestDefaultConfig/#47 === RUN TestDefaultConfig/#48 === PAUSE TestDefaultConfig/#48 === RUN TestDefaultConfig/#49 === PAUSE TestDefaultConfig/#49 === RUN TestDefaultConfig/#50 === PAUSE TestDefaultConfig/#50 === RUN TestDefaultConfig/#51 === PAUSE TestDefaultConfig/#51 === RUN TestDefaultConfig/#52 === PAUSE TestDefaultConfig/#52 === RUN TestDefaultConfig/#53 === PAUSE TestDefaultConfig/#53 === RUN TestDefaultConfig/#54 === PAUSE TestDefaultConfig/#54 === RUN TestDefaultConfig/#55 === PAUSE TestDefaultConfig/#55 === RUN TestDefaultConfig/#56 === PAUSE TestDefaultConfig/#56 === RUN TestDefaultConfig/#57 === PAUSE TestDefaultConfig/#57 === RUN TestDefaultConfig/#58 === PAUSE TestDefaultConfig/#58 === RUN TestDefaultConfig/#59 === PAUSE TestDefaultConfig/#59 === RUN TestDefaultConfig/#60 === PAUSE TestDefaultConfig/#60 === RUN TestDefaultConfig/#61 === PAUSE TestDefaultConfig/#61 === RUN TestDefaultConfig/#62 === PAUSE TestDefaultConfig/#62 === RUN TestDefaultConfig/#63 === PAUSE TestDefaultConfig/#63 === RUN TestDefaultConfig/#64 === PAUSE TestDefaultConfig/#64 === RUN TestDefaultConfig/#65 === PAUSE TestDefaultConfig/#65 === RUN TestDefaultConfig/#66 === PAUSE TestDefaultConfig/#66 === RUN TestDefaultConfig/#67 === PAUSE TestDefaultConfig/#67 === RUN TestDefaultConfig/#68 === PAUSE TestDefaultConfig/#68 === RUN TestDefaultConfig/#69 === PAUSE TestDefaultConfig/#69 === RUN TestDefaultConfig/#70 === PAUSE TestDefaultConfig/#70 === RUN TestDefaultConfig/#71 === PAUSE TestDefaultConfig/#71 === RUN TestDefaultConfig/#72 === PAUSE TestDefaultConfig/#72 === RUN TestDefaultConfig/#73 === PAUSE TestDefaultConfig/#73 === RUN TestDefaultConfig/#74 === PAUSE TestDefaultConfig/#74 === RUN TestDefaultConfig/#75 === PAUSE TestDefaultConfig/#75 === RUN TestDefaultConfig/#76 === PAUSE TestDefaultConfig/#76 === RUN TestDefaultConfig/#77 === PAUSE TestDefaultConfig/#77 === RUN TestDefaultConfig/#78 === PAUSE TestDefaultConfig/#78 === RUN TestDefaultConfig/#79 === PAUSE TestDefaultConfig/#79 === RUN TestDefaultConfig/#80 === PAUSE TestDefaultConfig/#80 === RUN TestDefaultConfig/#81 === PAUSE TestDefaultConfig/#81 === RUN TestDefaultConfig/#82 === PAUSE TestDefaultConfig/#82 === RUN TestDefaultConfig/#83 === PAUSE TestDefaultConfig/#83 === RUN TestDefaultConfig/#84 === PAUSE TestDefaultConfig/#84 === RUN TestDefaultConfig/#85 === PAUSE TestDefaultConfig/#85 === RUN TestDefaultConfig/#86 === PAUSE TestDefaultConfig/#86 === RUN TestDefaultConfig/#87 === PAUSE TestDefaultConfig/#87 === RUN TestDefaultConfig/#88 === PAUSE TestDefaultConfig/#88 === RUN TestDefaultConfig/#89 === PAUSE TestDefaultConfig/#89 === RUN TestDefaultConfig/#90 === PAUSE TestDefaultConfig/#90 === RUN TestDefaultConfig/#91 === PAUSE TestDefaultConfig/#91 === RUN TestDefaultConfig/#92 === PAUSE TestDefaultConfig/#92 === RUN TestDefaultConfig/#93 === PAUSE TestDefaultConfig/#93 === RUN TestDefaultConfig/#94 === PAUSE TestDefaultConfig/#94 === RUN TestDefaultConfig/#95 === PAUSE TestDefaultConfig/#95 === RUN TestDefaultConfig/#96 === PAUSE TestDefaultConfig/#96 === RUN TestDefaultConfig/#97 === PAUSE TestDefaultConfig/#97 === RUN TestDefaultConfig/#98 === PAUSE TestDefaultConfig/#98 === RUN TestDefaultConfig/#99 === PAUSE TestDefaultConfig/#99 === RUN TestDefaultConfig/#100 === PAUSE TestDefaultConfig/#100 === RUN TestDefaultConfig/#101 === PAUSE TestDefaultConfig/#101 === RUN TestDefaultConfig/#102 === PAUSE TestDefaultConfig/#102 === RUN TestDefaultConfig/#103 === PAUSE TestDefaultConfig/#103 === RUN TestDefaultConfig/#104 === PAUSE TestDefaultConfig/#104 === RUN TestDefaultConfig/#105 === PAUSE TestDefaultConfig/#105 === RUN TestDefaultConfig/#106 === PAUSE TestDefaultConfig/#106 === RUN TestDefaultConfig/#107 === PAUSE TestDefaultConfig/#107 === RUN TestDefaultConfig/#108 === PAUSE TestDefaultConfig/#108 === RUN TestDefaultConfig/#109 === PAUSE TestDefaultConfig/#109 === RUN TestDefaultConfig/#110 === PAUSE TestDefaultConfig/#110 === RUN TestDefaultConfig/#111 === PAUSE TestDefaultConfig/#111 === RUN TestDefaultConfig/#112 === PAUSE TestDefaultConfig/#112 === RUN TestDefaultConfig/#113 === PAUSE TestDefaultConfig/#113 === RUN TestDefaultConfig/#114 === PAUSE TestDefaultConfig/#114 === RUN TestDefaultConfig/#115 === PAUSE TestDefaultConfig/#115 === RUN TestDefaultConfig/#116 === PAUSE TestDefaultConfig/#116 === RUN TestDefaultConfig/#117 === PAUSE TestDefaultConfig/#117 === RUN TestDefaultConfig/#118 === PAUSE TestDefaultConfig/#118 === RUN TestDefaultConfig/#119 === PAUSE TestDefaultConfig/#119 === RUN TestDefaultConfig/#120 === PAUSE TestDefaultConfig/#120 === RUN TestDefaultConfig/#121 === PAUSE TestDefaultConfig/#121 === RUN TestDefaultConfig/#122 === PAUSE TestDefaultConfig/#122 === RUN TestDefaultConfig/#123 === PAUSE TestDefaultConfig/#123 === RUN TestDefaultConfig/#124 === PAUSE TestDefaultConfig/#124 === RUN TestDefaultConfig/#125 === PAUSE TestDefaultConfig/#125 === RUN TestDefaultConfig/#126 === PAUSE TestDefaultConfig/#126 === RUN TestDefaultConfig/#127 === PAUSE TestDefaultConfig/#127 === RUN TestDefaultConfig/#128 === PAUSE TestDefaultConfig/#128 === RUN TestDefaultConfig/#129 === PAUSE TestDefaultConfig/#129 === RUN TestDefaultConfig/#130 === PAUSE TestDefaultConfig/#130 === RUN TestDefaultConfig/#131 === PAUSE TestDefaultConfig/#131 === RUN TestDefaultConfig/#132 === PAUSE TestDefaultConfig/#132 === RUN TestDefaultConfig/#133 === PAUSE TestDefaultConfig/#133 === RUN TestDefaultConfig/#134 === PAUSE TestDefaultConfig/#134 === RUN TestDefaultConfig/#135 === PAUSE TestDefaultConfig/#135 === RUN TestDefaultConfig/#136 === PAUSE TestDefaultConfig/#136 === RUN TestDefaultConfig/#137 === PAUSE TestDefaultConfig/#137 === RUN TestDefaultConfig/#138 === PAUSE TestDefaultConfig/#138 === RUN TestDefaultConfig/#139 === PAUSE TestDefaultConfig/#139 === RUN TestDefaultConfig/#140 === PAUSE TestDefaultConfig/#140 === RUN TestDefaultConfig/#141 === PAUSE TestDefaultConfig/#141 === RUN TestDefaultConfig/#142 === PAUSE TestDefaultConfig/#142 === RUN TestDefaultConfig/#143 === PAUSE TestDefaultConfig/#143 === RUN TestDefaultConfig/#144 === PAUSE TestDefaultConfig/#144 === RUN TestDefaultConfig/#145 === PAUSE TestDefaultConfig/#145 === RUN TestDefaultConfig/#146 === PAUSE TestDefaultConfig/#146 === RUN TestDefaultConfig/#147 === PAUSE TestDefaultConfig/#147 === RUN TestDefaultConfig/#148 === PAUSE TestDefaultConfig/#148 === RUN TestDefaultConfig/#149 === PAUSE TestDefaultConfig/#149 === RUN TestDefaultConfig/#150 === PAUSE TestDefaultConfig/#150 === RUN TestDefaultConfig/#151 === PAUSE TestDefaultConfig/#151 === RUN TestDefaultConfig/#152 === PAUSE TestDefaultConfig/#152 === RUN TestDefaultConfig/#153 === PAUSE TestDefaultConfig/#153 === RUN TestDefaultConfig/#154 === PAUSE TestDefaultConfig/#154 === RUN TestDefaultConfig/#155 === PAUSE TestDefaultConfig/#155 === RUN TestDefaultConfig/#156 === PAUSE TestDefaultConfig/#156 === RUN TestDefaultConfig/#157 === PAUSE TestDefaultConfig/#157 === RUN TestDefaultConfig/#158 === PAUSE TestDefaultConfig/#158 === RUN TestDefaultConfig/#159 === PAUSE TestDefaultConfig/#159 === RUN TestDefaultConfig/#160 === PAUSE TestDefaultConfig/#160 === RUN TestDefaultConfig/#161 === PAUSE TestDefaultConfig/#161 === RUN TestDefaultConfig/#162 === PAUSE TestDefaultConfig/#162 === RUN TestDefaultConfig/#163 === PAUSE TestDefaultConfig/#163 === RUN TestDefaultConfig/#164 === PAUSE TestDefaultConfig/#164 === RUN TestDefaultConfig/#165 === PAUSE TestDefaultConfig/#165 === RUN TestDefaultConfig/#166 === PAUSE TestDefaultConfig/#166 === RUN TestDefaultConfig/#167 === PAUSE TestDefaultConfig/#167 === RUN TestDefaultConfig/#168 === PAUSE TestDefaultConfig/#168 === RUN TestDefaultConfig/#169 === PAUSE TestDefaultConfig/#169 === RUN TestDefaultConfig/#170 === PAUSE TestDefaultConfig/#170 === RUN TestDefaultConfig/#171 === PAUSE TestDefaultConfig/#171 === RUN TestDefaultConfig/#172 === PAUSE TestDefaultConfig/#172 === RUN TestDefaultConfig/#173 === PAUSE TestDefaultConfig/#173 === RUN TestDefaultConfig/#174 === PAUSE TestDefaultConfig/#174 === RUN TestDefaultConfig/#175 === PAUSE TestDefaultConfig/#175 === RUN TestDefaultConfig/#176 === PAUSE TestDefaultConfig/#176 === RUN TestDefaultConfig/#177 === PAUSE TestDefaultConfig/#177 === RUN TestDefaultConfig/#178 === PAUSE TestDefaultConfig/#178 === RUN TestDefaultConfig/#179 === PAUSE TestDefaultConfig/#179 === RUN TestDefaultConfig/#180 === PAUSE TestDefaultConfig/#180 === RUN TestDefaultConfig/#181 === PAUSE TestDefaultConfig/#181 === RUN TestDefaultConfig/#182 === PAUSE TestDefaultConfig/#182 === RUN TestDefaultConfig/#183 === PAUSE TestDefaultConfig/#183 === RUN TestDefaultConfig/#184 === PAUSE TestDefaultConfig/#184 === RUN TestDefaultConfig/#185 === PAUSE TestDefaultConfig/#185 === RUN TestDefaultConfig/#186 === PAUSE TestDefaultConfig/#186 === RUN TestDefaultConfig/#187 === PAUSE TestDefaultConfig/#187 === RUN TestDefaultConfig/#188 === PAUSE TestDefaultConfig/#188 === RUN TestDefaultConfig/#189 === PAUSE TestDefaultConfig/#189 === RUN TestDefaultConfig/#190 === PAUSE TestDefaultConfig/#190 === RUN TestDefaultConfig/#191 === PAUSE TestDefaultConfig/#191 === RUN TestDefaultConfig/#192 === PAUSE TestDefaultConfig/#192 === RUN TestDefaultConfig/#193 === PAUSE TestDefaultConfig/#193 === RUN TestDefaultConfig/#194 === PAUSE TestDefaultConfig/#194 === RUN TestDefaultConfig/#195 === PAUSE TestDefaultConfig/#195 === RUN TestDefaultConfig/#196 === PAUSE TestDefaultConfig/#196 === RUN TestDefaultConfig/#197 === PAUSE TestDefaultConfig/#197 === RUN TestDefaultConfig/#198 === PAUSE TestDefaultConfig/#198 === RUN TestDefaultConfig/#199 === PAUSE TestDefaultConfig/#199 === RUN TestDefaultConfig/#200 === PAUSE TestDefaultConfig/#200 === RUN TestDefaultConfig/#201 === PAUSE TestDefaultConfig/#201 === RUN TestDefaultConfig/#202 === PAUSE TestDefaultConfig/#202 === RUN TestDefaultConfig/#203 === PAUSE TestDefaultConfig/#203 === RUN TestDefaultConfig/#204 === PAUSE TestDefaultConfig/#204 === RUN TestDefaultConfig/#205 === PAUSE TestDefaultConfig/#205 === RUN TestDefaultConfig/#206 === PAUSE TestDefaultConfig/#206 === RUN TestDefaultConfig/#207 === PAUSE TestDefaultConfig/#207 === RUN TestDefaultConfig/#208 === PAUSE TestDefaultConfig/#208 === RUN TestDefaultConfig/#209 === PAUSE TestDefaultConfig/#209 === RUN TestDefaultConfig/#210 === PAUSE TestDefaultConfig/#210 === RUN TestDefaultConfig/#211 === PAUSE TestDefaultConfig/#211 === RUN TestDefaultConfig/#212 === PAUSE TestDefaultConfig/#212 === RUN TestDefaultConfig/#213 === PAUSE TestDefaultConfig/#213 === RUN TestDefaultConfig/#214 === PAUSE TestDefaultConfig/#214 === RUN TestDefaultConfig/#215 === PAUSE TestDefaultConfig/#215 === RUN TestDefaultConfig/#216 === PAUSE TestDefaultConfig/#216 === RUN TestDefaultConfig/#217 === PAUSE TestDefaultConfig/#217 === RUN TestDefaultConfig/#218 === PAUSE TestDefaultConfig/#218 === RUN TestDefaultConfig/#219 === PAUSE TestDefaultConfig/#219 === RUN TestDefaultConfig/#220 === PAUSE TestDefaultConfig/#220 === RUN TestDefaultConfig/#221 === PAUSE TestDefaultConfig/#221 === RUN TestDefaultConfig/#222 === PAUSE TestDefaultConfig/#222 === RUN TestDefaultConfig/#223 === PAUSE TestDefaultConfig/#223 === RUN TestDefaultConfig/#224 === PAUSE TestDefaultConfig/#224 === RUN TestDefaultConfig/#225 === PAUSE TestDefaultConfig/#225 === RUN TestDefaultConfig/#226 === PAUSE TestDefaultConfig/#226 === RUN TestDefaultConfig/#227 === PAUSE TestDefaultConfig/#227 === RUN TestDefaultConfig/#228 === PAUSE TestDefaultConfig/#228 === RUN TestDefaultConfig/#229 === PAUSE TestDefaultConfig/#229 === RUN TestDefaultConfig/#230 === PAUSE TestDefaultConfig/#230 === RUN TestDefaultConfig/#231 === PAUSE TestDefaultConfig/#231 === RUN TestDefaultConfig/#232 === PAUSE TestDefaultConfig/#232 === RUN TestDefaultConfig/#233 === PAUSE TestDefaultConfig/#233 === RUN TestDefaultConfig/#234 === PAUSE TestDefaultConfig/#234 === RUN TestDefaultConfig/#235 === PAUSE TestDefaultConfig/#235 === RUN TestDefaultConfig/#236 === PAUSE TestDefaultConfig/#236 === RUN TestDefaultConfig/#237 === PAUSE TestDefaultConfig/#237 === RUN TestDefaultConfig/#238 === PAUSE TestDefaultConfig/#238 === RUN TestDefaultConfig/#239 === PAUSE TestDefaultConfig/#239 === RUN TestDefaultConfig/#240 === PAUSE TestDefaultConfig/#240 === RUN TestDefaultConfig/#241 === PAUSE TestDefaultConfig/#241 === RUN TestDefaultConfig/#242 === PAUSE TestDefaultConfig/#242 === RUN TestDefaultConfig/#243 === PAUSE TestDefaultConfig/#243 === RUN TestDefaultConfig/#244 === PAUSE TestDefaultConfig/#244 === RUN TestDefaultConfig/#245 === PAUSE TestDefaultConfig/#245 === RUN TestDefaultConfig/#246 === PAUSE TestDefaultConfig/#246 === RUN TestDefaultConfig/#247 === PAUSE TestDefaultConfig/#247 === RUN TestDefaultConfig/#248 === PAUSE TestDefaultConfig/#248 === RUN TestDefaultConfig/#249 === PAUSE TestDefaultConfig/#249 === RUN TestDefaultConfig/#250 === PAUSE TestDefaultConfig/#250 === RUN TestDefaultConfig/#251 === PAUSE TestDefaultConfig/#251 === RUN TestDefaultConfig/#252 === PAUSE TestDefaultConfig/#252 === RUN TestDefaultConfig/#253 === PAUSE TestDefaultConfig/#253 === RUN TestDefaultConfig/#254 === PAUSE TestDefaultConfig/#254 === RUN TestDefaultConfig/#255 === PAUSE TestDefaultConfig/#255 === RUN TestDefaultConfig/#256 === PAUSE TestDefaultConfig/#256 === RUN TestDefaultConfig/#257 === PAUSE TestDefaultConfig/#257 === RUN TestDefaultConfig/#258 === PAUSE TestDefaultConfig/#258 === RUN TestDefaultConfig/#259 === PAUSE TestDefaultConfig/#259 === RUN TestDefaultConfig/#260 === PAUSE TestDefaultConfig/#260 === RUN TestDefaultConfig/#261 === PAUSE TestDefaultConfig/#261 === RUN TestDefaultConfig/#262 === PAUSE TestDefaultConfig/#262 === RUN TestDefaultConfig/#263 === PAUSE TestDefaultConfig/#263 === RUN TestDefaultConfig/#264 === PAUSE TestDefaultConfig/#264 === RUN TestDefaultConfig/#265 === PAUSE TestDefaultConfig/#265 === RUN TestDefaultConfig/#266 === PAUSE TestDefaultConfig/#266 === RUN TestDefaultConfig/#267 === PAUSE TestDefaultConfig/#267 === RUN TestDefaultConfig/#268 === PAUSE TestDefaultConfig/#268 === RUN TestDefaultConfig/#269 === PAUSE TestDefaultConfig/#269 === RUN TestDefaultConfig/#270 === PAUSE TestDefaultConfig/#270 === RUN TestDefaultConfig/#271 === PAUSE TestDefaultConfig/#271 === RUN TestDefaultConfig/#272 === PAUSE TestDefaultConfig/#272 === RUN TestDefaultConfig/#273 === PAUSE TestDefaultConfig/#273 === RUN TestDefaultConfig/#274 === PAUSE TestDefaultConfig/#274 === RUN TestDefaultConfig/#275 === PAUSE TestDefaultConfig/#275 === RUN TestDefaultConfig/#276 === PAUSE TestDefaultConfig/#276 === RUN TestDefaultConfig/#277 === PAUSE TestDefaultConfig/#277 === RUN TestDefaultConfig/#278 === PAUSE TestDefaultConfig/#278 === RUN TestDefaultConfig/#279 === PAUSE TestDefaultConfig/#279 === RUN TestDefaultConfig/#280 === PAUSE TestDefaultConfig/#280 === RUN TestDefaultConfig/#281 === PAUSE TestDefaultConfig/#281 === RUN TestDefaultConfig/#282 === PAUSE TestDefaultConfig/#282 === RUN TestDefaultConfig/#283 === PAUSE TestDefaultConfig/#283 === RUN TestDefaultConfig/#284 === PAUSE TestDefaultConfig/#284 === RUN TestDefaultConfig/#285 === PAUSE TestDefaultConfig/#285 === RUN TestDefaultConfig/#286 === PAUSE TestDefaultConfig/#286 === RUN TestDefaultConfig/#287 === PAUSE TestDefaultConfig/#287 === RUN TestDefaultConfig/#288 === PAUSE TestDefaultConfig/#288 === RUN TestDefaultConfig/#289 === PAUSE TestDefaultConfig/#289 === RUN TestDefaultConfig/#290 === PAUSE TestDefaultConfig/#290 === RUN TestDefaultConfig/#291 === PAUSE TestDefaultConfig/#291 === RUN TestDefaultConfig/#292 === PAUSE TestDefaultConfig/#292 === RUN TestDefaultConfig/#293 === PAUSE TestDefaultConfig/#293 === RUN TestDefaultConfig/#294 === PAUSE TestDefaultConfig/#294 === RUN TestDefaultConfig/#295 === PAUSE TestDefaultConfig/#295 === RUN TestDefaultConfig/#296 === PAUSE TestDefaultConfig/#296 === RUN TestDefaultConfig/#297 === PAUSE TestDefaultConfig/#297 === RUN TestDefaultConfig/#298 === PAUSE TestDefaultConfig/#298 === RUN TestDefaultConfig/#299 === PAUSE TestDefaultConfig/#299 === RUN TestDefaultConfig/#300 === PAUSE TestDefaultConfig/#300 === RUN TestDefaultConfig/#301 === PAUSE TestDefaultConfig/#301 === RUN TestDefaultConfig/#302 === PAUSE TestDefaultConfig/#302 === RUN TestDefaultConfig/#303 === PAUSE TestDefaultConfig/#303 === RUN TestDefaultConfig/#304 === PAUSE TestDefaultConfig/#304 === RUN TestDefaultConfig/#305 === PAUSE TestDefaultConfig/#305 === RUN TestDefaultConfig/#306 === PAUSE TestDefaultConfig/#306 === RUN TestDefaultConfig/#307 === PAUSE TestDefaultConfig/#307 === RUN TestDefaultConfig/#308 === PAUSE TestDefaultConfig/#308 === RUN TestDefaultConfig/#309 === PAUSE TestDefaultConfig/#309 === RUN TestDefaultConfig/#310 === PAUSE TestDefaultConfig/#310 === RUN TestDefaultConfig/#311 === PAUSE TestDefaultConfig/#311 === RUN TestDefaultConfig/#312 === PAUSE TestDefaultConfig/#312 === RUN TestDefaultConfig/#313 === PAUSE TestDefaultConfig/#313 === RUN TestDefaultConfig/#314 === PAUSE TestDefaultConfig/#314 === RUN TestDefaultConfig/#315 === PAUSE TestDefaultConfig/#315 === RUN TestDefaultConfig/#316 === PAUSE TestDefaultConfig/#316 === RUN TestDefaultConfig/#317 === PAUSE TestDefaultConfig/#317 === RUN TestDefaultConfig/#318 === PAUSE TestDefaultConfig/#318 === RUN TestDefaultConfig/#319 === PAUSE TestDefaultConfig/#319 === RUN TestDefaultConfig/#320 === PAUSE TestDefaultConfig/#320 === RUN TestDefaultConfig/#321 === PAUSE TestDefaultConfig/#321 === RUN TestDefaultConfig/#322 === PAUSE TestDefaultConfig/#322 === RUN TestDefaultConfig/#323 === PAUSE TestDefaultConfig/#323 === RUN TestDefaultConfig/#324 === PAUSE TestDefaultConfig/#324 === RUN TestDefaultConfig/#325 === PAUSE TestDefaultConfig/#325 === RUN TestDefaultConfig/#326 === PAUSE TestDefaultConfig/#326 === RUN TestDefaultConfig/#327 === PAUSE TestDefaultConfig/#327 === RUN TestDefaultConfig/#328 === PAUSE TestDefaultConfig/#328 === RUN TestDefaultConfig/#329 === PAUSE TestDefaultConfig/#329 === RUN TestDefaultConfig/#330 === PAUSE TestDefaultConfig/#330 === RUN TestDefaultConfig/#331 === PAUSE TestDefaultConfig/#331 === RUN TestDefaultConfig/#332 === PAUSE TestDefaultConfig/#332 === RUN TestDefaultConfig/#333 === PAUSE TestDefaultConfig/#333 === RUN TestDefaultConfig/#334 === PAUSE TestDefaultConfig/#334 === RUN TestDefaultConfig/#335 === PAUSE TestDefaultConfig/#335 === RUN TestDefaultConfig/#336 === PAUSE TestDefaultConfig/#336 === RUN TestDefaultConfig/#337 === PAUSE TestDefaultConfig/#337 === RUN TestDefaultConfig/#338 === PAUSE TestDefaultConfig/#338 === RUN TestDefaultConfig/#339 === PAUSE TestDefaultConfig/#339 === RUN TestDefaultConfig/#340 === PAUSE TestDefaultConfig/#340 === RUN TestDefaultConfig/#341 === PAUSE TestDefaultConfig/#341 === RUN TestDefaultConfig/#342 === PAUSE TestDefaultConfig/#342 === RUN TestDefaultConfig/#343 === PAUSE TestDefaultConfig/#343 === RUN TestDefaultConfig/#344 === PAUSE TestDefaultConfig/#344 === RUN TestDefaultConfig/#345 === PAUSE TestDefaultConfig/#345 === RUN TestDefaultConfig/#346 === PAUSE TestDefaultConfig/#346 === RUN TestDefaultConfig/#347 === PAUSE TestDefaultConfig/#347 === RUN TestDefaultConfig/#348 === PAUSE TestDefaultConfig/#348 === RUN TestDefaultConfig/#349 === PAUSE TestDefaultConfig/#349 === RUN TestDefaultConfig/#350 === PAUSE TestDefaultConfig/#350 === RUN TestDefaultConfig/#351 === PAUSE TestDefaultConfig/#351 === RUN TestDefaultConfig/#352 === PAUSE TestDefaultConfig/#352 === RUN TestDefaultConfig/#353 === PAUSE TestDefaultConfig/#353 === RUN TestDefaultConfig/#354 === PAUSE TestDefaultConfig/#354 === RUN TestDefaultConfig/#355 === PAUSE TestDefaultConfig/#355 === RUN TestDefaultConfig/#356 === PAUSE TestDefaultConfig/#356 === RUN TestDefaultConfig/#357 === PAUSE TestDefaultConfig/#357 === RUN TestDefaultConfig/#358 === PAUSE TestDefaultConfig/#358 === RUN TestDefaultConfig/#359 === PAUSE TestDefaultConfig/#359 === RUN TestDefaultConfig/#360 === PAUSE TestDefaultConfig/#360 === RUN TestDefaultConfig/#361 === PAUSE TestDefaultConfig/#361 === RUN TestDefaultConfig/#362 === PAUSE TestDefaultConfig/#362 === RUN TestDefaultConfig/#363 === PAUSE TestDefaultConfig/#363 === RUN TestDefaultConfig/#364 === PAUSE TestDefaultConfig/#364 === RUN TestDefaultConfig/#365 === PAUSE TestDefaultConfig/#365 === RUN TestDefaultConfig/#366 === PAUSE TestDefaultConfig/#366 === RUN TestDefaultConfig/#367 === PAUSE TestDefaultConfig/#367 === RUN TestDefaultConfig/#368 === PAUSE TestDefaultConfig/#368 === RUN TestDefaultConfig/#369 === PAUSE TestDefaultConfig/#369 === RUN TestDefaultConfig/#370 === PAUSE TestDefaultConfig/#370 === RUN TestDefaultConfig/#371 === PAUSE TestDefaultConfig/#371 === RUN TestDefaultConfig/#372 === PAUSE TestDefaultConfig/#372 === RUN TestDefaultConfig/#373 === PAUSE TestDefaultConfig/#373 === RUN TestDefaultConfig/#374 === PAUSE TestDefaultConfig/#374 === RUN TestDefaultConfig/#375 === PAUSE TestDefaultConfig/#375 === RUN TestDefaultConfig/#376 === PAUSE TestDefaultConfig/#376 === RUN TestDefaultConfig/#377 === PAUSE TestDefaultConfig/#377 === RUN TestDefaultConfig/#378 === PAUSE TestDefaultConfig/#378 === RUN TestDefaultConfig/#379 === PAUSE TestDefaultConfig/#379 === RUN TestDefaultConfig/#380 === PAUSE TestDefaultConfig/#380 === RUN TestDefaultConfig/#381 === PAUSE TestDefaultConfig/#381 === RUN TestDefaultConfig/#382 === PAUSE TestDefaultConfig/#382 === RUN TestDefaultConfig/#383 === PAUSE TestDefaultConfig/#383 === RUN TestDefaultConfig/#384 === PAUSE TestDefaultConfig/#384 === RUN TestDefaultConfig/#385 === PAUSE TestDefaultConfig/#385 === RUN TestDefaultConfig/#386 === PAUSE TestDefaultConfig/#386 === RUN TestDefaultConfig/#387 === PAUSE TestDefaultConfig/#387 === RUN TestDefaultConfig/#388 === PAUSE TestDefaultConfig/#388 === RUN TestDefaultConfig/#389 === PAUSE TestDefaultConfig/#389 === RUN TestDefaultConfig/#390 === PAUSE TestDefaultConfig/#390 === RUN TestDefaultConfig/#391 === PAUSE TestDefaultConfig/#391 === RUN TestDefaultConfig/#392 === PAUSE TestDefaultConfig/#392 === RUN TestDefaultConfig/#393 === PAUSE TestDefaultConfig/#393 === RUN TestDefaultConfig/#394 === PAUSE TestDefaultConfig/#394 === RUN TestDefaultConfig/#395 === PAUSE TestDefaultConfig/#395 === RUN TestDefaultConfig/#396 === PAUSE TestDefaultConfig/#396 === RUN TestDefaultConfig/#397 === PAUSE TestDefaultConfig/#397 === RUN TestDefaultConfig/#398 === PAUSE TestDefaultConfig/#398 === RUN TestDefaultConfig/#399 === PAUSE TestDefaultConfig/#399 === RUN TestDefaultConfig/#400 === PAUSE TestDefaultConfig/#400 === RUN TestDefaultConfig/#401 === PAUSE TestDefaultConfig/#401 === RUN TestDefaultConfig/#402 === PAUSE TestDefaultConfig/#402 === RUN TestDefaultConfig/#403 === PAUSE TestDefaultConfig/#403 === RUN TestDefaultConfig/#404 === PAUSE TestDefaultConfig/#404 === RUN TestDefaultConfig/#405 === PAUSE TestDefaultConfig/#405 === RUN TestDefaultConfig/#406 === PAUSE TestDefaultConfig/#406 === RUN TestDefaultConfig/#407 === PAUSE TestDefaultConfig/#407 === RUN TestDefaultConfig/#408 === PAUSE TestDefaultConfig/#408 === RUN TestDefaultConfig/#409 === PAUSE TestDefaultConfig/#409 === RUN TestDefaultConfig/#410 === PAUSE TestDefaultConfig/#410 === RUN TestDefaultConfig/#411 === PAUSE TestDefaultConfig/#411 === RUN TestDefaultConfig/#412 === PAUSE TestDefaultConfig/#412 === RUN TestDefaultConfig/#413 === PAUSE TestDefaultConfig/#413 === RUN TestDefaultConfig/#414 === PAUSE TestDefaultConfig/#414 === RUN TestDefaultConfig/#415 === PAUSE TestDefaultConfig/#415 === RUN TestDefaultConfig/#416 === PAUSE TestDefaultConfig/#416 === RUN TestDefaultConfig/#417 === PAUSE TestDefaultConfig/#417 === RUN TestDefaultConfig/#418 === PAUSE TestDefaultConfig/#418 === RUN TestDefaultConfig/#419 === PAUSE TestDefaultConfig/#419 === RUN TestDefaultConfig/#420 === PAUSE TestDefaultConfig/#420 === RUN TestDefaultConfig/#421 === PAUSE TestDefaultConfig/#421 === RUN TestDefaultConfig/#422 === PAUSE TestDefaultConfig/#422 === RUN TestDefaultConfig/#423 === PAUSE TestDefaultConfig/#423 === RUN TestDefaultConfig/#424 === PAUSE TestDefaultConfig/#424 === RUN TestDefaultConfig/#425 === PAUSE TestDefaultConfig/#425 === RUN TestDefaultConfig/#426 === PAUSE TestDefaultConfig/#426 === RUN TestDefaultConfig/#427 === PAUSE TestDefaultConfig/#427 === RUN TestDefaultConfig/#428 === PAUSE TestDefaultConfig/#428 === RUN TestDefaultConfig/#429 === PAUSE TestDefaultConfig/#429 === RUN TestDefaultConfig/#430 === PAUSE TestDefaultConfig/#430 === RUN TestDefaultConfig/#431 === PAUSE TestDefaultConfig/#431 === RUN TestDefaultConfig/#432 === PAUSE TestDefaultConfig/#432 === RUN TestDefaultConfig/#433 === PAUSE TestDefaultConfig/#433 === RUN TestDefaultConfig/#434 === PAUSE TestDefaultConfig/#434 === RUN TestDefaultConfig/#435 === PAUSE TestDefaultConfig/#435 === RUN TestDefaultConfig/#436 === PAUSE TestDefaultConfig/#436 === RUN TestDefaultConfig/#437 === PAUSE TestDefaultConfig/#437 === RUN TestDefaultConfig/#438 === PAUSE TestDefaultConfig/#438 === RUN TestDefaultConfig/#439 === PAUSE TestDefaultConfig/#439 === RUN TestDefaultConfig/#440 === PAUSE TestDefaultConfig/#440 === RUN TestDefaultConfig/#441 === PAUSE TestDefaultConfig/#441 === RUN TestDefaultConfig/#442 === PAUSE TestDefaultConfig/#442 === RUN TestDefaultConfig/#443 === PAUSE TestDefaultConfig/#443 === RUN TestDefaultConfig/#444 === PAUSE TestDefaultConfig/#444 === RUN TestDefaultConfig/#445 === PAUSE TestDefaultConfig/#445 === RUN TestDefaultConfig/#446 === PAUSE TestDefaultConfig/#446 === RUN TestDefaultConfig/#447 === PAUSE TestDefaultConfig/#447 === RUN TestDefaultConfig/#448 === PAUSE TestDefaultConfig/#448 === RUN TestDefaultConfig/#449 === PAUSE TestDefaultConfig/#449 === RUN TestDefaultConfig/#450 === PAUSE TestDefaultConfig/#450 === RUN TestDefaultConfig/#451 === PAUSE TestDefaultConfig/#451 === RUN TestDefaultConfig/#452 === PAUSE TestDefaultConfig/#452 === RUN TestDefaultConfig/#453 === PAUSE TestDefaultConfig/#453 === RUN TestDefaultConfig/#454 === PAUSE TestDefaultConfig/#454 === RUN TestDefaultConfig/#455 === PAUSE TestDefaultConfig/#455 === RUN TestDefaultConfig/#456 === PAUSE TestDefaultConfig/#456 === RUN TestDefaultConfig/#457 === PAUSE TestDefaultConfig/#457 === RUN TestDefaultConfig/#458 === PAUSE TestDefaultConfig/#458 === RUN TestDefaultConfig/#459 === PAUSE TestDefaultConfig/#459 === RUN TestDefaultConfig/#460 === PAUSE TestDefaultConfig/#460 === RUN TestDefaultConfig/#461 === PAUSE TestDefaultConfig/#461 === RUN TestDefaultConfig/#462 === PAUSE TestDefaultConfig/#462 === RUN TestDefaultConfig/#463 === PAUSE TestDefaultConfig/#463 === RUN TestDefaultConfig/#464 === PAUSE TestDefaultConfig/#464 === RUN TestDefaultConfig/#465 === PAUSE TestDefaultConfig/#465 === RUN TestDefaultConfig/#466 === PAUSE TestDefaultConfig/#466 === RUN TestDefaultConfig/#467 === PAUSE TestDefaultConfig/#467 === RUN TestDefaultConfig/#468 === PAUSE TestDefaultConfig/#468 === RUN TestDefaultConfig/#469 === PAUSE TestDefaultConfig/#469 === RUN TestDefaultConfig/#470 === PAUSE TestDefaultConfig/#470 === RUN TestDefaultConfig/#471 === PAUSE TestDefaultConfig/#471 === RUN TestDefaultConfig/#472 === PAUSE TestDefaultConfig/#472 === RUN TestDefaultConfig/#473 === PAUSE TestDefaultConfig/#473 === RUN TestDefaultConfig/#474 === PAUSE TestDefaultConfig/#474 === RUN TestDefaultConfig/#475 === PAUSE TestDefaultConfig/#475 === RUN TestDefaultConfig/#476 === PAUSE TestDefaultConfig/#476 === RUN TestDefaultConfig/#477 === PAUSE TestDefaultConfig/#477 === RUN TestDefaultConfig/#478 === PAUSE TestDefaultConfig/#478 === RUN TestDefaultConfig/#479 === PAUSE TestDefaultConfig/#479 === RUN TestDefaultConfig/#480 === PAUSE TestDefaultConfig/#480 === RUN TestDefaultConfig/#481 === PAUSE TestDefaultConfig/#481 === RUN TestDefaultConfig/#482 === PAUSE TestDefaultConfig/#482 === RUN TestDefaultConfig/#483 === PAUSE TestDefaultConfig/#483 === RUN TestDefaultConfig/#484 === PAUSE TestDefaultConfig/#484 === RUN TestDefaultConfig/#485 === PAUSE TestDefaultConfig/#485 === RUN TestDefaultConfig/#486 === PAUSE TestDefaultConfig/#486 === RUN TestDefaultConfig/#487 === PAUSE TestDefaultConfig/#487 === RUN TestDefaultConfig/#488 === PAUSE TestDefaultConfig/#488 === RUN TestDefaultConfig/#489 === PAUSE TestDefaultConfig/#489 === RUN TestDefaultConfig/#490 === PAUSE TestDefaultConfig/#490 === RUN TestDefaultConfig/#491 === PAUSE TestDefaultConfig/#491 === RUN TestDefaultConfig/#492 === PAUSE TestDefaultConfig/#492 === RUN TestDefaultConfig/#493 === PAUSE TestDefaultConfig/#493 === RUN TestDefaultConfig/#494 === PAUSE TestDefaultConfig/#494 === RUN TestDefaultConfig/#495 === PAUSE TestDefaultConfig/#495 === RUN TestDefaultConfig/#496 === PAUSE TestDefaultConfig/#496 === RUN TestDefaultConfig/#497 === PAUSE TestDefaultConfig/#497 === RUN TestDefaultConfig/#498 === PAUSE TestDefaultConfig/#498 === RUN TestDefaultConfig/#499 === PAUSE TestDefaultConfig/#499 === CONT TestDefaultConfig/#00 === CONT TestDefaultConfig/#499 === CONT TestDefaultConfig/#434 === CONT TestDefaultConfig/#385 === CONT TestDefaultConfig/#433 === CONT TestDefaultConfig/#384 === CONT TestDefaultConfig/#498 === CONT TestDefaultConfig/#242 === CONT TestDefaultConfig/#241 === CONT TestDefaultConfig/#240 === CONT TestDefaultConfig/#239 === CONT TestDefaultConfig/#238 === CONT TestDefaultConfig/#237 === CONT TestDefaultConfig/#236 === CONT TestDefaultConfig/#235 === CONT TestDefaultConfig/#234 === CONT TestDefaultConfig/#233 === CONT TestDefaultConfig/#232 === CONT TestDefaultConfig/#231 === CONT TestDefaultConfig/#230 === CONT TestDefaultConfig/#229 === CONT TestDefaultConfig/#228 === CONT TestDefaultConfig/#227 === CONT TestDefaultConfig/#226 === CONT TestDefaultConfig/#225 === CONT TestDefaultConfig/#224 === CONT TestDefaultConfig/#223 === CONT TestDefaultConfig/#222 === CONT TestDefaultConfig/#221 === CONT TestDefaultConfig/#220 === CONT TestDefaultConfig/#219 === CONT TestDefaultConfig/#218 === CONT TestDefaultConfig/#217 === CONT TestDefaultConfig/#216 === CONT TestDefaultConfig/#215 === CONT TestDefaultConfig/#214 === CONT TestDefaultConfig/#213 === CONT TestDefaultConfig/#212 === CONT TestDefaultConfig/#211 === CONT TestDefaultConfig/#210 === CONT TestDefaultConfig/#209 === CONT TestDefaultConfig/#208 === CONT TestDefaultConfig/#207 === CONT TestDefaultConfig/#206 === CONT TestDefaultConfig/#205 === CONT TestDefaultConfig/#204 === CONT TestDefaultConfig/#203 === CONT TestDefaultConfig/#202 === CONT TestDefaultConfig/#201 === CONT TestDefaultConfig/#200 === CONT TestDefaultConfig/#199 === CONT TestDefaultConfig/#198 === CONT TestDefaultConfig/#197 === CONT TestDefaultConfig/#196 === CONT TestDefaultConfig/#195 === CONT TestDefaultConfig/#194 === CONT TestDefaultConfig/#193 === CONT TestDefaultConfig/#349 === CONT TestDefaultConfig/#192 === CONT TestDefaultConfig/#191 === CONT TestDefaultConfig/#190 === CONT TestDefaultConfig/#189 === CONT TestDefaultConfig/#188 === CONT TestDefaultConfig/#187 === CONT TestDefaultConfig/#186 === CONT TestDefaultConfig/#185 === CONT TestDefaultConfig/#184 === CONT TestDefaultConfig/#183 === CONT TestDefaultConfig/#182 === CONT TestDefaultConfig/#181 === CONT TestDefaultConfig/#180 === CONT TestDefaultConfig/#179 === CONT TestDefaultConfig/#178 === CONT TestDefaultConfig/#177 === CONT TestDefaultConfig/#176 === CONT TestDefaultConfig/#175 === CONT TestDefaultConfig/#174 === CONT TestDefaultConfig/#173 === CONT TestDefaultConfig/#171 === CONT TestDefaultConfig/#170 === CONT TestDefaultConfig/#172 === CONT TestDefaultConfig/#169 === CONT TestDefaultConfig/#168 === CONT TestDefaultConfig/#167 === CONT TestDefaultConfig/#166 === CONT TestDefaultConfig/#165 === CONT TestDefaultConfig/#164 === CONT TestDefaultConfig/#163 === CONT TestDefaultConfig/#162 === CONT TestDefaultConfig/#161 === CONT TestDefaultConfig/#160 === CONT TestDefaultConfig/#159 === CONT TestDefaultConfig/#158 === CONT TestDefaultConfig/#157 === CONT TestDefaultConfig/#156 === CONT TestDefaultConfig/#155 === CONT TestDefaultConfig/#154 === CONT TestDefaultConfig/#153 === CONT TestDefaultConfig/#152 === CONT TestDefaultConfig/#151 === CONT TestDefaultConfig/#150 === CONT TestDefaultConfig/#149 === CONT TestDefaultConfig/#148 === CONT TestDefaultConfig/#147 === CONT TestDefaultConfig/#146 === CONT TestDefaultConfig/#145 === CONT TestDefaultConfig/#144 === CONT TestDefaultConfig/#143 === CONT TestDefaultConfig/#142 === CONT TestDefaultConfig/#141 === CONT TestDefaultConfig/#140 === CONT TestDefaultConfig/#139 === CONT TestDefaultConfig/#138 === CONT TestDefaultConfig/#137 === CONT TestDefaultConfig/#136 === CONT TestDefaultConfig/#135 === CONT TestDefaultConfig/#134 === CONT TestDefaultConfig/#133 === CONT TestDefaultConfig/#348 === CONT TestDefaultConfig/#132 === CONT TestDefaultConfig/#131 === CONT TestDefaultConfig/#130 === CONT TestDefaultConfig/#129 === CONT TestDefaultConfig/#128 === CONT TestDefaultConfig/#127 === CONT TestDefaultConfig/#126 === CONT TestDefaultConfig/#125 === CONT TestDefaultConfig/#124 === CONT TestDefaultConfig/#123 === CONT TestDefaultConfig/#122 === CONT TestDefaultConfig/#121 === CONT TestDefaultConfig/#120 === CONT TestDefaultConfig/#119 === CONT TestDefaultConfig/#118 === CONT TestDefaultConfig/#117 === CONT TestDefaultConfig/#116 === CONT TestDefaultConfig/#113 === CONT TestDefaultConfig/#115 === CONT TestDefaultConfig/#112 === CONT TestDefaultConfig/#111 === CONT TestDefaultConfig/#110 === CONT TestDefaultConfig/#109 === CONT TestDefaultConfig/#108 === CONT TestDefaultConfig/#107 === CONT TestDefaultConfig/#106 === CONT TestDefaultConfig/#105 === CONT TestDefaultConfig/#104 === CONT TestDefaultConfig/#103 === CONT TestDefaultConfig/#102 === CONT TestDefaultConfig/#101 === CONT TestDefaultConfig/#100 === CONT TestDefaultConfig/#99 === CONT TestDefaultConfig/#98 === CONT TestDefaultConfig/#97 === CONT TestDefaultConfig/#96 === CONT TestDefaultConfig/#95 === CONT TestDefaultConfig/#93 === CONT TestDefaultConfig/#94 === CONT TestDefaultConfig/#92 === CONT TestDefaultConfig/#91 === CONT TestDefaultConfig/#90 === CONT TestDefaultConfig/#89 === CONT TestDefaultConfig/#88 === CONT TestDefaultConfig/#87 === CONT TestDefaultConfig/#86 === CONT TestDefaultConfig/#85 === CONT TestDefaultConfig/#84 === CONT TestDefaultConfig/#83 === CONT TestDefaultConfig/#82 === CONT TestDefaultConfig/#81 === CONT TestDefaultConfig/#80 === CONT TestDefaultConfig/#79 === CONT TestDefaultConfig/#78 === CONT TestDefaultConfig/#77 === CONT TestDefaultConfig/#76 === CONT TestDefaultConfig/#75 === CONT TestDefaultConfig/#74 === CONT TestDefaultConfig/#73 === CONT TestDefaultConfig/#72 === CONT TestDefaultConfig/#347 === CONT TestDefaultConfig/#71 === CONT TestDefaultConfig/#70 === CONT TestDefaultConfig/#69 === CONT TestDefaultConfig/#68 === CONT TestDefaultConfig/#67 === CONT TestDefaultConfig/#66 === CONT TestDefaultConfig/#65 === CONT TestDefaultConfig/#64 === CONT TestDefaultConfig/#63 === CONT TestDefaultConfig/#62 === CONT TestDefaultConfig/#61 === CONT TestDefaultConfig/#60 === CONT TestDefaultConfig/#58 === CONT TestDefaultConfig/#59 === CONT TestDefaultConfig/#57 === CONT TestDefaultConfig/#56 === CONT TestDefaultConfig/#55 === CONT TestDefaultConfig/#54 === CONT TestDefaultConfig/#53 === CONT TestDefaultConfig/#52 === CONT TestDefaultConfig/#51 === CONT TestDefaultConfig/#50 === CONT TestDefaultConfig/#49 === CONT TestDefaultConfig/#48 === CONT TestDefaultConfig/#47 === CONT TestDefaultConfig/#46 === CONT TestDefaultConfig/#45 === CONT TestDefaultConfig/#44 === CONT TestDefaultConfig/#43 === CONT TestDefaultConfig/#42 === CONT TestDefaultConfig/#41 === CONT TestDefaultConfig/#40 === CONT TestDefaultConfig/#39 === CONT TestDefaultConfig/#38 === CONT TestDefaultConfig/#37 === CONT TestDefaultConfig/#36 === CONT TestDefaultConfig/#35 === CONT TestDefaultConfig/#34 === CONT TestDefaultConfig/#33 === CONT TestDefaultConfig/#32 === CONT TestDefaultConfig/#31 === CONT TestDefaultConfig/#30 === CONT TestDefaultConfig/#29 === CONT TestDefaultConfig/#28 === CONT TestDefaultConfig/#27 === CONT TestDefaultConfig/#26 === CONT TestDefaultConfig/#25 === CONT TestDefaultConfig/#24 === CONT TestDefaultConfig/#23 === CONT TestDefaultConfig/#22 === CONT TestDefaultConfig/#21 === CONT TestDefaultConfig/#20 === CONT TestDefaultConfig/#19 === CONT TestDefaultConfig/#18 === CONT TestDefaultConfig/#17 === CONT TestDefaultConfig/#16 === CONT TestDefaultConfig/#15 === CONT TestDefaultConfig/#14 === CONT TestDefaultConfig/#13 === CONT TestDefaultConfig/#12 === CONT TestDefaultConfig/#346 === CONT TestDefaultConfig/#11 === CONT TestDefaultConfig/#10 === CONT TestDefaultConfig/#09 === CONT TestDefaultConfig/#08 === CONT TestDefaultConfig/#07 === CONT TestDefaultConfig/#06 === CONT TestDefaultConfig/#05 === CONT TestDefaultConfig/#04 === CONT TestDefaultConfig/#02 === CONT TestDefaultConfig/#03 === CONT TestDefaultConfig/#01 === CONT TestDefaultConfig/#345 === CONT TestDefaultConfig/#344 === CONT TestDefaultConfig/#343 === CONT TestDefaultConfig/#342 === CONT TestDefaultConfig/#341 === CONT TestDefaultConfig/#340 === CONT TestDefaultConfig/#339 === CONT TestDefaultConfig/#338 === CONT TestDefaultConfig/#337 === CONT TestDefaultConfig/#336 === CONT TestDefaultConfig/#335 === CONT TestDefaultConfig/#334 === CONT TestDefaultConfig/#333 === CONT TestDefaultConfig/#332 === CONT TestDefaultConfig/#331 === CONT TestDefaultConfig/#330 === CONT TestDefaultConfig/#329 === CONT TestDefaultConfig/#328 === CONT TestDefaultConfig/#327 === CONT TestDefaultConfig/#326 === CONT TestDefaultConfig/#325 === CONT TestDefaultConfig/#324 === CONT TestDefaultConfig/#323 === CONT TestDefaultConfig/#322 === CONT TestDefaultConfig/#321 === CONT TestDefaultConfig/#320 === CONT TestDefaultConfig/#319 === CONT TestDefaultConfig/#318 === CONT TestDefaultConfig/#317 === CONT TestDefaultConfig/#316 === CONT TestDefaultConfig/#315 === CONT TestDefaultConfig/#314 === CONT TestDefaultConfig/#313 === CONT TestDefaultConfig/#312 === CONT TestDefaultConfig/#311 === CONT TestDefaultConfig/#310 === CONT TestDefaultConfig/#309 === CONT TestDefaultConfig/#308 === CONT TestDefaultConfig/#307 === CONT TestDefaultConfig/#306 === CONT TestDefaultConfig/#305 === CONT TestDefaultConfig/#304 === CONT TestDefaultConfig/#303 === CONT TestDefaultConfig/#302 === CONT TestDefaultConfig/#301 === CONT TestDefaultConfig/#300 === CONT TestDefaultConfig/#299 === CONT TestDefaultConfig/#298 === CONT TestDefaultConfig/#297 === CONT TestDefaultConfig/#283 === CONT TestDefaultConfig/#296 === CONT TestDefaultConfig/#295 === CONT TestDefaultConfig/#294 === CONT TestDefaultConfig/#293 === CONT TestDefaultConfig/#292 === CONT TestDefaultConfig/#291 === CONT TestDefaultConfig/#290 === CONT TestDefaultConfig/#289 === CONT TestDefaultConfig/#288 === CONT TestDefaultConfig/#287 === CONT TestDefaultConfig/#286 === CONT TestDefaultConfig/#285 === CONT TestDefaultConfig/#284 === CONT TestDefaultConfig/#282 === CONT TestDefaultConfig/#281 === CONT TestDefaultConfig/#280 === CONT TestDefaultConfig/#279 === CONT TestDefaultConfig/#278 === CONT TestDefaultConfig/#277 === CONT TestDefaultConfig/#276 === CONT TestDefaultConfig/#275 === CONT TestDefaultConfig/#274 === CONT TestDefaultConfig/#273 === CONT TestDefaultConfig/#272 === CONT TestDefaultConfig/#271 === CONT TestDefaultConfig/#270 === CONT TestDefaultConfig/#269 === CONT TestDefaultConfig/#268 === CONT TestDefaultConfig/#267 === CONT TestDefaultConfig/#266 === CONT TestDefaultConfig/#265 === CONT TestDefaultConfig/#264 === CONT TestDefaultConfig/#263 === CONT TestDefaultConfig/#262 === CONT TestDefaultConfig/#261 === CONT TestDefaultConfig/#260 === CONT TestDefaultConfig/#259 === CONT TestDefaultConfig/#258 === CONT TestDefaultConfig/#257 === CONT TestDefaultConfig/#256 === CONT TestDefaultConfig/#255 === CONT TestDefaultConfig/#254 === CONT TestDefaultConfig/#253 === CONT TestDefaultConfig/#252 === CONT TestDefaultConfig/#251 === CONT TestDefaultConfig/#250 === CONT TestDefaultConfig/#249 === CONT TestDefaultConfig/#248 === CONT TestDefaultConfig/#247 === CONT TestDefaultConfig/#246 === CONT TestDefaultConfig/#245 === CONT TestDefaultConfig/#244 === CONT TestDefaultConfig/#114 === CONT TestDefaultConfig/#432 === CONT TestDefaultConfig/#431 === CONT TestDefaultConfig/#430 === CONT TestDefaultConfig/#429 === CONT TestDefaultConfig/#428 === CONT TestDefaultConfig/#427 === CONT TestDefaultConfig/#426 === CONT TestDefaultConfig/#425 === CONT TestDefaultConfig/#424 === CONT TestDefaultConfig/#423 === CONT TestDefaultConfig/#422 === CONT TestDefaultConfig/#421 === CONT TestDefaultConfig/#420 === CONT TestDefaultConfig/#419 === CONT TestDefaultConfig/#418 === CONT TestDefaultConfig/#417 === CONT TestDefaultConfig/#416 === CONT TestDefaultConfig/#415 === CONT TestDefaultConfig/#414 === CONT TestDefaultConfig/#413 === CONT TestDefaultConfig/#412 === CONT TestDefaultConfig/#411 === CONT TestDefaultConfig/#410 === CONT TestDefaultConfig/#409 === CONT TestDefaultConfig/#408 === CONT TestDefaultConfig/#407 === CONT TestDefaultConfig/#406 === CONT TestDefaultConfig/#405 === CONT TestDefaultConfig/#404 === CONT TestDefaultConfig/#403 === CONT TestDefaultConfig/#402 === CONT TestDefaultConfig/#401 === CONT TestDefaultConfig/#400 === CONT TestDefaultConfig/#399 === CONT TestDefaultConfig/#398 === CONT TestDefaultConfig/#397 === CONT TestDefaultConfig/#396 === CONT TestDefaultConfig/#395 === CONT TestDefaultConfig/#383 === CONT TestDefaultConfig/#394 === CONT TestDefaultConfig/#382 === CONT TestDefaultConfig/#381 === CONT TestDefaultConfig/#380 === CONT TestDefaultConfig/#379 === CONT TestDefaultConfig/#378 === CONT TestDefaultConfig/#377 === CONT TestDefaultConfig/#376 === CONT TestDefaultConfig/#375 === CONT TestDefaultConfig/#374 === CONT TestDefaultConfig/#373 === CONT TestDefaultConfig/#372 === CONT TestDefaultConfig/#243 === CONT TestDefaultConfig/#371 === CONT TestDefaultConfig/#370 === CONT TestDefaultConfig/#369 === CONT TestDefaultConfig/#368 === CONT TestDefaultConfig/#367 === CONT TestDefaultConfig/#366 === CONT TestDefaultConfig/#365 === CONT TestDefaultConfig/#364 === CONT TestDefaultConfig/#363 === CONT TestDefaultConfig/#362 === CONT TestDefaultConfig/#361 === CONT TestDefaultConfig/#360 === CONT TestDefaultConfig/#359 === CONT TestDefaultConfig/#358 === CONT TestDefaultConfig/#357 === CONT TestDefaultConfig/#356 === CONT TestDefaultConfig/#355 === CONT TestDefaultConfig/#354 === CONT TestDefaultConfig/#353 === CONT TestDefaultConfig/#352 === CONT TestDefaultConfig/#351 === CONT TestDefaultConfig/#497 === CONT TestDefaultConfig/#350 === CONT TestDefaultConfig/#496 === CONT TestDefaultConfig/#495 === CONT TestDefaultConfig/#494 === CONT TestDefaultConfig/#493 === CONT TestDefaultConfig/#492 === CONT TestDefaultConfig/#491 === CONT TestDefaultConfig/#490 === CONT TestDefaultConfig/#489 === CONT TestDefaultConfig/#488 === CONT TestDefaultConfig/#487 === CONT TestDefaultConfig/#486 === CONT TestDefaultConfig/#485 === CONT TestDefaultConfig/#484 === CONT TestDefaultConfig/#483 === CONT TestDefaultConfig/#481 === CONT TestDefaultConfig/#480 === CONT TestDefaultConfig/#482 === CONT TestDefaultConfig/#479 === CONT TestDefaultConfig/#478 === CONT TestDefaultConfig/#477 === CONT TestDefaultConfig/#476 === CONT TestDefaultConfig/#475 === CONT TestDefaultConfig/#474 === CONT TestDefaultConfig/#473 === CONT TestDefaultConfig/#472 === CONT TestDefaultConfig/#471 === CONT TestDefaultConfig/#470 === CONT TestDefaultConfig/#469 === CONT TestDefaultConfig/#468 === CONT TestDefaultConfig/#467 === CONT TestDefaultConfig/#466 === CONT TestDefaultConfig/#465 === CONT TestDefaultConfig/#464 === CONT TestDefaultConfig/#463 === CONT TestDefaultConfig/#462 === CONT TestDefaultConfig/#461 === CONT TestDefaultConfig/#460 === CONT TestDefaultConfig/#459 === CONT TestDefaultConfig/#458 === CONT TestDefaultConfig/#457 === CONT TestDefaultConfig/#456 === CONT TestDefaultConfig/#455 === CONT TestDefaultConfig/#454 === CONT TestDefaultConfig/#453 === CONT TestDefaultConfig/#452 === CONT TestDefaultConfig/#451 === CONT TestDefaultConfig/#450 === CONT TestDefaultConfig/#449 === CONT TestDefaultConfig/#448 === CONT TestDefaultConfig/#447 === CONT TestDefaultConfig/#446 === CONT TestDefaultConfig/#445 === CONT TestDefaultConfig/#444 === CONT TestDefaultConfig/#443 === CONT TestDefaultConfig/#442 === CONT TestDefaultConfig/#441 === CONT TestDefaultConfig/#440 === CONT TestDefaultConfig/#439 === CONT TestDefaultConfig/#437 === CONT TestDefaultConfig/#436 === CONT TestDefaultConfig/#438 === CONT TestDefaultConfig/#435 === CONT TestDefaultConfig/#393 === CONT TestDefaultConfig/#392 === CONT TestDefaultConfig/#391 === CONT TestDefaultConfig/#390 === CONT TestDefaultConfig/#389 === CONT TestDefaultConfig/#388 === CONT TestDefaultConfig/#387 === CONT TestDefaultConfig/#386 --- PASS: TestDefaultConfig (0.32s) --- PASS: TestDefaultConfig/#434 (0.01s) --- PASS: TestDefaultConfig/#385 (0.01s) --- PASS: TestDefaultConfig/#499 (0.01s) --- PASS: TestDefaultConfig/#00 (0.02s) --- PASS: TestDefaultConfig/#498 (0.01s) --- PASS: TestDefaultConfig/#433 (0.02s) --- PASS: TestDefaultConfig/#384 (0.02s) --- PASS: TestDefaultConfig/#240 (0.02s) --- PASS: TestDefaultConfig/#241 (0.02s) --- PASS: TestDefaultConfig/#242 (0.14s) --- PASS: TestDefaultConfig/#239 (0.27s) --- PASS: TestDefaultConfig/#238 (0.44s) --- PASS: TestDefaultConfig/#237 (0.44s) --- PASS: TestDefaultConfig/#235 (0.18s) --- PASS: TestDefaultConfig/#236 (0.32s) --- PASS: TestDefaultConfig/#231 (0.01s) --- PASS: TestDefaultConfig/#232 (0.01s) --- PASS: TestDefaultConfig/#230 (0.01s) --- PASS: TestDefaultConfig/#229 (0.01s) --- PASS: TestDefaultConfig/#228 (0.01s) --- PASS: TestDefaultConfig/#227 (0.01s) --- PASS: TestDefaultConfig/#233 (0.02s) --- PASS: TestDefaultConfig/#234 (0.03s) --- PASS: TestDefaultConfig/#225 (0.21s) --- PASS: TestDefaultConfig/#224 (0.22s) --- PASS: TestDefaultConfig/#226 (0.22s) --- PASS: TestDefaultConfig/#221 (0.01s) --- PASS: TestDefaultConfig/#223 (0.23s) --- PASS: TestDefaultConfig/#219 (0.01s) --- PASS: TestDefaultConfig/#222 (0.03s) --- PASS: TestDefaultConfig/#220 (0.02s) --- PASS: TestDefaultConfig/#217 (0.01s) --- PASS: TestDefaultConfig/#218 (0.02s) --- PASS: TestDefaultConfig/#215 (0.03s) --- PASS: TestDefaultConfig/#214 (0.04s) --- PASS: TestDefaultConfig/#213 (0.04s) --- PASS: TestDefaultConfig/#212 (0.08s) --- PASS: TestDefaultConfig/#216 (0.12s) --- PASS: TestDefaultConfig/#210 (0.09s) --- PASS: TestDefaultConfig/#208 (0.17s) --- PASS: TestDefaultConfig/#211 (0.24s) --- PASS: TestDefaultConfig/#205 (0.01s) --- PASS: TestDefaultConfig/#207 (0.15s) --- PASS: TestDefaultConfig/#206 (0.02s) --- PASS: TestDefaultConfig/#203 (0.01s) --- PASS: TestDefaultConfig/#209 (0.20s) --- PASS: TestDefaultConfig/#204 (0.02s) --- PASS: TestDefaultConfig/#202 (0.02s) --- PASS: TestDefaultConfig/#200 (0.01s) --- PASS: TestDefaultConfig/#201 (0.02s) --- PASS: TestDefaultConfig/#199 (0.04s) --- PASS: TestDefaultConfig/#198 (0.04s) --- PASS: TestDefaultConfig/#197 (0.06s) --- PASS: TestDefaultConfig/#195 (0.07s) --- PASS: TestDefaultConfig/#193 (0.05s) --- PASS: TestDefaultConfig/#196 (0.16s) --- PASS: TestDefaultConfig/#194 (0.18s) --- PASS: TestDefaultConfig/#349 (0.33s) --- PASS: TestDefaultConfig/#192 (0.31s) --- PASS: TestDefaultConfig/#190 (0.19s) --- PASS: TestDefaultConfig/#189 (0.01s) --- PASS: TestDefaultConfig/#188 (0.02s) --- PASS: TestDefaultConfig/#191 (0.29s) --- PASS: TestDefaultConfig/#186 (0.03s) --- PASS: TestDefaultConfig/#185 (0.03s) --- PASS: TestDefaultConfig/#187 (0.05s) --- PASS: TestDefaultConfig/#183 (0.02s) --- PASS: TestDefaultConfig/#184 (0.04s) --- PASS: TestDefaultConfig/#181 (0.05s) --- PASS: TestDefaultConfig/#180 (0.07s) --- PASS: TestDefaultConfig/#182 (0.12s) --- PASS: TestDefaultConfig/#179 (0.15s) --- PASS: TestDefaultConfig/#177 (0.39s) --- PASS: TestDefaultConfig/#178 (0.42s) --- PASS: TestDefaultConfig/#176 (0.37s) --- PASS: TestDefaultConfig/#175 (0.32s) --- PASS: TestDefaultConfig/#173 (0.03s) --- PASS: TestDefaultConfig/#171 (0.01s) --- PASS: TestDefaultConfig/#174 (0.04s) --- PASS: TestDefaultConfig/#172 (0.02s) --- PASS: TestDefaultConfig/#170 (0.02s) --- PASS: TestDefaultConfig/#167 (0.01s) --- PASS: TestDefaultConfig/#168 (0.02s) --- PASS: TestDefaultConfig/#164 (0.01s) --- PASS: TestDefaultConfig/#166 (0.02s) --- PASS: TestDefaultConfig/#165 (0.02s) --- PASS: TestDefaultConfig/#169 (0.05s) --- PASS: TestDefaultConfig/#161 (0.01s) --- PASS: TestDefaultConfig/#162 (0.05s) --- PASS: TestDefaultConfig/#160 (0.05s) --- PASS: TestDefaultConfig/#163 (0.09s) --- PASS: TestDefaultConfig/#159 (0.14s) --- PASS: TestDefaultConfig/#158 (0.11s) --- PASS: TestDefaultConfig/#157 (0.25s) --- PASS: TestDefaultConfig/#156 (0.26s) --- PASS: TestDefaultConfig/#154 (0.34s) --- PASS: TestDefaultConfig/#153 (0.18s) --- PASS: TestDefaultConfig/#152 (0.02s) --- PASS: TestDefaultConfig/#151 (0.01s) --- PASS: TestDefaultConfig/#155 (0.35s) --- PASS: TestDefaultConfig/#150 (0.01s) --- PASS: TestDefaultConfig/#149 (0.02s) --- PASS: TestDefaultConfig/#148 (0.02s) --- PASS: TestDefaultConfig/#147 (0.02s) --- PASS: TestDefaultConfig/#146 (0.02s) --- PASS: TestDefaultConfig/#145 (0.01s) --- PASS: TestDefaultConfig/#143 (0.04s) --- PASS: TestDefaultConfig/#144 (0.04s) --- PASS: TestDefaultConfig/#142 (0.11s) --- PASS: TestDefaultConfig/#139 (0.09s) --- PASS: TestDefaultConfig/#141 (0.13s) --- PASS: TestDefaultConfig/#140 (0.11s) --- PASS: TestDefaultConfig/#136 (0.15s) --- PASS: TestDefaultConfig/#135 (0.29s) --- PASS: TestDefaultConfig/#138 (0.34s) --- PASS: TestDefaultConfig/#133 (0.01s) --- PASS: TestDefaultConfig/#134 (0.17s) --- PASS: TestDefaultConfig/#137 (0.33s) --- PASS: TestDefaultConfig/#132 (0.01s) --- PASS: TestDefaultConfig/#348 (0.02s) --- PASS: TestDefaultConfig/#129 (0.01s) --- PASS: TestDefaultConfig/#130 (0.01s) --- PASS: TestDefaultConfig/#128 (0.01s) --- PASS: TestDefaultConfig/#127 (0.01s) --- PASS: TestDefaultConfig/#126 (0.01s) --- PASS: TestDefaultConfig/#124 (0.01s) --- PASS: TestDefaultConfig/#131 (0.04s) --- PASS: TestDefaultConfig/#123 (0.04s) --- PASS: TestDefaultConfig/#122 (0.06s) --- PASS: TestDefaultConfig/#125 (0.10s) --- PASS: TestDefaultConfig/#121 (0.11s) --- PASS: TestDefaultConfig/#120 (0.11s) --- PASS: TestDefaultConfig/#119 (0.40s) --- PASS: TestDefaultConfig/#117 (0.34s) --- PASS: TestDefaultConfig/#116 (0.32s) --- PASS: TestDefaultConfig/#118 (0.41s) --- PASS: TestDefaultConfig/#115 (0.02s) --- PASS: TestDefaultConfig/#113 (0.04s) --- PASS: TestDefaultConfig/#111 (0.01s) --- PASS: TestDefaultConfig/#109 (0.01s) --- PASS: TestDefaultConfig/#112 (0.04s) --- PASS: TestDefaultConfig/#108 (0.01s) --- PASS: TestDefaultConfig/#105 (0.01s) --- PASS: TestDefaultConfig/#107 (0.02s) --- PASS: TestDefaultConfig/#104 (0.01s) --- PASS: TestDefaultConfig/#110 (0.03s) --- PASS: TestDefaultConfig/#102 (0.01s) --- PASS: TestDefaultConfig/#106 (0.03s) --- PASS: TestDefaultConfig/#103 (0.06s) --- PASS: TestDefaultConfig/#101 (0.08s) --- PASS: TestDefaultConfig/#100 (0.12s) --- PASS: TestDefaultConfig/#99 (0.15s) --- PASS: TestDefaultConfig/#98 (0.17s) --- PASS: TestDefaultConfig/#95 (0.07s) --- PASS: TestDefaultConfig/#96 (0.09s) --- PASS: TestDefaultConfig/#97 (0.32s) --- PASS: TestDefaultConfig/#92 (0.17s) --- PASS: TestDefaultConfig/#94 (0.17s) --- PASS: TestDefaultConfig/#93 (0.19s) --- PASS: TestDefaultConfig/#91 (0.02s) --- PASS: TestDefaultConfig/#89 (0.01s) --- PASS: TestDefaultConfig/#86 (0.01s) --- PASS: TestDefaultConfig/#87 (0.01s) --- PASS: TestDefaultConfig/#88 (0.02s) --- PASS: TestDefaultConfig/#85 (0.01s) --- PASS: TestDefaultConfig/#90 (0.03s) --- PASS: TestDefaultConfig/#84 (0.01s) --- PASS: TestDefaultConfig/#80 (0.02s) --- PASS: TestDefaultConfig/#82 (0.04s) --- PASS: TestDefaultConfig/#83 (0.08s) --- PASS: TestDefaultConfig/#81 (0.07s) --- PASS: TestDefaultConfig/#77 (0.04s) --- PASS: TestDefaultConfig/#79 (0.08s) --- PASS: TestDefaultConfig/#76 (0.27s) --- PASS: TestDefaultConfig/#78 (0.31s) --- PASS: TestDefaultConfig/#72 (0.03s) --- PASS: TestDefaultConfig/#75 (0.27s) --- PASS: TestDefaultConfig/#74 (0.27s) --- PASS: TestDefaultConfig/#347 (0.02s) --- PASS: TestDefaultConfig/#70 (0.01s) --- PASS: TestDefaultConfig/#73 (0.05s) --- PASS: TestDefaultConfig/#68 (0.01s) --- PASS: TestDefaultConfig/#66 (0.01s) --- PASS: TestDefaultConfig/#71 (0.05s) --- PASS: TestDefaultConfig/#65 (0.01s) --- PASS: TestDefaultConfig/#67 (0.04s) --- PASS: TestDefaultConfig/#69 (0.05s) --- PASS: TestDefaultConfig/#63 (0.03s) --- PASS: TestDefaultConfig/#62 (0.07s) --- PASS: TestDefaultConfig/#64 (0.09s) --- PASS: TestDefaultConfig/#60 (0.07s) --- PASS: TestDefaultConfig/#61 (0.08s) --- PASS: TestDefaultConfig/#59 (0.24s) --- PASS: TestDefaultConfig/#56 (0.25s) --- PASS: TestDefaultConfig/#57 (0.28s) --- PASS: TestDefaultConfig/#55 (0.05s) --- PASS: TestDefaultConfig/#58 (0.29s) --- PASS: TestDefaultConfig/#53 (0.01s) --- PASS: TestDefaultConfig/#50 (0.01s) --- PASS: TestDefaultConfig/#52 (0.02s) --- PASS: TestDefaultConfig/#51 (0.02s) --- PASS: TestDefaultConfig/#49 (0.01s) --- PASS: TestDefaultConfig/#48 (0.01s) --- PASS: TestDefaultConfig/#47 (0.01s) --- PASS: TestDefaultConfig/#46 (0.01s) --- PASS: TestDefaultConfig/#54 (0.06s) --- PASS: TestDefaultConfig/#42 (0.01s) --- PASS: TestDefaultConfig/#44 (0.02s) --- PASS: TestDefaultConfig/#43 (0.07s) --- PASS: TestDefaultConfig/#45 (0.08s) --- PASS: TestDefaultConfig/#40 (0.08s) --- PASS: TestDefaultConfig/#41 (0.08s) --- PASS: TestDefaultConfig/#39 (0.19s) --- PASS: TestDefaultConfig/#38 (0.18s) --- PASS: TestDefaultConfig/#37 (0.20s) --- PASS: TestDefaultConfig/#35 (0.04s) --- PASS: TestDefaultConfig/#33 (0.01s) --- PASS: TestDefaultConfig/#36 (0.21s) --- PASS: TestDefaultConfig/#32 (0.01s) --- PASS: TestDefaultConfig/#30 (0.01s) --- PASS: TestDefaultConfig/#31 (0.02s) --- PASS: TestDefaultConfig/#29 (0.01s) --- PASS: TestDefaultConfig/#28 (0.01s) --- PASS: TestDefaultConfig/#26 (0.01s) --- PASS: TestDefaultConfig/#34 (0.07s) --- PASS: TestDefaultConfig/#25 (0.01s) --- PASS: TestDefaultConfig/#27 (0.07s) --- PASS: TestDefaultConfig/#22 (0.07s) --- PASS: TestDefaultConfig/#23 (0.20s) --- PASS: TestDefaultConfig/#20 (0.13s) --- PASS: TestDefaultConfig/#24 (0.24s) --- PASS: TestDefaultConfig/#19 (0.26s) --- PASS: TestDefaultConfig/#21 (0.42s) --- PASS: TestDefaultConfig/#18 (0.24s) --- PASS: TestDefaultConfig/#16 (0.02s) --- PASS: TestDefaultConfig/#17 (0.23s) --- PASS: TestDefaultConfig/#15 (0.02s) --- PASS: TestDefaultConfig/#14 (0.02s) --- PASS: TestDefaultConfig/#346 (0.01s) --- PASS: TestDefaultConfig/#11 (0.01s) --- PASS: TestDefaultConfig/#12 (0.02s) --- PASS: TestDefaultConfig/#10 (0.01s) --- PASS: TestDefaultConfig/#08 (0.01s) --- PASS: TestDefaultConfig/#13 (0.04s) --- PASS: TestDefaultConfig/#06 (0.05s) --- PASS: TestDefaultConfig/#07 (0.08s) --- PASS: TestDefaultConfig/#09 (0.09s) --- PASS: TestDefaultConfig/#05 (0.14s) --- PASS: TestDefaultConfig/#04 (0.14s) --- PASS: TestDefaultConfig/#02 (0.18s) --- PASS: TestDefaultConfig/#01 (0.15s) --- PASS: TestDefaultConfig/#03 (0.31s) --- PASS: TestDefaultConfig/#345 (0.21s) --- PASS: TestDefaultConfig/#343 (0.11s) --- PASS: TestDefaultConfig/#341 (0.01s) --- PASS: TestDefaultConfig/#340 (0.01s) --- PASS: TestDefaultConfig/#344 (0.14s) --- PASS: TestDefaultConfig/#339 (0.01s) --- PASS: TestDefaultConfig/#342 (0.02s) --- PASS: TestDefaultConfig/#337 (0.01s) --- PASS: TestDefaultConfig/#338 (0.02s) --- PASS: TestDefaultConfig/#336 (0.02s) --- PASS: TestDefaultConfig/#335 (0.04s) --- PASS: TestDefaultConfig/#332 (0.04s) --- PASS: TestDefaultConfig/#334 (0.05s) --- PASS: TestDefaultConfig/#331 (0.06s) --- PASS: TestDefaultConfig/#330 (0.06s) --- PASS: TestDefaultConfig/#333 (0.19s) --- PASS: TestDefaultConfig/#329 (0.17s) --- PASS: TestDefaultConfig/#328 (0.13s) --- PASS: TestDefaultConfig/#327 (0.21s) --- PASS: TestDefaultConfig/#326 (0.13s) --- PASS: TestDefaultConfig/#323 (0.01s) --- PASS: TestDefaultConfig/#325 (0.11s) --- PASS: TestDefaultConfig/#324 (0.10s) --- PASS: TestDefaultConfig/#322 (0.02s) --- PASS: TestDefaultConfig/#321 (0.01s) --- PASS: TestDefaultConfig/#320 (0.01s) --- PASS: TestDefaultConfig/#319 (0.01s) --- PASS: TestDefaultConfig/#316 (0.01s) --- PASS: TestDefaultConfig/#318 (0.02s) --- PASS: TestDefaultConfig/#315 (0.01s) --- PASS: TestDefaultConfig/#313 (0.01s) --- PASS: TestDefaultConfig/#312 (0.03s) --- PASS: TestDefaultConfig/#317 (0.05s) --- PASS: TestDefaultConfig/#314 (0.05s) --- PASS: TestDefaultConfig/#311 (0.07s) --- PASS: TestDefaultConfig/#309 (0.08s) --- PASS: TestDefaultConfig/#308 (0.06s) --- PASS: TestDefaultConfig/#306 (0.16s) --- PASS: TestDefaultConfig/#305 (0.17s) --- PASS: TestDefaultConfig/#310 (0.26s) --- PASS: TestDefaultConfig/#303 (0.01s) --- PASS: TestDefaultConfig/#307 (0.21s) --- PASS: TestDefaultConfig/#301 (0.01s) --- PASS: TestDefaultConfig/#304 (0.02s) --- PASS: TestDefaultConfig/#300 (0.01s) --- PASS: TestDefaultConfig/#299 (0.01s) --- PASS: TestDefaultConfig/#302 (0.02s) --- PASS: TestDefaultConfig/#283 (0.01s) --- PASS: TestDefaultConfig/#295 (0.01s) --- PASS: TestDefaultConfig/#297 (0.02s) --- PASS: TestDefaultConfig/#298 (0.03s) --- PASS: TestDefaultConfig/#293 (0.05s) --- PASS: TestDefaultConfig/#296 (0.06s) --- PASS: TestDefaultConfig/#294 (0.06s) --- PASS: TestDefaultConfig/#292 (0.08s) --- PASS: TestDefaultConfig/#290 (0.05s) --- PASS: TestDefaultConfig/#291 (0.14s) --- PASS: TestDefaultConfig/#289 (0.23s) --- PASS: TestDefaultConfig/#287 (0.20s) --- PASS: TestDefaultConfig/#285 (0.03s) --- PASS: TestDefaultConfig/#286 (0.12s) --- PASS: TestDefaultConfig/#288 (0.24s) --- PASS: TestDefaultConfig/#282 (0.01s) --- PASS: TestDefaultConfig/#284 (0.02s) --- PASS: TestDefaultConfig/#281 (0.02s) --- PASS: TestDefaultConfig/#279 (0.01s) --- PASS: TestDefaultConfig/#278 (0.01s) --- PASS: TestDefaultConfig/#280 (0.03s) --- PASS: TestDefaultConfig/#277 (0.02s) --- PASS: TestDefaultConfig/#274 (0.01s) --- PASS: TestDefaultConfig/#275 (0.03s) --- PASS: TestDefaultConfig/#273 (0.07s) --- PASS: TestDefaultConfig/#276 (0.09s) --- PASS: TestDefaultConfig/#271 (0.09s) --- PASS: TestDefaultConfig/#272 (0.11s) --- PASS: TestDefaultConfig/#270 (0.18s) --- PASS: TestDefaultConfig/#268 (0.25s) --- PASS: TestDefaultConfig/#267 (0.26s) --- PASS: TestDefaultConfig/#265 (0.02s) --- PASS: TestDefaultConfig/#269 (0.31s) --- PASS: TestDefaultConfig/#264 (0.01s) --- PASS: TestDefaultConfig/#266 (0.13s) --- PASS: TestDefaultConfig/#262 (0.02s) --- PASS: TestDefaultConfig/#261 (0.02s) --- PASS: TestDefaultConfig/#259 (0.01s) --- PASS: TestDefaultConfig/#263 (0.04s) --- PASS: TestDefaultConfig/#258 (0.01s) --- PASS: TestDefaultConfig/#256 (0.01s) --- PASS: TestDefaultConfig/#255 (0.01s) --- PASS: TestDefaultConfig/#260 (0.05s) --- PASS: TestDefaultConfig/#254 (0.03s) --- PASS: TestDefaultConfig/#252 (0.03s) --- PASS: TestDefaultConfig/#257 (0.08s) --- PASS: TestDefaultConfig/#253 (0.07s) --- PASS: TestDefaultConfig/#251 (0.09s) --- PASS: TestDefaultConfig/#249 (0.14s) --- PASS: TestDefaultConfig/#248 (0.37s) --- PASS: TestDefaultConfig/#246 (0.24s) --- PASS: TestDefaultConfig/#247 (0.33s) --- PASS: TestDefaultConfig/#250 (0.42s) --- PASS: TestDefaultConfig/#114 (0.02s) --- PASS: TestDefaultConfig/#431 (0.01s) --- PASS: TestDefaultConfig/#244 (0.03s) --- PASS: TestDefaultConfig/#245 (0.04s) --- PASS: TestDefaultConfig/#429 (0.01s) --- PASS: TestDefaultConfig/#432 (0.03s) --- PASS: TestDefaultConfig/#428 (0.01s) --- PASS: TestDefaultConfig/#430 (0.04s) --- PASS: TestDefaultConfig/#427 (0.03s) --- PASS: TestDefaultConfig/#425 (0.01s) --- PASS: TestDefaultConfig/#422 (0.02s) --- PASS: TestDefaultConfig/#426 (0.08s) --- PASS: TestDefaultConfig/#423 (0.08s) --- PASS: TestDefaultConfig/#424 (0.09s) --- PASS: TestDefaultConfig/#419 (0.21s) --- PASS: TestDefaultConfig/#421 (0.27s) --- PASS: TestDefaultConfig/#420 (0.39s) --- PASS: TestDefaultConfig/#418 (0.45s) --- PASS: TestDefaultConfig/#414 (0.02s) --- PASS: TestDefaultConfig/#417 (0.28s) --- PASS: TestDefaultConfig/#415 (0.09s) --- PASS: TestDefaultConfig/#416 (0.27s) --- PASS: TestDefaultConfig/#413 (0.01s) --- PASS: TestDefaultConfig/#409 (0.01s) --- PASS: TestDefaultConfig/#411 (0.02s) --- PASS: TestDefaultConfig/#408 (0.01s) --- PASS: TestDefaultConfig/#412 (0.03s) --- PASS: TestDefaultConfig/#406 (0.01s) --- PASS: TestDefaultConfig/#410 (0.03s) --- PASS: TestDefaultConfig/#407 (0.02s) --- PASS: TestDefaultConfig/#404 (0.05s) --- PASS: TestDefaultConfig/#405 (0.06s) --- PASS: TestDefaultConfig/#403 (0.09s) --- PASS: TestDefaultConfig/#402 (0.12s) --- PASS: TestDefaultConfig/#401 (0.15s) --- PASS: TestDefaultConfig/#400 (0.20s) --- PASS: TestDefaultConfig/#398 (0.24s) --- PASS: TestDefaultConfig/#399 (0.27s) --- PASS: TestDefaultConfig/#397 (0.18s) --- PASS: TestDefaultConfig/#395 (0.03s) --- PASS: TestDefaultConfig/#383 (0.01s) --- PASS: TestDefaultConfig/#396 (0.14s) --- PASS: TestDefaultConfig/#394 (0.02s) --- PASS: TestDefaultConfig/#381 (0.01s) --- PASS: TestDefaultConfig/#380 (0.02s) --- PASS: TestDefaultConfig/#382 (0.03s) --- PASS: TestDefaultConfig/#378 (0.03s) --- PASS: TestDefaultConfig/#379 (0.04s) --- PASS: TestDefaultConfig/#376 (0.02s) --- PASS: TestDefaultConfig/#375 (0.03s) --- PASS: TestDefaultConfig/#377 (0.05s) --- PASS: TestDefaultConfig/#374 (0.03s) --- PASS: TestDefaultConfig/#372 (0.06s) --- PASS: TestDefaultConfig/#373 (0.08s) --- PASS: TestDefaultConfig/#243 (0.18s) --- PASS: TestDefaultConfig/#371 (0.33s) --- PASS: TestDefaultConfig/#370 (0.30s) --- PASS: TestDefaultConfig/#366 (0.01s) --- PASS: TestDefaultConfig/#369 (0.31s) --- PASS: TestDefaultConfig/#367 (0.02s) --- PASS: TestDefaultConfig/#368 (0.15s) --- PASS: TestDefaultConfig/#364 (0.03s) --- PASS: TestDefaultConfig/#362 (0.02s) --- PASS: TestDefaultConfig/#365 (0.03s) --- PASS: TestDefaultConfig/#363 (0.03s) --- PASS: TestDefaultConfig/#361 (0.02s) --- PASS: TestDefaultConfig/#360 (0.02s) --- PASS: TestDefaultConfig/#357 (0.02s) --- PASS: TestDefaultConfig/#359 (0.04s) --- PASS: TestDefaultConfig/#358 (0.05s) --- PASS: TestDefaultConfig/#355 (0.08s) --- PASS: TestDefaultConfig/#356 (0.12s) --- PASS: TestDefaultConfig/#354 (0.10s) --- PASS: TestDefaultConfig/#353 (0.10s) --- PASS: TestDefaultConfig/#352 (0.07s) --- PASS: TestDefaultConfig/#351 (0.04s) --- PASS: TestDefaultConfig/#497 (0.14s) --- PASS: TestDefaultConfig/#495 (0.10s) --- PASS: TestDefaultConfig/#493 (0.01s) --- PASS: TestDefaultConfig/#494 (0.02s) --- PASS: TestDefaultConfig/#491 (0.01s) --- PASS: TestDefaultConfig/#496 (0.13s) --- PASS: TestDefaultConfig/#350 (0.13s) --- PASS: TestDefaultConfig/#489 (0.02s) --- PASS: TestDefaultConfig/#492 (0.03s) --- PASS: TestDefaultConfig/#490 (0.03s) --- PASS: TestDefaultConfig/#486 (0.01s) --- PASS: TestDefaultConfig/#488 (0.03s) --- PASS: TestDefaultConfig/#483 (0.01s) --- PASS: TestDefaultConfig/#484 (0.02s) --- PASS: TestDefaultConfig/#485 (0.03s) --- PASS: TestDefaultConfig/#480 (0.06s) --- PASS: TestDefaultConfig/#487 (0.11s) --- PASS: TestDefaultConfig/#481 (0.08s) --- PASS: TestDefaultConfig/#482 (0.11s) --- PASS: TestDefaultConfig/#479 (0.27s) --- PASS: TestDefaultConfig/#477 (0.25s) --- PASS: TestDefaultConfig/#478 (0.27s) --- PASS: TestDefaultConfig/#476 (0.22s) --- PASS: TestDefaultConfig/#474 (0.04s) --- PASS: TestDefaultConfig/#473 (0.03s) --- PASS: TestDefaultConfig/#471 (0.01s) --- PASS: TestDefaultConfig/#475 (0.07s) --- PASS: TestDefaultConfig/#472 (0.04s) --- PASS: TestDefaultConfig/#469 (0.02s) --- PASS: TestDefaultConfig/#466 (0.01s) --- PASS: TestDefaultConfig/#470 (0.04s) --- PASS: TestDefaultConfig/#465 (0.01s) --- PASS: TestDefaultConfig/#463 (0.01s) --- PASS: TestDefaultConfig/#462 (0.01s) --- PASS: TestDefaultConfig/#468 (0.05s) --- PASS: TestDefaultConfig/#467 (0.05s) --- PASS: TestDefaultConfig/#464 (0.09s) --- PASS: TestDefaultConfig/#461 (0.07s) --- PASS: TestDefaultConfig/#460 (0.07s) --- PASS: TestDefaultConfig/#459 (0.16s) --- PASS: TestDefaultConfig/#457 (0.18s) --- PASS: TestDefaultConfig/#456 (0.18s) --- PASS: TestDefaultConfig/#458 (0.20s) --- PASS: TestDefaultConfig/#453 (0.01s) --- PASS: TestDefaultConfig/#455 (0.09s) --- PASS: TestDefaultConfig/#451 (0.01s) --- PASS: TestDefaultConfig/#454 (0.02s) --- PASS: TestDefaultConfig/#452 (0.02s) --- PASS: TestDefaultConfig/#449 (0.01s) --- PASS: TestDefaultConfig/#450 (0.02s) --- PASS: TestDefaultConfig/#446 (0.01s) --- PASS: TestDefaultConfig/#448 (0.02s) --- PASS: TestDefaultConfig/#444 (0.01s) --- PASS: TestDefaultConfig/#447 (0.02s) --- PASS: TestDefaultConfig/#442 (0.01s) --- PASS: TestDefaultConfig/#445 (0.04s) --- PASS: TestDefaultConfig/#443 (0.06s) --- PASS: TestDefaultConfig/#441 (0.06s) --- PASS: TestDefaultConfig/#440 (0.05s) --- PASS: TestDefaultConfig/#439 (0.07s) --- PASS: TestDefaultConfig/#436 (0.16s) --- PASS: TestDefaultConfig/#437 (0.17s) --- PASS: TestDefaultConfig/#435 (0.13s) --- PASS: TestDefaultConfig/#393 (0.01s) --- PASS: TestDefaultConfig/#391 (0.01s) --- PASS: TestDefaultConfig/#438 (0.17s) --- PASS: TestDefaultConfig/#389 (0.01s) --- PASS: TestDefaultConfig/#390 (0.01s) --- PASS: TestDefaultConfig/#387 (0.01s) --- PASS: TestDefaultConfig/#392 (0.03s) --- PASS: TestDefaultConfig/#388 (0.02s) --- PASS: TestDefaultConfig/#386 (0.01s) === RUN TestTxnEndpoint_Bad_JSON === PAUSE TestTxnEndpoint_Bad_JSON === RUN TestTxnEndpoint_Bad_Size_Item === PAUSE TestTxnEndpoint_Bad_Size_Item === RUN TestTxnEndpoint_Bad_Size_Net === PAUSE TestTxnEndpoint_Bad_Size_Net === RUN TestTxnEndpoint_Bad_Size_Ops === PAUSE TestTxnEndpoint_Bad_Size_Ops === RUN TestTxnEndpoint_KV_Actions === PAUSE TestTxnEndpoint_KV_Actions === RUN TestTxnEndpoint_UpdateCheck === PAUSE TestTxnEndpoint_UpdateCheck === RUN TestUiIndex === PAUSE TestUiIndex === RUN TestUiNodes === PAUSE TestUiNodes === RUN TestUiNodes_Filter === PAUSE TestUiNodes_Filter === RUN TestUiNodeInfo === PAUSE TestUiNodeInfo === RUN TestUiServices === PAUSE TestUiServices === RUN TestValidateUserEventParams === PAUSE TestValidateUserEventParams === RUN TestShouldProcessUserEvent === PAUSE TestShouldProcessUserEvent === RUN TestIngestUserEvent === PAUSE TestIngestUserEvent === RUN TestFireReceiveEvent === PAUSE TestFireReceiveEvent === RUN TestUserEventToken === PAUSE TestUserEventToken === RUN TestStringHash === PAUSE TestStringHash === RUN TestSetFilePermissions === PAUSE TestSetFilePermissions === RUN TestDurationFixer --- PASS: TestDurationFixer (0.00s) === RUN TestHelperProcess --- PASS: TestHelperProcess (0.00s) === RUN TestForwardSignals util_test.go:230: DM-skipped --- SKIP: TestForwardSignals (0.00s) === RUN TestMakeWatchHandler === PAUSE TestMakeWatchHandler === RUN TestMakeHTTPWatchHandler === PAUSE TestMakeHTTPWatchHandler === CONT TestStatusPeers === CONT TestUiNodeInfo === CONT TestStatusLeaderSecondary === CONT TestMakeHTTPWatchHandler writer.go:29: 2021-01-29T19:32:52.662Z [TRACE] TestMakeHTTPWatchHandler: http watch handler output: watch=http://127.0.0.1:46391 output="Ok, i see" --- PASS: TestMakeHTTPWatchHandler (0.02s) === CONT TestMakeWatchHandler === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:52.680Z [WARN] TestStatusLeaderSecondary: bootstrap = true: do not enable unless necessary === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:52.685Z [WARN] TestStatusPeers: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:52.687Z [DEBUG] TestStatusPeers.tlsutil: Update: version=1 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:52.685Z [DEBUG] TestStatusLeaderSecondary.tlsutil: Update: version=1 === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:52.709Z [DEBUG] TestStatusPeers.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:52.719Z [WARN] TestUiNodeInfo: bootstrap = true: do not enable unless necessary === CONT TestMakeWatchHandler writer.go:29: 2021-01-29T19:32:52.723Z [DEBUG] TestMakeWatchHandler: watch handler output: watch_handler="bash -c 'echo $CONSUL_INDEX >> handler_index_out && cat >> handler_out'" output= --- PASS: TestMakeWatchHandler (0.06s) === CONT TestSetFilePermissions --- PASS: TestSetFilePermissions (0.00s) === CONT TestStringHash --- PASS: TestStringHash (0.00s) === CONT TestUserEventToken === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:52.724Z [DEBUG] TestUiNodeInfo.tlsutil: Update: version=1 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:52.725Z [DEBUG] TestStatusLeaderSecondary.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:52.729Z [DEBUG] TestUiNodeInfo.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:52.739Z [INFO] TestStatusLeaderSecondary.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3d32d85f-168b-a9bd-f41d-f0ed2ba74472 Address:127.0.0.1:30221}]" writer.go:29: 2021-01-29T19:32:52.742Z [INFO] TestStatusLeaderSecondary.server.raft: entering follower state: follower="Node at 127.0.0.1:30221 [Follower]" leader= === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:52.744Z [INFO] TestUiNodeInfo.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9d296f67-d0d9-9218-7f8c-9736c00b8b27 Address:127.0.0.1:30227}]" === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:52.746Z [INFO] TestStatusLeaderSecondary.server.serf.wan: serf: EventMemberJoin: Node-3d32d85f-168b-a9bd-f41d-f0ed2ba74472.primary 127.0.0.1 === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:52.747Z [INFO] TestUiNodeInfo.server.serf.wan: serf: EventMemberJoin: Node-9d296f67-d0d9-9218-7f8c-9736c00b8b27.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:52.778Z [INFO] TestUiNodeInfo.server.serf.lan: serf: EventMemberJoin: Node-9d296f67-d0d9-9218-7f8c-9736c00b8b27 127.0.0.1 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:52.783Z [INFO] TestStatusLeaderSecondary.server.serf.lan: serf: EventMemberJoin: Node-3d32d85f-168b-a9bd-f41d-f0ed2ba74472 127.0.0.1 writer.go:29: 2021-01-29T19:32:52.790Z [INFO] TestStatusLeaderSecondary: Started DNS server: address=127.0.0.1:30216 network=udp === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:52.784Z [INFO] TestUiNodeInfo: Started DNS server: address=127.0.0.1:30222 network=udp === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:52.760Z [WARN] TestUserEventToken: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:32:52.942Z [WARN] TestUserEventToken: bootstrap = true: do not enable unless necessary === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:52.765Z [INFO] TestStatusPeers.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b91e0612-61ed-4065-2752-819518a989d8 Address:127.0.0.1:30215}]" === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:52.821Z [WARN] TestStatusLeaderSecondary.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:52.974Z [INFO] TestStatusLeaderSecondary.server.raft: entering candidate state: node="Node at 127.0.0.1:30221 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:52.822Z [INFO] TestStatusLeaderSecondary.server: Adding LAN server: server="Node-3d32d85f-168b-a9bd-f41d-f0ed2ba74472 (Addr: tcp/127.0.0.1:30221) (DC: primary)" writer.go:29: 2021-01-29T19:32:52.823Z [INFO] TestStatusLeaderSecondary.server: Handled event for server in area: event=member-join server=Node-3d32d85f-168b-a9bd-f41d-f0ed2ba74472.primary area=wan writer.go:29: 2021-01-29T19:32:52.953Z [INFO] TestStatusLeaderSecondary: Started DNS server: address=127.0.0.1:30216 network=tcp === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:52.749Z [INFO] TestUiNodeInfo.server.raft: entering follower state: follower="Node at 127.0.0.1:30227 [Follower]" leader= writer.go:29: 2021-01-29T19:32:52.938Z [INFO] TestUiNodeInfo.server: Adding LAN server: server="Node-9d296f67-d0d9-9218-7f8c-9736c00b8b27 (Addr: tcp/127.0.0.1:30227) (DC: dc1)" writer.go:29: 2021-01-29T19:32:52.940Z [INFO] TestUiNodeInfo.server: Handled event for server in area: event=member-join server=Node-9d296f67-d0d9-9218-7f8c-9736c00b8b27.dc1 area=wan writer.go:29: 2021-01-29T19:32:52.993Z [INFO] TestUiNodeInfo: Started DNS server: address=127.0.0.1:30222 network=tcp === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.048Z [INFO] TestStatusPeers.server.raft: entering follower state: follower="Node at 127.0.0.1:30215 [Follower]" leader= === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.064Z [DEBUG] TestStatusLeaderSecondary.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:53.065Z [DEBUG] TestStatusLeaderSecondary.server.raft: vote granted: from=3d32d85f-168b-a9bd-f41d-f0ed2ba74472 term=2 tally=1 writer.go:29: 2021-01-29T19:32:53.066Z [INFO] TestStatusLeaderSecondary.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:53.068Z [INFO] TestStatusLeaderSecondary.server.raft: entering leader state: leader="Node at 127.0.0.1:30221 [Leader]" writer.go:29: 2021-01-29T19:32:53.070Z [INFO] TestStatusLeaderSecondary.server: cluster leadership acquired === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.076Z [DEBUG] TestUserEventToken.tlsutil: Update: version=1 === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.077Z [INFO] TestStatusPeers.server.serf.wan: serf: EventMemberJoin: Node-b91e0612-61ed-4065-2752-819518a989d8.dc1 127.0.0.1 === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.080Z [DEBUG] TestUserEventToken.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.087Z [INFO] TestStatusPeers.server.serf.lan: serf: EventMemberJoin: Node-b91e0612-61ed-4065-2752-819518a989d8 127.0.0.1 === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.080Z [WARN] TestUiNodeInfo.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:53.101Z [INFO] TestUiNodeInfo.server.raft: entering candidate state: node="Node at 127.0.0.1:30227 [Candidate]" term=2 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.104Z [INFO] TestStatusLeaderSecondary: Started HTTP server: address=127.0.0.1:30217 network=tcp writer.go:29: 2021-01-29T19:32:53.105Z [INFO] TestStatusLeaderSecondary: started state syncer === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.107Z [INFO] TestStatusPeers.server: Handled event for server in area: event=member-join server=Node-b91e0612-61ed-4065-2752-819518a989d8.dc1 area=wan writer.go:29: 2021-01-29T19:32:53.109Z [INFO] TestStatusPeers.server: Adding LAN server: server="Node-b91e0612-61ed-4065-2752-819518a989d8 (Addr: tcp/127.0.0.1:30215) (DC: dc1)" === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.111Z [INFO] TestStatusLeaderSecondary.server: New leader elected: payload=Node-3d32d85f-168b-a9bd-f41d-f0ed2ba74472 === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.110Z [INFO] TestStatusPeers: Started DNS server: address=127.0.0.1:30210 network=tcp === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.112Z [INFO] TestUiNodeInfo: Started HTTP server: address=127.0.0.1:30223 network=tcp === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.112Z [INFO] TestStatusPeers: Started DNS server: address=127.0.0.1:30210 network=udp === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.112Z [INFO] TestUiNodeInfo: started state syncer === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.114Z [WARN] TestStatusPeers.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.115Z [DEBUG] TestUiNodeInfo.server.raft: votes: needed=1 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.116Z [DEBUG] TestStatusLeaderSecondary.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30221 === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.116Z [DEBUG] TestUiNodeInfo.server.raft: vote granted: from=9d296f67-d0d9-9218-7f8c-9736c00b8b27 term=2 tally=1 writer.go:29: 2021-01-29T19:32:53.121Z [INFO] TestUiNodeInfo.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:53.122Z [INFO] TestUiNodeInfo.server.raft: entering leader state: leader="Node at 127.0.0.1:30227 [Leader]" writer.go:29: 2021-01-29T19:32:53.123Z [INFO] TestUiNodeInfo.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:53.125Z [INFO] TestUiNodeInfo.server: New leader elected: payload=Node-9d296f67-d0d9-9218-7f8c-9736c00b8b27 === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.115Z [INFO] TestStatusPeers.server.raft: entering candidate state: node="Node at 127.0.0.1:30215 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:53.114Z [INFO] TestStatusPeers: Started HTTP server: address=127.0.0.1:30211 network=tcp === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.132Z [INFO] TestStatusLeaderSecondary: Synced node info === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.133Z [INFO] TestStatusPeers: started state syncer === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.133Z [DEBUG] TestStatusLeaderSecondary: Node info in sync === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.135Z [DEBUG] TestStatusPeers.server.raft: votes: needed=1 === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.135Z [INFO] TestUserEventToken.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:475c5575-ea7c-05b3-c864-32fc8533c186 Address:127.0.0.1:30233}]" === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.136Z [DEBUG] TestStatusPeers.server.raft: vote granted: from=b91e0612-61ed-4065-2752-819518a989d8 term=2 tally=1 writer.go:29: 2021-01-29T19:32:53.137Z [INFO] TestStatusPeers.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:53.138Z [INFO] TestStatusPeers.server.raft: entering leader state: leader="Node at 127.0.0.1:30215 [Leader]" === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.137Z [INFO] TestUserEventToken.server.serf.wan: serf: EventMemberJoin: Node-475c5575-ea7c-05b3-c864-32fc8533c186.dc1 127.0.0.1 === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.139Z [INFO] TestStatusPeers.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:53.140Z [INFO] TestStatusPeers.server: New leader elected: payload=Node-b91e0612-61ed-4065-2752-819518a989d8 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.142Z [WARN] TestStatusLeaderSecondary: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:53.143Z [DEBUG] TestStatusLeaderSecondary.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:53.144Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.140Z [INFO] TestUserEventToken.server.serf.lan: serf: EventMemberJoin: Node-475c5575-ea7c-05b3-c864-32fc8533c186 127.0.0.1 === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.145Z [DEBUG] TestStatusPeers.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30215 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.144Z [DEBUG] TestStatusLeaderSecondary.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.135Z [DEBUG] TestUiNodeInfo.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30227 === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.143Z [INFO] TestUserEventToken.server.raft: entering follower state: follower="Node at 127.0.0.1:30233 [Follower]" leader= === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.155Z [INFO] TestUiNodeInfo: Synced node info writer.go:29: 2021-01-29T19:32:53.158Z [DEBUG] TestUiNodeInfo: Node info in sync === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.160Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.162Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.151Z [INFO] TestUserEventToken.server: Adding LAN server: server="Node-475c5575-ea7c-05b3-c864-32fc8533c186 (Addr: tcp/127.0.0.1:30233) (DC: dc1)" === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.162Z [INFO] TestStatusLeaderSecondary.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.152Z [INFO] TestUserEventToken.server: Handled event for server in area: event=member-join server=Node-475c5575-ea7c-05b3-c864-32fc8533c186.dc1 area=wan writer.go:29: 2021-01-29T19:32:53.158Z [INFO] TestUserEventToken: Started DNS server: address=127.0.0.1:30228 network=udp === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.165Z [INFO] TestStatusLeaderSecondary.leader: started routine: routine="CA root pruning" === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.167Z [INFO] TestUserEventToken: Started DNS server: address=127.0.0.1:30228 network=tcp === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.167Z [INFO] TestStatusPeers.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.169Z [DEBUG] TestStatusLeaderSecondary.server: Skipping self join check for node since the cluster is too small: node=Node-3d32d85f-168b-a9bd-f41d-f0ed2ba74472 === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.170Z [INFO] TestStatusPeers.leader: started routine: routine="CA root pruning" === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.172Z [INFO] TestUserEventToken: Started HTTP server: address=127.0.0.1:30229 network=tcp === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.172Z [INFO] TestStatusLeaderSecondary.server: member joined, marking health alive: member=Node-3d32d85f-168b-a9bd-f41d-f0ed2ba74472 === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.173Z [DEBUG] TestStatusPeers.server: Skipping self join check for node since the cluster is too small: node=Node-b91e0612-61ed-4065-2752-819518a989d8 === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.173Z [INFO] TestUserEventToken: started state syncer === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.171Z [INFO] TestStatusLeaderSecondary.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bab73fe2-9c92-3a10-5bfc-db4ab2992d4a Address:127.0.0.1:30239}]" === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.176Z [INFO] TestStatusPeers.server: member joined, marking health alive: member=Node-b91e0612-61ed-4065-2752-819518a989d8 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.178Z [INFO] TestStatusLeaderSecondary.server.raft: entering follower state: follower="Node at 127.0.0.1:30239 [Follower]" leader= === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.179Z [INFO] TestUiNodeInfo.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.178Z [INFO] TestStatusLeaderSecondary.server.serf.wan: serf: EventMemberJoin: Node-bab73fe2-9c92-3a10-5bfc-db4ab2992d4a.secondary 127.0.0.1 === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.181Z [INFO] TestUiNodeInfo.leader: started routine: routine="CA root pruning" === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.182Z [INFO] TestStatusLeaderSecondary.server.serf.lan: serf: EventMemberJoin: Node-bab73fe2-9c92-3a10-5bfc-db4ab2992d4a 127.0.0.1 === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.182Z [DEBUG] TestUiNodeInfo.server: Skipping self join check for node since the cluster is too small: node=Node-9d296f67-d0d9-9218-7f8c-9736c00b8b27 writer.go:29: 2021-01-29T19:32:53.184Z [INFO] TestUiNodeInfo.server: member joined, marking health alive: member=Node-9d296f67-d0d9-9218-7f8c-9736c00b8b27 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.185Z [INFO] TestStatusLeaderSecondary: Started DNS server: address=127.0.0.1:30234 network=udp writer.go:29: 2021-01-29T19:32:53.185Z [INFO] TestStatusLeaderSecondary.server: Adding LAN server: server="Node-bab73fe2-9c92-3a10-5bfc-db4ab2992d4a (Addr: tcp/127.0.0.1:30239) (DC: secondary)" writer.go:29: 2021-01-29T19:32:53.186Z [INFO] TestStatusLeaderSecondary.server: Handled event for server in area: event=member-join server=Node-bab73fe2-9c92-3a10-5bfc-db4ab2992d4a.secondary area=wan writer.go:29: 2021-01-29T19:32:53.189Z [INFO] TestStatusLeaderSecondary: Started DNS server: address=127.0.0.1:30234 network=tcp writer.go:29: 2021-01-29T19:32:53.193Z [INFO] TestStatusLeaderSecondary: Started HTTP server: address=127.0.0.1:30235 network=tcp writer.go:29: 2021-01-29T19:32:53.194Z [INFO] TestStatusLeaderSecondary: started state syncer === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.203Z [DEBUG] TestUserEventToken.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:53.208Z [WARN] TestUserEventToken.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:53.211Z [INFO] TestUserEventToken.server.raft: entering candidate state: node="Node at 127.0.0.1:30233 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:53.214Z [DEBUG] TestUserEventToken.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:53.215Z [DEBUG] TestUserEventToken.server.raft: vote granted: from=475c5575-ea7c-05b3-c864-32fc8533c186 term=2 tally=1 writer.go:29: 2021-01-29T19:32:53.217Z [INFO] TestUserEventToken.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:53.218Z [INFO] TestUserEventToken.server.raft: entering leader state: leader="Node at 127.0.0.1:30233 [Leader]" writer.go:29: 2021-01-29T19:32:53.220Z [INFO] TestUserEventToken.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:53.221Z [INFO] TestUserEventToken.server: New leader elected: payload=Node-475c5575-ea7c-05b3-c864-32fc8533c186 writer.go:29: 2021-01-29T19:32:53.222Z [INFO] TestUserEventToken.server: initializing acls writer.go:29: 2021-01-29T19:32:53.224Z [INFO] TestUserEventToken.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:53.225Z [WARN] TestUserEventToken.server: Configuring a non-UUID master token is deprecated === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.229Z [WARN] TestStatusLeaderSecondary.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.230Z [INFO] TestUserEventToken.server: Bootstrapped ACL master token from configuration === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.231Z [INFO] TestStatusLeaderSecondary.server.raft: entering candidate state: node="Node at 127.0.0.1:30239 [Candidate]" term=2 === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.234Z [INFO] TestUserEventToken.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:53.236Z [INFO] TestUserEventToken.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:53.237Z [INFO] TestUserEventToken.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:53.239Z [INFO] TestUserEventToken.server.serf.lan: serf: EventMemberUpdate: Node-475c5575-ea7c-05b3-c864-32fc8533c186 writer.go:29: 2021-01-29T19:32:53.242Z [INFO] TestUserEventToken.server: Updating LAN server: server="Node-475c5575-ea7c-05b3-c864-32fc8533c186 (Addr: tcp/127.0.0.1:30233) (DC: dc1)" writer.go:29: 2021-01-29T19:32:53.243Z [INFO] TestUserEventToken.server.serf.wan: serf: EventMemberUpdate: Node-475c5575-ea7c-05b3-c864-32fc8533c186.dc1 writer.go:29: 2021-01-29T19:32:53.245Z [INFO] TestUserEventToken.server: Handled event for server in area: event=member-update server=Node-475c5575-ea7c-05b3-c864-32fc8533c186.dc1 area=wan === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.247Z [DEBUG] TestStatusLeaderSecondary.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:53.248Z [DEBUG] TestStatusLeaderSecondary.server.raft: vote granted: from=bab73fe2-9c92-3a10-5bfc-db4ab2992d4a term=2 tally=1 writer.go:29: 2021-01-29T19:32:53.249Z [INFO] TestStatusLeaderSecondary.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:53.251Z [INFO] TestStatusLeaderSecondary.server.raft: entering leader state: leader="Node at 127.0.0.1:30239 [Leader]" writer.go:29: 2021-01-29T19:32:53.253Z [INFO] TestStatusLeaderSecondary.server: cluster leadership acquired === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.253Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.254Z [INFO] TestStatusLeaderSecondary.server: New leader elected: payload=Node-bab73fe2-9c92-3a10-5bfc-db4ab2992d4a === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.258Z [DEBUG] TestStatusPeers: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:53.261Z [INFO] TestStatusPeers: Synced node info writer.go:29: 2021-01-29T19:32:53.263Z [DEBUG] TestStatusPeers: Node info in sync === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.255Z [DEBUG] TestStatusLeaderSecondary.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30239 writer.go:29: 2021-01-29T19:32:53.269Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.276Z [INFO] TestUserEventToken.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:53.277Z [INFO] TestUserEventToken.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:53.279Z [DEBUG] TestUserEventToken.server: Skipping self join check for node since the cluster is too small: node=Node-475c5575-ea7c-05b3-c864-32fc8533c186 writer.go:29: 2021-01-29T19:32:53.277Z [WARN] TestUserEventToken.server.internal: user event blocked by ACLs: event=foo accessorID= writer.go:29: 2021-01-29T19:32:53.282Z [WARN] TestUserEventToken.server.internal: user event blocked by ACLs: event=bar accessorID= writer.go:29: 2021-01-29T19:32:53.281Z [INFO] TestUserEventToken.server: member joined, marking health alive: member=Node-475c5575-ea7c-05b3-c864-32fc8533c186 writer.go:29: 2021-01-29T19:32:53.285Z [WARN] TestUserEventToken.server.internal: user event blocked by ACLs: event=zip accessorID= writer.go:29: 2021-01-29T19:32:53.287Z [INFO] TestUserEventToken: Requesting shutdown writer.go:29: 2021-01-29T19:32:53.285Z [DEBUG] TestUserEventToken.server: User event: event=baz writer.go:29: 2021-01-29T19:32:53.291Z [DEBUG] TestUserEventToken: new event: event_name=baz event_id=1e13c9d3-22a9-719d-243b-ecae0b2a8c74 writer.go:29: 2021-01-29T19:32:53.291Z [INFO] TestUserEventToken.server: shutting down server === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.282Z [INFO] TestStatusLeaderSecondary.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.293Z [DEBUG] TestUserEventToken.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.293Z [INFO] TestStatusLeaderSecondary.leader: started routine: routine="CA root pruning" === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.294Z [DEBUG] TestUserEventToken.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:53.292Z [DEBUG] TestUserEventToken.server: Skipping self join check for node since the cluster is too small: node=Node-475c5575-ea7c-05b3-c864-32fc8533c186 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.312Z [DEBUG] TestStatusLeaderSecondary.server: Skipping self join check for node since the cluster is too small: node=Node-bab73fe2-9c92-3a10-5bfc-db4ab2992d4a writer.go:29: 2021-01-29T19:32:53.313Z [INFO] TestStatusLeaderSecondary.server: member joined, marking health alive: member=Node-bab73fe2-9c92-3a10-5bfc-db4ab2992d4a === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.293Z [ERROR] TestUserEventToken.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:53.294Z [DEBUG] TestUserEventToken.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:53.312Z [DEBUG] TestUserEventToken.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:53.312Z [DEBUG] TestUserEventToken.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:53.317Z [WARN] TestUserEventToken.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:53.317Z [DEBUG] TestUserEventToken.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:53.321Z [WARN] TestUserEventToken.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:53.323Z [INFO] TestUserEventToken.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:53.323Z [INFO] TestUserEventToken: consul server down writer.go:29: 2021-01-29T19:32:53.326Z [INFO] TestUserEventToken: shutdown complete writer.go:29: 2021-01-29T19:32:53.328Z [INFO] TestUserEventToken: Stopping server: protocol=DNS address=127.0.0.1:30228 network=tcp writer.go:29: 2021-01-29T19:32:53.329Z [INFO] TestUserEventToken: Stopping server: protocol=DNS address=127.0.0.1:30228 network=udp writer.go:29: 2021-01-29T19:32:53.330Z [INFO] TestUserEventToken: Stopping server: protocol=HTTP address=127.0.0.1:30229 network=tcp === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.454Z [INFO] TestStatusPeers: Requesting shutdown writer.go:29: 2021-01-29T19:32:53.456Z [INFO] TestStatusPeers.server: shutting down server writer.go:29: 2021-01-29T19:32:53.457Z [DEBUG] TestStatusPeers.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:53.457Z [DEBUG] TestStatusPeers.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:53.457Z [WARN] TestStatusPeers.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:53.460Z [WARN] TestStatusPeers.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:53.462Z [INFO] TestStatusPeers.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:53.462Z [INFO] TestStatusPeers: consul server down writer.go:29: 2021-01-29T19:32:53.464Z [INFO] TestStatusPeers: shutdown complete writer.go:29: 2021-01-29T19:32:53.465Z [INFO] TestStatusPeers: Stopping server: protocol=DNS address=127.0.0.1:30210 network=tcp writer.go:29: 2021-01-29T19:32:53.466Z [INFO] TestStatusPeers: Stopping server: protocol=DNS address=127.0.0.1:30210 network=udp writer.go:29: 2021-01-29T19:32:53.466Z [INFO] TestStatusPeers: Stopping server: protocol=HTTP address=127.0.0.1:30211 network=tcp === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:53.507Z [INFO] TestUiNodeInfo: Requesting shutdown writer.go:29: 2021-01-29T19:32:53.508Z [INFO] TestUiNodeInfo.server: shutting down server writer.go:29: 2021-01-29T19:32:53.509Z [DEBUG] TestUiNodeInfo.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:53.510Z [WARN] TestUiNodeInfo.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:53.510Z [DEBUG] TestUiNodeInfo.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:53.512Z [WARN] TestUiNodeInfo.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:53.516Z [INFO] TestUiNodeInfo.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:53.516Z [INFO] TestUiNodeInfo: consul server down writer.go:29: 2021-01-29T19:32:53.519Z [INFO] TestUiNodeInfo: shutdown complete writer.go:29: 2021-01-29T19:32:53.521Z [INFO] TestUiNodeInfo: Stopping server: protocol=DNS address=127.0.0.1:30222 network=tcp writer.go:29: 2021-01-29T19:32:53.522Z [INFO] TestUiNodeInfo: Stopping server: protocol=DNS address=127.0.0.1:30222 network=udp writer.go:29: 2021-01-29T19:32:53.524Z [INFO] TestUiNodeInfo: Stopping server: protocol=HTTP address=127.0.0.1:30223 network=tcp === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:53.535Z [INFO] TestStatusLeaderSecondary: (WAN) joining: wan_addresses=[127.0.0.1:30220] writer.go:29: 2021-01-29T19:32:53.538Z [DEBUG] TestStatusLeaderSecondary.server.memberlist.wan: memberlist: Initiating push/pull sync with: 127.0.0.1:30220 writer.go:29: 2021-01-29T19:32:53.538Z [DEBUG] TestStatusLeaderSecondary.server.memberlist.wan: memberlist: Stream connection from=127.0.0.1:55798 writer.go:29: 2021-01-29T19:32:53.546Z [INFO] TestStatusLeaderSecondary.server.serf.wan: serf: EventMemberJoin: Node-3d32d85f-168b-a9bd-f41d-f0ed2ba74472.primary 127.0.0.1 writer.go:29: 2021-01-29T19:32:53.549Z [INFO] TestStatusLeaderSecondary: (WAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:32:53.549Z [INFO] TestStatusLeaderSecondary.server: Handled event for server in area: event=member-join server=Node-3d32d85f-168b-a9bd-f41d-f0ed2ba74472.primary area=wan writer.go:29: 2021-01-29T19:32:53.552Z [INFO] TestStatusLeaderSecondary.server.serf.wan: serf: EventMemberJoin: Node-bab73fe2-9c92-3a10-5bfc-db4ab2992d4a.secondary 127.0.0.1 writer.go:29: 2021-01-29T19:32:53.558Z [INFO] TestStatusLeaderSecondary.server: Handled event for server in area: event=member-join server=Node-bab73fe2-9c92-3a10-5bfc-db4ab2992d4a.secondary area=wan writer.go:29: 2021-01-29T19:32:53.560Z [DEBUG] TestStatusLeaderSecondary: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:53.570Z [INFO] TestStatusLeaderSecondary: Synced node info writer.go:29: 2021-01-29T19:32:53.573Z [DEBUG] TestStatusLeaderSecondary: Node info in sync writer.go:29: 2021-01-29T19:32:53.585Z [DEBUG] TestStatusLeaderSecondary.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:53.594Z [DEBUG] TestStatusLeaderSecondary.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:53.601Z [INFO] TestStatusLeaderSecondary: Requesting shutdown writer.go:29: 2021-01-29T19:32:53.602Z [INFO] TestStatusLeaderSecondary.server: shutting down server writer.go:29: 2021-01-29T19:32:53.604Z [DEBUG] TestStatusLeaderSecondary.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:53.605Z [WARN] TestStatusLeaderSecondary.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:53.606Z [DEBUG] TestStatusLeaderSecondary.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:53.609Z [WARN] TestStatusLeaderSecondary.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:53.612Z [INFO] TestStatusLeaderSecondary.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:53.612Z [INFO] TestStatusLeaderSecondary.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:53.613Z [INFO] TestStatusLeaderSecondary: consul server down writer.go:29: 2021-01-29T19:32:53.615Z [INFO] TestStatusLeaderSecondary: shutdown complete writer.go:29: 2021-01-29T19:32:53.616Z [INFO] TestStatusLeaderSecondary: Stopping server: protocol=DNS address=127.0.0.1:30234 network=tcp writer.go:29: 2021-01-29T19:32:53.617Z [INFO] TestStatusLeaderSecondary: Stopping server: protocol=DNS address=127.0.0.1:30234 network=udp writer.go:29: 2021-01-29T19:32:53.619Z [INFO] TestStatusLeaderSecondary: Stopping server: protocol=HTTP address=127.0.0.1:30235 network=tcp === CONT TestUserEventToken writer.go:29: 2021-01-29T19:32:53.831Z [INFO] TestUserEventToken: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:53.832Z [INFO] TestUserEventToken: Endpoints down --- PASS: TestUserEventToken (1.11s) === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:53.853Z [WARN] TestFireReceiveEvent: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:53.855Z [DEBUG] TestFireReceiveEvent.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:53.856Z [DEBUG] TestFireReceiveEvent.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:53.863Z [INFO] TestFireReceiveEvent.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9c5374b5-9035-55c3-c774-d8561fc1ca0f Address:127.0.0.1:30245}]" writer.go:29: 2021-01-29T19:32:53.865Z [INFO] TestFireReceiveEvent.server.raft: entering follower state: follower="Node at 127.0.0.1:30245 [Follower]" leader= writer.go:29: 2021-01-29T19:32:53.867Z [INFO] TestFireReceiveEvent.server.serf.wan: serf: EventMemberJoin: Node-9c5374b5-9035-55c3-c774-d8561fc1ca0f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:53.869Z [INFO] TestFireReceiveEvent.server.serf.lan: serf: EventMemberJoin: Node-9c5374b5-9035-55c3-c774-d8561fc1ca0f 127.0.0.1 writer.go:29: 2021-01-29T19:32:53.872Z [INFO] TestFireReceiveEvent.server: Adding LAN server: server="Node-9c5374b5-9035-55c3-c774-d8561fc1ca0f (Addr: tcp/127.0.0.1:30245) (DC: dc1)" writer.go:29: 2021-01-29T19:32:53.873Z [INFO] TestFireReceiveEvent.server: Handled event for server in area: event=member-join server=Node-9c5374b5-9035-55c3-c774-d8561fc1ca0f.dc1 area=wan writer.go:29: 2021-01-29T19:32:53.874Z [INFO] TestFireReceiveEvent: Started DNS server: address=127.0.0.1:30240 network=udp writer.go:29: 2021-01-29T19:32:53.876Z [INFO] TestFireReceiveEvent: Started DNS server: address=127.0.0.1:30240 network=tcp writer.go:29: 2021-01-29T19:32:53.877Z [INFO] TestFireReceiveEvent: Started HTTP server: address=127.0.0.1:30241 network=tcp writer.go:29: 2021-01-29T19:32:53.878Z [INFO] TestFireReceiveEvent: started state syncer writer.go:29: 2021-01-29T19:32:53.936Z [WARN] TestFireReceiveEvent.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:53.937Z [INFO] TestFireReceiveEvent.server.raft: entering candidate state: node="Node at 127.0.0.1:30245 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:53.939Z [DEBUG] TestFireReceiveEvent.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:53.940Z [DEBUG] TestFireReceiveEvent.server.raft: vote granted: from=9c5374b5-9035-55c3-c774-d8561fc1ca0f term=2 tally=1 writer.go:29: 2021-01-29T19:32:53.941Z [INFO] TestFireReceiveEvent.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:53.941Z [INFO] TestFireReceiveEvent.server.raft: entering leader state: leader="Node at 127.0.0.1:30245 [Leader]" writer.go:29: 2021-01-29T19:32:53.942Z [INFO] TestFireReceiveEvent.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:53.943Z [INFO] TestFireReceiveEvent.server: New leader elected: payload=Node-9c5374b5-9035-55c3-c774-d8561fc1ca0f writer.go:29: 2021-01-29T19:32:53.952Z [DEBUG] TestFireReceiveEvent.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30245 writer.go:29: 2021-01-29T19:32:53.956Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:53.967Z [INFO] TestFireReceiveEvent.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestStatusPeers writer.go:29: 2021-01-29T19:32:53.968Z [INFO] TestStatusPeers: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:53.969Z [INFO] TestStatusPeers: Endpoints down --- PASS: TestStatusPeers (1.33s) === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:53.976Z [WARN] TestIngestUserEvent: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:53.977Z [DEBUG] TestIngestUserEvent.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:53.978Z [DEBUG] TestIngestUserEvent.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:53.968Z [INFO] TestFireReceiveEvent.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:53.980Z [DEBUG] TestFireReceiveEvent.server: Skipping self join check for node since the cluster is too small: node=Node-9c5374b5-9035-55c3-c774-d8561fc1ca0f writer.go:29: 2021-01-29T19:32:53.981Z [INFO] TestFireReceiveEvent.server: member joined, marking health alive: member=Node-9c5374b5-9035-55c3-c774-d8561fc1ca0f === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:53.984Z [INFO] TestIngestUserEvent.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7e4f16e6-4509-7f0b-ad40-5593a9e17e0f Address:127.0.0.1:30251}]" writer.go:29: 2021-01-29T19:32:53.986Z [INFO] TestIngestUserEvent.server.serf.wan: serf: EventMemberJoin: Node-7e4f16e6-4509-7f0b-ad40-5593a9e17e0f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:53.986Z [INFO] TestIngestUserEvent.server.raft: entering follower state: follower="Node at 127.0.0.1:30251 [Follower]" leader= writer.go:29: 2021-01-29T19:32:53.989Z [INFO] TestIngestUserEvent.server.serf.lan: serf: EventMemberJoin: Node-7e4f16e6-4509-7f0b-ad40-5593a9e17e0f 127.0.0.1 writer.go:29: 2021-01-29T19:32:53.994Z [INFO] TestIngestUserEvent.server: Adding LAN server: server="Node-7e4f16e6-4509-7f0b-ad40-5593a9e17e0f (Addr: tcp/127.0.0.1:30251) (DC: dc1)" writer.go:29: 2021-01-29T19:32:53.994Z [INFO] TestIngestUserEvent.server: Handled event for server in area: event=member-join server=Node-7e4f16e6-4509-7f0b-ad40-5593a9e17e0f.dc1 area=wan writer.go:29: 2021-01-29T19:32:53.997Z [INFO] TestIngestUserEvent: Started DNS server: address=127.0.0.1:30246 network=udp writer.go:29: 2021-01-29T19:32:53.998Z [INFO] TestIngestUserEvent: Started DNS server: address=127.0.0.1:30246 network=tcp writer.go:29: 2021-01-29T19:32:53.999Z [INFO] TestIngestUserEvent: Started HTTP server: address=127.0.0.1:30247 network=tcp writer.go:29: 2021-01-29T19:32:54.000Z [INFO] TestIngestUserEvent: started state syncer === CONT TestUiNodeInfo writer.go:29: 2021-01-29T19:32:54.025Z [INFO] TestUiNodeInfo: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:54.026Z [INFO] TestUiNodeInfo: Endpoints down --- PASS: TestUiNodeInfo (1.38s) === CONT TestShouldProcessUserEvent === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.043Z [WARN] TestIngestUserEvent.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:54.044Z [INFO] TestIngestUserEvent.server.raft: entering candidate state: node="Node at 127.0.0.1:30251 [Candidate]" term=2 === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.044Z [WARN] TestShouldProcessUserEvent: bootstrap = true: do not enable unless necessary === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.046Z [DEBUG] TestIngestUserEvent.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:54.047Z [DEBUG] TestIngestUserEvent.server.raft: vote granted: from=7e4f16e6-4509-7f0b-ad40-5593a9e17e0f term=2 tally=1 === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.046Z [DEBUG] TestShouldProcessUserEvent.tlsutil: Update: version=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.048Z [INFO] TestIngestUserEvent.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:54.049Z [INFO] TestIngestUserEvent.server.raft: entering leader state: leader="Node at 127.0.0.1:30251 [Leader]" writer.go:29: 2021-01-29T19:32:54.050Z [INFO] TestIngestUserEvent.server: cluster leadership acquired === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.049Z [DEBUG] TestShouldProcessUserEvent.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.051Z [INFO] TestIngestUserEvent.server: New leader elected: payload=Node-7e4f16e6-4509-7f0b-ad40-5593a9e17e0f writer.go:29: 2021-01-29T19:32:54.053Z [DEBUG] TestIngestUserEvent.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30251 === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.059Z [INFO] TestShouldProcessUserEvent.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a5740f99-651d-7b7e-24b9-0a19c78d9d32 Address:127.0.0.1:30257}]" writer.go:29: 2021-01-29T19:32:54.062Z [INFO] TestShouldProcessUserEvent.server.serf.wan: serf: EventMemberJoin: Node-a5740f99-651d-7b7e-24b9-0a19c78d9d32.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:54.062Z [INFO] TestShouldProcessUserEvent.server.raft: entering follower state: follower="Node at 127.0.0.1:30257 [Follower]" leader= writer.go:29: 2021-01-29T19:32:54.064Z [INFO] TestShouldProcessUserEvent.server.serf.lan: serf: EventMemberJoin: Node-a5740f99-651d-7b7e-24b9-0a19c78d9d32 127.0.0.1 writer.go:29: 2021-01-29T19:32:54.070Z [INFO] TestShouldProcessUserEvent.server: Adding LAN server: server="Node-a5740f99-651d-7b7e-24b9-0a19c78d9d32 (Addr: tcp/127.0.0.1:30257) (DC: dc1)" writer.go:29: 2021-01-29T19:32:54.070Z [INFO] TestShouldProcessUserEvent.server: Handled event for server in area: event=member-join server=Node-a5740f99-651d-7b7e-24b9-0a19c78d9d32.dc1 area=wan writer.go:29: 2021-01-29T19:32:54.070Z [INFO] TestShouldProcessUserEvent: Started DNS server: address=127.0.0.1:30252 network=udp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.062Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.072Z [INFO] TestShouldProcessUserEvent: Started DNS server: address=127.0.0.1:30252 network=tcp writer.go:29: 2021-01-29T19:32:54.074Z [INFO] TestShouldProcessUserEvent: Started HTTP server: address=127.0.0.1:30253 network=tcp writer.go:29: 2021-01-29T19:32:54.076Z [INFO] TestShouldProcessUserEvent: started state syncer === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.081Z [INFO] TestIngestUserEvent.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:54.083Z [INFO] TestIngestUserEvent.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:54.085Z [DEBUG] TestIngestUserEvent.server: Skipping self join check for node since the cluster is too small: node=Node-7e4f16e6-4509-7f0b-ad40-5593a9e17e0f writer.go:29: 2021-01-29T19:32:54.086Z [INFO] TestIngestUserEvent.server: member joined, marking health alive: member=Node-7e4f16e6-4509-7f0b-ad40-5593a9e17e0f === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.111Z [DEBUG] TestFireReceiveEvent.server: User event: event=deploy writer.go:29: 2021-01-29T19:32:54.112Z [DEBUG] TestFireReceiveEvent.server: User event: event=deploy writer.go:29: 2021-01-29T19:32:54.113Z [DEBUG] TestFireReceiveEvent: new event: event_name=deploy event_id=20598aa3-e6fb-f18d-45ce-d56c6b5ee17d === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:54.121Z [INFO] TestStatusLeaderSecondary: Waiting for endpoints to shut down === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.123Z [WARN] TestShouldProcessUserEvent.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:54.123Z [INFO] TestStatusLeaderSecondary: Endpoints down === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.124Z [INFO] TestShouldProcessUserEvent.server.raft: entering candidate state: node="Node at 127.0.0.1:30257 [Candidate]" term=2 === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:54.125Z [INFO] TestStatusLeaderSecondary: Requesting shutdown writer.go:29: 2021-01-29T19:32:54.126Z [INFO] TestStatusLeaderSecondary.server: shutting down server writer.go:29: 2021-01-29T19:32:54.128Z [DEBUG] TestStatusLeaderSecondary.leader: stopping routine: routine="CA root pruning" === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.131Z [DEBUG] TestFireReceiveEvent: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.131Z [DEBUG] TestShouldProcessUserEvent.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:54.132Z [DEBUG] TestShouldProcessUserEvent.server.raft: vote granted: from=a5740f99-651d-7b7e-24b9-0a19c78d9d32 term=2 tally=1 writer.go:29: 2021-01-29T19:32:54.133Z [INFO] TestShouldProcessUserEvent.server.raft: election won: tally=1 === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.133Z [INFO] TestFireReceiveEvent: Synced node info === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.134Z [INFO] TestShouldProcessUserEvent.server.raft: entering leader state: leader="Node at 127.0.0.1:30257 [Leader]" === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.134Z [INFO] TestFireReceiveEvent: Synced service: service=mysql === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.135Z [INFO] TestShouldProcessUserEvent.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:54.136Z [INFO] TestShouldProcessUserEvent.server: New leader elected: payload=Node-a5740f99-651d-7b7e-24b9-0a19c78d9d32 === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.137Z [INFO] TestFireReceiveEvent: Requesting shutdown === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.137Z [DEBUG] TestShouldProcessUserEvent.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30257 === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.137Z [INFO] TestFireReceiveEvent.server: shutting down server writer.go:29: 2021-01-29T19:32:54.138Z [DEBUG] TestFireReceiveEvent.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:54.139Z [WARN] TestFireReceiveEvent.server.serf.lan: serf: Shutdown without a Leave === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.140Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.140Z [DEBUG] TestFireReceiveEvent.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:54.143Z [WARN] TestFireReceiveEvent.server.serf.wan: serf: Shutdown without a Leave === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.144Z [INFO] TestShouldProcessUserEvent.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:54.145Z [INFO] TestShouldProcessUserEvent.leader: started routine: routine="CA root pruning" === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.145Z [INFO] TestFireReceiveEvent.server.router.manager: shutting down === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.145Z [DEBUG] TestShouldProcessUserEvent.server: Skipping self join check for node since the cluster is too small: node=Node-a5740f99-651d-7b7e-24b9-0a19c78d9d32 === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.146Z [INFO] TestFireReceiveEvent: consul server down === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.146Z [INFO] TestShouldProcessUserEvent.server: member joined, marking health alive: member=Node-a5740f99-651d-7b7e-24b9-0a19c78d9d32 === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.147Z [INFO] TestFireReceiveEvent: shutdown complete writer.go:29: 2021-01-29T19:32:54.147Z [INFO] TestFireReceiveEvent: Stopping server: protocol=DNS address=127.0.0.1:30240 network=tcp writer.go:29: 2021-01-29T19:32:54.148Z [INFO] TestFireReceiveEvent: Stopping server: protocol=DNS address=127.0.0.1:30240 network=udp writer.go:29: 2021-01-29T19:32:54.149Z [INFO] TestFireReceiveEvent: Stopping server: protocol=HTTP address=127.0.0.1:30241 network=tcp === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:54.130Z [WARN] TestStatusLeaderSecondary.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:54.132Z [DEBUG] TestStatusLeaderSecondary.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:54.153Z [WARN] TestStatusLeaderSecondary.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:54.155Z [INFO] TestStatusLeaderSecondary.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:54.155Z [INFO] TestStatusLeaderSecondary.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:54.156Z [INFO] TestStatusLeaderSecondary: consul server down writer.go:29: 2021-01-29T19:32:54.158Z [INFO] TestStatusLeaderSecondary: shutdown complete writer.go:29: 2021-01-29T19:32:54.159Z [INFO] TestStatusLeaderSecondary: Stopping server: protocol=DNS address=127.0.0.1:30216 network=tcp writer.go:29: 2021-01-29T19:32:54.160Z [INFO] TestStatusLeaderSecondary: Stopping server: protocol=DNS address=127.0.0.1:30216 network=udp writer.go:29: 2021-01-29T19:32:54.161Z [INFO] TestStatusLeaderSecondary: Stopping server: protocol=HTTP address=127.0.0.1:30217 network=tcp === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.177Z [DEBUG] TestShouldProcessUserEvent: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:54.179Z [INFO] TestShouldProcessUserEvent: Synced node info === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.194Z [DEBUG] TestIngestUserEvent: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:54.196Z [INFO] TestIngestUserEvent: Synced node info writer.go:29: 2021-01-29T19:32:54.309Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.311Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.311Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.312Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.313Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.314Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.315Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.316Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.317Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.317Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.318Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.332Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.335Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.336Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.337Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.338Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.339Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.340Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.341Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.342Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.342Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.343Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.344Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.345Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.346Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.347Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.347Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.348Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.349Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.350Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.351Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.372Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.373Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.374Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.374Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.375Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.376Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.377Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.378Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.379Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.380Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.380Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.402Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.403Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.404Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.405Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.405Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.406Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.407Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.408Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.409Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.410Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.411Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.432Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.433Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.434Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.434Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.435Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.436Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.437Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.438Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.439Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.439Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.440Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.442Z [INFO] TestShouldProcessUserEvent: Requesting shutdown writer.go:29: 2021-01-29T19:32:54.443Z [INFO] TestShouldProcessUserEvent.server: shutting down server writer.go:29: 2021-01-29T19:32:54.445Z [DEBUG] TestShouldProcessUserEvent.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:54.447Z [WARN] TestShouldProcessUserEvent.server.serf.lan: serf: Shutdown without a Leave === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.441Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.451Z [DEBUG] TestShouldProcessUserEvent.leader: stopped routine: routine="CA root pruning" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.454Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.456Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.457Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.453Z [WARN] TestShouldProcessUserEvent.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:54.459Z [INFO] TestShouldProcessUserEvent.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:54.461Z [INFO] TestShouldProcessUserEvent: consul server down writer.go:29: 2021-01-29T19:32:54.464Z [INFO] TestShouldProcessUserEvent: shutdown complete writer.go:29: 2021-01-29T19:32:54.465Z [INFO] TestShouldProcessUserEvent: Stopping server: protocol=DNS address=127.0.0.1:30252 network=tcp writer.go:29: 2021-01-29T19:32:54.466Z [INFO] TestShouldProcessUserEvent: Stopping server: protocol=DNS address=127.0.0.1:30252 network=udp writer.go:29: 2021-01-29T19:32:54.466Z [INFO] TestShouldProcessUserEvent: Stopping server: protocol=HTTP address=127.0.0.1:30253 network=tcp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.458Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.482Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.483Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.484Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.485Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.485Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.486Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.487Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.488Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.489Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.490Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.491Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.512Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.513Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.513Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.514Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.515Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.516Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.517Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.518Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.519Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.519Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.520Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.521Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.542Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.543Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.544Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.544Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.545Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.546Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.547Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.548Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.549Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.550Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.551Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.572Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.573Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.575Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.577Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.577Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.578Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.579Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.580Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.581Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.602Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.603Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.605Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.606Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.608Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.609Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.611Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.636Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.637Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.638Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.639Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.639Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.640Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.641Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.642Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.644Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.646Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.648Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.649Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.650Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.651Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.653Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.654Z [INFO] TestFireReceiveEvent: Waiting for endpoints to shut down === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.655Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestFireReceiveEvent writer.go:29: 2021-01-29T19:32:54.656Z [INFO] TestFireReceiveEvent: Endpoints down --- PASS: TestFireReceiveEvent (0.82s) === CONT TestValidateUserEventParams --- PASS: TestValidateUserEventParams (0.00s) === CONT TestUiServices === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.657Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.659Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.661Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:54.662Z [INFO] TestStatusLeaderSecondary: Waiting for endpoints to shut down === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.663Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestStatusLeaderSecondary writer.go:29: 2021-01-29T19:32:54.664Z [INFO] TestStatusLeaderSecondary: Endpoints down --- PASS: TestStatusLeaderSecondary (2.02s) === CONT TestSnapshot_Options === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.664Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === RUN TestSnapshot_Options/GET === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.666Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.668Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.669Z [WARN] TestUiServices: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:54.670Z [DEBUG] TestUiServices.tlsutil: Update: version=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.670Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.672Z [DEBUG] TestUiServices.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.672Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.674Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.675Z [INFO] TestUiServices.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7551c734-11a1-29ca-7808-0b7465def295 Address:127.0.0.1:30263}]" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.676Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.677Z [INFO] TestUiServices.server.raft: entering follower state: follower="Node at 127.0.0.1:30263 [Follower]" leader= === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.678Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.677Z [INFO] TestUiServices.server.serf.wan: serf: EventMemberJoin: Node-7551c734-11a1-29ca-7808-0b7465def295.dc1 127.0.0.1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.679Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.681Z [INFO] TestUiServices.server.serf.lan: serf: EventMemberJoin: Node-7551c734-11a1-29ca-7808-0b7465def295 127.0.0.1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.681Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.683Z [INFO] TestUiServices.server: Adding LAN server: server="Node-7551c734-11a1-29ca-7808-0b7465def295 (Addr: tcp/127.0.0.1:30263) (DC: dc1)" === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.683Z [WARN] TestSnapshot_Options/GET: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.684Z [INFO] TestUiServices: Started DNS server: address=127.0.0.1:30258 network=udp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.683Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.684Z [INFO] TestUiServices.server: Handled event for server in area: event=member-join server=Node-7551c734-11a1-29ca-7808-0b7465def295.dc1 area=wan writer.go:29: 2021-01-29T19:32:54.685Z [INFO] TestUiServices: Started DNS server: address=127.0.0.1:30258 network=tcp === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.685Z [WARN] TestSnapshot_Options/GET: bootstrap = true: do not enable unless necessary === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.685Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.687Z [INFO] TestUiServices: Started HTTP server: address=127.0.0.1:30259 network=tcp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.687Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.688Z [INFO] TestUiServices: started state syncer === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.687Z [DEBUG] TestSnapshot_Options/GET.tlsutil: Update: version=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.689Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.691Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.690Z [DEBUG] TestSnapshot_Options/GET.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.692Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.694Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.698Z [INFO] TestSnapshot_Options/GET.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5687b7da-aa6b-fa0d-64da-4379355be376 Address:127.0.0.1:30269}]" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.698Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.701Z [INFO] TestSnapshot_Options/GET.server.raft: entering follower state: follower="Node at 127.0.0.1:30269 [Follower]" leader= writer.go:29: 2021-01-29T19:32:54.702Z [INFO] TestSnapshot_Options/GET.server.serf.wan: serf: EventMemberJoin: Node-5687b7da-aa6b-fa0d-64da-4379355be376.dc1 127.0.0.1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.702Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.705Z [INFO] TestSnapshot_Options/GET.server.serf.lan: serf: EventMemberJoin: Node-5687b7da-aa6b-fa0d-64da-4379355be376 127.0.0.1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.705Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.707Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.709Z [INFO] TestSnapshot_Options/GET.server: Adding LAN server: server="Node-5687b7da-aa6b-fa0d-64da-4379355be376 (Addr: tcp/127.0.0.1:30269) (DC: dc1)" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.709Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.710Z [INFO] TestSnapshot_Options/GET.server: Handled event for server in area: event=member-join server=Node-5687b7da-aa6b-fa0d-64da-4379355be376.dc1 area=wan === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.711Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.713Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.714Z [INFO] TestSnapshot_Options/GET: Started DNS server: address=127.0.0.1:30264 network=udp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.715Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.715Z [INFO] TestSnapshot_Options/GET: Started DNS server: address=127.0.0.1:30264 network=tcp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.717Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.718Z [INFO] TestSnapshot_Options/GET: Started HTTP server: address=127.0.0.1:30265 network=tcp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.719Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.721Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.723Z [WARN] TestUiServices.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.723Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.724Z [INFO] TestUiServices.server.raft: entering candidate state: node="Node at 127.0.0.1:30263 [Candidate]" term=2 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.725Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.727Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.727Z [DEBUG] TestUiServices.server.raft: votes: needed=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.729Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.730Z [DEBUG] TestUiServices.server.raft: vote granted: from=7551c734-11a1-29ca-7808-0b7465def295 term=2 tally=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.730Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.721Z [INFO] TestSnapshot_Options/GET: started state syncer === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.732Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.731Z [INFO] TestUiServices.server.raft: election won: tally=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.734Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.734Z [INFO] TestUiServices.server.raft: entering leader state: leader="Node at 127.0.0.1:30263 [Leader]" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.736Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.737Z [INFO] TestUiServices.server: cluster leadership acquired === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.738Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.740Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.742Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.737Z [INFO] TestUiServices.server: New leader elected: payload=Node-7551c734-11a1-29ca-7808-0b7465def295 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.744Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.746Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.743Z [DEBUG] TestUiServices.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30263 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.747Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.749Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.751Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.752Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.753Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.756Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.757Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.759Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.760Z [INFO] TestUiServices.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.761Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.763Z [DEBUG] TestSnapshot_Options/GET.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:54.764Z [WARN] TestSnapshot_Options/GET.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.763Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.764Z [INFO] TestUiServices.leader: started routine: routine="CA root pruning" === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.765Z [INFO] TestSnapshot_Options/GET.server.raft: entering candidate state: node="Node at 127.0.0.1:30269 [Candidate]" term=2 === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.766Z [DEBUG] TestUiServices.server: Skipping self join check for node since the cluster is too small: node=Node-7551c734-11a1-29ca-7808-0b7465def295 === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.767Z [DEBUG] TestSnapshot_Options/GET.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:54.768Z [DEBUG] TestSnapshot_Options/GET.server.raft: vote granted: from=5687b7da-aa6b-fa0d-64da-4379355be376 term=2 tally=1 writer.go:29: 2021-01-29T19:32:54.769Z [INFO] TestSnapshot_Options/GET.server.raft: election won: tally=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.766Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.768Z [INFO] TestUiServices.server: member joined, marking health alive: member=Node-7551c734-11a1-29ca-7808-0b7465def295 === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.770Z [INFO] TestSnapshot_Options/GET.server.raft: entering leader state: leader="Node at 127.0.0.1:30269 [Leader]" writer.go:29: 2021-01-29T19:32:54.770Z [INFO] TestSnapshot_Options/GET.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:54.772Z [INFO] TestSnapshot_Options/GET.server: New leader elected: payload=Node-5687b7da-aa6b-fa0d-64da-4379355be376 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.770Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.772Z [INFO] TestSnapshot_Options/GET.server: initializing acls === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.774Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.776Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.777Z [INFO] TestSnapshot_Options/GET.server: Created ACL 'global-management' policy === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.778Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.779Z [WARN] TestSnapshot_Options/GET.server: Configuring a non-UUID master token is deprecated === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.780Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.779Z [ERROR] TestSnapshot_Options/GET.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.783Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.783Z [INFO] TestSnapshot_Options/GET.server: Bootstrapped ACL master token from configuration === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.784Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.786Z [INFO] TestSnapshot_Options/GET.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:54.787Z [INFO] TestSnapshot_Options/GET.leader: started routine: routine="legacy ACL token upgrade" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.786Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.788Z [INFO] TestSnapshot_Options/GET.leader: started routine: routine="acl token reaping" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.789Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.790Z [INFO] TestSnapshot_Options/GET.server.serf.lan: serf: EventMemberUpdate: Node-5687b7da-aa6b-fa0d-64da-4379355be376 writer.go:29: 2021-01-29T19:32:54.791Z [INFO] TestSnapshot_Options/GET.server.serf.wan: serf: EventMemberUpdate: Node-5687b7da-aa6b-fa0d-64da-4379355be376.dc1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.791Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.792Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.791Z [INFO] TestSnapshot_Options/GET.server: Updating LAN server: server="Node-5687b7da-aa6b-fa0d-64da-4379355be376 (Addr: tcp/127.0.0.1:30269) (DC: dc1)" writer.go:29: 2021-01-29T19:32:54.792Z [INFO] TestSnapshot_Options/GET.server: Handled event for server in area: event=member-update server=Node-5687b7da-aa6b-fa0d-64da-4379355be376.dc1 area=wan === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.794Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.797Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.796Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.798Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.800Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.802Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.804Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.807Z [INFO] TestSnapshot_Options/GET.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.806Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.807Z [INFO] TestSnapshot_Options/GET.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:54.808Z [DEBUG] TestSnapshot_Options/GET.server: Skipping self join check for node since the cluster is too small: node=Node-5687b7da-aa6b-fa0d-64da-4379355be376 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.808Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.809Z [INFO] TestSnapshot_Options/GET.server: member joined, marking health alive: member=Node-5687b7da-aa6b-fa0d-64da-4379355be376 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.810Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.812Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:54.814Z [DEBUG] TestSnapshot_Options/GET.server: Skipping self join check for node since the cluster is too small: node=Node-5687b7da-aa6b-fa0d-64da-4379355be376 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.814Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.816Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.819Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.821Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.823Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.825Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.827Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.829Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.831Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.833Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.836Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.839Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.843Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.847Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.849Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.852Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.854Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.856Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.859Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.862Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.864Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.866Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.869Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.871Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.873Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.875Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.877Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.880Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.882Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.884Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.886Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.889Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.891Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.893Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.895Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.898Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.902Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.905Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === RUN TestUiServices/No_Filter === PAUSE TestUiServices/No_Filter === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.914Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === RUN TestUiServices/Filtered === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.921Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.934Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.938Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.941Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.945Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.948Z [INFO] TestUiServices: Requesting shutdown writer.go:29: 2021-01-29T19:32:54.949Z [INFO] TestUiServices.server: shutting down server === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.948Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.950Z [DEBUG] TestUiServices.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:54.950Z [WARN] TestUiServices.server.serf.lan: serf: Shutdown without a Leave === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.950Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.952Z [ERROR] TestUiServices.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:54.953Z [DEBUG] TestUiServices.leader: stopped routine: routine="CA root pruning" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.952Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.954Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.956Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.958Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.960Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.962Z [WARN] TestUiServices.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:54.963Z [INFO] TestUiServices.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:54.963Z [INFO] TestUiServices: consul server down writer.go:29: 2021-01-29T19:32:54.967Z [INFO] TestUiServices: shutdown complete writer.go:29: 2021-01-29T19:32:54.968Z [INFO] TestUiServices: Stopping server: protocol=DNS address=127.0.0.1:30258 network=tcp === CONT TestShouldProcessUserEvent writer.go:29: 2021-01-29T19:32:54.969Z [INFO] TestShouldProcessUserEvent: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:54.970Z [INFO] TestShouldProcessUserEvent: Endpoints down --- PASS: TestShouldProcessUserEvent (0.94s) === CONT TestSessionDeleteDestroy === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.970Z [INFO] TestUiServices: Stopping server: protocol=DNS address=127.0.0.1:30258 network=udp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.962Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestUiServices writer.go:29: 2021-01-29T19:32:54.971Z [INFO] TestUiServices: Stopping server: protocol=HTTP address=127.0.0.1:30259 network=tcp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.975Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.977Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:54.979Z [WARN] TestSessionDeleteDestroy: bootstrap = true: do not enable unless necessary === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.979Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:54.980Z [DEBUG] TestSessionDeleteDestroy.tlsutil: Update: version=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.981Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.983Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:54.981Z [DEBUG] TestSessionDeleteDestroy.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.985Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.987Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.989Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:54.990Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:54.992Z [INFO] TestSessionDeleteDestroy.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:dd9c5cfb-aa03-0fd7-697d-e977232d99b6 Address:127.0.0.1:30275}]" writer.go:29: 2021-01-29T19:32:54.994Z [INFO] TestSessionDeleteDestroy.server.raft: entering follower state: follower="Node at 127.0.0.1:30275 [Follower]" leader= writer.go:29: 2021-01-29T19:32:54.994Z [INFO] TestSessionDeleteDestroy.server.serf.wan: serf: EventMemberJoin: Node-dd9c5cfb-aa03-0fd7-697d-e977232d99b6.dc1 127.0.0.1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:54.992Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.003Z [INFO] TestSessionDeleteDestroy.server.serf.lan: serf: EventMemberJoin: Node-dd9c5cfb-aa03-0fd7-697d-e977232d99b6 127.0.0.1 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.006Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.014Z [INFO] TestSessionDeleteDestroy.server: Adding LAN server: server="Node-dd9c5cfb-aa03-0fd7-697d-e977232d99b6 (Addr: tcp/127.0.0.1:30275) (DC: dc1)" writer.go:29: 2021-01-29T19:32:55.017Z [INFO] TestSessionDeleteDestroy.server: Handled event for server in area: event=member-join server=Node-dd9c5cfb-aa03-0fd7-697d-e977232d99b6.dc1 area=wan === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.013Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.019Z [INFO] TestSessionDeleteDestroy: Started DNS server: address=127.0.0.1:30270 network=tcp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.021Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.023Z [INFO] TestSessionDeleteDestroy: Started DNS server: address=127.0.0.1:30270 network=udp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.024Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.028Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.029Z [INFO] TestSessionDeleteDestroy: Started HTTP server: address=127.0.0.1:30271 network=tcp writer.go:29: 2021-01-29T19:32:55.035Z [INFO] TestSessionDeleteDestroy: started state syncer === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.033Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.038Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.040Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.042Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.045Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.047Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.050Z [WARN] TestSessionDeleteDestroy.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:55.052Z [INFO] TestSessionDeleteDestroy.server.raft: entering candidate state: node="Node at 127.0.0.1:30275 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:55.054Z [DEBUG] TestSessionDeleteDestroy.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:55.056Z [DEBUG] TestSessionDeleteDestroy.server.raft: vote granted: from=dd9c5cfb-aa03-0fd7-697d-e977232d99b6 term=2 tally=1 writer.go:29: 2021-01-29T19:32:55.058Z [INFO] TestSessionDeleteDestroy.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:55.060Z [INFO] TestSessionDeleteDestroy.server.raft: entering leader state: leader="Node at 127.0.0.1:30275 [Leader]" writer.go:29: 2021-01-29T19:32:55.062Z [INFO] TestSessionDeleteDestroy.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:55.063Z [INFO] TestSessionDeleteDestroy.server: New leader elected: payload=Node-dd9c5cfb-aa03-0fd7-697d-e977232d99b6 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.049Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.069Z [DEBUG] TestSessionDeleteDestroy.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30275 writer.go:29: 2021-01-29T19:32:55.077Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.067Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.079Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.081Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.087Z [INFO] TestSessionDeleteDestroy.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.086Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.090Z [INFO] TestSessionDeleteDestroy.leader: started routine: routine="CA root pruning" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.092Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.093Z [DEBUG] TestSessionDeleteDestroy.server: Skipping self join check for node since the cluster is too small: node=Node-dd9c5cfb-aa03-0fd7-697d-e977232d99b6 writer.go:29: 2021-01-29T19:32:55.097Z [INFO] TestSessionDeleteDestroy.server: member joined, marking health alive: member=Node-dd9c5cfb-aa03-0fd7-697d-e977232d99b6 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.097Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.100Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:55.105Z [DEBUG] TestSnapshot_Options/GET.acl: dropping node from result due to ACLs: node=Node-5687b7da-aa6b-fa0d-64da-4379355be376 === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.104Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:55.112Z [INFO] TestSnapshot_Options/GET: Requesting shutdown writer.go:29: 2021-01-29T19:32:55.113Z [INFO] TestSnapshot_Options/GET.server: shutting down server writer.go:29: 2021-01-29T19:32:55.113Z [DEBUG] TestSnapshot_Options/GET.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:55.114Z [DEBUG] TestSnapshot_Options/GET.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:55.115Z [DEBUG] TestSnapshot_Options/GET.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.116Z [WARN] TestSnapshot_Options/GET.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:55.114Z [DEBUG] TestSnapshot_Options/GET.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:55.115Z [DEBUG] TestSnapshot_Options/GET.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:55.116Z [DEBUG] TestSnapshot_Options/GET.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.120Z [WARN] TestSnapshot_Options/GET.server.serf.wan: serf: Shutdown without a Leave === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.111Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.123Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.125Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:55.125Z [INFO] TestSnapshot_Options/GET.server.router.manager: shutting down === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.126Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:55.125Z [INFO] TestSnapshot_Options/GET: consul server down === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.128Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:55.129Z [INFO] TestSnapshot_Options/GET: shutdown complete === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.130Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:55.130Z [INFO] TestSnapshot_Options/GET: Stopping server: protocol=DNS address=127.0.0.1:30264 network=tcp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.132Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:55.132Z [INFO] TestSnapshot_Options/GET: Stopping server: protocol=DNS address=127.0.0.1:30264 network=udp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.134Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:55.134Z [INFO] TestSnapshot_Options/GET: Stopping server: protocol=HTTP address=127.0.0.1:30265 network=tcp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.136Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.138Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.140Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.141Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.142Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.143Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.144Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.145Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.146Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.147Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.148Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.148Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.149Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.150Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.151Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.152Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.153Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.154Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.155Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.156Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.157Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.158Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.159Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.160Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.161Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.162Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.163Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.164Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.165Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.166Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.167Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.168Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.169Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.170Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.171Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.172Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.173Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.174Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.174Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.175Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.176Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.177Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.178Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.179Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.180Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.182Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.183Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.185Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.186Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.188Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.190Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.191Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.192Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.193Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.194Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.195Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.196Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.197Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.198Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.198Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.199Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.200Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.201Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.203Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.204Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.205Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.206Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.207Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.208Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.208Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.209Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.210Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.211Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.212Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.213Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.214Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.215Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.216Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.217Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.218Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.219Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.220Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.221Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.222Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.221Z [INFO] TestSessionDeleteDestroy: Requesting shutdown === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.223Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.223Z [INFO] TestSessionDeleteDestroy.server: shutting down server === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.223Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.224Z [DEBUG] TestSessionDeleteDestroy.leader: stopping routine: routine="CA root pruning" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.224Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.225Z [WARN] TestSessionDeleteDestroy.server.serf.lan: serf: Shutdown without a Leave === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.225Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.226Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.226Z [ERROR] TestSessionDeleteDestroy.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.227Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.226Z [DEBUG] TestSessionDeleteDestroy.leader: stopped routine: routine="CA root pruning" === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.228Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.228Z [WARN] TestSessionDeleteDestroy.server.serf.wan: serf: Shutdown without a Leave === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.229Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.230Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.231Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.232Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.233Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.234Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.235Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.236Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.236Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.237Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.238Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.239Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.240Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.241Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.242Z [INFO] TestSessionDeleteDestroy.server.router.manager: shutting down === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.242Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.243Z [INFO] TestSessionDeleteDestroy: consul server down === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.243Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.244Z [INFO] TestSessionDeleteDestroy: shutdown complete === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.244Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.245Z [INFO] TestSessionDeleteDestroy: Stopping server: protocol=DNS address=127.0.0.1:30270 network=tcp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.245Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.246Z [INFO] TestSessionDeleteDestroy: Stopping server: protocol=DNS address=127.0.0.1:30270 network=udp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.246Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.247Z [INFO] TestSessionDeleteDestroy: Stopping server: protocol=HTTP address=127.0.0.1:30271 network=tcp === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.247Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.248Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.249Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.250Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.251Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.252Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.253Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.254Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.255Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.255Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.257Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.255Z [DEBUG] TestIngestUserEvent: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:55.258Z [DEBUG] TestIngestUserEvent: Node info in sync writer.go:29: 2021-01-29T19:32:55.259Z [DEBUG] TestIngestUserEvent: Node info in sync writer.go:29: 2021-01-29T19:32:55.258Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.262Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.262Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.263Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.264Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.265Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.266Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.267Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.268Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.269Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.270Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.271Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.272Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.273Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.275Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.276Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.277Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.278Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.279Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.280Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.281Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.282Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.283Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.283Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.284Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.285Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.286Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.287Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.288Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.289Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.290Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.290Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.291Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.292Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.293Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.294Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.295Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.296Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.297Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.300Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.301Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.302Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.303Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.304Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.305Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.305Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.306Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.307Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.308Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.309Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.310Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.311Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.312Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.314Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.315Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.317Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.318Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.320Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.322Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.323Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.324Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.326Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.327Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.329Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.331Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.332Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.333Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.335Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.338Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.340Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.341Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.343Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.344Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.346Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.347Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.349Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.350Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.351Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.352Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.353Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.354Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.355Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.356Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.356Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.357Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.358Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.359Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.361Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.363Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.364Z [DEBUG] TestIngestUserEvent: new event: event_name=test event_id= writer.go:29: 2021-01-29T19:32:55.366Z [INFO] TestIngestUserEvent: Requesting shutdown writer.go:29: 2021-01-29T19:32:55.368Z [INFO] TestIngestUserEvent.server: shutting down server writer.go:29: 2021-01-29T19:32:55.370Z [DEBUG] TestIngestUserEvent.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.371Z [WARN] TestIngestUserEvent.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:55.371Z [DEBUG] TestIngestUserEvent.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.383Z [WARN] TestIngestUserEvent.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:55.387Z [INFO] TestIngestUserEvent.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:55.389Z [INFO] TestIngestUserEvent: consul server down writer.go:29: 2021-01-29T19:32:55.390Z [INFO] TestIngestUserEvent: shutdown complete writer.go:29: 2021-01-29T19:32:55.393Z [INFO] TestIngestUserEvent: Stopping server: protocol=DNS address=127.0.0.1:30246 network=tcp writer.go:29: 2021-01-29T19:32:55.395Z [INFO] TestIngestUserEvent: Stopping server: protocol=DNS address=127.0.0.1:30246 network=udp writer.go:29: 2021-01-29T19:32:55.396Z [INFO] TestIngestUserEvent: Stopping server: protocol=HTTP address=127.0.0.1:30247 network=tcp === CONT TestUiServices writer.go:29: 2021-01-29T19:32:55.475Z [INFO] TestUiServices: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:55.477Z [INFO] TestUiServices: Endpoints down === CONT TestSessionGet === RUN TestSessionGet/#00 writer.go:29: 2021-01-29T19:32:55.506Z [WARN] TestSessionGet/#00: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:55.509Z [DEBUG] TestSessionGet/#00.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:55.513Z [DEBUG] TestSessionGet/#00.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:55.534Z [INFO] TestSessionGet/#00.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1d6c2000-cece-4c2b-88ef-b2b6b06c1d4f Address:127.0.0.1:30281}]" writer.go:29: 2021-01-29T19:32:55.538Z [INFO] TestSessionGet/#00.server.raft: entering follower state: follower="Node at 127.0.0.1:30281 [Follower]" leader= writer.go:29: 2021-01-29T19:32:55.569Z [INFO] TestSessionGet/#00.server.serf.wan: serf: EventMemberJoin: Node-1d6c2000-cece-4c2b-88ef-b2b6b06c1d4f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:55.574Z [INFO] TestSessionGet/#00.server.serf.lan: serf: EventMemberJoin: Node-1d6c2000-cece-4c2b-88ef-b2b6b06c1d4f 127.0.0.1 writer.go:29: 2021-01-29T19:32:55.578Z [INFO] TestSessionGet/#00.server: Adding LAN server: server="Node-1d6c2000-cece-4c2b-88ef-b2b6b06c1d4f (Addr: tcp/127.0.0.1:30281) (DC: dc1)" writer.go:29: 2021-01-29T19:32:55.580Z [INFO] TestSessionGet/#00.server: Handled event for server in area: event=member-join server=Node-1d6c2000-cece-4c2b-88ef-b2b6b06c1d4f.dc1 area=wan writer.go:29: 2021-01-29T19:32:55.584Z [INFO] TestSessionGet/#00: Started DNS server: address=127.0.0.1:30276 network=tcp writer.go:29: 2021-01-29T19:32:55.587Z [INFO] TestSessionGet/#00: Started DNS server: address=127.0.0.1:30276 network=udp writer.go:29: 2021-01-29T19:32:55.590Z [INFO] TestSessionGet/#00: Started HTTP server: address=127.0.0.1:30277 network=tcp writer.go:29: 2021-01-29T19:32:55.591Z [INFO] TestSessionGet/#00: started state syncer writer.go:29: 2021-01-29T19:32:55.608Z [WARN] TestSessionGet/#00.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:55.610Z [INFO] TestSessionGet/#00.server.raft: entering candidate state: node="Node at 127.0.0.1:30281 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:55.612Z [DEBUG] TestSessionGet/#00.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:55.614Z [DEBUG] TestSessionGet/#00.server.raft: vote granted: from=1d6c2000-cece-4c2b-88ef-b2b6b06c1d4f term=2 tally=1 writer.go:29: 2021-01-29T19:32:55.617Z [INFO] TestSessionGet/#00.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:55.619Z [INFO] TestSessionGet/#00.server.raft: entering leader state: leader="Node at 127.0.0.1:30281 [Leader]" writer.go:29: 2021-01-29T19:32:55.621Z [INFO] TestSessionGet/#00.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:55.622Z [INFO] TestSessionGet/#00.server: New leader elected: payload=Node-1d6c2000-cece-4c2b-88ef-b2b6b06c1d4f writer.go:29: 2021-01-29T19:32:55.625Z [DEBUG] TestSessionGet/#00.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30281 writer.go:29: 2021-01-29T19:32:55.631Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestSnapshot_Options/GET writer.go:29: 2021-01-29T19:32:55.636Z [INFO] TestSnapshot_Options/GET: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:55.638Z [INFO] TestSnapshot_Options/GET: Endpoints down === RUN TestSnapshot_Options/GET#01 === CONT TestSessionGet/#00 writer.go:29: 2021-01-29T19:32:55.652Z [INFO] TestSessionGet/#00.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:55.654Z [INFO] TestSessionGet/#00.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.658Z [DEBUG] TestSessionGet/#00.server: Skipping self join check for node since the cluster is too small: node=Node-1d6c2000-cece-4c2b-88ef-b2b6b06c1d4f writer.go:29: 2021-01-29T19:32:55.660Z [INFO] TestSessionGet/#00.server: member joined, marking health alive: member=Node-1d6c2000-cece-4c2b-88ef-b2b6b06c1d4f === CONT TestSnapshot_Options/GET#01 writer.go:29: 2021-01-29T19:32:55.661Z [WARN] TestSnapshot_Options/GET#01: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:32:55.662Z [WARN] TestSnapshot_Options/GET#01: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:55.663Z [DEBUG] TestSnapshot_Options/GET#01.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:55.664Z [DEBUG] TestSnapshot_Options/GET#01.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:55.668Z [INFO] TestSnapshot_Options/GET#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:03c08f9b-9b96-337e-aa1c-746d495e4473 Address:127.0.0.1:30287}]" writer.go:29: 2021-01-29T19:32:55.670Z [INFO] TestSnapshot_Options/GET#01.server.raft: entering follower state: follower="Node at 127.0.0.1:30287 [Follower]" leader= writer.go:29: 2021-01-29T19:32:55.671Z [INFO] TestSnapshot_Options/GET#01.server.serf.wan: serf: EventMemberJoin: Node-03c08f9b-9b96-337e-aa1c-746d495e4473.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:55.675Z [INFO] TestSnapshot_Options/GET#01.server.serf.lan: serf: EventMemberJoin: Node-03c08f9b-9b96-337e-aa1c-746d495e4473 127.0.0.1 writer.go:29: 2021-01-29T19:32:55.678Z [INFO] TestSnapshot_Options/GET#01.server: Adding LAN server: server="Node-03c08f9b-9b96-337e-aa1c-746d495e4473 (Addr: tcp/127.0.0.1:30287) (DC: dc1)" writer.go:29: 2021-01-29T19:32:55.679Z [INFO] TestSnapshot_Options/GET#01.server: Handled event for server in area: event=member-join server=Node-03c08f9b-9b96-337e-aa1c-746d495e4473.dc1 area=wan writer.go:29: 2021-01-29T19:32:55.680Z [INFO] TestSnapshot_Options/GET#01: Started DNS server: address=127.0.0.1:30282 network=udp writer.go:29: 2021-01-29T19:32:55.683Z [INFO] TestSnapshot_Options/GET#01: Started DNS server: address=127.0.0.1:30282 network=tcp writer.go:29: 2021-01-29T19:32:55.684Z [INFO] TestSnapshot_Options/GET#01: Started HTTP server: address=127.0.0.1:30283 network=tcp writer.go:29: 2021-01-29T19:32:55.685Z [INFO] TestSnapshot_Options/GET#01: started state syncer === CONT TestSessionGet/#00 writer.go:29: 2021-01-29T19:32:55.710Z [DEBUG] TestSessionGet/#00: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:55.712Z [INFO] TestSessionGet/#00: Synced node info === CONT TestSnapshot_Options/GET#01 writer.go:29: 2021-01-29T19:32:55.730Z [DEBUG] TestSnapshot_Options/GET#01.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:55.731Z [WARN] TestSnapshot_Options/GET#01.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:55.731Z [INFO] TestSnapshot_Options/GET#01.server.raft: entering candidate state: node="Node at 127.0.0.1:30287 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:55.734Z [DEBUG] TestSnapshot_Options/GET#01.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:55.734Z [DEBUG] TestSnapshot_Options/GET#01.server.raft: vote granted: from=03c08f9b-9b96-337e-aa1c-746d495e4473 term=2 tally=1 writer.go:29: 2021-01-29T19:32:55.735Z [INFO] TestSnapshot_Options/GET#01.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:55.736Z [INFO] TestSnapshot_Options/GET#01.server.raft: entering leader state: leader="Node at 127.0.0.1:30287 [Leader]" writer.go:29: 2021-01-29T19:32:55.738Z [INFO] TestSnapshot_Options/GET#01.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:55.738Z [INFO] TestSnapshot_Options/GET#01.server: New leader elected: payload=Node-03c08f9b-9b96-337e-aa1c-746d495e4473 writer.go:29: 2021-01-29T19:32:55.743Z [INFO] TestSnapshot_Options/GET#01.server: initializing acls writer.go:29: 2021-01-29T19:32:55.746Z [INFO] TestSnapshot_Options/GET#01.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:55.746Z [WARN] TestSnapshot_Options/GET#01.server: Configuring a non-UUID master token is deprecated === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.748Z [INFO] TestSessionDeleteDestroy: Waiting for endpoints to shut down === CONT TestSnapshot_Options/GET#01 writer.go:29: 2021-01-29T19:32:55.748Z [INFO] TestSnapshot_Options/GET#01.server: Bootstrapped ACL master token from configuration === CONT TestSessionDeleteDestroy writer.go:29: 2021-01-29T19:32:55.749Z [INFO] TestSessionDeleteDestroy: Endpoints down --- PASS: TestSessionDeleteDestroy (0.78s) === CONT TestSessionCustomTTL === CONT TestSnapshot_Options/GET#01 writer.go:29: 2021-01-29T19:32:55.751Z [INFO] TestSnapshot_Options/GET#01.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:55.753Z [INFO] TestSnapshot_Options/GET#01.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:55.755Z [INFO] TestSnapshot_Options/GET#01.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:55.756Z [INFO] TestSnapshot_Options/GET#01.server.serf.lan: serf: EventMemberUpdate: Node-03c08f9b-9b96-337e-aa1c-746d495e4473 writer.go:29: 2021-01-29T19:32:55.759Z [INFO] TestSnapshot_Options/GET#01.server: Updating LAN server: server="Node-03c08f9b-9b96-337e-aa1c-746d495e4473 (Addr: tcp/127.0.0.1:30287) (DC: dc1)" writer.go:29: 2021-01-29T19:32:55.760Z [INFO] TestSnapshot_Options/GET#01.server.serf.wan: serf: EventMemberUpdate: Node-03c08f9b-9b96-337e-aa1c-746d495e4473.dc1 writer.go:29: 2021-01-29T19:32:55.762Z [INFO] TestSnapshot_Options/GET#01.server: Handled event for server in area: event=member-update server=Node-03c08f9b-9b96-337e-aa1c-746d495e4473.dc1 area=wan writer.go:29: 2021-01-29T19:32:55.764Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:55.768Z [INFO] TestSnapshot_Options/GET#01.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:55.769Z [INFO] TestSnapshot_Options/GET#01.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.770Z [DEBUG] TestSnapshot_Options/GET#01.server: Skipping self join check for node since the cluster is too small: node=Node-03c08f9b-9b96-337e-aa1c-746d495e4473 writer.go:29: 2021-01-29T19:32:55.770Z [INFO] TestSnapshot_Options/GET#01.server: member joined, marking health alive: member=Node-03c08f9b-9b96-337e-aa1c-746d495e4473 writer.go:29: 2021-01-29T19:32:55.791Z [DEBUG] TestSnapshot_Options/GET#01.server: Skipping self join check for node since the cluster is too small: node=Node-03c08f9b-9b96-337e-aa1c-746d495e4473 writer.go:29: 2021-01-29T19:32:55.792Z [DEBUG] TestSnapshot_Options/GET#01.acl: dropping node from result due to ACLs: node=Node-03c08f9b-9b96-337e-aa1c-746d495e4473 writer.go:29: 2021-01-29T19:32:55.794Z [INFO] TestSnapshot_Options/GET#01: Requesting shutdown writer.go:29: 2021-01-29T19:32:55.795Z [INFO] TestSnapshot_Options/GET#01.server: shutting down server writer.go:29: 2021-01-29T19:32:55.796Z [DEBUG] TestSnapshot_Options/GET#01.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.797Z [DEBUG] TestSnapshot_Options/GET#01.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:55.798Z [DEBUG] TestSnapshot_Options/GET#01.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:55.798Z [WARN] TestSnapshot_Options/GET#01.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:55.800Z [ERROR] TestSnapshot_Options/GET#01.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:55.800Z [WARN] TestSessionCustomTTL: bootstrap = true: do not enable unless necessary === CONT TestSnapshot_Options/GET#01 writer.go:29: 2021-01-29T19:32:55.801Z [DEBUG] TestSnapshot_Options/GET#01.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.801Z [DEBUG] TestSnapshot_Options/GET#01.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:55.802Z [DEBUG] TestSnapshot_Options/GET#01.leader: stopped routine: routine="acl token reaping" === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:55.804Z [DEBUG] TestSessionCustomTTL.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:55.810Z [DEBUG] TestSessionCustomTTL.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestSnapshot_Options/GET#01 writer.go:29: 2021-01-29T19:32:55.842Z [WARN] TestSnapshot_Options/GET#01.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:55.845Z [INFO] TestSnapshot_Options/GET#01.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:55.846Z [INFO] TestSnapshot_Options/GET#01: consul server down writer.go:29: 2021-01-29T19:32:55.846Z [INFO] TestSnapshot_Options/GET#01: shutdown complete writer.go:29: 2021-01-29T19:32:55.847Z [INFO] TestSnapshot_Options/GET#01: Stopping server: protocol=DNS address=127.0.0.1:30282 network=tcp writer.go:29: 2021-01-29T19:32:55.848Z [INFO] TestSnapshot_Options/GET#01: Stopping server: protocol=DNS address=127.0.0.1:30282 network=udp writer.go:29: 2021-01-29T19:32:55.849Z [INFO] TestSnapshot_Options/GET#01: Stopping server: protocol=HTTP address=127.0.0.1:30283 network=tcp === CONT TestSessionGet/#00 writer.go:29: 2021-01-29T19:32:55.852Z [INFO] TestSessionGet/#00: Requesting shutdown writer.go:29: 2021-01-29T19:32:55.856Z [INFO] TestSessionGet/#00.server: shutting down server writer.go:29: 2021-01-29T19:32:55.858Z [DEBUG] TestSessionGet/#00.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.860Z [WARN] TestSessionGet/#00.server.serf.lan: serf: Shutdown without a Leave === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:55.873Z [INFO] TestSessionCustomTTL.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:819f020a-de2e-19ec-f9bc-59ef6fa504fc Address:127.0.0.1:30293}]" === CONT TestSessionGet/#00 writer.go:29: 2021-01-29T19:32:55.863Z [DEBUG] TestSessionGet/#00.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.872Z [WARN] TestSessionGet/#00.server.serf.wan: serf: Shutdown without a Leave === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:55.877Z [INFO] TestSessionCustomTTL.server.serf.wan: serf: EventMemberJoin: Node-819f020a-de2e-19ec-f9bc-59ef6fa504fc.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:55.880Z [INFO] TestSessionCustomTTL.server.raft: entering follower state: follower="Node at 127.0.0.1:30293 [Follower]" leader= === CONT TestSessionGet/#00 writer.go:29: 2021-01-29T19:32:55.881Z [INFO] TestSessionGet/#00.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:55.882Z [INFO] TestSessionGet/#00: consul server down === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:55.880Z [INFO] TestSessionCustomTTL.server.serf.lan: serf: EventMemberJoin: Node-819f020a-de2e-19ec-f9bc-59ef6fa504fc 127.0.0.1 === CONT TestSessionGet/#00 writer.go:29: 2021-01-29T19:32:55.883Z [INFO] TestSessionGet/#00: shutdown complete writer.go:29: 2021-01-29T19:32:55.884Z [INFO] TestSessionGet/#00: Stopping server: protocol=DNS address=127.0.0.1:30276 network=tcp writer.go:29: 2021-01-29T19:32:55.885Z [INFO] TestSessionGet/#00: Stopping server: protocol=DNS address=127.0.0.1:30276 network=udp writer.go:29: 2021-01-29T19:32:55.886Z [INFO] TestSessionGet/#00: Stopping server: protocol=HTTP address=127.0.0.1:30277 network=tcp === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:55.887Z [INFO] TestSessionCustomTTL.server: Adding LAN server: server="Node-819f020a-de2e-19ec-f9bc-59ef6fa504fc (Addr: tcp/127.0.0.1:30293) (DC: dc1)" writer.go:29: 2021-01-29T19:32:55.887Z [INFO] TestSessionCustomTTL.server: Handled event for server in area: event=member-join server=Node-819f020a-de2e-19ec-f9bc-59ef6fa504fc.dc1 area=wan writer.go:29: 2021-01-29T19:32:55.889Z [INFO] TestSessionCustomTTL: Started DNS server: address=127.0.0.1:30288 network=udp writer.go:29: 2021-01-29T19:32:55.891Z [INFO] TestSessionCustomTTL: Started DNS server: address=127.0.0.1:30288 network=tcp writer.go:29: 2021-01-29T19:32:55.894Z [INFO] TestSessionCustomTTL: Started HTTP server: address=127.0.0.1:30289 network=tcp writer.go:29: 2021-01-29T19:32:55.895Z [INFO] TestSessionCustomTTL: started state syncer === CONT TestIngestUserEvent writer.go:29: 2021-01-29T19:32:55.900Z [INFO] TestIngestUserEvent: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:55.901Z [INFO] TestIngestUserEvent: Endpoints down --- PASS: TestIngestUserEvent (1.93s) === CONT TestSessionDestroy writer.go:29: 2021-01-29T19:32:55.930Z [WARN] TestSessionDestroy: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:55.931Z [DEBUG] TestSessionDestroy.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:55.935Z [DEBUG] TestSessionDestroy.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:55.952Z [WARN] TestSessionCustomTTL.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:55.953Z [INFO] TestSessionCustomTTL.server.raft: entering candidate state: node="Node at 127.0.0.1:30293 [Candidate]" term=2 === CONT TestSessionDestroy writer.go:29: 2021-01-29T19:32:55.972Z [INFO] TestSessionDestroy.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c28e5597-07f4-f24a-454e-292ce0895516 Address:127.0.0.1:30299}]" === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:55.973Z [DEBUG] TestSessionCustomTTL.server.raft: votes: needed=1 === CONT TestSessionDestroy writer.go:29: 2021-01-29T19:32:55.975Z [INFO] TestSessionDestroy.server.serf.wan: serf: EventMemberJoin: Node-c28e5597-07f4-f24a-454e-292ce0895516.dc1 127.0.0.1 === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:55.975Z [DEBUG] TestSessionCustomTTL.server.raft: vote granted: from=819f020a-de2e-19ec-f9bc-59ef6fa504fc term=2 tally=1 writer.go:29: 2021-01-29T19:32:55.977Z [INFO] TestSessionCustomTTL.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:55.979Z [INFO] TestSessionCustomTTL.server.raft: entering leader state: leader="Node at 127.0.0.1:30293 [Leader]" writer.go:29: 2021-01-29T19:32:55.981Z [INFO] TestSessionCustomTTL.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:55.982Z [INFO] TestSessionCustomTTL.server: New leader elected: payload=Node-819f020a-de2e-19ec-f9bc-59ef6fa504fc === CONT TestSessionDestroy writer.go:29: 2021-01-29T19:32:55.980Z [INFO] TestSessionDestroy.server.serf.lan: serf: EventMemberJoin: Node-c28e5597-07f4-f24a-454e-292ce0895516 127.0.0.1 writer.go:29: 2021-01-29T19:32:55.995Z [INFO] TestSessionDestroy: Started DNS server: address=127.0.0.1:30294 network=udp writer.go:29: 2021-01-29T19:32:55.982Z [INFO] TestSessionDestroy.server.raft: entering follower state: follower="Node at 127.0.0.1:30299 [Follower]" leader= writer.go:29: 2021-01-29T19:32:55.998Z [INFO] TestSessionDestroy.server: Adding LAN server: server="Node-c28e5597-07f4-f24a-454e-292ce0895516 (Addr: tcp/127.0.0.1:30299) (DC: dc1)" writer.go:29: 2021-01-29T19:32:55.999Z [INFO] TestSessionDestroy.server: Handled event for server in area: event=member-join server=Node-c28e5597-07f4-f24a-454e-292ce0895516.dc1 area=wan === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:56.002Z [DEBUG] TestSessionCustomTTL.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30293 === CONT TestSessionDestroy writer.go:29: 2021-01-29T19:32:56.001Z [INFO] TestSessionDestroy: Started DNS server: address=127.0.0.1:30294 network=tcp writer.go:29: 2021-01-29T19:32:56.014Z [INFO] TestSessionDestroy: Started HTTP server: address=127.0.0.1:30295 network=tcp writer.go:29: 2021-01-29T19:32:56.015Z [INFO] TestSessionDestroy: started state syncer === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:56.017Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:56.024Z [INFO] TestSessionCustomTTL.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:56.026Z [INFO] TestSessionCustomTTL.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.028Z [DEBUG] TestSessionCustomTTL.server: Skipping self join check for node since the cluster is too small: node=Node-819f020a-de2e-19ec-f9bc-59ef6fa504fc writer.go:29: 2021-01-29T19:32:56.029Z [INFO] TestSessionCustomTTL.server: member joined, marking health alive: member=Node-819f020a-de2e-19ec-f9bc-59ef6fa504fc === CONT TestSessionDestroy writer.go:29: 2021-01-29T19:32:56.046Z [WARN] TestSessionDestroy.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:56.047Z [INFO] TestSessionDestroy.server.raft: entering candidate state: node="Node at 127.0.0.1:30299 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:56.049Z [DEBUG] TestSessionDestroy.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:56.050Z [DEBUG] TestSessionDestroy.server.raft: vote granted: from=c28e5597-07f4-f24a-454e-292ce0895516 term=2 tally=1 writer.go:29: 2021-01-29T19:32:56.051Z [INFO] TestSessionDestroy.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:56.052Z [INFO] TestSessionDestroy.server.raft: entering leader state: leader="Node at 127.0.0.1:30299 [Leader]" writer.go:29: 2021-01-29T19:32:56.053Z [INFO] TestSessionDestroy.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:56.053Z [INFO] TestSessionDestroy.server: New leader elected: payload=Node-c28e5597-07f4-f24a-454e-292ce0895516 writer.go:29: 2021-01-29T19:32:56.054Z [DEBUG] TestSessionDestroy.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30299 writer.go:29: 2021-01-29T19:32:56.059Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:56.064Z [INFO] TestSessionDestroy.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:56.065Z [INFO] TestSessionDestroy.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.066Z [DEBUG] TestSessionDestroy.server: Skipping self join check for node since the cluster is too small: node=Node-c28e5597-07f4-f24a-454e-292ce0895516 writer.go:29: 2021-01-29T19:32:56.066Z [INFO] TestSessionDestroy.server: member joined, marking health alive: member=Node-c28e5597-07f4-f24a-454e-292ce0895516 === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:56.139Z [DEBUG] TestSessionCustomTTL: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestSessionDestroy writer.go:29: 2021-01-29T19:32:56.140Z [DEBUG] TestSessionDestroy: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:56.141Z [INFO] TestSessionCustomTTL: Synced node info writer.go:29: 2021-01-29T19:32:56.141Z [DEBUG] TestSessionCustomTTL: Node info in sync === CONT TestSessionDestroy writer.go:29: 2021-01-29T19:32:56.142Z [INFO] TestSessionDestroy: Synced node info === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:56.175Z [DEBUG] TestSessionCustomTTL: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:56.177Z [DEBUG] TestSessionCustomTTL: Node info in sync === CONT TestSessionDestroy writer.go:29: 2021-01-29T19:32:56.291Z [INFO] TestSessionDestroy: Requesting shutdown writer.go:29: 2021-01-29T19:32:56.293Z [INFO] TestSessionDestroy.server: shutting down server writer.go:29: 2021-01-29T19:32:56.294Z [DEBUG] TestSessionDestroy.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.296Z [WARN] TestSessionDestroy.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:56.299Z [DEBUG] TestSessionDestroy.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.301Z [WARN] TestSessionDestroy.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:56.303Z [INFO] TestSessionDestroy.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:56.303Z [INFO] TestSessionDestroy: consul server down writer.go:29: 2021-01-29T19:32:56.305Z [INFO] TestSessionDestroy: shutdown complete writer.go:29: 2021-01-29T19:32:56.306Z [INFO] TestSessionDestroy: Stopping server: protocol=DNS address=127.0.0.1:30294 network=tcp writer.go:29: 2021-01-29T19:32:56.306Z [INFO] TestSessionDestroy: Stopping server: protocol=DNS address=127.0.0.1:30294 network=udp writer.go:29: 2021-01-29T19:32:56.307Z [INFO] TestSessionDestroy: Stopping server: protocol=HTTP address=127.0.0.1:30295 network=tcp === CONT TestSnapshot_Options/GET#01 writer.go:29: 2021-01-29T19:32:56.350Z [INFO] TestSnapshot_Options/GET#01: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:56.351Z [INFO] TestSnapshot_Options/GET#01: Endpoints down === RUN TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.363Z [WARN] TestSnapshot_Options/GET#02: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:32:56.365Z [WARN] TestSnapshot_Options/GET#02: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:56.370Z [DEBUG] TestSnapshot_Options/GET#02.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:56.371Z [DEBUG] TestSnapshot_Options/GET#02.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:56.376Z [INFO] TestSnapshot_Options/GET#02.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:936466cb-10b8-7066-3a0c-4c54e9da9ac7 Address:127.0.0.1:30305}]" writer.go:29: 2021-01-29T19:32:56.379Z [INFO] TestSnapshot_Options/GET#02.server.raft: entering follower state: follower="Node at 127.0.0.1:30305 [Follower]" leader= writer.go:29: 2021-01-29T19:32:56.380Z [INFO] TestSnapshot_Options/GET#02.server.serf.wan: serf: EventMemberJoin: Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7.dc1 127.0.0.1 === CONT TestSessionGet/#00 writer.go:29: 2021-01-29T19:32:56.387Z [INFO] TestSessionGet/#00: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:56.388Z [INFO] TestSessionGet/#00: Endpoints down === RUN TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.400Z [WARN] TestSessionGet/#01: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:56.403Z [DEBUG] TestSessionGet/#01.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:56.405Z [DEBUG] TestSessionGet/#01.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.386Z [INFO] TestSnapshot_Options/GET#02.server.serf.lan: serf: EventMemberJoin: Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7 127.0.0.1 writer.go:29: 2021-01-29T19:32:56.408Z [INFO] TestSnapshot_Options/GET#02.server: Adding LAN server: server="Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7 (Addr: tcp/127.0.0.1:30305) (DC: dc1)" writer.go:29: 2021-01-29T19:32:56.408Z [INFO] TestSnapshot_Options/GET#02.server: Handled event for server in area: event=member-join server=Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7.dc1 area=wan writer.go:29: 2021-01-29T19:32:56.410Z [INFO] TestSnapshot_Options/GET#02: Started DNS server: address=127.0.0.1:30300 network=tcp writer.go:29: 2021-01-29T19:32:56.411Z [INFO] TestSnapshot_Options/GET#02: Started DNS server: address=127.0.0.1:30300 network=udp writer.go:29: 2021-01-29T19:32:56.413Z [INFO] TestSnapshot_Options/GET#02: Started HTTP server: address=127.0.0.1:30301 network=tcp === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.416Z [INFO] TestSessionGet/#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a324f6a4-a400-1fb4-d2be-bcd29c2186d6 Address:127.0.0.1:30311}]" === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.414Z [INFO] TestSnapshot_Options/GET#02: started state syncer === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.418Z [INFO] TestSessionGet/#01.server.raft: entering follower state: follower="Node at 127.0.0.1:30311 [Follower]" leader= writer.go:29: 2021-01-29T19:32:56.423Z [INFO] TestSessionGet/#01.server.serf.wan: serf: EventMemberJoin: Node-a324f6a4-a400-1fb4-d2be-bcd29c2186d6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:56.426Z [INFO] TestSessionGet/#01.server.serf.lan: serf: EventMemberJoin: Node-a324f6a4-a400-1fb4-d2be-bcd29c2186d6 127.0.0.1 writer.go:29: 2021-01-29T19:32:56.428Z [INFO] TestSessionGet/#01.server: Handled event for server in area: event=member-join server=Node-a324f6a4-a400-1fb4-d2be-bcd29c2186d6.dc1 area=wan writer.go:29: 2021-01-29T19:32:56.429Z [INFO] TestSessionGet/#01.server: Adding LAN server: server="Node-a324f6a4-a400-1fb4-d2be-bcd29c2186d6 (Addr: tcp/127.0.0.1:30311) (DC: dc1)" writer.go:29: 2021-01-29T19:32:56.429Z [INFO] TestSessionGet/#01: Started DNS server: address=127.0.0.1:30306 network=udp writer.go:29: 2021-01-29T19:32:56.431Z [INFO] TestSessionGet/#01: Started DNS server: address=127.0.0.1:30306 network=tcp writer.go:29: 2021-01-29T19:32:56.433Z [INFO] TestSessionGet/#01: Started HTTP server: address=127.0.0.1:30307 network=tcp writer.go:29: 2021-01-29T19:32:56.435Z [INFO] TestSessionGet/#01: started state syncer === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.451Z [WARN] TestSnapshot_Options/GET#02.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:56.452Z [INFO] TestSnapshot_Options/GET#02.server.raft: entering candidate state: node="Node at 127.0.0.1:30305 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:56.456Z [DEBUG] TestSnapshot_Options/GET#02.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:56.458Z [DEBUG] TestSnapshot_Options/GET#02.server.raft: vote granted: from=936466cb-10b8-7066-3a0c-4c54e9da9ac7 term=2 tally=1 writer.go:29: 2021-01-29T19:32:56.460Z [INFO] TestSnapshot_Options/GET#02.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:56.461Z [INFO] TestSnapshot_Options/GET#02.server.raft: entering leader state: leader="Node at 127.0.0.1:30305 [Leader]" === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.472Z [WARN] TestSessionGet/#01.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:56.474Z [INFO] TestSessionGet/#01.server.raft: entering candidate state: node="Node at 127.0.0.1:30311 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:56.476Z [DEBUG] TestSessionGet/#01.server.raft: votes: needed=1 === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.459Z [DEBUG] TestSnapshot_Options/GET#02.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.477Z [DEBUG] TestSessionGet/#01.server.raft: vote granted: from=a324f6a4-a400-1fb4-d2be-bcd29c2186d6 term=2 tally=1 writer.go:29: 2021-01-29T19:32:56.482Z [INFO] TestSessionGet/#01.server.raft: election won: tally=1 === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.481Z [INFO] TestSnapshot_Options/GET#02.server: cluster leadership acquired === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.484Z [INFO] TestSessionGet/#01.server.raft: entering leader state: leader="Node at 127.0.0.1:30311 [Leader]" writer.go:29: 2021-01-29T19:32:56.485Z [INFO] TestSessionGet/#01.server: cluster leadership acquired === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.482Z [INFO] TestSnapshot_Options/GET#02.server: New leader elected: payload=Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7 === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.485Z [INFO] TestSessionGet/#01.server: New leader elected: payload=Node-a324f6a4-a400-1fb4-d2be-bcd29c2186d6 === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.482Z [INFO] TestSnapshot_Options/GET#02.server: initializing acls === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.489Z [DEBUG] TestSessionGet/#01.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30311 === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.492Z [INFO] TestSnapshot_Options/GET#02.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:56.492Z [WARN] TestSnapshot_Options/GET#02.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:56.494Z [INFO] TestSnapshot_Options/GET#02.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:56.496Z [INFO] TestSnapshot_Options/GET#02.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:56.498Z [INFO] TestSnapshot_Options/GET#02.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:56.499Z [INFO] TestSnapshot_Options/GET#02.leader: started routine: routine="acl token reaping" === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.497Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.500Z [INFO] TestSnapshot_Options/GET#02.server.serf.lan: serf: EventMemberUpdate: Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7 === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.515Z [INFO] TestSessionGet/#01.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:56.518Z [INFO] TestSessionGet/#01.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.520Z [DEBUG] TestSessionGet/#01.server: Skipping self join check for node since the cluster is too small: node=Node-a324f6a4-a400-1fb4-d2be-bcd29c2186d6 writer.go:29: 2021-01-29T19:32:56.522Z [INFO] TestSessionGet/#01.server: member joined, marking health alive: member=Node-a324f6a4-a400-1fb4-d2be-bcd29c2186d6 === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:56.512Z [INFO] TestSnapshot_Options/GET#02.server.serf.wan: serf: EventMemberUpdate: Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7.dc1 writer.go:29: 2021-01-29T19:32:56.513Z [INFO] TestSnapshot_Options/GET#02.server: Updating LAN server: server="Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7 (Addr: tcp/127.0.0.1:30305) (DC: dc1)" writer.go:29: 2021-01-29T19:32:56.525Z [INFO] TestSnapshot_Options/GET#02.server: Handled event for server in area: event=member-update server=Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7.dc1 area=wan writer.go:29: 2021-01-29T19:32:56.527Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:56.528Z [INFO] TestSnapshot_Options/GET#02.server.fsm: snapshot created: duration=47.96µs writer.go:29: 2021-01-29T19:32:56.533Z [INFO] TestSnapshot_Options/GET#02.server.raft: starting snapshot up to: index=10 writer.go:29: 2021-01-29T19:32:56.535Z [INFO] snapshot: creating new snapshot: path=/tmp/TestSnapshot_Options_GET#02-agent156009944/raft/snapshots/2-10-1611948776535.tmp writer.go:29: 2021-01-29T19:32:56.533Z [INFO] TestSnapshot_Options/GET#02.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:56.536Z [INFO] TestSnapshot_Options/GET#02.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.537Z [DEBUG] TestSnapshot_Options/GET#02.server: Skipping self join check for node since the cluster is too small: node=Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7 writer.go:29: 2021-01-29T19:32:56.538Z [INFO] TestSnapshot_Options/GET#02.server: member joined, marking health alive: member=Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7 writer.go:29: 2021-01-29T19:32:56.540Z [DEBUG] TestSnapshot_Options/GET#02.server: Skipping self join check for node since the cluster is too small: node=Node-936466cb-10b8-7066-3a0c-4c54e9da9ac7 writer.go:29: 2021-01-29T19:32:56.547Z [INFO] TestSnapshot_Options/GET#02.server.raft: snapshot complete up to: index=10 writer.go:29: 2021-01-29T19:32:56.553Z [INFO] TestSnapshot_Options/GET#02: Requesting shutdown writer.go:29: 2021-01-29T19:32:56.554Z [INFO] TestSnapshot_Options/GET#02.server: shutting down server writer.go:29: 2021-01-29T19:32:56.555Z [DEBUG] TestSnapshot_Options/GET#02.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:56.556Z [DEBUG] TestSnapshot_Options/GET#02.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:56.557Z [DEBUG] TestSnapshot_Options/GET#02.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.557Z [WARN] TestSnapshot_Options/GET#02.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:56.557Z [DEBUG] TestSnapshot_Options/GET#02.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:56.557Z [ERROR] TestSnapshot_Options/GET#02.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:56.557Z [DEBUG] TestSnapshot_Options/GET#02.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:56.558Z [DEBUG] TestSnapshot_Options/GET#02.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.559Z [WARN] TestSnapshot_Options/GET#02.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:56.566Z [INFO] TestSnapshot_Options/GET#02.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:56.566Z [INFO] TestSnapshot_Options/GET#02: consul server down writer.go:29: 2021-01-29T19:32:56.568Z [INFO] TestSnapshot_Options/GET#02: shutdown complete writer.go:29: 2021-01-29T19:32:56.569Z [INFO] TestSnapshot_Options/GET#02: Stopping server: protocol=DNS address=127.0.0.1:30300 network=tcp writer.go:29: 2021-01-29T19:32:56.570Z [INFO] TestSnapshot_Options/GET#02: Stopping server: protocol=DNS address=127.0.0.1:30300 network=udp writer.go:29: 2021-01-29T19:32:56.570Z [INFO] TestSnapshot_Options/GET#02: Stopping server: protocol=HTTP address=127.0.0.1:30301 network=tcp === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:56.710Z [DEBUG] TestSessionCustomTTL.server: Session TTL expired: session=136479f1-232c-ec18-2fe7-3973df350e8f === CONT TestSessionDestroy writer.go:29: 2021-01-29T19:32:56.808Z [INFO] TestSessionDestroy: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:56.810Z [INFO] TestSessionDestroy: Endpoints down --- PASS: TestSessionDestroy (0.91s) === CONT TestSessionCreate_NoCheck === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.828Z [DEBUG] TestSessionGet/#01: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:56.830Z [INFO] TestSessionGet/#01: Requesting shutdown writer.go:29: 2021-01-29T19:32:56.830Z [INFO] TestSessionGet/#01: Synced node info writer.go:29: 2021-01-29T19:32:56.833Z [DEBUG] TestSessionGet/#01: Node info in sync writer.go:29: 2021-01-29T19:32:56.832Z [INFO] TestSessionGet/#01.server: shutting down server === CONT TestSessionCreate_NoCheck writer.go:29: 2021-01-29T19:32:56.836Z [WARN] TestSessionCreate_NoCheck: bootstrap = true: do not enable unless necessary === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.836Z [DEBUG] TestSessionGet/#01.leader: stopping routine: routine="CA root pruning" === CONT TestSessionCreate_NoCheck writer.go:29: 2021-01-29T19:32:56.837Z [DEBUG] TestSessionCreate_NoCheck.tlsutil: Update: version=1 === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.837Z [WARN] TestSessionGet/#01.server.serf.lan: serf: Shutdown without a Leave === CONT TestSessionCreate_NoCheck writer.go:29: 2021-01-29T19:32:56.839Z [DEBUG] TestSessionCreate_NoCheck.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.837Z [DEBUG] TestSessionGet/#01.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.840Z [WARN] TestSessionGet/#01.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:56.845Z [INFO] TestSessionGet/#01.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:56.845Z [INFO] TestSessionGet/#01: consul server down writer.go:29: 2021-01-29T19:32:56.849Z [INFO] TestSessionGet/#01: shutdown complete writer.go:29: 2021-01-29T19:32:56.851Z [INFO] TestSessionGet/#01: Stopping server: protocol=DNS address=127.0.0.1:30306 network=tcp === CONT TestSessionCreate_NoCheck writer.go:29: 2021-01-29T19:32:56.853Z [INFO] TestSessionCreate_NoCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:19a1778f-d325-ab8c-f739-6e3bcf22be28 Address:127.0.0.1:30317}]" === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:56.853Z [INFO] TestSessionGet/#01: Stopping server: protocol=DNS address=127.0.0.1:30306 network=udp writer.go:29: 2021-01-29T19:32:56.855Z [INFO] TestSessionGet/#01: Stopping server: protocol=HTTP address=127.0.0.1:30307 network=tcp === CONT TestSessionCreate_NoCheck writer.go:29: 2021-01-29T19:32:56.855Z [INFO] TestSessionCreate_NoCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:30317 [Follower]" leader= writer.go:29: 2021-01-29T19:32:56.858Z [INFO] TestSessionCreate_NoCheck.server.serf.wan: serf: EventMemberJoin: Node-19a1778f-d325-ab8c-f739-6e3bcf22be28.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:56.861Z [INFO] TestSessionCreate_NoCheck.server.serf.lan: serf: EventMemberJoin: Node-19a1778f-d325-ab8c-f739-6e3bcf22be28 127.0.0.1 writer.go:29: 2021-01-29T19:32:56.863Z [INFO] TestSessionCreate_NoCheck.server: Handled event for server in area: event=member-join server=Node-19a1778f-d325-ab8c-f739-6e3bcf22be28.dc1 area=wan writer.go:29: 2021-01-29T19:32:56.864Z [INFO] TestSessionCreate_NoCheck.server: Adding LAN server: server="Node-19a1778f-d325-ab8c-f739-6e3bcf22be28 (Addr: tcp/127.0.0.1:30317) (DC: dc1)" writer.go:29: 2021-01-29T19:32:56.867Z [INFO] TestSessionCreate_NoCheck: Started DNS server: address=127.0.0.1:30312 network=tcp writer.go:29: 2021-01-29T19:32:56.868Z [INFO] TestSessionCreate_NoCheck: Started DNS server: address=127.0.0.1:30312 network=udp writer.go:29: 2021-01-29T19:32:56.870Z [INFO] TestSessionCreate_NoCheck: Started HTTP server: address=127.0.0.1:30313 network=tcp writer.go:29: 2021-01-29T19:32:56.872Z [INFO] TestSessionCreate_NoCheck: started state syncer writer.go:29: 2021-01-29T19:32:56.916Z [WARN] TestSessionCreate_NoCheck.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:56.917Z [INFO] TestSessionCreate_NoCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:30317 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:56.920Z [DEBUG] TestSessionCreate_NoCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:56.921Z [DEBUG] TestSessionCreate_NoCheck.server.raft: vote granted: from=19a1778f-d325-ab8c-f739-6e3bcf22be28 term=2 tally=1 writer.go:29: 2021-01-29T19:32:56.927Z [INFO] TestSessionCreate_NoCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:56.931Z [INFO] TestSessionCreate_NoCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:30317 [Leader]" writer.go:29: 2021-01-29T19:32:56.939Z [INFO] TestSessionCreate_NoCheck.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:56.940Z [INFO] TestSessionCreate_NoCheck.server: New leader elected: payload=Node-19a1778f-d325-ab8c-f739-6e3bcf22be28 writer.go:29: 2021-01-29T19:32:56.941Z [DEBUG] TestSessionCreate_NoCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30317 writer.go:29: 2021-01-29T19:32:56.948Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:56.954Z [INFO] TestSessionCreate_NoCheck.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:56.955Z [INFO] TestSessionCreate_NoCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.957Z [DEBUG] TestSessionCreate_NoCheck.server: Skipping self join check for node since the cluster is too small: node=Node-19a1778f-d325-ab8c-f739-6e3bcf22be28 writer.go:29: 2021-01-29T19:32:56.959Z [INFO] TestSessionCreate_NoCheck.server: member joined, marking health alive: member=Node-19a1778f-d325-ab8c-f739-6e3bcf22be28 === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:56.958Z [INFO] TestSessionCustomTTL: Requesting shutdown writer.go:29: 2021-01-29T19:32:56.961Z [INFO] TestSessionCustomTTL.server: shutting down server writer.go:29: 2021-01-29T19:32:56.962Z [DEBUG] TestSessionCustomTTL.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.962Z [WARN] TestSessionCustomTTL.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:56.962Z [DEBUG] TestSessionCustomTTL.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.964Z [WARN] TestSessionCustomTTL.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:56.969Z [INFO] TestSessionCustomTTL.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:56.970Z [INFO] TestSessionCustomTTL: consul server down writer.go:29: 2021-01-29T19:32:56.972Z [INFO] TestSessionCustomTTL: shutdown complete writer.go:29: 2021-01-29T19:32:56.972Z [INFO] TestSessionCustomTTL: Stopping server: protocol=DNS address=127.0.0.1:30288 network=tcp writer.go:29: 2021-01-29T19:32:56.973Z [INFO] TestSessionCustomTTL: Stopping server: protocol=DNS address=127.0.0.1:30288 network=udp writer.go:29: 2021-01-29T19:32:56.974Z [INFO] TestSessionCustomTTL: Stopping server: protocol=HTTP address=127.0.0.1:30289 network=tcp === RUN TestSessionCreate_NoCheck/no_check_fields_should_yield_default_serfHealth === RUN TestSessionCreate_NoCheck/overwrite_nodechecks_to_associate_with_no_checks === RUN TestSessionCreate_NoCheck/overwrite_checks_to_associate_with_no_checks === CONT TestSessionCreate_NoCheck writer.go:29: 2021-01-29T19:32:56.987Z [INFO] TestSessionCreate_NoCheck: Requesting shutdown writer.go:29: 2021-01-29T19:32:56.988Z [INFO] TestSessionCreate_NoCheck.server: shutting down server writer.go:29: 2021-01-29T19:32:56.989Z [DEBUG] TestSessionCreate_NoCheck.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.991Z [WARN] TestSessionCreate_NoCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:56.991Z [DEBUG] TestSessionCreate_NoCheck.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.990Z [ERROR] TestSessionCreate_NoCheck.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:56.996Z [WARN] TestSessionCreate_NoCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:57.000Z [INFO] TestSessionCreate_NoCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:57.000Z [INFO] TestSessionCreate_NoCheck: consul server down writer.go:29: 2021-01-29T19:32:57.003Z [INFO] TestSessionCreate_NoCheck: shutdown complete writer.go:29: 2021-01-29T19:32:57.004Z [INFO] TestSessionCreate_NoCheck: Stopping server: protocol=DNS address=127.0.0.1:30312 network=tcp writer.go:29: 2021-01-29T19:32:57.006Z [INFO] TestSessionCreate_NoCheck: Stopping server: protocol=DNS address=127.0.0.1:30312 network=udp writer.go:29: 2021-01-29T19:32:57.008Z [INFO] TestSessionCreate_NoCheck: Stopping server: protocol=HTTP address=127.0.0.1:30313 network=tcp === CONT TestSnapshot_Options/GET#02 writer.go:29: 2021-01-29T19:32:57.071Z [INFO] TestSnapshot_Options/GET#02: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:57.073Z [INFO] TestSnapshot_Options/GET#02: Endpoints down === RUN TestSnapshot_Options/PUT writer.go:29: 2021-01-29T19:32:57.091Z [WARN] TestSnapshot_Options/PUT: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:32:57.091Z [WARN] TestSnapshot_Options/PUT: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:57.092Z [DEBUG] TestSnapshot_Options/PUT.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:57.094Z [DEBUG] TestSnapshot_Options/PUT.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:57.112Z [INFO] TestSnapshot_Options/PUT.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:89eaeb0b-0c61-9f8d-1987-bb5a554d703d Address:127.0.0.1:30323}]" writer.go:29: 2021-01-29T19:32:57.114Z [INFO] TestSnapshot_Options/PUT.server.raft: entering follower state: follower="Node at 127.0.0.1:30323 [Follower]" leader= writer.go:29: 2021-01-29T19:32:57.114Z [INFO] TestSnapshot_Options/PUT.server.serf.wan: serf: EventMemberJoin: Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:57.118Z [INFO] TestSnapshot_Options/PUT.server.serf.lan: serf: EventMemberJoin: Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d 127.0.0.1 writer.go:29: 2021-01-29T19:32:57.121Z [INFO] TestSnapshot_Options/PUT.server: Adding LAN server: server="Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d (Addr: tcp/127.0.0.1:30323) (DC: dc1)" writer.go:29: 2021-01-29T19:32:57.122Z [INFO] TestSnapshot_Options/PUT: Started DNS server: address=127.0.0.1:30318 network=udp writer.go:29: 2021-01-29T19:32:57.125Z [INFO] TestSnapshot_Options/PUT: Started DNS server: address=127.0.0.1:30318 network=tcp writer.go:29: 2021-01-29T19:32:57.123Z [INFO] TestSnapshot_Options/PUT.server: Handled event for server in area: event=member-join server=Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d.dc1 area=wan writer.go:29: 2021-01-29T19:32:57.129Z [INFO] TestSnapshot_Options/PUT: Started HTTP server: address=127.0.0.1:30319 network=tcp writer.go:29: 2021-01-29T19:32:57.133Z [INFO] TestSnapshot_Options/PUT: started state syncer writer.go:29: 2021-01-29T19:32:57.165Z [WARN] TestSnapshot_Options/PUT.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:57.167Z [INFO] TestSnapshot_Options/PUT.server.raft: entering candidate state: node="Node at 127.0.0.1:30323 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:57.169Z [DEBUG] TestSnapshot_Options/PUT.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:57.171Z [DEBUG] TestSnapshot_Options/PUT.server.raft: vote granted: from=89eaeb0b-0c61-9f8d-1987-bb5a554d703d term=2 tally=1 writer.go:29: 2021-01-29T19:32:57.172Z [INFO] TestSnapshot_Options/PUT.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:57.174Z [INFO] TestSnapshot_Options/PUT.server.raft: entering leader state: leader="Node at 127.0.0.1:30323 [Leader]" writer.go:29: 2021-01-29T19:32:57.172Z [DEBUG] TestSnapshot_Options/PUT.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:57.176Z [INFO] TestSnapshot_Options/PUT.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:57.177Z [INFO] TestSnapshot_Options/PUT.server: New leader elected: payload=Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d writer.go:29: 2021-01-29T19:32:57.177Z [INFO] TestSnapshot_Options/PUT.server: initializing acls writer.go:29: 2021-01-29T19:32:57.182Z [INFO] TestSnapshot_Options/PUT.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:57.183Z [WARN] TestSnapshot_Options/PUT.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:57.186Z [INFO] TestSnapshot_Options/PUT.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:57.189Z [INFO] TestSnapshot_Options/PUT.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:57.191Z [INFO] TestSnapshot_Options/PUT.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:57.192Z [INFO] TestSnapshot_Options/PUT.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:57.194Z [INFO] TestSnapshot_Options/PUT.server.serf.lan: serf: EventMemberUpdate: Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d writer.go:29: 2021-01-29T19:32:57.196Z [INFO] TestSnapshot_Options/PUT.server.serf.wan: serf: EventMemberUpdate: Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d.dc1 writer.go:29: 2021-01-29T19:32:57.197Z [INFO] TestSnapshot_Options/PUT.server: Updating LAN server: server="Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d (Addr: tcp/127.0.0.1:30323) (DC: dc1)" writer.go:29: 2021-01-29T19:32:57.199Z [INFO] TestSnapshot_Options/PUT.server: Handled event for server in area: event=member-update server=Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d.dc1 area=wan writer.go:29: 2021-01-29T19:32:57.203Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:57.209Z [INFO] TestSnapshot_Options/PUT.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:57.210Z [INFO] TestSnapshot_Options/PUT.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.212Z [DEBUG] TestSnapshot_Options/PUT.server: Skipping self join check for node since the cluster is too small: node=Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d writer.go:29: 2021-01-29T19:32:57.214Z [INFO] TestSnapshot_Options/PUT.server: member joined, marking health alive: member=Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d writer.go:29: 2021-01-29T19:32:57.216Z [DEBUG] TestSnapshot_Options/PUT.server: Skipping self join check for node since the cluster is too small: node=Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d writer.go:29: 2021-01-29T19:32:57.261Z [DEBUG] TestSnapshot_Options/PUT: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:57.267Z [INFO] TestSnapshot_Options/PUT: Synced node info writer.go:29: 2021-01-29T19:32:57.268Z [DEBUG] TestSnapshot_Options/PUT: Node info in sync === CONT TestSessionGet/#01 writer.go:29: 2021-01-29T19:32:57.356Z [INFO] TestSessionGet/#01: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:57.358Z [INFO] TestSessionGet/#01: Endpoints down --- PASS: TestSessionGet (1.88s) --- PASS: TestSessionGet/#00 (0.91s) --- PASS: TestSessionGet/#01 (0.97s) === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.374Z [WARN] TestSessionCreate_DefaultCheck: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:57.377Z [DEBUG] TestSessionCreate_DefaultCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:57.380Z [DEBUG] TestSessionCreate_DefaultCheck.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestSnapshot_Options/PUT writer.go:29: 2021-01-29T19:32:57.387Z [DEBUG] TestSnapshot_Options/PUT.acl: dropping node from result due to ACLs: node=Node-89eaeb0b-0c61-9f8d-1987-bb5a554d703d === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.394Z [INFO] TestSessionCreate_DefaultCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bed8f143-d358-b374-79d2-c5c100cc8d6d Address:127.0.0.1:30329}]" === CONT TestSnapshot_Options/PUT writer.go:29: 2021-01-29T19:32:57.394Z [INFO] TestSnapshot_Options/PUT: Requesting shutdown writer.go:29: 2021-01-29T19:32:57.397Z [INFO] TestSnapshot_Options/PUT.server: shutting down server writer.go:29: 2021-01-29T19:32:57.398Z [DEBUG] TestSnapshot_Options/PUT.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.399Z [DEBUG] TestSnapshot_Options/PUT.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:57.400Z [DEBUG] TestSnapshot_Options/PUT.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:57.401Z [WARN] TestSnapshot_Options/PUT.server.serf.lan: serf: Shutdown without a Leave === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.396Z [INFO] TestSessionCreate_DefaultCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:30329 [Follower]" leader= writer.go:29: 2021-01-29T19:32:57.396Z [INFO] TestSessionCreate_DefaultCheck.server.serf.wan: serf: EventMemberJoin: Node-bed8f143-d358-b374-79d2-c5c100cc8d6d.dc1 127.0.0.1 === CONT TestSnapshot_Options/PUT writer.go:29: 2021-01-29T19:32:57.399Z [DEBUG] TestSnapshot_Options/PUT.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.401Z [DEBUG] TestSnapshot_Options/PUT.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:57.400Z [DEBUG] TestSnapshot_Options/PUT.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:57.408Z [WARN] TestSnapshot_Options/PUT.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:57.415Z [INFO] TestSnapshot_Options/PUT.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:57.416Z [INFO] TestSnapshot_Options/PUT: consul server down writer.go:29: 2021-01-29T19:32:57.418Z [INFO] TestSnapshot_Options/PUT: shutdown complete writer.go:29: 2021-01-29T19:32:57.419Z [INFO] TestSnapshot_Options/PUT: Stopping server: protocol=DNS address=127.0.0.1:30318 network=tcp === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.410Z [INFO] TestSessionCreate_DefaultCheck.server.serf.lan: serf: EventMemberJoin: Node-bed8f143-d358-b374-79d2-c5c100cc8d6d 127.0.0.1 writer.go:29: 2021-01-29T19:32:57.423Z [INFO] TestSessionCreate_DefaultCheck.server: Adding LAN server: server="Node-bed8f143-d358-b374-79d2-c5c100cc8d6d (Addr: tcp/127.0.0.1:30329) (DC: dc1)" writer.go:29: 2021-01-29T19:32:57.424Z [INFO] TestSessionCreate_DefaultCheck: Started DNS server: address=127.0.0.1:30324 network=udp === CONT TestSnapshot_Options/PUT writer.go:29: 2021-01-29T19:32:57.420Z [INFO] TestSnapshot_Options/PUT: Stopping server: protocol=DNS address=127.0.0.1:30318 network=udp === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.426Z [INFO] TestSessionCreate_DefaultCheck: Started DNS server: address=127.0.0.1:30324 network=tcp === CONT TestSnapshot_Options/PUT writer.go:29: 2021-01-29T19:32:57.426Z [INFO] TestSnapshot_Options/PUT: Stopping server: protocol=HTTP address=127.0.0.1:30319 network=tcp === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.429Z [INFO] TestSessionCreate_DefaultCheck: Started HTTP server: address=127.0.0.1:30325 network=tcp writer.go:29: 2021-01-29T19:32:57.431Z [INFO] TestSessionCreate_DefaultCheck: started state syncer writer.go:29: 2021-01-29T19:32:57.426Z [INFO] TestSessionCreate_DefaultCheck.server: Handled event for server in area: event=member-join server=Node-bed8f143-d358-b374-79d2-c5c100cc8d6d.dc1 area=wan === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:57.476Z [INFO] TestSessionCustomTTL: Waiting for endpoints to shut down === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.476Z [WARN] TestSessionCreate_DefaultCheck.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestSessionCustomTTL writer.go:29: 2021-01-29T19:32:57.476Z [INFO] TestSessionCustomTTL: Endpoints down --- PASS: TestSessionCustomTTL (1.73s) === CONT TestSessionCreate_Delete === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.477Z [INFO] TestSessionCreate_DefaultCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:30329 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:57.480Z [DEBUG] TestSessionCreate_DefaultCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:57.481Z [DEBUG] TestSessionCreate_DefaultCheck.server.raft: vote granted: from=bed8f143-d358-b374-79d2-c5c100cc8d6d term=2 tally=1 writer.go:29: 2021-01-29T19:32:57.483Z [INFO] TestSessionCreate_DefaultCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:57.484Z [INFO] TestSessionCreate_DefaultCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:30329 [Leader]" writer.go:29: 2021-01-29T19:32:57.486Z [INFO] TestSessionCreate_DefaultCheck.server: cluster leadership acquired === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.486Z [WARN] TestSessionCreate_Delete: bootstrap = true: do not enable unless necessary === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.487Z [INFO] TestSessionCreate_DefaultCheck.server: New leader elected: payload=Node-bed8f143-d358-b374-79d2-c5c100cc8d6d === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.487Z [DEBUG] TestSessionCreate_Delete.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:57.488Z [DEBUG] TestSessionCreate_Delete.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:57.493Z [INFO] TestSessionCreate_Delete.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e9247d08-225d-c798-9540-be7413123b2e Address:127.0.0.1:30335}]" writer.go:29: 2021-01-29T19:32:57.496Z [INFO] TestSessionCreate_Delete.server.raft: entering follower state: follower="Node at 127.0.0.1:30335 [Follower]" leader= writer.go:29: 2021-01-29T19:32:57.499Z [INFO] TestSessionCreate_Delete.server.serf.wan: serf: EventMemberJoin: Node-e9247d08-225d-c798-9540-be7413123b2e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:57.503Z [INFO] TestSessionCreate_Delete.server.serf.lan: serf: EventMemberJoin: Node-e9247d08-225d-c798-9540-be7413123b2e 127.0.0.1 writer.go:29: 2021-01-29T19:32:57.508Z [INFO] TestSessionCreate_Delete.server: Adding LAN server: server="Node-e9247d08-225d-c798-9540-be7413123b2e (Addr: tcp/127.0.0.1:30335) (DC: dc1)" === CONT TestSessionCreate_NoCheck writer.go:29: 2021-01-29T19:32:57.509Z [INFO] TestSessionCreate_NoCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:57.511Z [INFO] TestSessionCreate_NoCheck: Endpoints down --- PASS: TestSessionCreate_NoCheck (0.70s) --- PASS: TestSessionCreate_NoCheck/no_check_fields_should_yield_default_serfHealth (0.00s) --- PASS: TestSessionCreate_NoCheck/overwrite_nodechecks_to_associate_with_no_checks (0.00s) --- PASS: TestSessionCreate_NoCheck/overwrite_checks_to_associate_with_no_checks (0.00s) === CONT TestSessionCreate_NodeChecks === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.517Z [DEBUG] TestSessionCreate_DefaultCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30329 writer.go:29: 2021-01-29T19:32:57.518Z [INFO] TestSessionCreate_DefaultCheck: Synced node info === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.519Z [INFO] TestSessionCreate_Delete.server: Handled event for server in area: event=member-join server=Node-e9247d08-225d-c798-9540-be7413123b2e.dc1 area=wan === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.520Z [DEBUG] TestSessionCreate_DefaultCheck: Node info in sync writer.go:29: 2021-01-29T19:32:57.521Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.523Z [INFO] TestSessionCreate_Delete: Started DNS server: address=127.0.0.1:30330 network=udp writer.go:29: 2021-01-29T19:32:57.525Z [INFO] TestSessionCreate_Delete: Started DNS server: address=127.0.0.1:30330 network=tcp writer.go:29: 2021-01-29T19:32:57.527Z [INFO] TestSessionCreate_Delete: Started HTTP server: address=127.0.0.1:30331 network=tcp writer.go:29: 2021-01-29T19:32:57.528Z [INFO] TestSessionCreate_Delete: started state syncer === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.530Z [INFO] TestSessionCreate_DefaultCheck.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:57.531Z [INFO] TestSessionCreate_DefaultCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.532Z [DEBUG] TestSessionCreate_DefaultCheck.server: Skipping self join check for node since the cluster is too small: node=Node-bed8f143-d358-b374-79d2-c5c100cc8d6d writer.go:29: 2021-01-29T19:32:57.533Z [INFO] TestSessionCreate_DefaultCheck.server: member joined, marking health alive: member=Node-bed8f143-d358-b374-79d2-c5c100cc8d6d === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.556Z [WARN] TestSessionCreate_NodeChecks: bootstrap = true: do not enable unless necessary === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.559Z [WARN] TestSessionCreate_Delete.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:57.562Z [INFO] TestSessionCreate_Delete.server.raft: entering candidate state: node="Node at 127.0.0.1:30335 [Candidate]" term=2 === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.564Z [DEBUG] TestSessionCreate_NodeChecks.tlsutil: Update: version=1 === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.568Z [DEBUG] TestSessionCreate_Delete.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:57.571Z [DEBUG] TestSessionCreate_Delete.server.raft: vote granted: from=e9247d08-225d-c798-9540-be7413123b2e term=2 tally=1 === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.573Z [DEBUG] TestSessionCreate_NodeChecks.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.574Z [INFO] TestSessionCreate_Delete.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:57.582Z [INFO] TestSessionCreate_Delete.server.raft: entering leader state: leader="Node at 127.0.0.1:30335 [Leader]" writer.go:29: 2021-01-29T19:32:57.584Z [INFO] TestSessionCreate_Delete.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:57.585Z [INFO] TestSessionCreate_Delete.server: New leader elected: payload=Node-e9247d08-225d-c798-9540-be7413123b2e === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.596Z [INFO] TestSessionCreate_NodeChecks.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8c16cb26-18ef-6173-bdba-407fce823689 Address:127.0.0.1:30341}]" === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.588Z [DEBUG] TestSessionCreate_Delete.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30335 === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.601Z [INFO] TestSessionCreate_NodeChecks.server.raft: entering follower state: follower="Node at 127.0.0.1:30341 [Follower]" leader= writer.go:29: 2021-01-29T19:32:57.606Z [INFO] TestSessionCreate_NodeChecks.server.serf.wan: serf: EventMemberJoin: Node-8c16cb26-18ef-6173-bdba-407fce823689.dc1 127.0.0.1 === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.614Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.622Z [INFO] TestSessionCreate_NodeChecks.server.serf.lan: serf: EventMemberJoin: Node-8c16cb26-18ef-6173-bdba-407fce823689 127.0.0.1 writer.go:29: 2021-01-29T19:32:57.630Z [INFO] TestSessionCreate_NodeChecks.server: Handled event for server in area: event=member-join server=Node-8c16cb26-18ef-6173-bdba-407fce823689.dc1 area=wan writer.go:29: 2021-01-29T19:32:57.635Z [INFO] TestSessionCreate_NodeChecks.server: Adding LAN server: server="Node-8c16cb26-18ef-6173-bdba-407fce823689 (Addr: tcp/127.0.0.1:30341) (DC: dc1)" === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.639Z [INFO] TestSessionCreate_Delete.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:57.639Z [INFO] TestSessionCreate_Delete.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.640Z [DEBUG] TestSessionCreate_Delete.server: Skipping self join check for node since the cluster is too small: node=Node-e9247d08-225d-c798-9540-be7413123b2e writer.go:29: 2021-01-29T19:32:57.641Z [INFO] TestSessionCreate_Delete.server: member joined, marking health alive: member=Node-e9247d08-225d-c798-9540-be7413123b2e === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.640Z [INFO] TestSessionCreate_NodeChecks: Started DNS server: address=127.0.0.1:30336 network=udp writer.go:29: 2021-01-29T19:32:57.645Z [INFO] TestSessionCreate_NodeChecks: Started DNS server: address=127.0.0.1:30336 network=tcp === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.653Z [INFO] TestSessionCreate_Delete: Synced node info === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.654Z [INFO] TestSessionCreate_NodeChecks: Started HTTP server: address=127.0.0.1:30337 network=tcp writer.go:29: 2021-01-29T19:32:57.655Z [INFO] TestSessionCreate_NodeChecks: started state syncer writer.go:29: 2021-01-29T19:32:57.675Z [WARN] TestSessionCreate_NodeChecks.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:57.677Z [INFO] TestSessionCreate_NodeChecks.server.raft: entering candidate state: node="Node at 127.0.0.1:30341 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:57.702Z [DEBUG] TestSessionCreate_NodeChecks.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:57.703Z [DEBUG] TestSessionCreate_NodeChecks.server.raft: vote granted: from=8c16cb26-18ef-6173-bdba-407fce823689 term=2 tally=1 writer.go:29: 2021-01-29T19:32:57.704Z [INFO] TestSessionCreate_NodeChecks.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:57.705Z [INFO] TestSessionCreate_NodeChecks.server.raft: entering leader state: leader="Node at 127.0.0.1:30341 [Leader]" writer.go:29: 2021-01-29T19:32:57.706Z [INFO] TestSessionCreate_NodeChecks.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:57.709Z [INFO] TestSessionCreate_NodeChecks.server: New leader elected: payload=Node-8c16cb26-18ef-6173-bdba-407fce823689 writer.go:29: 2021-01-29T19:32:57.711Z [DEBUG] TestSessionCreate_NodeChecks.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30341 === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.729Z [DEBUG] TestSessionCreate_Delete: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:57.731Z [DEBUG] TestSessionCreate_Delete: Node info in sync writer.go:29: 2021-01-29T19:32:57.733Z [DEBUG] TestSessionCreate_Delete: Node info in sync === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.787Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:57.796Z [INFO] TestSessionCreate_NodeChecks.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:57.796Z [INFO] TestSessionCreate_NodeChecks.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.797Z [DEBUG] TestSessionCreate_NodeChecks.server: Skipping self join check for node since the cluster is too small: node=Node-8c16cb26-18ef-6173-bdba-407fce823689 writer.go:29: 2021-01-29T19:32:57.798Z [INFO] TestSessionCreate_NodeChecks.server: member joined, marking health alive: member=Node-8c16cb26-18ef-6173-bdba-407fce823689 writer.go:29: 2021-01-29T19:32:57.822Z [INFO] TestSessionCreate_NodeChecks: Requesting shutdown writer.go:29: 2021-01-29T19:32:57.823Z [INFO] TestSessionCreate_NodeChecks.server: shutting down server writer.go:29: 2021-01-29T19:32:57.827Z [DEBUG] TestSessionCreate_NodeChecks.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.828Z [WARN] TestSessionCreate_NodeChecks.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:57.829Z [ERROR] TestSessionCreate_NodeChecks.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:57.829Z [DEBUG] TestSessionCreate_NodeChecks.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.830Z [WARN] TestSessionCreate_NodeChecks.server.serf.wan: serf: Shutdown without a Leave === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.832Z [INFO] TestSessionCreate_DefaultCheck: Requesting shutdown === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.833Z [INFO] TestSessionCreate_NodeChecks.server.router.manager: shutting down === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.835Z [INFO] TestSessionCreate_DefaultCheck.server: shutting down server === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.833Z [INFO] TestSessionCreate_NodeChecks: consul server down === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.836Z [DEBUG] TestSessionCreate_DefaultCheck.leader: stopping routine: routine="CA root pruning" === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.836Z [INFO] TestSessionCreate_NodeChecks: shutdown complete === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.836Z [DEBUG] TestSessionCreate_DefaultCheck.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.836Z [WARN] TestSessionCreate_DefaultCheck.server.serf.lan: serf: Shutdown without a Leave === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.837Z [INFO] TestSessionCreate_NodeChecks: Stopping server: protocol=DNS address=127.0.0.1:30336 network=tcp === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.839Z [WARN] TestSessionCreate_DefaultCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:57.841Z [INFO] TestSessionCreate_DefaultCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:57.841Z [INFO] TestSessionCreate_DefaultCheck: consul server down === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.839Z [INFO] TestSessionCreate_NodeChecks: Stopping server: protocol=DNS address=127.0.0.1:30336 network=udp === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.845Z [INFO] TestSessionCreate_DefaultCheck: shutdown complete === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:57.848Z [INFO] TestSessionCreate_NodeChecks: Stopping server: protocol=HTTP address=127.0.0.1:30337 network=tcp === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:57.848Z [INFO] TestSessionCreate_DefaultCheck: Stopping server: protocol=DNS address=127.0.0.1:30324 network=tcp writer.go:29: 2021-01-29T19:32:57.852Z [INFO] TestSessionCreate_DefaultCheck: Stopping server: protocol=DNS address=127.0.0.1:30324 network=udp writer.go:29: 2021-01-29T19:32:57.853Z [INFO] TestSessionCreate_DefaultCheck: Stopping server: protocol=HTTP address=127.0.0.1:30325 network=tcp === CONT TestSnapshot_Options/PUT writer.go:29: 2021-01-29T19:32:57.927Z [INFO] TestSnapshot_Options/PUT: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:57.931Z [INFO] TestSnapshot_Options/PUT: Endpoints down === RUN TestSnapshot_Options/PUT#01 writer.go:29: 2021-01-29T19:32:57.959Z [WARN] TestSnapshot_Options/PUT#01: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:32:57.960Z [WARN] TestSnapshot_Options/PUT#01: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:57.962Z [DEBUG] TestSnapshot_Options/PUT#01.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:57.965Z [DEBUG] TestSnapshot_Options/PUT#01.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.968Z [INFO] TestSessionCreate_Delete: Requesting shutdown writer.go:29: 2021-01-29T19:32:57.969Z [INFO] TestSessionCreate_Delete.server: shutting down server writer.go:29: 2021-01-29T19:32:57.970Z [DEBUG] TestSessionCreate_Delete.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.971Z [WARN] TestSessionCreate_Delete.server.serf.lan: serf: Shutdown without a Leave === CONT TestSnapshot_Options/PUT#01 writer.go:29: 2021-01-29T19:32:57.970Z [INFO] TestSnapshot_Options/PUT#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d1aea3fe-9a85-b44c-3b3b-6992c054a82b Address:127.0.0.1:30347}]" === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.971Z [DEBUG] TestSessionCreate_Delete.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.973Z [WARN] TestSessionCreate_Delete.server.serf.wan: serf: Shutdown without a Leave === CONT TestSnapshot_Options/PUT#01 writer.go:29: 2021-01-29T19:32:57.973Z [INFO] TestSnapshot_Options/PUT#01.server.raft: entering follower state: follower="Node at 127.0.0.1:30347 [Follower]" leader= === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.975Z [INFO] TestSessionCreate_Delete.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:57.976Z [INFO] TestSessionCreate_Delete: consul server down writer.go:29: 2021-01-29T19:32:57.977Z [INFO] TestSessionCreate_Delete: shutdown complete === CONT TestSnapshot_Options/PUT#01 writer.go:29: 2021-01-29T19:32:57.976Z [INFO] TestSnapshot_Options/PUT#01.server.serf.wan: serf: EventMemberJoin: Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b.dc1 127.0.0.1 === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:57.978Z [INFO] TestSessionCreate_Delete: Stopping server: protocol=DNS address=127.0.0.1:30330 network=tcp writer.go:29: 2021-01-29T19:32:57.979Z [INFO] TestSessionCreate_Delete: Stopping server: protocol=DNS address=127.0.0.1:30330 network=udp writer.go:29: 2021-01-29T19:32:57.980Z [INFO] TestSessionCreate_Delete: Stopping server: protocol=HTTP address=127.0.0.1:30331 network=tcp === CONT TestSnapshot_Options/PUT#01 writer.go:29: 2021-01-29T19:32:57.981Z [INFO] TestSnapshot_Options/PUT#01.server.serf.lan: serf: EventMemberJoin: Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b 127.0.0.1 writer.go:29: 2021-01-29T19:32:57.985Z [INFO] TestSnapshot_Options/PUT#01.server: Handled event for server in area: event=member-join server=Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b.dc1 area=wan writer.go:29: 2021-01-29T19:32:57.985Z [INFO] TestSnapshot_Options/PUT#01.server: Adding LAN server: server="Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b (Addr: tcp/127.0.0.1:30347) (DC: dc1)" writer.go:29: 2021-01-29T19:32:57.990Z [INFO] TestSnapshot_Options/PUT#01: Started DNS server: address=127.0.0.1:30342 network=tcp writer.go:29: 2021-01-29T19:32:57.991Z [INFO] TestSnapshot_Options/PUT#01: Started DNS server: address=127.0.0.1:30342 network=udp writer.go:29: 2021-01-29T19:32:57.993Z [INFO] TestSnapshot_Options/PUT#01: Started HTTP server: address=127.0.0.1:30343 network=tcp writer.go:29: 2021-01-29T19:32:57.994Z [INFO] TestSnapshot_Options/PUT#01: started state syncer writer.go:29: 2021-01-29T19:32:58.015Z [WARN] TestSnapshot_Options/PUT#01.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:58.016Z [INFO] TestSnapshot_Options/PUT#01.server.raft: entering candidate state: node="Node at 127.0.0.1:30347 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:58.019Z [DEBUG] TestSnapshot_Options/PUT#01.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:58.020Z [DEBUG] TestSnapshot_Options/PUT#01.server.raft: vote granted: from=d1aea3fe-9a85-b44c-3b3b-6992c054a82b term=2 tally=1 writer.go:29: 2021-01-29T19:32:58.021Z [INFO] TestSnapshot_Options/PUT#01.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:58.022Z [INFO] TestSnapshot_Options/PUT#01.server.raft: entering leader state: leader="Node at 127.0.0.1:30347 [Leader]" writer.go:29: 2021-01-29T19:32:58.023Z [INFO] TestSnapshot_Options/PUT#01.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:58.024Z [INFO] TestSnapshot_Options/PUT#01.server: New leader elected: payload=Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b writer.go:29: 2021-01-29T19:32:58.024Z [INFO] TestSnapshot_Options/PUT#01.server: initializing acls writer.go:29: 2021-01-29T19:32:58.027Z [INFO] TestSnapshot_Options/PUT#01.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:58.028Z [WARN] TestSnapshot_Options/PUT#01.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:58.032Z [INFO] TestSnapshot_Options/PUT#01.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:58.035Z [INFO] TestSnapshot_Options/PUT#01.server: initializing acls writer.go:29: 2021-01-29T19:32:58.036Z [WARN] TestSnapshot_Options/PUT#01.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:58.037Z [INFO] TestSnapshot_Options/PUT#01.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:58.038Z [INFO] TestSnapshot_Options/PUT#01.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:58.039Z [DEBUG] TestSnapshot_Options/PUT#01.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:32:58.040Z [INFO] TestSnapshot_Options/PUT#01.server.serf.lan: serf: EventMemberUpdate: Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b writer.go:29: 2021-01-29T19:32:58.041Z [INFO] TestSnapshot_Options/PUT#01.server.serf.wan: serf: EventMemberUpdate: Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b.dc1 writer.go:29: 2021-01-29T19:32:58.037Z [INFO] TestSnapshot_Options/PUT#01.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:58.042Z [INFO] TestSnapshot_Options/PUT#01.server: Updating LAN server: server="Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b (Addr: tcp/127.0.0.1:30347) (DC: dc1)" writer.go:29: 2021-01-29T19:32:58.043Z [INFO] TestSnapshot_Options/PUT#01.server: Handled event for server in area: event=member-update server=Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b.dc1 area=wan writer.go:29: 2021-01-29T19:32:58.044Z [INFO] TestSnapshot_Options/PUT#01.server.serf.lan: serf: EventMemberUpdate: Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b writer.go:29: 2021-01-29T19:32:58.050Z [INFO] TestSnapshot_Options/PUT#01.server: Updating LAN server: server="Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b (Addr: tcp/127.0.0.1:30347) (DC: dc1)" writer.go:29: 2021-01-29T19:32:58.050Z [INFO] TestSnapshot_Options/PUT#01.server.serf.wan: serf: EventMemberUpdate: Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b.dc1 writer.go:29: 2021-01-29T19:32:58.053Z [INFO] TestSnapshot_Options/PUT#01.server: Handled event for server in area: event=member-update server=Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b.dc1 area=wan writer.go:29: 2021-01-29T19:32:58.055Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:58.059Z [INFO] TestSnapshot_Options/PUT#01.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:58.060Z [INFO] TestSnapshot_Options/PUT#01.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:58.061Z [DEBUG] TestSnapshot_Options/PUT#01.server: Skipping self join check for node since the cluster is too small: node=Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b writer.go:29: 2021-01-29T19:32:58.062Z [INFO] TestSnapshot_Options/PUT#01.server: member joined, marking health alive: member=Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b writer.go:29: 2021-01-29T19:32:58.070Z [DEBUG] TestSnapshot_Options/PUT#01.server: Skipping self join check for node since the cluster is too small: node=Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b writer.go:29: 2021-01-29T19:32:58.071Z [DEBUG] TestSnapshot_Options/PUT#01.server: Skipping self join check for node since the cluster is too small: node=Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b writer.go:29: 2021-01-29T19:32:58.071Z [DEBUG] TestSnapshot_Options/PUT#01: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:58.074Z [INFO] TestSnapshot_Options/PUT#01: Synced node info writer.go:29: 2021-01-29T19:32:58.075Z [DEBUG] TestSnapshot_Options/PUT#01: Node info in sync writer.go:29: 2021-01-29T19:32:58.116Z [DEBUG] TestSnapshot_Options/PUT#01.acl: dropping node from result due to ACLs: node=Node-d1aea3fe-9a85-b44c-3b3b-6992c054a82b writer.go:29: 2021-01-29T19:32:58.119Z [INFO] TestSnapshot_Options/PUT#01: Requesting shutdown writer.go:29: 2021-01-29T19:32:58.120Z [INFO] TestSnapshot_Options/PUT#01.server: shutting down server writer.go:29: 2021-01-29T19:32:58.122Z [DEBUG] TestSnapshot_Options/PUT#01.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:58.123Z [DEBUG] TestSnapshot_Options/PUT#01.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:58.125Z [DEBUG] TestSnapshot_Options/PUT#01.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:58.123Z [DEBUG] TestSnapshot_Options/PUT#01.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:58.125Z [DEBUG] TestSnapshot_Options/PUT#01.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:58.126Z [WARN] TestSnapshot_Options/PUT#01.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:58.126Z [DEBUG] TestSnapshot_Options/PUT#01.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:58.133Z [WARN] TestSnapshot_Options/PUT#01.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:58.147Z [INFO] TestSnapshot_Options/PUT#01.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:58.147Z [INFO] TestSnapshot_Options/PUT#01: consul server down writer.go:29: 2021-01-29T19:32:58.153Z [INFO] TestSnapshot_Options/PUT#01: shutdown complete writer.go:29: 2021-01-29T19:32:58.154Z [INFO] TestSnapshot_Options/PUT#01: Stopping server: protocol=DNS address=127.0.0.1:30342 network=tcp writer.go:29: 2021-01-29T19:32:58.156Z [INFO] TestSnapshot_Options/PUT#01: Stopping server: protocol=DNS address=127.0.0.1:30342 network=udp writer.go:29: 2021-01-29T19:32:58.158Z [INFO] TestSnapshot_Options/PUT#01: Stopping server: protocol=HTTP address=127.0.0.1:30343 network=tcp === CONT TestSessionCreate_NodeChecks writer.go:29: 2021-01-29T19:32:58.349Z [INFO] TestSessionCreate_NodeChecks: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:58.351Z [INFO] TestSessionCreate_NodeChecks: Endpoints down --- PASS: TestSessionCreate_NodeChecks (0.84s) === CONT TestSessionCreate === CONT TestSessionCreate_DefaultCheck writer.go:29: 2021-01-29T19:32:58.354Z [INFO] TestSessionCreate_DefaultCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:58.356Z [INFO] TestSessionCreate_DefaultCheck: Endpoints down --- PASS: TestSessionCreate_DefaultCheck (1.00s) === CONT TestServiceManager_PersistService_ConfigFiles === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.375Z [WARN] TestSessionCreate: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:58.376Z [DEBUG] TestSessionCreate.tlsutil: Update: version=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.376Z [WARN] TestServiceManager_PersistService_ConfigFiles: bootstrap = true: do not enable unless necessary === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.377Z [DEBUG] TestSessionCreate.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.377Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:58.380Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.386Z [INFO] TestSessionCreate.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:abcb203b-cd9b-a876-9a87-e53bddf49a9a Address:127.0.0.1:30353}]" === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.387Z [INFO] TestServiceManager_PersistService_ConfigFiles.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:62b5279b-1bf5-1b76-e0c3-f44b4f62394e Address:127.0.0.1:30359}]" === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.388Z [INFO] TestSessionCreate.server.raft: entering follower state: follower="Node at 127.0.0.1:30353 [Follower]" leader= === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.390Z [INFO] TestServiceManager_PersistService_ConfigFiles.server.raft: entering follower state: follower="Node at 127.0.0.1:30359 [Follower]" leader= writer.go:29: 2021-01-29T19:32:58.391Z [INFO] TestServiceManager_PersistService_ConfigFiles.server.serf.wan: serf: EventMemberJoin: Node-62b5279b-1bf5-1b76-e0c3-f44b4f62394e.dc1 127.0.0.1 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.389Z [INFO] TestSessionCreate.server.serf.wan: serf: EventMemberJoin: Node-abcb203b-cd9b-a876-9a87-e53bddf49a9a.dc1 127.0.0.1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.395Z [INFO] TestServiceManager_PersistService_ConfigFiles.server.serf.lan: serf: EventMemberJoin: Node-62b5279b-1bf5-1b76-e0c3-f44b4f62394e 127.0.0.1 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.395Z [INFO] TestSessionCreate.server.serf.lan: serf: EventMemberJoin: Node-abcb203b-cd9b-a876-9a87-e53bddf49a9a 127.0.0.1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.399Z [INFO] TestServiceManager_PersistService_ConfigFiles.server: Handled event for server in area: event=member-join server=Node-62b5279b-1bf5-1b76-e0c3-f44b4f62394e.dc1 area=wan writer.go:29: 2021-01-29T19:32:58.399Z [INFO] TestServiceManager_PersistService_ConfigFiles.server: Adding LAN server: server="Node-62b5279b-1bf5-1b76-e0c3-f44b4f62394e (Addr: tcp/127.0.0.1:30359) (DC: dc1)" === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.401Z [INFO] TestSessionCreate: Started DNS server: address=127.0.0.1:30348 network=udp writer.go:29: 2021-01-29T19:32:58.401Z [INFO] TestSessionCreate.server: Adding LAN server: server="Node-abcb203b-cd9b-a876-9a87-e53bddf49a9a (Addr: tcp/127.0.0.1:30353) (DC: dc1)" writer.go:29: 2021-01-29T19:32:58.401Z [INFO] TestSessionCreate.server: Handled event for server in area: event=member-join server=Node-abcb203b-cd9b-a876-9a87-e53bddf49a9a.dc1 area=wan writer.go:29: 2021-01-29T19:32:58.403Z [INFO] TestSessionCreate: Started DNS server: address=127.0.0.1:30348 network=tcp writer.go:29: 2021-01-29T19:32:58.412Z [INFO] TestSessionCreate: Started HTTP server: address=127.0.0.1:30349 network=tcp === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.400Z [INFO] TestServiceManager_PersistService_ConfigFiles: Started DNS server: address=127.0.0.1:30354 network=udp === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.419Z [INFO] TestSessionCreate: started state syncer === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.419Z [INFO] TestServiceManager_PersistService_ConfigFiles: Started DNS server: address=127.0.0.1:30354 network=tcp writer.go:29: 2021-01-29T19:32:58.425Z [INFO] TestServiceManager_PersistService_ConfigFiles: Started HTTP server: address=127.0.0.1:30355 network=tcp writer.go:29: 2021-01-29T19:32:58.426Z [INFO] TestServiceManager_PersistService_ConfigFiles: started state syncer === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.431Z [WARN] TestSessionCreate.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:58.435Z [INFO] TestSessionCreate.server.raft: entering candidate state: node="Node at 127.0.0.1:30353 [Candidate]" term=2 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.445Z [WARN] TestServiceManager_PersistService_ConfigFiles.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:58.450Z [INFO] TestServiceManager_PersistService_ConfigFiles.server.raft: entering candidate state: node="Node at 127.0.0.1:30359 [Candidate]" term=2 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.444Z [DEBUG] TestSessionCreate.server.raft: votes: needed=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.458Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.server.raft: votes: needed=1 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.459Z [DEBUG] TestSessionCreate.server.raft: vote granted: from=abcb203b-cd9b-a876-9a87-e53bddf49a9a term=2 tally=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.467Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.server.raft: vote granted: from=62b5279b-1bf5-1b76-e0c3-f44b4f62394e term=2 tally=1 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.469Z [INFO] TestSessionCreate.server.raft: election won: tally=1 === CONT TestSessionCreate_Delete writer.go:29: 2021-01-29T19:32:58.481Z [INFO] TestSessionCreate_Delete: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:58.483Z [INFO] TestSessionCreate_Delete: Endpoints down --- PASS: TestSessionCreate_Delete (1.01s) === CONT TestHandleRemoteExecFailed === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.476Z [INFO] TestServiceManager_PersistService_ConfigFiles.server.raft: election won: tally=1 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.478Z [INFO] TestSessionCreate.server.raft: entering leader state: leader="Node at 127.0.0.1:30353 [Leader]" writer.go:29: 2021-01-29T19:32:58.487Z [INFO] TestSessionCreate.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:58.488Z [INFO] TestSessionCreate.server: New leader elected: payload=Node-abcb203b-cd9b-a876-9a87-e53bddf49a9a === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.497Z [WARN] TestHandleRemoteExecFailed: bootstrap = true: do not enable unless necessary === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.485Z [INFO] TestServiceManager_PersistService_ConfigFiles.server.raft: entering leader state: leader="Node at 127.0.0.1:30359 [Leader]" writer.go:29: 2021-01-29T19:32:58.499Z [INFO] TestServiceManager_PersistService_ConfigFiles.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:58.500Z [INFO] TestServiceManager_PersistService_ConfigFiles.server: New leader elected: payload=Node-62b5279b-1bf5-1b76-e0c3-f44b4f62394e === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.499Z [DEBUG] TestHandleRemoteExecFailed.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:58.506Z [DEBUG] TestHandleRemoteExecFailed.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:58.513Z [INFO] TestHandleRemoteExecFailed.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:97d20984-1008-5ae3-d4f2-fef3a217e902 Address:127.0.0.1:30365}]" === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.496Z [DEBUG] TestSessionCreate.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30353 === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.515Z [INFO] TestHandleRemoteExecFailed.server.raft: entering follower state: follower="Node at 127.0.0.1:30365 [Follower]" leader= writer.go:29: 2021-01-29T19:32:58.518Z [INFO] TestHandleRemoteExecFailed.server.serf.wan: serf: EventMemberJoin: Node-97d20984-1008-5ae3-d4f2-fef3a217e902.dc1 127.0.0.1 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.520Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.523Z [INFO] TestHandleRemoteExecFailed.server.serf.lan: serf: EventMemberJoin: Node-97d20984-1008-5ae3-d4f2-fef3a217e902 127.0.0.1 writer.go:29: 2021-01-29T19:32:58.527Z [INFO] TestHandleRemoteExecFailed.server: Adding LAN server: server="Node-97d20984-1008-5ae3-d4f2-fef3a217e902 (Addr: tcp/127.0.0.1:30365) (DC: dc1)" === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.504Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30359 === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.529Z [INFO] TestHandleRemoteExecFailed: Started DNS server: address=127.0.0.1:30360 network=udp writer.go:29: 2021-01-29T19:32:58.529Z [INFO] TestHandleRemoteExecFailed.server: Handled event for server in area: event=member-join server=Node-97d20984-1008-5ae3-d4f2-fef3a217e902.dc1 area=wan writer.go:29: 2021-01-29T19:32:58.533Z [INFO] TestHandleRemoteExecFailed: Started DNS server: address=127.0.0.1:30360 network=tcp === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.533Z [INFO] TestSessionCreate.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.533Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.535Z [INFO] TestHandleRemoteExecFailed: Started HTTP server: address=127.0.0.1:30361 network=tcp === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.535Z [INFO] TestSessionCreate.leader: started routine: routine="CA root pruning" === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.536Z [INFO] TestHandleRemoteExecFailed: started state syncer === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.536Z [DEBUG] TestSessionCreate.server: Skipping self join check for node since the cluster is too small: node=Node-abcb203b-cd9b-a876-9a87-e53bddf49a9a writer.go:29: 2021-01-29T19:32:58.537Z [INFO] TestSessionCreate.server: member joined, marking health alive: member=Node-abcb203b-cd9b-a876-9a87-e53bddf49a9a === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.541Z [INFO] TestServiceManager_PersistService_ConfigFiles.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:58.542Z [INFO] TestServiceManager_PersistService_ConfigFiles.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:58.543Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.server: Skipping self join check for node since the cluster is too small: node=Node-62b5279b-1bf5-1b76-e0c3-f44b4f62394e writer.go:29: 2021-01-29T19:32:58.544Z [INFO] TestServiceManager_PersistService_ConfigFiles.server: member joined, marking health alive: member=Node-62b5279b-1bf5-1b76-e0c3-f44b4f62394e === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.571Z [WARN] TestHandleRemoteExecFailed.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:58.576Z [INFO] TestHandleRemoteExecFailed.server.raft: entering candidate state: node="Node at 127.0.0.1:30365 [Candidate]" term=2 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.581Z [INFO] TestSessionCreate: Requesting shutdown writer.go:29: 2021-01-29T19:32:58.583Z [INFO] TestSessionCreate.server: shutting down server writer.go:29: 2021-01-29T19:32:58.584Z [DEBUG] TestSessionCreate.leader: stopping routine: routine="CA root pruning" === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.581Z [DEBUG] TestHandleRemoteExecFailed.server.raft: votes: needed=1 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.584Z [ERROR] TestSessionCreate.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.586Z [DEBUG] TestHandleRemoteExecFailed.server.raft: vote granted: from=97d20984-1008-5ae3-d4f2-fef3a217e902 term=2 tally=1 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.586Z [DEBUG] TestSessionCreate.leader: stopped routine: routine="CA root pruning" === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.588Z [INFO] TestHandleRemoteExecFailed.server.raft: election won: tally=1 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.586Z [WARN] TestSessionCreate.server.serf.lan: serf: Shutdown without a Leave === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.589Z [INFO] TestHandleRemoteExecFailed.server.raft: entering leader state: leader="Node at 127.0.0.1:30365 [Leader]" writer.go:29: 2021-01-29T19:32:58.591Z [INFO] TestHandleRemoteExecFailed.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:58.592Z [INFO] TestHandleRemoteExecFailed.server: New leader elected: payload=Node-97d20984-1008-5ae3-d4f2-fef3a217e902 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.594Z [WARN] TestSessionCreate.server.serf.wan: serf: Shutdown without a Leave === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.596Z [DEBUG] TestHandleRemoteExecFailed.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30365 === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:58.599Z [INFO] TestSessionCreate.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:58.599Z [INFO] TestSessionCreate: consul server down writer.go:29: 2021-01-29T19:32:58.602Z [INFO] TestSessionCreate: shutdown complete writer.go:29: 2021-01-29T19:32:58.603Z [INFO] TestSessionCreate: Stopping server: protocol=DNS address=127.0.0.1:30348 network=tcp writer.go:29: 2021-01-29T19:32:58.604Z [INFO] TestSessionCreate: Stopping server: protocol=DNS address=127.0.0.1:30348 network=udp writer.go:29: 2021-01-29T19:32:58.605Z [INFO] TestSessionCreate: Stopping server: protocol=HTTP address=127.0.0.1:30349 network=tcp === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.614Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:58.620Z [INFO] TestHandleRemoteExecFailed.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:58.622Z [INFO] TestHandleRemoteExecFailed.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:58.624Z [DEBUG] TestHandleRemoteExecFailed.server: Skipping self join check for node since the cluster is too small: node=Node-97d20984-1008-5ae3-d4f2-fef3a217e902 writer.go:29: 2021-01-29T19:32:58.625Z [INFO] TestHandleRemoteExecFailed.server: member joined, marking health alive: member=Node-97d20984-1008-5ae3-d4f2-fef3a217e902 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.650Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.tlsutil: Update: version=1 === CONT TestSnapshot_Options/PUT#01 writer.go:29: 2021-01-29T19:32:58.659Z [INFO] TestSnapshot_Options/PUT#01: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:58.660Z [INFO] TestSnapshot_Options/PUT#01: Endpoints down === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.656Z [INFO] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: EventMemberJoin: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 127.0.0.1 === RUN TestSnapshot_Options/PUT#02 === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.662Z [DEBUG] TestHandleRemoteExecFailed: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:58.664Z [INFO] TestHandleRemoteExecFailed: Synced node info === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.676Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.681Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.684Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.685Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: added local registration for service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:32:58.704Z [ERROR] TestServiceManager_PersistService_ConfigFiles: error handling service update: error="error watching service config: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.710Z [ERROR] TestServiceManager_PersistService_ConfigFiles: error handling service update: error="error watching service config: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.710Z [ERROR] TestServiceManager_PersistService_ConfigFiles: error handling service update: error="error watching service config: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.712Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.716Z [ERROR] TestServiceManager_PersistService_ConfigFiles: error handling service update: error="error watching service config: No known Consul servers" === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.719Z [WARN] TestSnapshot_Options/PUT#02: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:32:58.720Z [WARN] TestSnapshot_Options/PUT#02: bootstrap = true: do not enable unless necessary === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.712Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.714Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.721Z [DEBUG] TestSnapshot_Options/PUT#02.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:58.731Z [DEBUG] TestSnapshot_Options/PUT#02.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.715Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.716Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:32:58.744Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=roots error="error filling agent cache: No known Consul servers" === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.746Z [INFO] TestSnapshot_Options/PUT#02.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4bef944b-4474-fe3d-7926-48b9974d151f Address:127.0.0.1:30377}]" === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.716Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.748Z [INFO] TestSnapshot_Options/PUT#02.server.raft: entering follower state: follower="Node at 127.0.0.1:30377 [Follower]" leader= === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.719Z [INFO] TestServiceManager_PersistService_ConfigFiles: Started DNS server: address=127.0.0.1:30366 network=tcp === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.748Z [INFO] TestSnapshot_Options/PUT#02.server.serf.wan: serf: EventMemberJoin: Node-4bef944b-4474-fe3d-7926-48b9974d151f.dc1 127.0.0.1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.733Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.733Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.741Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.744Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=leaf error="error filling agent cache: No known Consul servers" === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.753Z [INFO] TestSnapshot_Options/PUT#02.server.serf.lan: serf: EventMemberJoin: Node-4bef944b-4474-fe3d-7926-48b9974d151f 127.0.0.1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.750Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.758Z [INFO] TestSnapshot_Options/PUT#02.server: Handled event for server in area: event=member-join server=Node-4bef944b-4474-fe3d-7926-48b9974d151f.dc1 area=wan writer.go:29: 2021-01-29T19:32:58.759Z [INFO] TestSnapshot_Options/PUT#02.server: Adding LAN server: server="Node-4bef944b-4474-fe3d-7926-48b9974d151f (Addr: tcp/127.0.0.1:30377) (DC: dc1)" === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.751Z [INFO] TestServiceManager_PersistService_ConfigFiles: Started DNS server: address=127.0.0.1:30366 network=udp === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.761Z [INFO] TestSnapshot_Options/PUT#02: Started DNS server: address=127.0.0.1:30372 network=udp writer.go:29: 2021-01-29T19:32:58.761Z [INFO] TestSnapshot_Options/PUT#02: Started DNS server: address=127.0.0.1:30372 network=tcp writer.go:29: 2021-01-29T19:32:58.763Z [INFO] TestSnapshot_Options/PUT#02: Started HTTP server: address=127.0.0.1:30373 network=tcp === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.752Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.764Z [INFO] TestSnapshot_Options/PUT#02: started state syncer === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.754Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.755Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.756Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=discovery-chain:redis error="error filling agent cache: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.758Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.762Z [INFO] TestServiceManager_PersistService_ConfigFiles: Started HTTP server: address=127.0.0.1:30367 network=tcp writer.go:29: 2021-01-29T19:32:58.766Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.768Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.769Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.770Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=intentions error="error filling agent cache: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.772Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.772Z [INFO] TestServiceManager_PersistService_ConfigFiles: started state syncer writer.go:29: 2021-01-29T19:32:58.773Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:32:58.777Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=roots error="error filling agent cache: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.779Z [INFO] TestServiceManager_PersistService_ConfigFiles: (LAN) joining: lan_addresses=[127.0.0.1:30357] writer.go:29: 2021-01-29T19:32:58.780Z [ERROR] TestServiceManager_PersistService_ConfigFiles.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:32:58.780Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=leaf error="error filling agent cache: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.781Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.memberlist.lan: memberlist: Initiating push/pull sync with: 127.0.0.1:30357 writer.go:29: 2021-01-29T19:32:58.782Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.server.memberlist.lan: memberlist: Stream connection from=127.0.0.1:44666 === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.788Z [WARN] TestSnapshot_Options/PUT#02.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.788Z [INFO] TestServiceManager_PersistService_ConfigFiles.server.serf.lan: serf: EventMemberJoin: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 127.0.0.1 === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.789Z [INFO] TestSnapshot_Options/PUT#02.server.raft: entering candidate state: node="Node at 127.0.0.1:30377 [Candidate]" term=2 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.790Z [INFO] TestServiceManager_PersistService_ConfigFiles.server: member joined, marking health alive: member=Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 writer.go:29: 2021-01-29T19:32:58.783Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=discovery-chain:redis error="error filling agent cache: No known Consul servers" === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.792Z [DEBUG] TestSnapshot_Options/PUT#02.server.raft: votes: needed=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.787Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.793Z [DEBUG] TestSnapshot_Options/PUT#02.server.raft: vote granted: from=4bef944b-4474-fe3d-7926-48b9974d151f term=2 tally=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.793Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=intentions error="error filling agent cache: No known Consul servers" === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.795Z [INFO] TestSnapshot_Options/PUT#02.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:58.796Z [INFO] TestSnapshot_Options/PUT#02.server.raft: entering leader state: leader="Node at 127.0.0.1:30377 [Leader]" writer.go:29: 2021-01-29T19:32:58.798Z [INFO] TestSnapshot_Options/PUT#02.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:58.800Z [INFO] TestSnapshot_Options/PUT#02.server: New leader elected: payload=Node-4bef944b-4474-fe3d-7926-48b9974d151f writer.go:29: 2021-01-29T19:32:58.800Z [INFO] TestSnapshot_Options/PUT#02.server: initializing acls === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.792Z [INFO] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: EventMemberJoin: Node-62b5279b-1bf5-1b76-e0c3-f44b4f62394e 127.0.0.1 === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.803Z [INFO] TestSnapshot_Options/PUT#02.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:58.803Z [WARN] TestSnapshot_Options/PUT#02.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:58.805Z [INFO] TestSnapshot_Options/PUT#02.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:58.808Z [INFO] TestSnapshot_Options/PUT#02.server: Created ACL anonymous token from configuration === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.804Z [INFO] TestServiceManager_PersistService_ConfigFiles: (LAN) joined: number_of_nodes=1 === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.809Z [INFO] TestSnapshot_Options/PUT#02.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:58.810Z [INFO] TestSnapshot_Options/PUT#02.leader: started routine: routine="acl token reaping" === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.796Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=roots error="error filling agent cache: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.804Z [INFO] TestServiceManager_PersistService_ConfigFiles.client: adding server: server="Node-62b5279b-1bf5-1b76-e0c3-f44b4f62394e (Addr: tcp/127.0.0.1:30359) (DC: dc1)" writer.go:29: 2021-01-29T19:32:58.806Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: messageUserEventType: consul:new-leader === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.810Z [INFO] TestSnapshot_Options/PUT#02.server.serf.lan: serf: EventMemberUpdate: Node-4bef944b-4474-fe3d-7926-48b9974d151f === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.809Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: systemd notify failed: error="No socket" === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.809Z [INFO] TestSnapshot_Options/PUT#02.server: initializing acls === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.795Z [INFO] TestServiceManager_PersistService_ConfigFiles: Synced node info === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.825Z [WARN] TestSnapshot_Options/PUT#02.server: Configuring a non-UUID master token is deprecated === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.825Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: Node info in sync === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.827Z [INFO] TestSnapshot_Options/PUT#02.server: Updating LAN server: server="Node-4bef944b-4474-fe3d-7926-48b9974d151f (Addr: tcp/127.0.0.1:30377) (DC: dc1)" writer.go:29: 2021-01-29T19:32:58.827Z [INFO] TestSnapshot_Options/PUT#02.server.serf.wan: serf: EventMemberUpdate: Node-4bef944b-4474-fe3d-7926-48b9974d151f.dc1 writer.go:29: 2021-01-29T19:32:58.829Z [INFO] TestSnapshot_Options/PUT#02.server: Handled event for server in area: event=member-update server=Node-4bef944b-4474-fe3d-7926-48b9974d151f.dc1 area=wan writer.go:29: 2021-01-29T19:32:58.829Z [DEBUG] TestSnapshot_Options/PUT#02.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:32:58.830Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.813Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=leaf error="error filling agent cache: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.832Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:58.849Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=discovery-chain:redis error="error filling agent cache: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.851Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=intentions error="error filling agent cache: No known Consul servers" === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.837Z [INFO] TestSnapshot_Options/PUT#02.server.serf.lan: serf: EventMemberUpdate: Node-4bef944b-4474-fe3d-7926-48b9974d151f === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.853Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=roots error="error filling agent cache: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.854Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=leaf error="error filling agent cache: No known Consul servers" writer.go:29: 2021-01-29T19:32:58.855Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=discovery-chain:redis error="error filling agent cache: No known Consul servers" === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.843Z [INFO] TestSnapshot_Options/PUT#02.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.857Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=intentions error="error filling agent cache: No known Consul servers" === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.854Z [INFO] TestSnapshot_Options/PUT#02.server.serf.wan: serf: EventMemberUpdate: Node-4bef944b-4474-fe3d-7926-48b9974d151f.dc1 writer.go:29: 2021-01-29T19:32:58.854Z [INFO] TestSnapshot_Options/PUT#02.server: Updating LAN server: server="Node-4bef944b-4474-fe3d-7926-48b9974d151f (Addr: tcp/127.0.0.1:30377) (DC: dc1)" writer.go:29: 2021-01-29T19:32:58.857Z [INFO] TestSnapshot_Options/PUT#02.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:58.867Z [DEBUG] TestSnapshot_Options/PUT#02.server: Skipping self join check for node since the cluster is too small: node=Node-4bef944b-4474-fe3d-7926-48b9974d151f writer.go:29: 2021-01-29T19:32:58.864Z [INFO] TestSnapshot_Options/PUT#02.server: Handled event for server in area: event=member-update server=Node-4bef944b-4474-fe3d-7926-48b9974d151f.dc1 area=wan writer.go:29: 2021-01-29T19:32:58.869Z [INFO] TestSnapshot_Options/PUT#02.server: member joined, marking health alive: member=Node-4bef944b-4474-fe3d-7926-48b9974d151f === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:58.881Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.server.serf.lan: serf: messageJoinType: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.884Z [DEBUG] TestSnapshot_Options/PUT#02.server: Skipping self join check for node since the cluster is too small: node=Node-4bef944b-4474-fe3d-7926-48b9974d151f writer.go:29: 2021-01-29T19:32:58.886Z [DEBUG] TestSnapshot_Options/PUT#02.server: Skipping self join check for node since the cluster is too small: node=Node-4bef944b-4474-fe3d-7926-48b9974d151f === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.955Z [DEBUG] TestHandleRemoteExecFailed: received remote exec event: id=ba3cc2a4-93e5-fac2-bc20-182fcc3f1cc9 writer.go:29: 2021-01-29T19:32:58.960Z [INFO] TestHandleRemoteExecFailed: remote exec script: script="echo failing;exit 2" === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:58.961Z [DEBUG] TestSnapshot_Options/PUT#02: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:58.963Z [INFO] TestSnapshot_Options/PUT#02: Synced node info === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:58.970Z [INFO] TestHandleRemoteExecFailed: Requesting shutdown writer.go:29: 2021-01-29T19:32:58.972Z [INFO] TestHandleRemoteExecFailed.server: shutting down server writer.go:29: 2021-01-29T19:32:58.974Z [DEBUG] TestHandleRemoteExecFailed.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:58.975Z [WARN] TestHandleRemoteExecFailed.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:58.975Z [DEBUG] TestHandleRemoteExecFailed.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:58.981Z [WARN] TestHandleRemoteExecFailed.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:58.983Z [INFO] TestHandleRemoteExecFailed.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:58.983Z [INFO] TestHandleRemoteExecFailed: consul server down writer.go:29: 2021-01-29T19:32:58.987Z [INFO] TestHandleRemoteExecFailed: shutdown complete writer.go:29: 2021-01-29T19:32:58.989Z [INFO] TestHandleRemoteExecFailed: Stopping server: protocol=DNS address=127.0.0.1:30360 network=tcp writer.go:29: 2021-01-29T19:32:58.991Z [INFO] TestHandleRemoteExecFailed: Stopping server: protocol=DNS address=127.0.0.1:30360 network=udp writer.go:29: 2021-01-29T19:32:58.995Z [INFO] TestHandleRemoteExecFailed: Stopping server: protocol=HTTP address=127.0.0.1:30361 network=tcp === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:59.013Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: messageUserEventType: consul:new-leader writer.go:29: 2021-01-29T19:32:59.014Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: messageJoinType: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:59.047Z [DEBUG] TestSnapshot_Options/PUT#02.acl: dropping node from result due to ACLs: node=Node-4bef944b-4474-fe3d-7926-48b9974d151f writer.go:29: 2021-01-29T19:32:59.052Z [INFO] TestSnapshot_Options/PUT#02: Requesting shutdown writer.go:29: 2021-01-29T19:32:59.053Z [INFO] TestSnapshot_Options/PUT#02.server: shutting down server writer.go:29: 2021-01-29T19:32:59.061Z [DEBUG] TestSnapshot_Options/PUT#02.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:59.065Z [DEBUG] TestSnapshot_Options/PUT#02.leader: stopping routine: routine="acl token reaping" === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:59.072Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.server.serf.lan: serf: messageJoinType: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:59.069Z [DEBUG] TestSnapshot_Options/PUT#02.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:59.068Z [DEBUG] TestSnapshot_Options/PUT#02.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:59.069Z [DEBUG] TestSnapshot_Options/PUT#02.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:59.077Z [WARN] TestSnapshot_Options/PUT#02.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:59.077Z [DEBUG] TestSnapshot_Options/PUT#02.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:59.079Z [WARN] TestSnapshot_Options/PUT#02.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:59.089Z [INFO] TestSnapshot_Options/PUT#02.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:59.091Z [INFO] TestSnapshot_Options/PUT#02: consul server down writer.go:29: 2021-01-29T19:32:59.095Z [INFO] TestSnapshot_Options/PUT#02: shutdown complete writer.go:29: 2021-01-29T19:32:59.099Z [INFO] TestSnapshot_Options/PUT#02: Stopping server: protocol=DNS address=127.0.0.1:30372 network=tcp writer.go:29: 2021-01-29T19:32:59.101Z [INFO] TestSnapshot_Options/PUT#02: Stopping server: protocol=DNS address=127.0.0.1:30372 network=udp writer.go:29: 2021-01-29T19:32:59.103Z [INFO] TestSnapshot_Options/PUT#02: Stopping server: protocol=HTTP address=127.0.0.1:30373 network=tcp === CONT TestSessionCreate writer.go:29: 2021-01-29T19:32:59.106Z [INFO] TestSessionCreate: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:59.107Z [INFO] TestSessionCreate: Endpoints down --- PASS: TestSessionCreate (0.76s) === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.123Z [WARN] TestHandleRemoteExec: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:59.125Z [DEBUG] TestHandleRemoteExec.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:59.140Z [DEBUG] TestHandleRemoteExec.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:59.209Z [INFO] TestHandleRemoteExec.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:49562470-d4ab-27fd-a97e-1568f6a8399c Address:127.0.0.1:30383}]" === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:59.210Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: messageUserEventType: consul:new-leader writer.go:29: 2021-01-29T19:32:59.212Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: messageJoinType: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.215Z [INFO] TestHandleRemoteExec.server.serf.wan: serf: EventMemberJoin: Node-49562470-d4ab-27fd-a97e-1568f6a8399c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:59.215Z [INFO] TestHandleRemoteExec.server.raft: entering follower state: follower="Node at 127.0.0.1:30383 [Follower]" leader= writer.go:29: 2021-01-29T19:32:59.227Z [INFO] TestHandleRemoteExec.server.serf.lan: serf: EventMemberJoin: Node-49562470-d4ab-27fd-a97e-1568f6a8399c 127.0.0.1 writer.go:29: 2021-01-29T19:32:59.242Z [INFO] TestHandleRemoteExec: Started DNS server: address=127.0.0.1:30378 network=udp writer.go:29: 2021-01-29T19:32:59.262Z [INFO] TestHandleRemoteExec: Started DNS server: address=127.0.0.1:30378 network=tcp writer.go:29: 2021-01-29T19:32:59.269Z [WARN] TestHandleRemoteExec.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:59.248Z [INFO] TestHandleRemoteExec.server: Adding LAN server: server="Node-49562470-d4ab-27fd-a97e-1568f6a8399c (Addr: tcp/127.0.0.1:30383) (DC: dc1)" writer.go:29: 2021-01-29T19:32:59.250Z [INFO] TestHandleRemoteExec.server: Handled event for server in area: event=member-join server=Node-49562470-d4ab-27fd-a97e-1568f6a8399c.dc1 area=wan writer.go:29: 2021-01-29T19:32:59.273Z [INFO] TestHandleRemoteExec.server.raft: entering candidate state: node="Node at 127.0.0.1:30383 [Candidate]" term=2 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:59.276Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.server.serf.lan: serf: messageJoinType: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.275Z [INFO] TestHandleRemoteExec: Started HTTP server: address=127.0.0.1:30379 network=tcp writer.go:29: 2021-01-29T19:32:59.278Z [INFO] TestHandleRemoteExec: started state syncer writer.go:29: 2021-01-29T19:32:59.313Z [DEBUG] TestHandleRemoteExec.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:59.315Z [DEBUG] TestHandleRemoteExec.server.raft: vote granted: from=49562470-d4ab-27fd-a97e-1568f6a8399c term=2 tally=1 writer.go:29: 2021-01-29T19:32:59.317Z [INFO] TestHandleRemoteExec.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:59.319Z [INFO] TestHandleRemoteExec.server.raft: entering leader state: leader="Node at 127.0.0.1:30383 [Leader]" writer.go:29: 2021-01-29T19:32:59.324Z [INFO] TestHandleRemoteExec.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:59.327Z [INFO] TestHandleRemoteExec.server: New leader elected: payload=Node-49562470-d4ab-27fd-a97e-1568f6a8399c writer.go:29: 2021-01-29T19:32:59.355Z [DEBUG] TestHandleRemoteExec.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30383 writer.go:29: 2021-01-29T19:32:59.404Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:32:59.418Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.server.serf.lan: serf: messageJoinType: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 writer.go:29: 2021-01-29T19:32:59.419Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: messageJoinType: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 writer.go:29: 2021-01-29T19:32:59.422Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: messageUserEventType: consul:new-leader writer.go:29: 2021-01-29T19:32:59.424Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: messageJoinType: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20 === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.465Z [INFO] TestHandleRemoteExec: Synced node info === CONT TestHandleRemoteExecFailed writer.go:29: 2021-01-29T19:32:59.502Z [INFO] TestHandleRemoteExecFailed: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:59.504Z [INFO] TestHandleRemoteExecFailed: Endpoints down --- PASS: TestHandleRemoteExecFailed (1.02s) === CONT TestRemoteExecWrites_ACLDeny === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.533Z [INFO] TestHandleRemoteExec.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:59.535Z [INFO] TestHandleRemoteExec.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:59.537Z [DEBUG] TestHandleRemoteExec.server: Skipping self join check for node since the cluster is too small: node=Node-49562470-d4ab-27fd-a97e-1568f6a8399c writer.go:29: 2021-01-29T19:32:59.538Z [INFO] TestHandleRemoteExec.server: member joined, marking health alive: member=Node-49562470-d4ab-27fd-a97e-1568f6a8399c === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.547Z [WARN] TestRemoteExecWrites_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:32:59.548Z [WARN] TestRemoteExecWrites_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:59.549Z [DEBUG] TestRemoteExecWrites_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:59.552Z [DEBUG] TestRemoteExecWrites_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:59.560Z [INFO] TestRemoteExecWrites_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1264304e-c01d-c230-5755-3eb6f1b28a12 Address:127.0.0.1:30389}]" writer.go:29: 2021-01-29T19:32:59.562Z [INFO] TestRemoteExecWrites_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-1264304e-c01d-c230-5755-3eb6f1b28a12.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:59.562Z [INFO] TestRemoteExecWrites_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30389 [Follower]" leader= writer.go:29: 2021-01-29T19:32:59.565Z [INFO] TestRemoteExecWrites_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-1264304e-c01d-c230-5755-3eb6f1b28a12 127.0.0.1 writer.go:29: 2021-01-29T19:32:59.573Z [INFO] TestRemoteExecWrites_ACLDeny.server: Handled event for server in area: event=member-join server=Node-1264304e-c01d-c230-5755-3eb6f1b28a12.dc1 area=wan writer.go:29: 2021-01-29T19:32:59.573Z [INFO] TestRemoteExecWrites_ACLDeny.server: Adding LAN server: server="Node-1264304e-c01d-c230-5755-3eb6f1b28a12 (Addr: tcp/127.0.0.1:30389) (DC: dc1)" writer.go:29: 2021-01-29T19:32:59.579Z [INFO] TestRemoteExecWrites_ACLDeny: Started DNS server: address=127.0.0.1:30384 network=tcp writer.go:29: 2021-01-29T19:32:59.581Z [INFO] TestRemoteExecWrites_ACLDeny: Started DNS server: address=127.0.0.1:30384 network=udp writer.go:29: 2021-01-29T19:32:59.582Z [INFO] TestRemoteExecWrites_ACLDeny: Started HTTP server: address=127.0.0.1:30385 network=tcp writer.go:29: 2021-01-29T19:32:59.585Z [INFO] TestRemoteExecWrites_ACLDeny: started state syncer === CONT TestSnapshot_Options/PUT#02 writer.go:29: 2021-01-29T19:32:59.604Z [INFO] TestSnapshot_Options/PUT#02: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:59.605Z [INFO] TestSnapshot_Options/PUT#02: Endpoints down --- PASS: TestSnapshot_Options (4.94s) --- PASS: TestSnapshot_Options/GET (0.97s) --- PASS: TestSnapshot_Options/GET#01 (0.71s) --- PASS: TestSnapshot_Options/GET#02 (0.72s) --- PASS: TestSnapshot_Options/PUT (0.86s) --- PASS: TestSnapshot_Options/PUT#01 (0.73s) --- PASS: TestSnapshot_Options/PUT#02 (0.94s) === CONT TestRemoteExecWrites_ACLAgentToken === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.627Z [DEBUG] TestRemoteExecWrites_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.635Z [DEBUG] TestHandleRemoteExec: received remote exec event: id=dfe0447b-7f05-0a01-7819-877eb20d9a90 writer.go:29: 2021-01-29T19:32:59.637Z [INFO] TestHandleRemoteExec: remote exec script: script=uptime === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.638Z [WARN] TestRemoteExecWrites_ACLAgentToken: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:32:59.639Z [WARN] TestRemoteExecWrites_ACLAgentToken: bootstrap = true: do not enable unless necessary === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.641Z [WARN] TestRemoteExecWrites_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:59.641Z [INFO] TestRemoteExecWrites_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30389 [Candidate]" term=2 === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.641Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.tlsutil: Update: version=1 === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.665Z [DEBUG] TestRemoteExecWrites_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:59.666Z [DEBUG] TestRemoteExecWrites_ACLDeny.server.raft: vote granted: from=1264304e-c01d-c230-5755-3eb6f1b28a12 term=2 tally=1 writer.go:29: 2021-01-29T19:32:59.667Z [INFO] TestRemoteExecWrites_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:59.668Z [INFO] TestRemoteExecWrites_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30389 [Leader]" writer.go:29: 2021-01-29T19:32:59.669Z [INFO] TestRemoteExecWrites_ACLDeny.server: cluster leadership acquired === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.649Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.672Z [INFO] TestRemoteExecWrites_ACLDeny.server: New leader elected: payload=Node-1264304e-c01d-c230-5755-3eb6f1b28a12 === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.676Z [INFO] TestHandleRemoteExec: Requesting shutdown === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.676Z [INFO] TestRemoteExecWrites_ACLDeny.server: initializing acls === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.678Z [INFO] TestHandleRemoteExec.server: shutting down server writer.go:29: 2021-01-29T19:32:59.679Z [DEBUG] TestHandleRemoteExec.leader: stopping routine: routine="CA root pruning" === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.680Z [INFO] TestRemoteExecWrites_ACLDeny.server: Created ACL 'global-management' policy === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.680Z [WARN] TestHandleRemoteExec.server.serf.lan: serf: Shutdown without a Leave === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.682Z [WARN] TestRemoteExecWrites_ACLDeny.server: Configuring a non-UUID master token is deprecated === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.680Z [DEBUG] TestHandleRemoteExec.leader: stopped routine: routine="CA root pruning" === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.684Z [INFO] TestRemoteExecWrites_ACLDeny.server: Bootstrapped ACL master token from configuration === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.684Z [WARN] TestHandleRemoteExec.server.serf.wan: serf: Shutdown without a Leave === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.686Z [INFO] TestRemoteExecWrites_ACLDeny.server: Created ACL anonymous token from configuration === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.682Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b844b6ce-7d74-6812-c209-7d04c2651bc6 Address:127.0.0.1:30395}]" writer.go:29: 2021-01-29T19:32:59.691Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.raft: entering follower state: follower="Node at 127.0.0.1:30395 [Follower]" leader= writer.go:29: 2021-01-29T19:32:59.692Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.serf.wan: serf: EventMemberJoin: Node-b844b6ce-7d74-6812-c209-7d04c2651bc6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:59.698Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.serf.lan: serf: EventMemberJoin: Node-b844b6ce-7d74-6812-c209-7d04c2651bc6 127.0.0.1 writer.go:29: 2021-01-29T19:32:59.701Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: Handled event for server in area: event=member-join server=Node-b844b6ce-7d74-6812-c209-7d04c2651bc6.dc1 area=wan === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.689Z [INFO] TestHandleRemoteExec.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:59.691Z [INFO] TestHandleRemoteExec: consul server down === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.703Z [INFO] TestRemoteExecWrites_ACLAgentToken: Started DNS server: address=127.0.0.1:30390 network=udp === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.688Z [INFO] TestRemoteExecWrites_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.706Z [INFO] TestHandleRemoteExec: shutdown complete === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.703Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: Adding LAN server: server="Node-b844b6ce-7d74-6812-c209-7d04c2651bc6 (Addr: tcp/127.0.0.1:30395) (DC: dc1)" === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.708Z [INFO] TestHandleRemoteExec: Stopping server: protocol=DNS address=127.0.0.1:30378 network=tcp === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.706Z [INFO] TestRemoteExecWrites_ACLDeny.leader: started routine: routine="acl token reaping" === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:32:59.711Z [INFO] TestHandleRemoteExec: Stopping server: protocol=DNS address=127.0.0.1:30378 network=udp writer.go:29: 2021-01-29T19:32:59.716Z [INFO] TestHandleRemoteExec: Stopping server: protocol=HTTP address=127.0.0.1:30379 network=tcp === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.715Z [INFO] TestRemoteExecWrites_ACLAgentToken: Started DNS server: address=127.0.0.1:30390 network=tcp === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.712Z [INFO] TestRemoteExecWrites_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-1264304e-c01d-c230-5755-3eb6f1b28a12 === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.726Z [INFO] TestRemoteExecWrites_ACLAgentToken: Started HTTP server: address=127.0.0.1:30391 network=tcp writer.go:29: 2021-01-29T19:32:59.729Z [INFO] TestRemoteExecWrites_ACLAgentToken: started state syncer === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.724Z [INFO] TestRemoteExecWrites_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-1264304e-c01d-c230-5755-3eb6f1b28a12.dc1 writer.go:29: 2021-01-29T19:32:59.725Z [INFO] TestRemoteExecWrites_ACLDeny.server: Updating LAN server: server="Node-1264304e-c01d-c230-5755-3eb6f1b28a12 (Addr: tcp/127.0.0.1:30389) (DC: dc1)" writer.go:29: 2021-01-29T19:32:59.731Z [INFO] TestRemoteExecWrites_ACLDeny.server: Handled event for server in area: event=member-update server=Node-1264304e-c01d-c230-5755-3eb6f1b28a12.dc1 area=wan writer.go:29: 2021-01-29T19:32:59.736Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:59.748Z [INFO] TestRemoteExecWrites_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.755Z [WARN] TestRemoteExecWrites_ACLAgentToken.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.751Z [INFO] TestRemoteExecWrites_ACLDeny.leader: started routine: routine="CA root pruning" === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.759Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.raft: entering candidate state: node="Node at 127.0.0.1:30395 [Candidate]" term=2 === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.763Z [DEBUG] TestRemoteExecWrites_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-1264304e-c01d-c230-5755-3eb6f1b28a12 writer.go:29: 2021-01-29T19:32:59.765Z [INFO] TestRemoteExecWrites_ACLDeny.server: member joined, marking health alive: member=Node-1264304e-c01d-c230-5755-3eb6f1b28a12 writer.go:29: 2021-01-29T19:32:59.768Z [DEBUG] TestRemoteExecWrites_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-1264304e-c01d-c230-5755-3eb6f1b28a12 === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.767Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:59.769Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:59.780Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.server.raft: vote granted: from=b844b6ce-7d74-6812-c209-7d04c2651bc6 term=2 tally=1 writer.go:29: 2021-01-29T19:32:59.783Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:59.787Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.raft: entering leader state: leader="Node at 127.0.0.1:30395 [Leader]" writer.go:29: 2021-01-29T19:32:59.792Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:59.793Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: New leader elected: payload=Node-b844b6ce-7d74-6812-c209-7d04c2651bc6 writer.go:29: 2021-01-29T19:32:59.797Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: initializing acls writer.go:29: 2021-01-29T19:32:59.802Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:59.803Z [WARN] TestRemoteExecWrites_ACLAgentToken.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:59.806Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:59.809Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:59.810Z [INFO] TestRemoteExecWrites_ACLAgentToken.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:59.813Z [INFO] TestRemoteExecWrites_ACLAgentToken.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:59.818Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.serf.lan: serf: EventMemberUpdate: Node-b844b6ce-7d74-6812-c209-7d04c2651bc6 writer.go:29: 2021-01-29T19:32:59.825Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: Updating LAN server: server="Node-b844b6ce-7d74-6812-c209-7d04c2651bc6 (Addr: tcp/127.0.0.1:30395) (DC: dc1)" writer.go:29: 2021-01-29T19:32:59.825Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.serf.wan: serf: EventMemberUpdate: Node-b844b6ce-7d74-6812-c209-7d04c2651bc6.dc1 writer.go:29: 2021-01-29T19:32:59.829Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: Handled event for server in area: event=member-update server=Node-b844b6ce-7d74-6812-c209-7d04c2651bc6.dc1 area=wan === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.822Z [DEBUG] TestRemoteExecWrites_ACLDeny.acl: dropping check from result due to ACLs: check=serfHealth === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.831Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.831Z [DEBUG] TestRemoteExecWrites_ACLDeny.acl: dropping node from result due to ACLs: node=Node-1264304e-c01d-c230-5755-3eb6f1b28a12 writer.go:29: 2021-01-29T19:32:59.833Z [WARN] TestRemoteExecWrites_ACLDeny: Node info update blocked by ACLs: node=1264304e-c01d-c230-5755-3eb6f1b28a12 accessorID= === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.844Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.846Z [DEBUG] TestRemoteExecWrites_ACLDeny.acl: dropping node from result due to ACLs: node=Node-1264304e-c01d-c230-5755-3eb6f1b28a12 === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.848Z [INFO] TestRemoteExecWrites_ACLAgentToken.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:59.852Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.server: Skipping self join check for node since the cluster is too small: node=Node-b844b6ce-7d74-6812-c209-7d04c2651bc6 writer.go:29: 2021-01-29T19:32:59.857Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: member joined, marking health alive: member=Node-b844b6ce-7d74-6812-c209-7d04c2651bc6 writer.go:29: 2021-01-29T19:32:59.864Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.server: Skipping self join check for node since the cluster is too small: node=Node-b844b6ce-7d74-6812-c209-7d04c2651bc6 === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.868Z [ERROR] TestRemoteExecWrites_ACLDeny: failed to ack remote exec job: error="Permission denied" writer.go:29: 2021-01-29T19:32:59.869Z [ERROR] TestRemoteExecWrites_ACLDeny: failed to write output for remote exec job: error="Permission denied" writer.go:29: 2021-01-29T19:32:59.873Z [ERROR] TestRemoteExecWrites_ACLDeny: failed to write output for remote exec job: error="Permission denied" writer.go:29: 2021-01-29T19:32:59.882Z [INFO] TestRemoteExecWrites_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:32:59.884Z [INFO] TestRemoteExecWrites_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:32:59.886Z [DEBUG] TestRemoteExecWrites_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:59.888Z [DEBUG] TestRemoteExecWrites_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:59.889Z [DEBUG] TestRemoteExecWrites_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:59.888Z [DEBUG] TestRemoteExecWrites_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:59.890Z [DEBUG] TestRemoteExecWrites_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:59.891Z [DEBUG] TestRemoteExecWrites_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:59.891Z [WARN] TestRemoteExecWrites_ACLDeny.server.serf.lan: serf: Shutdown without a Leave === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.898Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.acl: dropping node from result due to ACLs: node=Node-b844b6ce-7d74-6812-c209-7d04c2651bc6 === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.904Z [WARN] TestRemoteExecWrites_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:59.909Z [INFO] TestRemoteExecWrites_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:59.911Z [INFO] TestRemoteExecWrites_ACLDeny: consul server down writer.go:29: 2021-01-29T19:32:59.913Z [INFO] TestRemoteExecWrites_ACLDeny: shutdown complete === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.904Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.acl: dropping node from result due to ACLs: node=Node-b844b6ce-7d74-6812-c209-7d04c2651bc6 === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:32:59.922Z [INFO] TestRemoteExecWrites_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30384 network=tcp writer.go:29: 2021-01-29T19:32:59.930Z [INFO] TestRemoteExecWrites_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30384 network=udp writer.go:29: 2021-01-29T19:32:59.932Z [INFO] TestRemoteExecWrites_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30385 network=tcp === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:32:59.938Z [INFO] TestRemoteExecWrites_ACLAgentToken: Requesting shutdown writer.go:29: 2021-01-29T19:32:59.940Z [INFO] TestRemoteExecWrites_ACLAgentToken.server: shutting down server writer.go:29: 2021-01-29T19:32:59.944Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:59.948Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:59.944Z [ERROR] TestRemoteExecWrites_ACLAgentToken.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:59.948Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:59.952Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:59.952Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:59.959Z [WARN] TestRemoteExecWrites_ACLAgentToken.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:59.959Z [DEBUG] TestRemoteExecWrites_ACLAgentToken.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:59.970Z [WARN] TestRemoteExecWrites_ACLAgentToken.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:59.974Z [INFO] TestRemoteExecWrites_ACLAgentToken.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:59.975Z [INFO] TestRemoteExecWrites_ACLAgentToken: consul server down writer.go:29: 2021-01-29T19:32:59.978Z [INFO] TestRemoteExecWrites_ACLAgentToken: shutdown complete writer.go:29: 2021-01-29T19:32:59.979Z [INFO] TestRemoteExecWrites_ACLAgentToken: Stopping server: protocol=DNS address=127.0.0.1:30390 network=tcp writer.go:29: 2021-01-29T19:32:59.981Z [INFO] TestRemoteExecWrites_ACLAgentToken: Stopping server: protocol=DNS address=127.0.0.1:30390 network=udp writer.go:29: 2021-01-29T19:32:59.982Z [INFO] TestRemoteExecWrites_ACLAgentToken: Stopping server: protocol=HTTP address=127.0.0.1:30391 network=tcp === CONT TestHandleRemoteExec writer.go:29: 2021-01-29T19:33:00.218Z [INFO] TestHandleRemoteExec: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:00.220Z [INFO] TestHandleRemoteExec: Endpoints down --- PASS: TestHandleRemoteExec (1.12s) === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.236Z [WARN] TestRemoteExecWrites_ACLToken: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:00.237Z [WARN] TestRemoteExecWrites_ACLToken: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:00.238Z [DEBUG] TestRemoteExecWrites_ACLToken.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:00.240Z [DEBUG] TestRemoteExecWrites_ACLToken.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:00.250Z [INFO] TestRemoteExecWrites_ACLToken.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b Address:127.0.0.1:30401}]" writer.go:29: 2021-01-29T19:33:00.252Z [INFO] TestRemoteExecWrites_ACLToken.server.raft: entering follower state: follower="Node at 127.0.0.1:30401 [Follower]" leader= writer.go:29: 2021-01-29T19:33:00.253Z [INFO] TestRemoteExecWrites_ACLToken.server.serf.wan: serf: EventMemberJoin: Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:00.258Z [INFO] TestRemoteExecWrites_ACLToken.server.serf.lan: serf: EventMemberJoin: Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b 127.0.0.1 writer.go:29: 2021-01-29T19:33:00.260Z [INFO] TestRemoteExecWrites_ACLToken.server: Adding LAN server: server="Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b (Addr: tcp/127.0.0.1:30401) (DC: dc1)" writer.go:29: 2021-01-29T19:33:00.261Z [INFO] TestRemoteExecWrites_ACLToken: Started DNS server: address=127.0.0.1:30396 network=udp writer.go:29: 2021-01-29T19:33:00.263Z [INFO] TestRemoteExecWrites_ACLToken: Started DNS server: address=127.0.0.1:30396 network=tcp writer.go:29: 2021-01-29T19:33:00.262Z [INFO] TestRemoteExecWrites_ACLToken.server: Handled event for server in area: event=member-join server=Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b.dc1 area=wan writer.go:29: 2021-01-29T19:33:00.266Z [INFO] TestRemoteExecWrites_ACLToken: Started HTTP server: address=127.0.0.1:30397 network=tcp writer.go:29: 2021-01-29T19:33:00.268Z [INFO] TestRemoteExecWrites_ACLToken: started state syncer writer.go:29: 2021-01-29T19:33:00.301Z [WARN] TestRemoteExecWrites_ACLToken.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:00.302Z [INFO] TestRemoteExecWrites_ACLToken.server.raft: entering candidate state: node="Node at 127.0.0.1:30401 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:00.309Z [DEBUG] TestRemoteExecWrites_ACLToken.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:00.311Z [DEBUG] TestRemoteExecWrites_ACLToken.server.raft: vote granted: from=6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b term=2 tally=1 writer.go:29: 2021-01-29T19:33:00.312Z [INFO] TestRemoteExecWrites_ACLToken.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:00.314Z [INFO] TestRemoteExecWrites_ACLToken.server.raft: entering leader state: leader="Node at 127.0.0.1:30401 [Leader]" writer.go:29: 2021-01-29T19:33:00.312Z [DEBUG] TestRemoteExecWrites_ACLToken.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:33:00.316Z [INFO] TestRemoteExecWrites_ACLToken.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:00.317Z [INFO] TestRemoteExecWrites_ACLToken.server: New leader elected: payload=Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b writer.go:29: 2021-01-29T19:33:00.318Z [INFO] TestRemoteExecWrites_ACLToken.server: initializing acls writer.go:29: 2021-01-29T19:33:00.354Z [INFO] TestRemoteExecWrites_ACLToken.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:00.360Z [WARN] TestRemoteExecWrites_ACLToken.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:33:00.372Z [INFO] TestRemoteExecWrites_ACLToken.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:33:00.384Z [INFO] TestRemoteExecWrites_ACLToken.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:00.389Z [INFO] TestRemoteExecWrites_ACLToken.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:00.405Z [INFO] TestRemoteExecWrites_ACLToken.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:00.411Z [INFO] TestRemoteExecWrites_ACLToken.server.serf.lan: serf: EventMemberUpdate: Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b writer.go:29: 2021-01-29T19:33:00.417Z [INFO] TestRemoteExecWrites_ACLToken.server.serf.wan: serf: EventMemberUpdate: Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b.dc1 writer.go:29: 2021-01-29T19:33:00.421Z [INFO] TestRemoteExecWrites_ACLToken.server: Updating LAN server: server="Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b (Addr: tcp/127.0.0.1:30401) (DC: dc1)" writer.go:29: 2021-01-29T19:33:00.425Z [INFO] TestRemoteExecWrites_ACLToken.server: Handled event for server in area: event=member-update server=Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b.dc1 area=wan writer.go:29: 2021-01-29T19:33:00.431Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:00.438Z [INFO] TestRemoteExecWrites_ACLToken.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:33:00.433Z [INFO] TestRemoteExecWrites_ACLDeny: Waiting for endpoints to shut down === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.440Z [INFO] TestRemoteExecWrites_ACLToken.leader: started routine: routine="CA root pruning" === CONT TestRemoteExecWrites_ACLDeny writer.go:29: 2021-01-29T19:33:00.441Z [INFO] TestRemoteExecWrites_ACLDeny: Endpoints down --- PASS: TestRemoteExecWrites_ACLDeny (0.94s) === CONT TestRemoteExecWrites === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.445Z [DEBUG] TestRemoteExecWrites_ACLToken.server: Skipping self join check for node since the cluster is too small: node=Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b writer.go:29: 2021-01-29T19:33:00.451Z [INFO] TestRemoteExecWrites_ACLToken.server: member joined, marking health alive: member=Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b writer.go:29: 2021-01-29T19:33:00.466Z [DEBUG] TestRemoteExecWrites_ACLToken.server: Skipping self join check for node since the cluster is too small: node=Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b === CONT TestRemoteExecWrites_ACLAgentToken writer.go:29: 2021-01-29T19:33:00.483Z [INFO] TestRemoteExecWrites_ACLAgentToken: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:00.484Z [INFO] TestRemoteExecWrites_ACLAgentToken: Endpoints down --- PASS: TestRemoteExecWrites_ACLAgentToken (0.88s) === CONT TestRemoteExecGetSpec_ACLDeny === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.487Z [DEBUG] TestRemoteExecWrites_ACLToken: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:00.490Z [INFO] TestRemoteExecWrites_ACLToken: Synced node info writer.go:29: 2021-01-29T19:33:00.492Z [DEBUG] TestRemoteExecWrites_ACLToken: Node info in sync === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.497Z [WARN] TestRemoteExecGetSpec_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:00.498Z [WARN] TestRemoteExecGetSpec_ACLDeny: bootstrap = true: do not enable unless necessary === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.502Z [DEBUG] TestRemoteExecWrites_ACLToken.acl: dropping node from result due to ACLs: node=Node-6f67fd28-4bd0-58b6-4aa9-7f591df1cf6b === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.504Z [WARN] TestRemoteExecWrites: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:00.510Z [DEBUG] TestRemoteExecWrites.tlsutil: Update: version=1 === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.503Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:00.513Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.513Z [DEBUG] TestRemoteExecWrites.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.525Z [INFO] TestRemoteExecWrites_ACLToken: Requesting shutdown writer.go:29: 2021-01-29T19:33:00.526Z [INFO] TestRemoteExecWrites_ACLToken.server: shutting down server writer.go:29: 2021-01-29T19:33:00.527Z [DEBUG] TestRemoteExecWrites_ACLToken.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:00.528Z [DEBUG] TestRemoteExecWrites_ACLToken.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:00.529Z [DEBUG] TestRemoteExecWrites_ACLToken.leader: stopping routine: routine="CA root pruning" === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.529Z [INFO] TestRemoteExecWrites.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fa88675d-5813-8fcb-0391-4076ea8c5ef2 Address:127.0.0.1:30407}]" === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.529Z [WARN] TestRemoteExecWrites_ACLToken.server.serf.lan: serf: Shutdown without a Leave === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.532Z [INFO] TestRemoteExecWrites.server.serf.wan: serf: EventMemberJoin: Node-fa88675d-5813-8fcb-0391-4076ea8c5ef2.dc1 127.0.0.1 === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.534Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2f457ed8-8296-e8b5-643d-d6f18332421f Address:127.0.0.1:30413}]" === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.535Z [INFO] TestRemoteExecWrites.server.serf.lan: serf: EventMemberJoin: Node-fa88675d-5813-8fcb-0391-4076ea8c5ef2 127.0.0.1 === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.536Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-2f457ed8-8296-e8b5-643d-d6f18332421f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:00.538Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-2f457ed8-8296-e8b5-643d-d6f18332421f 127.0.0.1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:00.531Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.541Z [INFO] TestRemoteExecGetSpec_ACLDeny: Started DNS server: address=127.0.0.1:30408 network=udp writer.go:29: 2021-01-29T19:33:00.545Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30413 [Follower]" leader= === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.541Z [DEBUG] TestRemoteExecWrites_ACLToken.leader: stopped routine: routine="CA root pruning" === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.552Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: Adding LAN server: server="Node-2f457ed8-8296-e8b5-643d-d6f18332421f (Addr: tcp/127.0.0.1:30413) (DC: dc1)" === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.539Z [INFO] TestRemoteExecWrites.server.raft: entering follower state: follower="Node at 127.0.0.1:30407 [Follower]" leader= === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.528Z [DEBUG] TestRemoteExecWrites_ACLToken.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.554Z [INFO] TestRemoteExecWrites.server: Adding LAN server: server="Node-fa88675d-5813-8fcb-0391-4076ea8c5ef2 (Addr: tcp/127.0.0.1:30407) (DC: dc1)" === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.552Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: Handled event for server in area: event=member-join server=Node-2f457ed8-8296-e8b5-643d-d6f18332421f.dc1 area=wan === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.554Z [INFO] TestRemoteExecWrites.server: Handled event for server in area: event=member-join server=Node-fa88675d-5813-8fcb-0391-4076ea8c5ef2.dc1 area=wan === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.554Z [INFO] TestRemoteExecGetSpec_ACLDeny: Started DNS server: address=127.0.0.1:30408 network=tcp === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.555Z [INFO] TestRemoteExecWrites: Started DNS server: address=127.0.0.1:30402 network=udp writer.go:29: 2021-01-29T19:33:00.565Z [INFO] TestRemoteExecWrites: Started DNS server: address=127.0.0.1:30402 network=tcp writer.go:29: 2021-01-29T19:33:00.567Z [INFO] TestRemoteExecWrites: Started HTTP server: address=127.0.0.1:30403 network=tcp writer.go:29: 2021-01-29T19:33:00.569Z [INFO] TestRemoteExecWrites: started state syncer === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.566Z [INFO] TestRemoteExecGetSpec_ACLDeny: Started HTTP server: address=127.0.0.1:30409 network=tcp === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.545Z [WARN] TestRemoteExecWrites_ACLToken.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:00.529Z [DEBUG] TestRemoteExecWrites_ACLToken.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:00.574Z [INFO] TestRemoteExecWrites_ACLToken.server.router.manager: shutting down === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.570Z [INFO] TestRemoteExecGetSpec_ACLDeny: started state syncer === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:00.576Z [INFO] TestRemoteExecWrites_ACLToken: consul server down writer.go:29: 2021-01-29T19:33:00.577Z [INFO] TestRemoteExecWrites_ACLToken: shutdown complete writer.go:29: 2021-01-29T19:33:00.578Z [INFO] TestRemoteExecWrites_ACLToken: Stopping server: protocol=DNS address=127.0.0.1:30396 network=tcp writer.go:29: 2021-01-29T19:33:00.579Z [INFO] TestRemoteExecWrites_ACLToken: Stopping server: protocol=DNS address=127.0.0.1:30396 network=udp writer.go:29: 2021-01-29T19:33:00.582Z [INFO] TestRemoteExecWrites_ACLToken: Stopping server: protocol=HTTP address=127.0.0.1:30397 network=tcp === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.593Z [WARN] TestRemoteExecGetSpec_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:00.594Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30413 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:00.598Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:00.599Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.server.raft: vote granted: from=2f457ed8-8296-e8b5-643d-d6f18332421f term=2 tally=1 writer.go:29: 2021-01-29T19:33:00.601Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:00.607Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30413 [Leader]" === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.610Z [WARN] TestRemoteExecWrites.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:00.612Z [INFO] TestRemoteExecWrites.server.raft: entering candidate state: node="Node at 127.0.0.1:30407 [Candidate]" term=2 === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.603Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.615Z [DEBUG] TestRemoteExecWrites.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:00.617Z [DEBUG] TestRemoteExecWrites.server.raft: vote granted: from=fa88675d-5813-8fcb-0391-4076ea8c5ef2 term=2 tally=1 === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.615Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: cluster leadership acquired === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.618Z [INFO] TestRemoteExecWrites.server.raft: election won: tally=1 === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.617Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: New leader elected: payload=Node-2f457ed8-8296-e8b5-643d-d6f18332421f writer.go:29: 2021-01-29T19:33:00.618Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: initializing acls === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.620Z [INFO] TestRemoteExecWrites.server.raft: entering leader state: leader="Node at 127.0.0.1:30407 [Leader]" === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.622Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: Created ACL 'global-management' policy === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.622Z [INFO] TestRemoteExecWrites.server: cluster leadership acquired === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.623Z [WARN] TestRemoteExecGetSpec_ACLDeny.server: Configuring a non-UUID master token is deprecated === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.624Z [INFO] TestRemoteExecWrites.server: New leader elected: payload=Node-fa88675d-5813-8fcb-0391-4076ea8c5ef2 === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.625Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:33:00.631Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:00.632Z [INFO] TestRemoteExecGetSpec_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:00.633Z [INFO] TestRemoteExecGetSpec_ACLDeny.leader: started routine: routine="acl token reaping" === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.632Z [DEBUG] TestRemoteExecWrites.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30407 === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.634Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-2f457ed8-8296-e8b5-643d-d6f18332421f writer.go:29: 2021-01-29T19:33:00.636Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-2f457ed8-8296-e8b5-643d-d6f18332421f.dc1 writer.go:29: 2021-01-29T19:33:00.636Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: Updating LAN server: server="Node-2f457ed8-8296-e8b5-643d-d6f18332421f (Addr: tcp/127.0.0.1:30413) (DC: dc1)" writer.go:29: 2021-01-29T19:33:00.639Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: Handled event for server in area: event=member-update server=Node-2f457ed8-8296-e8b5-643d-d6f18332421f.dc1 area=wan writer.go:29: 2021-01-29T19:33:00.649Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:00.654Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:00.655Z [INFO] TestRemoteExecGetSpec_ACLDeny.leader: started routine: routine="CA root pruning" === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.648Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.659Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-2f457ed8-8296-e8b5-643d-d6f18332421f writer.go:29: 2021-01-29T19:33:00.660Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: member joined, marking health alive: member=Node-2f457ed8-8296-e8b5-643d-d6f18332421f writer.go:29: 2021-01-29T19:33:00.663Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-2f457ed8-8296-e8b5-643d-d6f18332421f === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.667Z [INFO] TestRemoteExecWrites.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:00.669Z [INFO] TestRemoteExecWrites.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:00.676Z [DEBUG] TestRemoteExecWrites.server: Skipping self join check for node since the cluster is too small: node=Node-fa88675d-5813-8fcb-0391-4076ea8c5ef2 writer.go:29: 2021-01-29T19:33:00.679Z [INFO] TestRemoteExecWrites.server: member joined, marking health alive: member=Node-fa88675d-5813-8fcb-0391-4076ea8c5ef2 === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.813Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.acl: dropping check from result due to ACLs: check=serfHealth writer.go:29: 2021-01-29T19:33:00.815Z [WARN] TestRemoteExecGetSpec_ACLDeny: Node info update blocked by ACLs: node=2f457ed8-8296-e8b5-643d-d6f18332421f accessorID= === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:00.872Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:00.877Z [INFO] TestServiceManager_PersistService_ConfigFiles: Synced node info writer.go:29: 2021-01-29T19:33:00.879Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:00.880Z [INFO] TestServiceManager_PersistService_ConfigFiles: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:33:00.881Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: Node info in sync writer.go:29: 2021-01-29T19:33:00.884Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: Node info in sync writer.go:29: 2021-01-29T19:33:00.887Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: Service in sync: service=web-sidecar-proxy === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.932Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.acl: dropping node from result due to ACLs: node=Node-2f457ed8-8296-e8b5-643d-d6f18332421f writer.go:29: 2021-01-29T19:33:00.936Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.acl: dropping node from result due to ACLs: node=Node-2f457ed8-8296-e8b5-643d-d6f18332421f === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.947Z [INFO] TestRemoteExecWrites: Requesting shutdown writer.go:29: 2021-01-29T19:33:00.948Z [INFO] TestRemoteExecWrites.server: shutting down server writer.go:29: 2021-01-29T19:33:00.949Z [DEBUG] TestRemoteExecWrites.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:00.950Z [WARN] TestRemoteExecWrites.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:00.949Z [ERROR] TestRemoteExecWrites.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.952Z [ERROR] TestRemoteExecGetSpec_ACLDeny: failed to get remote exec job: error="Permission denied" === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.950Z [DEBUG] TestRemoteExecWrites.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:00.954Z [WARN] TestRemoteExecWrites.server.serf.wan: serf: Shutdown without a Leave === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.955Z [INFO] TestRemoteExecGetSpec_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:33:00.956Z [INFO] TestRemoteExecGetSpec_ACLDeny.server: shutting down server === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.956Z [INFO] TestRemoteExecWrites.server.router.manager: shutting down === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.958Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:00.960Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.leader: stopping routine: routine="acl token reaping" === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.956Z [INFO] TestRemoteExecWrites: consul server down writer.go:29: 2021-01-29T19:33:00.961Z [INFO] TestRemoteExecWrites: shutdown complete writer.go:29: 2021-01-29T19:33:00.962Z [INFO] TestRemoteExecWrites: Stopping server: protocol=DNS address=127.0.0.1:30402 network=tcp writer.go:29: 2021-01-29T19:33:00.963Z [INFO] TestRemoteExecWrites: Stopping server: protocol=DNS address=127.0.0.1:30402 network=udp === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.961Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.leader: stopping routine: routine="CA root pruning" === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:00.964Z [INFO] TestRemoteExecWrites: Stopping server: protocol=HTTP address=127.0.0.1:30403 network=tcp === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:00.970Z [WARN] TestRemoteExecGetSpec_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:00.963Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:00.963Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:00.970Z [DEBUG] TestRemoteExecGetSpec_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:00.979Z [WARN] TestRemoteExecGetSpec_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:00.982Z [INFO] TestRemoteExecGetSpec_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:00.983Z [INFO] TestRemoteExecGetSpec_ACLDeny: consul server down writer.go:29: 2021-01-29T19:33:00.986Z [INFO] TestRemoteExecGetSpec_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:33:00.987Z [INFO] TestRemoteExecGetSpec_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30408 network=tcp writer.go:29: 2021-01-29T19:33:00.988Z [INFO] TestRemoteExecGetSpec_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30408 network=udp writer.go:29: 2021-01-29T19:33:00.989Z [INFO] TestRemoteExecGetSpec_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30409 network=tcp === CONT TestRemoteExecWrites_ACLToken writer.go:29: 2021-01-29T19:33:01.084Z [INFO] TestRemoteExecWrites_ACLToken: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:01.086Z [INFO] TestRemoteExecWrites_ACLToken: Endpoints down --- PASS: TestRemoteExecWrites_ACLToken (0.86s) === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.109Z [WARN] TestRemoteExecGetSpec_ACLAgentToken: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:01.111Z [WARN] TestRemoteExecGetSpec_ACLAgentToken: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:01.114Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:01.118Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:01.130Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9a4126ad-96bc-a060-c65e-af58dcf1921a Address:127.0.0.1:30419}]" writer.go:29: 2021-01-29T19:33:01.133Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.raft: entering follower state: follower="Node at 127.0.0.1:30419 [Follower]" leader= writer.go:29: 2021-01-29T19:33:01.134Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.serf.wan: serf: EventMemberJoin: Node-9a4126ad-96bc-a060-c65e-af58dcf1921a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:01.145Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.serf.lan: serf: EventMemberJoin: Node-9a4126ad-96bc-a060-c65e-af58dcf1921a 127.0.0.1 writer.go:29: 2021-01-29T19:33:01.173Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: Handled event for server in area: event=member-join server=Node-9a4126ad-96bc-a060-c65e-af58dcf1921a.dc1 area=wan writer.go:29: 2021-01-29T19:33:01.174Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: Adding LAN server: server="Node-9a4126ad-96bc-a060-c65e-af58dcf1921a (Addr: tcp/127.0.0.1:30419) (DC: dc1)" writer.go:29: 2021-01-29T19:33:01.174Z [WARN] TestRemoteExecGetSpec_ACLAgentToken.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:01.181Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.raft: entering candidate state: node="Node at 127.0.0.1:30419 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:01.194Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: Started DNS server: address=127.0.0.1:30414 network=udp writer.go:29: 2021-01-29T19:33:01.197Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: Started DNS server: address=127.0.0.1:30414 network=tcp writer.go:29: 2021-01-29T19:33:01.201Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: Started HTTP server: address=127.0.0.1:30415 network=tcp writer.go:29: 2021-01-29T19:33:01.202Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: started state syncer writer.go:29: 2021-01-29T19:33:01.223Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:01.226Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.server.raft: vote granted: from=9a4126ad-96bc-a060-c65e-af58dcf1921a term=2 tally=1 writer.go:29: 2021-01-29T19:33:01.229Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:01.232Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.raft: entering leader state: leader="Node at 127.0.0.1:30419 [Leader]" writer.go:29: 2021-01-29T19:33:01.230Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:33:01.238Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:01.239Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: New leader elected: payload=Node-9a4126ad-96bc-a060-c65e-af58dcf1921a writer.go:29: 2021-01-29T19:33:01.263Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: initializing acls writer.go:29: 2021-01-29T19:33:01.273Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:01.275Z [WARN] TestRemoteExecGetSpec_ACLAgentToken.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:33:01.285Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:33:01.321Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:01.323Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:01.334Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: initializing acls writer.go:29: 2021-01-29T19:33:01.345Z [WARN] TestRemoteExecGetSpec_ACLAgentToken.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:33:01.338Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:01.353Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.serf.lan: serf: EventMemberUpdate: Node-9a4126ad-96bc-a060-c65e-af58dcf1921a writer.go:29: 2021-01-29T19:33:01.361Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.serf.wan: serf: EventMemberUpdate: Node-9a4126ad-96bc-a060-c65e-af58dcf1921a.dc1 writer.go:29: 2021-01-29T19:33:01.363Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: Updating LAN server: server="Node-9a4126ad-96bc-a060-c65e-af58dcf1921a (Addr: tcp/127.0.0.1:30419) (DC: dc1)" writer.go:29: 2021-01-29T19:33:01.366Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: Handled event for server in area: event=member-update server=Node-9a4126ad-96bc-a060-c65e-af58dcf1921a.dc1 area=wan writer.go:29: 2021-01-29T19:33:01.369Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:33:01.378Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.serf.lan: serf: EventMemberUpdate: Node-9a4126ad-96bc-a060-c65e-af58dcf1921a writer.go:29: 2021-01-29T19:33:01.388Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.serf.wan: serf: EventMemberUpdate: Node-9a4126ad-96bc-a060-c65e-af58dcf1921a.dc1 writer.go:29: 2021-01-29T19:33:01.391Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: Updating LAN server: server="Node-9a4126ad-96bc-a060-c65e-af58dcf1921a (Addr: tcp/127.0.0.1:30419) (DC: dc1)" writer.go:29: 2021-01-29T19:33:01.384Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:01.434Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: Handled event for server in area: event=member-update server=Node-9a4126ad-96bc-a060-c65e-af58dcf1921a.dc1 area=wan === CONT TestRemoteExecWrites writer.go:29: 2021-01-29T19:33:01.472Z [INFO] TestRemoteExecWrites: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:01.473Z [INFO] TestRemoteExecWrites: Endpoints down --- PASS: TestRemoteExecWrites (1.03s) === CONT TestRemoteExecGetSpec_ACLToken === CONT TestRemoteExecGetSpec_ACLDeny writer.go:29: 2021-01-29T19:33:01.491Z [INFO] TestRemoteExecGetSpec_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:01.496Z [INFO] TestRemoteExecGetSpec_ACLDeny: Endpoints down --- PASS: TestRemoteExecGetSpec_ACLDeny (1.01s) === CONT TestRemoteExecGetSpec === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.501Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:01.505Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:01.508Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.server: Skipping self join check for node since the cluster is too small: node=Node-9a4126ad-96bc-a060-c65e-af58dcf1921a writer.go:29: 2021-01-29T19:33:01.519Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: member joined, marking health alive: member=Node-9a4126ad-96bc-a060-c65e-af58dcf1921a writer.go:29: 2021-01-29T19:33:01.524Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.server: Skipping self join check for node since the cluster is too small: node=Node-9a4126ad-96bc-a060-c65e-af58dcf1921a writer.go:29: 2021-01-29T19:33:01.526Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.server: Skipping self join check for node since the cluster is too small: node=Node-9a4126ad-96bc-a060-c65e-af58dcf1921a === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.597Z [WARN] TestRemoteExecGetSpec: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:01.625Z [DEBUG] TestRemoteExecGetSpec.tlsutil: Update: version=1 === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.590Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.acl: dropping node from result due to ACLs: node=Node-9a4126ad-96bc-a060-c65e-af58dcf1921a === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.634Z [DEBUG] TestRemoteExecGetSpec.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.638Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.acl: dropping node from result due to ACLs: node=Node-9a4126ad-96bc-a060-c65e-af58dcf1921a writer.go:29: 2021-01-29T19:33:01.634Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.666Z [WARN] TestRemoteExecGetSpec_ACLToken: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:01.667Z [WARN] TestRemoteExecGetSpec_ACLToken: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:01.670Z [DEBUG] TestRemoteExecGetSpec_ACLToken.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:01.674Z [DEBUG] TestRemoteExecGetSpec_ACLToken.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.684Z [INFO] TestRemoteExecGetSpec.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:829f84ab-4063-40ab-5e9d-961f9626f12f Address:127.0.0.1:29411}]" === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.683Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: Synced node info === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.687Z [INFO] TestRemoteExecGetSpec.server.serf.wan: serf: EventMemberJoin: Node-829f84ab-4063-40ab-5e9d-961f9626f12f.dc1 127.0.0.1 === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.685Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken: Node info in sync === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.689Z [INFO] TestRemoteExecGetSpec_ACLToken.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c3b01068-bcb2-32f0-95f3-29f46a2867c3 Address:127.0.0.1:29405}]" === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.687Z [INFO] TestRemoteExecGetSpec.server.raft: entering follower state: follower="Node at 127.0.0.1:29411 [Follower]" leader= === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.692Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: Requesting shutdown === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.690Z [INFO] TestRemoteExecGetSpec_ACLToken.server.raft: entering follower state: follower="Node at 127.0.0.1:29405 [Follower]" leader= === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.693Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server: shutting down server writer.go:29: 2021-01-29T19:33:01.694Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.690Z [INFO] TestRemoteExecGetSpec.server.serf.lan: serf: EventMemberJoin: Node-829f84ab-4063-40ab-5e9d-961f9626f12f 127.0.0.1 === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.695Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:01.696Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:01.697Z [WARN] TestRemoteExecGetSpec_ACLAgentToken.server.serf.lan: serf: Shutdown without a Leave === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.691Z [INFO] TestRemoteExecGetSpec_ACLToken.server.serf.wan: serf: EventMemberJoin: Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3.dc1 127.0.0.1 === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.698Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:01.699Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:01.700Z [DEBUG] TestRemoteExecGetSpec_ACLAgentToken.leader: stopped routine: routine="CA root pruning" === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.699Z [INFO] TestRemoteExecGetSpec.server: Adding LAN server: server="Node-829f84ab-4063-40ab-5e9d-961f9626f12f (Addr: tcp/127.0.0.1:29411) (DC: dc1)" writer.go:29: 2021-01-29T19:33:01.701Z [INFO] TestRemoteExecGetSpec.server: Handled event for server in area: event=member-join server=Node-829f84ab-4063-40ab-5e9d-961f9626f12f.dc1 area=wan === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.702Z [INFO] TestRemoteExecGetSpec_ACLToken.server.serf.lan: serf: EventMemberJoin: Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3 127.0.0.1 === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.703Z [INFO] TestRemoteExecGetSpec: Started DNS server: address=127.0.0.1:29406 network=udp writer.go:29: 2021-01-29T19:33:01.704Z [INFO] TestRemoteExecGetSpec: Started DNS server: address=127.0.0.1:29406 network=tcp === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.705Z [INFO] TestRemoteExecGetSpec_ACLToken.server: Handled event for server in area: event=member-join server=Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3.dc1 area=wan writer.go:29: 2021-01-29T19:33:01.705Z [INFO] TestRemoteExecGetSpec_ACLToken.server: Adding LAN server: server="Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3 (Addr: tcp/127.0.0.1:29405) (DC: dc1)" === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.706Z [INFO] TestRemoteExecGetSpec: Started HTTP server: address=127.0.0.1:29407 network=tcp === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.705Z [INFO] TestRemoteExecGetSpec_ACLToken: Started DNS server: address=127.0.0.1:29400 network=udp === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.707Z [INFO] TestRemoteExecGetSpec: started state syncer === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.707Z [INFO] TestRemoteExecGetSpec_ACLToken: Started DNS server: address=127.0.0.1:29400 network=tcp === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.706Z [WARN] TestRemoteExecGetSpec_ACLAgentToken.server.serf.wan: serf: Shutdown without a Leave === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.709Z [INFO] TestRemoteExecGetSpec_ACLToken: Started HTTP server: address=127.0.0.1:29401 network=tcp writer.go:29: 2021-01-29T19:33:01.710Z [INFO] TestRemoteExecGetSpec_ACLToken: started state syncer === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:01.711Z [INFO] TestRemoteExecGetSpec_ACLAgentToken.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:01.712Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: consul server down writer.go:29: 2021-01-29T19:33:01.715Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: shutdown complete writer.go:29: 2021-01-29T19:33:01.717Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: Stopping server: protocol=DNS address=127.0.0.1:30414 network=tcp writer.go:29: 2021-01-29T19:33:01.720Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: Stopping server: protocol=DNS address=127.0.0.1:30414 network=udp writer.go:29: 2021-01-29T19:33:01.722Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: Stopping server: protocol=HTTP address=127.0.0.1:30415 network=tcp === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.736Z [WARN] TestRemoteExecGetSpec.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:01.737Z [INFO] TestRemoteExecGetSpec.server.raft: entering candidate state: node="Node at 127.0.0.1:29411 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:01.738Z [DEBUG] TestRemoteExecGetSpec.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:01.739Z [DEBUG] TestRemoteExecGetSpec.server.raft: vote granted: from=829f84ab-4063-40ab-5e9d-961f9626f12f term=2 tally=1 writer.go:29: 2021-01-29T19:33:01.740Z [INFO] TestRemoteExecGetSpec.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:01.741Z [INFO] TestRemoteExecGetSpec.server.raft: entering leader state: leader="Node at 127.0.0.1:29411 [Leader]" writer.go:29: 2021-01-29T19:33:01.742Z [INFO] TestRemoteExecGetSpec.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:01.742Z [INFO] TestRemoteExecGetSpec.server: New leader elected: payload=Node-829f84ab-4063-40ab-5e9d-961f9626f12f writer.go:29: 2021-01-29T19:33:01.743Z [DEBUG] TestRemoteExecGetSpec.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29411 === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.748Z [WARN] TestRemoteExecGetSpec_ACLToken.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:01.749Z [INFO] TestRemoteExecGetSpec_ACLToken.server.raft: entering candidate state: node="Node at 127.0.0.1:29405 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:01.752Z [DEBUG] TestRemoteExecGetSpec_ACLToken.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:01.754Z [DEBUG] TestRemoteExecGetSpec_ACLToken.server.raft: vote granted: from=c3b01068-bcb2-32f0-95f3-29f46a2867c3 term=2 tally=1 === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.755Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.756Z [INFO] TestRemoteExecGetSpec_ACLToken.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:01.758Z [INFO] TestRemoteExecGetSpec_ACLToken.server.raft: entering leader state: leader="Node at 127.0.0.1:29405 [Leader]" writer.go:29: 2021-01-29T19:33:01.756Z [DEBUG] TestRemoteExecGetSpec_ACLToken.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:33:01.763Z [INFO] TestRemoteExecGetSpec_ACLToken.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:01.763Z [INFO] TestRemoteExecGetSpec_ACLToken.server: New leader elected: payload=Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3 writer.go:29: 2021-01-29T19:33:01.771Z [INFO] TestRemoteExecGetSpec_ACLToken.server: initializing acls === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.772Z [INFO] TestRemoteExecGetSpec.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.774Z [INFO] TestRemoteExecGetSpec_ACLToken.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:01.775Z [WARN] TestRemoteExecGetSpec_ACLToken.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:33:01.777Z [INFO] TestRemoteExecGetSpec_ACLToken.server: Bootstrapped ACL master token from configuration === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.774Z [INFO] TestRemoteExecGetSpec.leader: started routine: routine="CA root pruning" === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.779Z [INFO] TestRemoteExecGetSpec_ACLToken.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:01.780Z [INFO] TestRemoteExecGetSpec_ACLToken.leader: started routine: routine="legacy ACL token upgrade" === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.778Z [DEBUG] TestRemoteExecGetSpec.server: Skipping self join check for node since the cluster is too small: node=Node-829f84ab-4063-40ab-5e9d-961f9626f12f === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.781Z [INFO] TestRemoteExecGetSpec_ACLToken.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:01.781Z [INFO] TestRemoteExecGetSpec_ACLToken.server.serf.lan: serf: EventMemberUpdate: Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3 writer.go:29: 2021-01-29T19:33:01.783Z [INFO] TestRemoteExecGetSpec_ACLToken.server: Updating LAN server: server="Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3 (Addr: tcp/127.0.0.1:29405) (DC: dc1)" writer.go:29: 2021-01-29T19:33:01.783Z [INFO] TestRemoteExecGetSpec_ACLToken.server.serf.wan: serf: EventMemberUpdate: Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3.dc1 writer.go:29: 2021-01-29T19:33:01.785Z [INFO] TestRemoteExecGetSpec_ACLToken.server: Handled event for server in area: event=member-update server=Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3.dc1 area=wan writer.go:29: 2021-01-29T19:33:01.787Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.781Z [INFO] TestRemoteExecGetSpec.server: member joined, marking health alive: member=Node-829f84ab-4063-40ab-5e9d-961f9626f12f === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.793Z [INFO] TestRemoteExecGetSpec_ACLToken.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:01.794Z [INFO] TestRemoteExecGetSpec_ACLToken.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:01.796Z [DEBUG] TestRemoteExecGetSpec_ACLToken.server: Skipping self join check for node since the cluster is too small: node=Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3 writer.go:29: 2021-01-29T19:33:01.798Z [INFO] TestRemoteExecGetSpec_ACLToken.server: member joined, marking health alive: member=Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3 writer.go:29: 2021-01-29T19:33:01.824Z [DEBUG] TestRemoteExecGetSpec_ACLToken.server: Skipping self join check for node since the cluster is too small: node=Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3 === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.870Z [DEBUG] TestRemoteExecGetSpec: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:01.873Z [INFO] TestRemoteExecGetSpec: Synced node info writer.go:29: 2021-01-29T19:33:01.875Z [DEBUG] TestRemoteExecGetSpec: Node info in sync === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.880Z [DEBUG] TestRemoteExecGetSpec_ACLToken: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:01.883Z [INFO] TestRemoteExecGetSpec_ACLToken: Synced node info writer.go:29: 2021-01-29T19:33:01.886Z [DEBUG] TestRemoteExecGetSpec_ACLToken: Node info in sync === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:01.921Z [INFO] TestRemoteExecGetSpec: Requesting shutdown writer.go:29: 2021-01-29T19:33:01.922Z [INFO] TestRemoteExecGetSpec.server: shutting down server writer.go:29: 2021-01-29T19:33:01.922Z [DEBUG] TestRemoteExecGetSpec.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:01.923Z [WARN] TestRemoteExecGetSpec.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:01.925Z [DEBUG] TestRemoteExecGetSpec.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:01.932Z [WARN] TestRemoteExecGetSpec.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:01.942Z [INFO] TestRemoteExecGetSpec.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:01.944Z [INFO] TestRemoteExecGetSpec: consul server down writer.go:29: 2021-01-29T19:33:01.945Z [INFO] TestRemoteExecGetSpec: shutdown complete writer.go:29: 2021-01-29T19:33:01.946Z [INFO] TestRemoteExecGetSpec: Stopping server: protocol=DNS address=127.0.0.1:29406 network=tcp writer.go:29: 2021-01-29T19:33:01.948Z [INFO] TestRemoteExecGetSpec: Stopping server: protocol=DNS address=127.0.0.1:29406 network=udp writer.go:29: 2021-01-29T19:33:01.949Z [INFO] TestRemoteExecGetSpec: Stopping server: protocol=HTTP address=127.0.0.1:29407 network=tcp === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:01.964Z [DEBUG] TestRemoteExecGetSpec_ACLToken.acl: dropping node from result due to ACLs: node=Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3 writer.go:29: 2021-01-29T19:33:01.967Z [DEBUG] TestRemoteExecGetSpec_ACLToken.acl: dropping node from result due to ACLs: node=Node-c3b01068-bcb2-32f0-95f3-29f46a2867c3 writer.go:29: 2021-01-29T19:33:01.971Z [INFO] TestRemoteExecGetSpec_ACLToken: Requesting shutdown writer.go:29: 2021-01-29T19:33:01.971Z [INFO] TestRemoteExecGetSpec_ACLToken.server: shutting down server writer.go:29: 2021-01-29T19:33:01.973Z [DEBUG] TestRemoteExecGetSpec_ACLToken.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:01.974Z [DEBUG] TestRemoteExecGetSpec_ACLToken.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:01.975Z [DEBUG] TestRemoteExecGetSpec_ACLToken.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:01.976Z [WARN] TestRemoteExecGetSpec_ACLToken.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:01.974Z [DEBUG] TestRemoteExecGetSpec_ACLToken.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:01.975Z [DEBUG] TestRemoteExecGetSpec_ACLToken.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:01.976Z [DEBUG] TestRemoteExecGetSpec_ACLToken.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:01.978Z [WARN] TestRemoteExecGetSpec_ACLToken.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:02.012Z [INFO] TestRemoteExecGetSpec_ACLToken.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:02.013Z [INFO] TestRemoteExecGetSpec_ACLToken: consul server down writer.go:29: 2021-01-29T19:33:02.014Z [INFO] TestRemoteExecGetSpec_ACLToken: shutdown complete writer.go:29: 2021-01-29T19:33:02.015Z [INFO] TestRemoteExecGetSpec_ACLToken: Stopping server: protocol=DNS address=127.0.0.1:29400 network=tcp writer.go:29: 2021-01-29T19:33:02.017Z [INFO] TestRemoteExecGetSpec_ACLToken: Stopping server: protocol=DNS address=127.0.0.1:29400 network=udp writer.go:29: 2021-01-29T19:33:02.018Z [INFO] TestRemoteExecGetSpec_ACLToken: Stopping server: protocol=HTTP address=127.0.0.1:29401 network=tcp === CONT TestRemoteExecGetSpec_ACLAgentToken writer.go:29: 2021-01-29T19:33:02.223Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:02.225Z [INFO] TestRemoteExecGetSpec_ACLAgentToken: Endpoints down --- PASS: TestRemoteExecGetSpec_ACLAgentToken (1.14s) === CONT TestPreparedQuery_parseLimit === CONT TestPreparedQuery_Delete --- PASS: TestPreparedQuery_parseLimit (0.00s) === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:02.233Z [WARN] TestPreparedQuery_Delete: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:02.234Z [DEBUG] TestPreparedQuery_Delete.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:02.235Z [DEBUG] TestPreparedQuery_Delete.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:02.244Z [INFO] TestPreparedQuery_Delete.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:48dec0ad-c7fd-d74e-f590-9d1d6e73a8f7 Address:127.0.0.1:29417}]" writer.go:29: 2021-01-29T19:33:02.247Z [INFO] TestPreparedQuery_Delete.server.raft: entering follower state: follower="Node at 127.0.0.1:29417 [Follower]" leader= writer.go:29: 2021-01-29T19:33:02.248Z [INFO] TestPreparedQuery_Delete.server.serf.wan: serf: EventMemberJoin: Node-48dec0ad-c7fd-d74e-f590-9d1d6e73a8f7.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:02.253Z [INFO] TestPreparedQuery_Delete.server.serf.lan: serf: EventMemberJoin: Node-48dec0ad-c7fd-d74e-f590-9d1d6e73a8f7 127.0.0.1 writer.go:29: 2021-01-29T19:33:02.257Z [INFO] TestPreparedQuery_Delete.server: Adding LAN server: server="Node-48dec0ad-c7fd-d74e-f590-9d1d6e73a8f7 (Addr: tcp/127.0.0.1:29417) (DC: dc1)" writer.go:29: 2021-01-29T19:33:02.257Z [INFO] TestPreparedQuery_Delete.server: Handled event for server in area: event=member-join server=Node-48dec0ad-c7fd-d74e-f590-9d1d6e73a8f7.dc1 area=wan writer.go:29: 2021-01-29T19:33:02.258Z [INFO] TestPreparedQuery_Delete: Started DNS server: address=127.0.0.1:29412 network=tcp writer.go:29: 2021-01-29T19:33:02.261Z [INFO] TestPreparedQuery_Delete: Started DNS server: address=127.0.0.1:29412 network=udp writer.go:29: 2021-01-29T19:33:02.264Z [INFO] TestPreparedQuery_Delete: Started HTTP server: address=127.0.0.1:29413 network=tcp writer.go:29: 2021-01-29T19:33:02.265Z [INFO] TestPreparedQuery_Delete: started state syncer writer.go:29: 2021-01-29T19:33:02.285Z [WARN] TestPreparedQuery_Delete.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:02.286Z [INFO] TestPreparedQuery_Delete.server.raft: entering candidate state: node="Node at 127.0.0.1:29417 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:02.289Z [DEBUG] TestPreparedQuery_Delete.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:02.290Z [DEBUG] TestPreparedQuery_Delete.server.raft: vote granted: from=48dec0ad-c7fd-d74e-f590-9d1d6e73a8f7 term=2 tally=1 writer.go:29: 2021-01-29T19:33:02.291Z [INFO] TestPreparedQuery_Delete.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:02.293Z [INFO] TestPreparedQuery_Delete.server.raft: entering leader state: leader="Node at 127.0.0.1:29417 [Leader]" writer.go:29: 2021-01-29T19:33:02.295Z [INFO] TestPreparedQuery_Delete.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:02.296Z [INFO] TestPreparedQuery_Delete.server: New leader elected: payload=Node-48dec0ad-c7fd-d74e-f590-9d1d6e73a8f7 writer.go:29: 2021-01-29T19:33:02.299Z [DEBUG] TestPreparedQuery_Delete.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29417 writer.go:29: 2021-01-29T19:33:02.304Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:02.310Z [INFO] TestPreparedQuery_Delete.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:02.311Z [INFO] TestPreparedQuery_Delete.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:02.313Z [DEBUG] TestPreparedQuery_Delete.server: Skipping self join check for node since the cluster is too small: node=Node-48dec0ad-c7fd-d74e-f590-9d1d6e73a8f7 writer.go:29: 2021-01-29T19:33:02.315Z [INFO] TestPreparedQuery_Delete.server: member joined, marking health alive: member=Node-48dec0ad-c7fd-d74e-f590-9d1d6e73a8f7 === CONT TestRemoteExecGetSpec writer.go:29: 2021-01-29T19:33:02.451Z [INFO] TestRemoteExecGetSpec: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:02.453Z [INFO] TestRemoteExecGetSpec: Endpoints down --- PASS: TestRemoteExecGetSpec (0.96s) === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.471Z [WARN] TestPreparedQuery_Update: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:02.473Z [DEBUG] TestPreparedQuery_Update.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:02.475Z [DEBUG] TestPreparedQuery_Update.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:02.481Z [INFO] TestPreparedQuery_Update.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:77eeb1b0-98b0-ea25-d942-ceccd316423d Address:127.0.0.1:29423}]" writer.go:29: 2021-01-29T19:33:02.483Z [INFO] TestPreparedQuery_Update.server.serf.wan: serf: EventMemberJoin: Node-77eeb1b0-98b0-ea25-d942-ceccd316423d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:02.483Z [INFO] TestPreparedQuery_Update.server.raft: entering follower state: follower="Node at 127.0.0.1:29423 [Follower]" leader= writer.go:29: 2021-01-29T19:33:02.485Z [INFO] TestPreparedQuery_Update.server.serf.lan: serf: EventMemberJoin: Node-77eeb1b0-98b0-ea25-d942-ceccd316423d 127.0.0.1 writer.go:29: 2021-01-29T19:33:02.490Z [INFO] TestPreparedQuery_Update.server: Adding LAN server: server="Node-77eeb1b0-98b0-ea25-d942-ceccd316423d (Addr: tcp/127.0.0.1:29423) (DC: dc1)" writer.go:29: 2021-01-29T19:33:02.491Z [INFO] TestPreparedQuery_Update.server: Handled event for server in area: event=member-join server=Node-77eeb1b0-98b0-ea25-d942-ceccd316423d.dc1 area=wan writer.go:29: 2021-01-29T19:33:02.492Z [INFO] TestPreparedQuery_Update: Started DNS server: address=127.0.0.1:29418 network=udp writer.go:29: 2021-01-29T19:33:02.494Z [INFO] TestPreparedQuery_Update: Started DNS server: address=127.0.0.1:29418 network=tcp === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.500Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: Node info in sync === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.499Z [INFO] TestPreparedQuery_Update: Started HTTP server: address=127.0.0.1:29419 network=tcp writer.go:29: 2021-01-29T19:33:02.502Z [INFO] TestPreparedQuery_Update: started state syncer === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.505Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:33:02.514Z [INFO] TestServiceManager_PersistService_ConfigFiles: Synced service: service=web-sidecar-proxy === CONT TestRemoteExecGetSpec_ACLToken writer.go:29: 2021-01-29T19:33:02.521Z [INFO] TestRemoteExecGetSpec_ACLToken: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:02.522Z [INFO] TestRemoteExecGetSpec_ACLToken: Endpoints down --- PASS: TestRemoteExecGetSpec_ACLToken (1.04s) === CONT TestPreparedQuery_Get === RUN TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.530Z [WARN] TestPreparedQuery_Get/#00: bootstrap = true: do not enable unless necessary === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.530Z [INFO] TestServiceManager_PersistService_ConfigFiles: Requesting shutdown === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.530Z [DEBUG] TestPreparedQuery_Get/#00.tlsutil: Update: version=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.531Z [INFO] TestServiceManager_PersistService_ConfigFiles.client: shutting down client writer.go:29: 2021-01-29T19:33:02.532Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:02.532Z [INFO] TestServiceManager_PersistService_ConfigFiles.client.manager: shutting down writer.go:29: 2021-01-29T19:33:02.534Z [INFO] TestServiceManager_PersistService_ConfigFiles: consul client down === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.535Z [DEBUG] TestPreparedQuery_Get/#00.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.536Z [INFO] TestServiceManager_PersistService_ConfigFiles: shutdown complete === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.540Z [WARN] TestPreparedQuery_Update.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:02.542Z [INFO] TestPreparedQuery_Update.server.raft: entering candidate state: node="Node at 127.0.0.1:29423 [Candidate]" term=2 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.535Z [ERROR] TestServiceManager_PersistService_ConfigFiles.client: RPC failed to server: method=ConnectCA.Roots server=127.0.0.1:30359 error="rpc error making call: EOF" === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.545Z [DEBUG] TestPreparedQuery_Update.server.raft: votes: needed=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.535Z [ERROR] TestServiceManager_PersistService_ConfigFiles.client: RPC failed to server: method=DiscoveryChain.Get server=127.0.0.1:30359 error="rpc error making call: EOF" === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.546Z [INFO] TestPreparedQuery_Get/#00.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:45ee4836-05b1-2c74-c9cf-ef3bf644d538 Address:127.0.0.1:29429}]" === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.535Z [ERROR] TestServiceManager_PersistService_ConfigFiles.client: RPC failed to server: method=ConnectCA.Roots server=127.0.0.1:30359 error="rpc error making call: EOF" === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.547Z [DEBUG] TestPreparedQuery_Update.server.raft: vote granted: from=77eeb1b0-98b0-ea25-d942-ceccd316423d term=2 tally=1 === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.536Z [ERROR] TestServiceManager_PersistService_ConfigFiles.client: RPC failed to server: method=Intention.Match server=127.0.0.1:30359 error="rpc error making call: EOF" === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.550Z [INFO] TestPreparedQuery_Update.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:02.551Z [INFO] TestPreparedQuery_Update.server.raft: entering leader state: leader="Node at 127.0.0.1:29423 [Leader]" writer.go:29: 2021-01-29T19:33:02.553Z [INFO] TestPreparedQuery_Update.server: cluster leadership acquired === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.551Z [INFO] TestPreparedQuery_Get/#00.server.serf.wan: serf: EventMemberJoin: Node-45ee4836-05b1-2c74-c9cf-ef3bf644d538.dc1 127.0.0.1 === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.555Z [INFO] TestPreparedQuery_Update.server: New leader elected: payload=Node-77eeb1b0-98b0-ea25-d942-ceccd316423d === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.552Z [INFO] TestPreparedQuery_Get/#00.server.raft: entering follower state: follower="Node at 127.0.0.1:29429 [Follower]" leader= === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.536Z [ERROR] TestServiceManager_PersistService_ConfigFiles.client: RPC failed to server: method=DiscoveryChain.Get server=127.0.0.1:30359 error="rpc error making call: EOF" writer.go:29: 2021-01-29T19:33:02.536Z [ERROR] TestServiceManager_PersistService_ConfigFiles.client: RPC failed to server: method=ConfigEntry.ResolveServiceConfig server=127.0.0.1:30359 error="rpc error making call: EOF" === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.562Z [DEBUG] TestPreparedQuery_Update.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29423 === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.556Z [INFO] TestPreparedQuery_Get/#00.server.serf.lan: serf: EventMemberJoin: Node-45ee4836-05b1-2c74-c9cf-ef3bf644d538 127.0.0.1 writer.go:29: 2021-01-29T19:33:02.569Z [INFO] TestPreparedQuery_Get/#00.server: Adding LAN server: server="Node-45ee4836-05b1-2c74-c9cf-ef3bf644d538 (Addr: tcp/127.0.0.1:29429) (DC: dc1)" writer.go:29: 2021-01-29T19:33:02.569Z [INFO] TestPreparedQuery_Get/#00.server: Handled event for server in area: event=member-join server=Node-45ee4836-05b1-2c74-c9cf-ef3bf644d538.dc1 area=wan === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.536Z [ERROR] TestServiceManager_PersistService_ConfigFiles.client: RPC failed to server: method=Health.ServiceNodes server=127.0.0.1:30359 error="rpc error making call: EOF" === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.571Z [INFO] TestPreparedQuery_Get/#00: Started DNS server: address=127.0.0.1:29424 network=tcp === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.542Z [INFO] TestServiceManager_PersistService_ConfigFiles: Stopping server: protocol=DNS address=127.0.0.1:30366 network=tcp === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.572Z [INFO] TestPreparedQuery_Get/#00: Started DNS server: address=127.0.0.1:29424 network=udp writer.go:29: 2021-01-29T19:33:02.573Z [INFO] TestPreparedQuery_Get/#00: Started HTTP server: address=127.0.0.1:29425 network=tcp writer.go:29: 2021-01-29T19:33:02.574Z [INFO] TestPreparedQuery_Get/#00: started state syncer === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:02.554Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:02.572Z [INFO] TestServiceManager_PersistService_ConfigFiles: Stopping server: protocol=DNS address=127.0.0.1:30366 network=udp writer.go:29: 2021-01-29T19:33:02.577Z [INFO] TestServiceManager_PersistService_ConfigFiles: Stopping server: protocol=HTTP address=127.0.0.1:30367 network=tcp === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.586Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:02.595Z [DEBUG] TestPreparedQuery_Delete: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:02.601Z [INFO] TestPreparedQuery_Delete: Synced node info === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.614Z [INFO] TestPreparedQuery_Update.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:02.616Z [INFO] TestPreparedQuery_Update.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:02.618Z [DEBUG] TestPreparedQuery_Update.server: Skipping self join check for node since the cluster is too small: node=Node-77eeb1b0-98b0-ea25-d942-ceccd316423d writer.go:29: 2021-01-29T19:33:02.620Z [INFO] TestPreparedQuery_Update.server: member joined, marking health alive: member=Node-77eeb1b0-98b0-ea25-d942-ceccd316423d === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.621Z [WARN] TestPreparedQuery_Get/#00.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:02.622Z [INFO] TestPreparedQuery_Get/#00.server.raft: entering candidate state: node="Node at 127.0.0.1:29429 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:02.624Z [DEBUG] TestPreparedQuery_Get/#00.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:02.626Z [DEBUG] TestPreparedQuery_Get/#00.server.raft: vote granted: from=45ee4836-05b1-2c74-c9cf-ef3bf644d538 term=2 tally=1 writer.go:29: 2021-01-29T19:33:02.627Z [INFO] TestPreparedQuery_Get/#00.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:02.628Z [INFO] TestPreparedQuery_Get/#00.server.raft: entering leader state: leader="Node at 127.0.0.1:29429 [Leader]" writer.go:29: 2021-01-29T19:33:02.631Z [INFO] TestPreparedQuery_Get/#00.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:02.632Z [INFO] TestPreparedQuery_Get/#00.server: New leader elected: payload=Node-45ee4836-05b1-2c74-c9cf-ef3bf644d538 === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.644Z [WARN] TestPreparedQuery_Update.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:02.649Z [INFO] TestPreparedQuery_Update: Requesting shutdown writer.go:29: 2021-01-29T19:33:02.650Z [INFO] TestPreparedQuery_Update.server: shutting down server writer.go:29: 2021-01-29T19:33:02.652Z [DEBUG] TestPreparedQuery_Update.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:02.653Z [WARN] TestPreparedQuery_Update.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:02.652Z [ERROR] TestPreparedQuery_Update.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.655Z [DEBUG] TestPreparedQuery_Get/#00.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29429 === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:02.656Z [WARN] TestPreparedQuery_Delete.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:02.658Z [INFO] TestPreparedQuery_Delete: Requesting shutdown === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.662Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.653Z [DEBUG] TestPreparedQuery_Update.leader: stopped routine: routine="CA root pruning" === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:02.660Z [INFO] TestPreparedQuery_Delete.server: shutting down server writer.go:29: 2021-01-29T19:33:02.664Z [DEBUG] TestPreparedQuery_Delete.leader: stopping routine: routine="CA root pruning" === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.661Z [WARN] TestPreparedQuery_Update.server.serf.wan: serf: Shutdown without a Leave === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:02.664Z [WARN] TestPreparedQuery_Delete.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:02.664Z [DEBUG] TestPreparedQuery_Delete.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:02.670Z [WARN] TestPreparedQuery_Delete.server.serf.wan: serf: Shutdown without a Leave === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.670Z [INFO] TestPreparedQuery_Update.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:02.672Z [INFO] TestPreparedQuery_Update: consul server down writer.go:29: 2021-01-29T19:33:02.673Z [INFO] TestPreparedQuery_Update: shutdown complete writer.go:29: 2021-01-29T19:33:02.673Z [INFO] TestPreparedQuery_Update: Stopping server: protocol=DNS address=127.0.0.1:29418 network=tcp === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:02.674Z [INFO] TestPreparedQuery_Delete.server.router.manager: shutting down === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.674Z [INFO] TestPreparedQuery_Update: Stopping server: protocol=DNS address=127.0.0.1:29418 network=udp === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:02.674Z [INFO] TestPreparedQuery_Delete: consul server down === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:02.675Z [INFO] TestPreparedQuery_Update: Stopping server: protocol=HTTP address=127.0.0.1:29419 network=tcp === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.675Z [INFO] TestPreparedQuery_Get/#00.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:02.676Z [INFO] TestPreparedQuery_Delete: shutdown complete writer.go:29: 2021-01-29T19:33:02.677Z [INFO] TestPreparedQuery_Delete: Stopping server: protocol=DNS address=127.0.0.1:29412 network=tcp === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.676Z [INFO] TestPreparedQuery_Get/#00.leader: started routine: routine="CA root pruning" === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:02.677Z [INFO] TestPreparedQuery_Delete: Stopping server: protocol=DNS address=127.0.0.1:29412 network=udp === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.678Z [DEBUG] TestPreparedQuery_Get/#00.server: Skipping self join check for node since the cluster is too small: node=Node-45ee4836-05b1-2c74-c9cf-ef3bf644d538 === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:02.679Z [INFO] TestPreparedQuery_Delete: Stopping server: protocol=HTTP address=127.0.0.1:29413 network=tcp === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:02.678Z [WARN] TestPreparedQuery_Get/#00.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:02.682Z [INFO] TestPreparedQuery_Get/#00: Requesting shutdown writer.go:29: 2021-01-29T19:33:02.679Z [INFO] TestPreparedQuery_Get/#00.server: member joined, marking health alive: member=Node-45ee4836-05b1-2c74-c9cf-ef3bf644d538 writer.go:29: 2021-01-29T19:33:02.683Z [INFO] TestPreparedQuery_Get/#00.server: shutting down server writer.go:29: 2021-01-29T19:33:02.685Z [DEBUG] TestPreparedQuery_Get/#00.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:02.686Z [WARN] TestPreparedQuery_Get/#00.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:02.685Z [ERROR] TestPreparedQuery_Get/#00.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:02.688Z [DEBUG] TestPreparedQuery_Get/#00.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:02.691Z [WARN] TestPreparedQuery_Get/#00.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:02.694Z [INFO] TestPreparedQuery_Get/#00.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:02.694Z [INFO] TestPreparedQuery_Get/#00: consul server down writer.go:29: 2021-01-29T19:33:02.697Z [INFO] TestPreparedQuery_Get/#00: shutdown complete writer.go:29: 2021-01-29T19:33:02.697Z [INFO] TestPreparedQuery_Get/#00: Stopping server: protocol=DNS address=127.0.0.1:29424 network=tcp writer.go:29: 2021-01-29T19:33:02.698Z [INFO] TestPreparedQuery_Get/#00: Stopping server: protocol=DNS address=127.0.0.1:29424 network=udp writer.go:29: 2021-01-29T19:33:02.699Z [INFO] TestPreparedQuery_Get/#00: Stopping server: protocol=HTTP address=127.0.0.1:29425 network=tcp === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:03.079Z [INFO] TestServiceManager_PersistService_ConfigFiles: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:03.080Z [INFO] TestServiceManager_PersistService_ConfigFiles: Endpoints down writer.go:29: 2021-01-29T19:33:03.081Z [INFO] TestServiceManager_PersistService_ConfigFiles: Requesting shutdown writer.go:29: 2021-01-29T19:33:03.082Z [INFO] TestServiceManager_PersistService_ConfigFiles.server: shutting down server writer.go:29: 2021-01-29T19:33:03.083Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:03.084Z [WARN] TestServiceManager_PersistService_ConfigFiles.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:03.085Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:03.087Z [WARN] TestServiceManager_PersistService_ConfigFiles.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:03.090Z [INFO] TestServiceManager_PersistService_ConfigFiles.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:03.091Z [INFO] TestServiceManager_PersistService_ConfigFiles: consul server down writer.go:29: 2021-01-29T19:33:03.094Z [INFO] TestServiceManager_PersistService_ConfigFiles: shutdown complete writer.go:29: 2021-01-29T19:33:03.096Z [INFO] TestServiceManager_PersistService_ConfigFiles: Stopping server: protocol=DNS address=127.0.0.1:30354 network=tcp writer.go:29: 2021-01-29T19:33:03.097Z [INFO] TestServiceManager_PersistService_ConfigFiles: Stopping server: protocol=DNS address=127.0.0.1:30354 network=udp writer.go:29: 2021-01-29T19:33:03.099Z [INFO] TestServiceManager_PersistService_ConfigFiles: Stopping server: protocol=HTTP address=127.0.0.1:30355 network=tcp === CONT TestPreparedQuery_Update writer.go:29: 2021-01-29T19:33:03.176Z [INFO] TestPreparedQuery_Update: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:03.178Z [INFO] TestPreparedQuery_Update: Endpoints down --- PASS: TestPreparedQuery_Update (0.73s) === CONT TestPreparedQuery_Explain === RUN TestPreparedQuery_Explain/#00 === CONT TestPreparedQuery_Delete writer.go:29: 2021-01-29T19:33:03.181Z [INFO] TestPreparedQuery_Delete: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:03.182Z [INFO] TestPreparedQuery_Delete: Endpoints down --- PASS: TestPreparedQuery_Delete (0.96s) === CONT TestPreparedQuery_ExecuteCached === CONT TestPreparedQuery_Get/#00 writer.go:29: 2021-01-29T19:33:03.200Z [INFO] TestPreparedQuery_Get/#00: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:03.202Z [INFO] TestPreparedQuery_Get/#00: Endpoints down === RUN TestPreparedQuery_Get/#01 === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.223Z [WARN] TestPreparedQuery_ExecuteCached: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:03.225Z [DEBUG] TestPreparedQuery_ExecuteCached.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:03.227Z [DEBUG] TestPreparedQuery_ExecuteCached.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:03.232Z [INFO] TestPreparedQuery_ExecuteCached.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8bba7470-9133-9e42-de0e-2e1a2001c365 Address:127.0.0.1:29441}]" writer.go:29: 2021-01-29T19:33:03.235Z [INFO] TestPreparedQuery_ExecuteCached.server.serf.wan: serf: EventMemberJoin: Node-8bba7470-9133-9e42-de0e-2e1a2001c365.dc1 127.0.0.1 === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.251Z [WARN] TestPreparedQuery_Explain/#00: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:03.252Z [DEBUG] TestPreparedQuery_Explain/#00.tlsutil: Update: version=1 === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.238Z [INFO] TestPreparedQuery_ExecuteCached.server.raft: entering follower state: follower="Node at 127.0.0.1:29441 [Follower]" leader= === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.253Z [DEBUG] TestPreparedQuery_Explain/#00.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:03.257Z [INFO] TestPreparedQuery_Explain/#00.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f7984ec9-9664-4830-02ab-6b1955c0ef1c Address:127.0.0.1:29435}]" writer.go:29: 2021-01-29T19:33:03.258Z [INFO] TestPreparedQuery_Explain/#00.server.serf.wan: serf: EventMemberJoin: Node-f7984ec9-9664-4830-02ab-6b1955c0ef1c.dc1 127.0.0.1 === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.261Z [WARN] TestPreparedQuery_Get/#01: bootstrap = true: do not enable unless necessary === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.258Z [INFO] TestPreparedQuery_Explain/#00.server.raft: entering follower state: follower="Node at 127.0.0.1:29435 [Follower]" leader= writer.go:29: 2021-01-29T19:33:03.261Z [INFO] TestPreparedQuery_Explain/#00.server.serf.lan: serf: EventMemberJoin: Node-f7984ec9-9664-4830-02ab-6b1955c0ef1c 127.0.0.1 === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.256Z [INFO] TestPreparedQuery_ExecuteCached.server.serf.lan: serf: EventMemberJoin: Node-8bba7470-9133-9e42-de0e-2e1a2001c365 127.0.0.1 === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.274Z [INFO] TestPreparedQuery_Explain/#00.server: Handled event for server in area: event=member-join server=Node-f7984ec9-9664-4830-02ab-6b1955c0ef1c.dc1 area=wan writer.go:29: 2021-01-29T19:33:03.277Z [INFO] TestPreparedQuery_Explain/#00.server: Adding LAN server: server="Node-f7984ec9-9664-4830-02ab-6b1955c0ef1c (Addr: tcp/127.0.0.1:29435) (DC: dc1)" === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.271Z [DEBUG] TestPreparedQuery_Get/#01.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:03.279Z [DEBUG] TestPreparedQuery_Get/#01.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.279Z [INFO] TestPreparedQuery_ExecuteCached.server: Adding LAN server: server="Node-8bba7470-9133-9e42-de0e-2e1a2001c365 (Addr: tcp/127.0.0.1:29441) (DC: dc1)" writer.go:29: 2021-01-29T19:33:03.281Z [INFO] TestPreparedQuery_ExecuteCached.server: Handled event for server in area: event=member-join server=Node-8bba7470-9133-9e42-de0e-2e1a2001c365.dc1 area=wan writer.go:29: 2021-01-29T19:33:03.284Z [INFO] TestPreparedQuery_ExecuteCached: Started DNS server: address=127.0.0.1:29436 network=tcp writer.go:29: 2021-01-29T19:33:03.286Z [INFO] TestPreparedQuery_ExecuteCached: Started DNS server: address=127.0.0.1:29436 network=udp === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.286Z [INFO] TestPreparedQuery_Explain/#00: Started DNS server: address=127.0.0.1:29430 network=udp === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.290Z [INFO] TestPreparedQuery_Get/#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6c7eb2bf-6fd7-d58a-aa06-d144c055dff2 Address:127.0.0.1:29447}]" writer.go:29: 2021-01-29T19:33:03.291Z [INFO] TestPreparedQuery_Get/#01.server.raft: entering follower state: follower="Node at 127.0.0.1:29447 [Follower]" leader= === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.290Z [INFO] TestPreparedQuery_Explain/#00: Started DNS server: address=127.0.0.1:29430 network=tcp === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.291Z [INFO] TestPreparedQuery_ExecuteCached: Started HTTP server: address=127.0.0.1:29437 network=tcp === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.294Z [INFO] TestPreparedQuery_Get/#01.server.serf.wan: serf: EventMemberJoin: Node-6c7eb2bf-6fd7-d58a-aa06-d144c055dff2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:03.296Z [INFO] TestPreparedQuery_Get/#01.server.serf.lan: serf: EventMemberJoin: Node-6c7eb2bf-6fd7-d58a-aa06-d144c055dff2 127.0.0.1 === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.298Z [INFO] TestPreparedQuery_Explain/#00: Started HTTP server: address=127.0.0.1:29431 network=tcp writer.go:29: 2021-01-29T19:33:03.299Z [INFO] TestPreparedQuery_Explain/#00: started state syncer === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.297Z [INFO] TestPreparedQuery_ExecuteCached: started state syncer === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.305Z [INFO] TestPreparedQuery_Get/#01.server: Adding LAN server: server="Node-6c7eb2bf-6fd7-d58a-aa06-d144c055dff2 (Addr: tcp/127.0.0.1:29447) (DC: dc1)" writer.go:29: 2021-01-29T19:33:03.306Z [INFO] TestPreparedQuery_Get/#01.server: Handled event for server in area: event=member-join server=Node-6c7eb2bf-6fd7-d58a-aa06-d144c055dff2.dc1 area=wan writer.go:29: 2021-01-29T19:33:03.311Z [INFO] TestPreparedQuery_Get/#01: Started DNS server: address=127.0.0.1:29442 network=tcp writer.go:29: 2021-01-29T19:33:03.314Z [INFO] TestPreparedQuery_Get/#01: Started DNS server: address=127.0.0.1:29442 network=udp writer.go:29: 2021-01-29T19:33:03.316Z [INFO] TestPreparedQuery_Get/#01: Started HTTP server: address=127.0.0.1:29443 network=tcp writer.go:29: 2021-01-29T19:33:03.319Z [INFO] TestPreparedQuery_Get/#01: started state syncer === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.322Z [WARN] TestPreparedQuery_ExecuteCached.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:03.323Z [INFO] TestPreparedQuery_ExecuteCached.server.raft: entering candidate state: node="Node at 127.0.0.1:29441 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:03.325Z [DEBUG] TestPreparedQuery_ExecuteCached.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:03.326Z [DEBUG] TestPreparedQuery_ExecuteCached.server.raft: vote granted: from=8bba7470-9133-9e42-de0e-2e1a2001c365 term=2 tally=1 writer.go:29: 2021-01-29T19:33:03.327Z [INFO] TestPreparedQuery_ExecuteCached.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:03.328Z [INFO] TestPreparedQuery_ExecuteCached.server.raft: entering leader state: leader="Node at 127.0.0.1:29441 [Leader]" writer.go:29: 2021-01-29T19:33:03.329Z [INFO] TestPreparedQuery_ExecuteCached.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:03.329Z [INFO] TestPreparedQuery_ExecuteCached.server: New leader elected: payload=Node-8bba7470-9133-9e42-de0e-2e1a2001c365 === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.334Z [WARN] TestPreparedQuery_Explain/#00.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:03.335Z [INFO] TestPreparedQuery_Explain/#00.server.raft: entering candidate state: node="Node at 127.0.0.1:29435 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:03.337Z [DEBUG] TestPreparedQuery_Explain/#00.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:03.337Z [DEBUG] TestPreparedQuery_Explain/#00.server.raft: vote granted: from=f7984ec9-9664-4830-02ab-6b1955c0ef1c term=2 tally=1 writer.go:29: 2021-01-29T19:33:03.338Z [INFO] TestPreparedQuery_Explain/#00.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:03.339Z [INFO] TestPreparedQuery_Explain/#00.server.raft: entering leader state: leader="Node at 127.0.0.1:29435 [Leader]" writer.go:29: 2021-01-29T19:33:03.340Z [INFO] TestPreparedQuery_Explain/#00.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:03.341Z [INFO] TestPreparedQuery_Explain/#00.server: New leader elected: payload=Node-f7984ec9-9664-4830-02ab-6b1955c0ef1c writer.go:29: 2021-01-29T19:33:03.343Z [DEBUG] TestPreparedQuery_Explain/#00.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29435 writer.go:29: 2021-01-29T19:33:03.347Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.331Z [DEBUG] TestPreparedQuery_ExecuteCached.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29441 === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.361Z [WARN] TestPreparedQuery_Get/#01.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:03.363Z [INFO] TestPreparedQuery_Get/#01.server.raft: entering candidate state: node="Node at 127.0.0.1:29447 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:03.365Z [DEBUG] TestPreparedQuery_Get/#01.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:03.366Z [DEBUG] TestPreparedQuery_Get/#01.server.raft: vote granted: from=6c7eb2bf-6fd7-d58a-aa06-d144c055dff2 term=2 tally=1 writer.go:29: 2021-01-29T19:33:03.367Z [INFO] TestPreparedQuery_Get/#01.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:03.368Z [INFO] TestPreparedQuery_Get/#01.server.raft: entering leader state: leader="Node at 127.0.0.1:29447 [Leader]" writer.go:29: 2021-01-29T19:33:03.369Z [INFO] TestPreparedQuery_Get/#01.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:03.370Z [INFO] TestPreparedQuery_Get/#01.server: New leader elected: payload=Node-6c7eb2bf-6fd7-d58a-aa06-d144c055dff2 === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.371Z [INFO] TestPreparedQuery_Explain/#00.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:03.373Z [INFO] TestPreparedQuery_Explain/#00.leader: started routine: routine="CA root pruning" === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.372Z [DEBUG] TestPreparedQuery_Get/#01.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29447 === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.365Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.375Z [DEBUG] TestPreparedQuery_Explain/#00.server: Skipping self join check for node since the cluster is too small: node=Node-f7984ec9-9664-4830-02ab-6b1955c0ef1c === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.384Z [INFO] TestPreparedQuery_ExecuteCached.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:03.385Z [INFO] TestPreparedQuery_ExecuteCached.leader: started routine: routine="CA root pruning" === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.381Z [INFO] TestPreparedQuery_Explain/#00.server: member joined, marking health alive: member=Node-f7984ec9-9664-4830-02ab-6b1955c0ef1c === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.385Z [DEBUG] TestPreparedQuery_ExecuteCached.server: Skipping self join check for node since the cluster is too small: node=Node-8bba7470-9133-9e42-de0e-2e1a2001c365 writer.go:29: 2021-01-29T19:33:03.388Z [INFO] TestPreparedQuery_ExecuteCached.server: member joined, marking health alive: member=Node-8bba7470-9133-9e42-de0e-2e1a2001c365 === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.391Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:03.419Z [INFO] TestPreparedQuery_Get/#01.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:03.420Z [INFO] TestPreparedQuery_Get/#01.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:03.421Z [DEBUG] TestPreparedQuery_Get/#01.server: Skipping self join check for node since the cluster is too small: node=Node-6c7eb2bf-6fd7-d58a-aa06-d144c055dff2 writer.go:29: 2021-01-29T19:33:03.422Z [INFO] TestPreparedQuery_Get/#01.server: member joined, marking health alive: member=Node-6c7eb2bf-6fd7-d58a-aa06-d144c055dff2 === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.423Z [WARN] TestPreparedQuery_ExecuteCached.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:03.445Z [INFO] TestPreparedQuery_ExecuteCached: Requesting shutdown writer.go:29: 2021-01-29T19:33:03.446Z [INFO] TestPreparedQuery_ExecuteCached.server: shutting down server writer.go:29: 2021-01-29T19:33:03.452Z [DEBUG] TestPreparedQuery_ExecuteCached.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:03.453Z [WARN] TestPreparedQuery_ExecuteCached.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:03.454Z [ERROR] TestPreparedQuery_ExecuteCached.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:03.454Z [DEBUG] TestPreparedQuery_ExecuteCached.leader: stopped routine: routine="CA root pruning" === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.457Z [INFO] TestPreparedQuery_Get/#01: Requesting shutdown === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.459Z [WARN] TestPreparedQuery_ExecuteCached.server.serf.wan: serf: Shutdown without a Leave === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.460Z [INFO] TestPreparedQuery_Get/#01.server: shutting down server === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.461Z [INFO] TestPreparedQuery_ExecuteCached.server.router.manager: shutting down === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.464Z [DEBUG] TestPreparedQuery_Get/#01.leader: stopping routine: routine="CA root pruning" === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.465Z [INFO] TestPreparedQuery_ExecuteCached: consul server down writer.go:29: 2021-01-29T19:33:03.467Z [INFO] TestPreparedQuery_ExecuteCached: shutdown complete === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.467Z [WARN] TestPreparedQuery_Get/#01.server.serf.lan: serf: Shutdown without a Leave === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.470Z [INFO] TestPreparedQuery_ExecuteCached: Stopping server: protocol=DNS address=127.0.0.1:29436 network=tcp === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.465Z [ERROR] TestPreparedQuery_Get/#01.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:03.469Z [DEBUG] TestPreparedQuery_Get/#01.leader: stopped routine: routine="CA root pruning" === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.473Z [INFO] TestPreparedQuery_ExecuteCached: Stopping server: protocol=DNS address=127.0.0.1:29436 network=udp === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.477Z [WARN] TestPreparedQuery_Get/#01.server.serf.wan: serf: Shutdown without a Leave === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.477Z [INFO] TestPreparedQuery_ExecuteCached: Stopping server: protocol=HTTP address=127.0.0.1:29437 network=tcp === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.480Z [INFO] TestPreparedQuery_Get/#01.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:03.481Z [INFO] TestPreparedQuery_Get/#01: consul server down writer.go:29: 2021-01-29T19:33:03.484Z [INFO] TestPreparedQuery_Get/#01: shutdown complete writer.go:29: 2021-01-29T19:33:03.486Z [INFO] TestPreparedQuery_Get/#01: Stopping server: protocol=DNS address=127.0.0.1:29442 network=tcp writer.go:29: 2021-01-29T19:33:03.492Z [INFO] TestPreparedQuery_Get/#01: Stopping server: protocol=DNS address=127.0.0.1:29442 network=udp writer.go:29: 2021-01-29T19:33:03.494Z [INFO] TestPreparedQuery_Get/#01: Stopping server: protocol=HTTP address=127.0.0.1:29443 network=tcp === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.565Z [DEBUG] TestPreparedQuery_Explain/#00: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:03.570Z [INFO] TestPreparedQuery_Explain/#00: Synced node info writer.go:29: 2021-01-29T19:33:03.572Z [WARN] TestPreparedQuery_Explain/#00.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:03.582Z [INFO] TestPreparedQuery_Explain/#00: Requesting shutdown writer.go:29: 2021-01-29T19:33:03.585Z [INFO] TestPreparedQuery_Explain/#00.server: shutting down server writer.go:29: 2021-01-29T19:33:03.587Z [DEBUG] TestPreparedQuery_Explain/#00.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:03.592Z [WARN] TestPreparedQuery_Explain/#00.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:03.592Z [DEBUG] TestPreparedQuery_Explain/#00.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:03.595Z [WARN] TestPreparedQuery_Explain/#00.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:03.597Z [INFO] TestPreparedQuery_Explain/#00.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:03.598Z [INFO] TestPreparedQuery_Explain/#00: consul server down === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:03.600Z [INFO] TestServiceManager_PersistService_ConfigFiles: Waiting for endpoints to shut down === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.599Z [INFO] TestPreparedQuery_Explain/#00: shutdown complete === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:03.601Z [INFO] TestServiceManager_PersistService_ConfigFiles: Endpoints down === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:03.601Z [INFO] TestPreparedQuery_Explain/#00: Stopping server: protocol=DNS address=127.0.0.1:29430 network=tcp writer.go:29: 2021-01-29T19:33:03.603Z [INFO] TestPreparedQuery_Explain/#00: Stopping server: protocol=DNS address=127.0.0.1:29430 network=udp writer.go:29: 2021-01-29T19:33:03.604Z [INFO] TestPreparedQuery_Explain/#00: Stopping server: protocol=HTTP address=127.0.0.1:29431 network=tcp === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:03.611Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:03.614Z [INFO] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: EventMemberJoin: Node-f3d198b9-0409-5b14-4f63-e54d5200e73f 127.0.0.1 writer.go:29: 2021-01-29T19:33:03.616Z [INFO] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: Attempting re-join to previously known node: Node-62b5279b-1bf5-1b76-e0c3-f44b4f62394e: 127.0.0.1:30357 writer.go:29: 2021-01-29T19:33:03.618Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.619Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.621Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.621Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.memberlist.lan: memberlist: Failed to join 127.0.0.1: dial tcp 127.0.0.1:30357: connect: connection refused writer.go:29: 2021-01-29T19:33:03.623Z [INFO] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: Attempting re-join to previously known node: Node-ca69a67c-ebf9-2e0e-7b70-7ce1b4943b20: 127.0.0.1:30369 writer.go:29: 2021-01-29T19:33:03.622Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.623Z [ERROR] TestServiceManager_PersistService_ConfigFiles: error handling service update: error="error watching service config: No known Consul servers" writer.go:29: 2021-01-29T19:33:03.623Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: added local registration for service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:33:03.624Z [DEBUG] TestServiceManager_PersistService_ConfigFiles.client.memberlist.lan: memberlist: Failed to join 127.0.0.1: dial tcp 127.0.0.1:30369: connect: connection refused writer.go:29: 2021-01-29T19:33:03.626Z [ERROR] TestServiceManager_PersistService_ConfigFiles: error handling service update: error="error watching service config: No known Consul servers" writer.go:29: 2021-01-29T19:33:03.627Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.629Z [INFO] TestServiceManager_PersistService_ConfigFiles: Started DNS server: address=127.0.0.1:29448 network=udp writer.go:29: 2021-01-29T19:33:03.631Z [INFO] TestServiceManager_PersistService_ConfigFiles: Started DNS server: address=127.0.0.1:29448 network=tcp writer.go:29: 2021-01-29T19:33:03.629Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: Failed to re-join any previously known node writer.go:29: 2021-01-29T19:33:03.629Z [ERROR] TestServiceManager_PersistService_ConfigFiles: error handling service update: error="error watching service config: No known Consul servers" writer.go:29: 2021-01-29T19:33:03.631Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.633Z [INFO] TestServiceManager_PersistService_ConfigFiles: Started HTTP server: address=127.0.0.1:29449 network=tcp writer.go:29: 2021-01-29T19:33:03.634Z [ERROR] TestServiceManager_PersistService_ConfigFiles: error handling service update: error="error watching service config: No known Consul servers" writer.go:29: 2021-01-29T19:33:03.637Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.637Z [INFO] TestServiceManager_PersistService_ConfigFiles: started state syncer writer.go:29: 2021-01-29T19:33:03.639Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.640Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.643Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.643Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.643Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.645Z [ERROR] TestServiceManager_PersistService_ConfigFiles.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:33:03.645Z [ERROR] TestServiceManager_PersistService_ConfigFiles.proxycfg: watch error: id=service-http-checks:web error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:33:03.645Z [DEBUG] TestServiceManager_PersistService_ConfigFiles: removed service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:33:03.647Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.649Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.651Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.652Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.656Z [INFO] TestServiceManager_PersistService_ConfigFiles: Requesting shutdown writer.go:29: 2021-01-29T19:33:03.658Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.659Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.661Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.662Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.manager: No servers available writer.go:29: 2021-01-29T19:33:03.664Z [INFO] TestServiceManager_PersistService_ConfigFiles.client: shutting down client writer.go:29: 2021-01-29T19:33:03.671Z [WARN] TestServiceManager_PersistService_ConfigFiles.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:03.671Z [INFO] TestServiceManager_PersistService_ConfigFiles.client.manager: shutting down writer.go:29: 2021-01-29T19:33:03.682Z [INFO] TestServiceManager_PersistService_ConfigFiles: consul client down writer.go:29: 2021-01-29T19:33:03.683Z [INFO] TestServiceManager_PersistService_ConfigFiles: shutdown complete writer.go:29: 2021-01-29T19:33:03.684Z [INFO] TestServiceManager_PersistService_ConfigFiles: Stopping server: protocol=DNS address=127.0.0.1:29448 network=tcp writer.go:29: 2021-01-29T19:33:03.685Z [INFO] TestServiceManager_PersistService_ConfigFiles: Stopping server: protocol=DNS address=127.0.0.1:29448 network=udp writer.go:29: 2021-01-29T19:33:03.686Z [INFO] TestServiceManager_PersistService_ConfigFiles: Stopping server: protocol=HTTP address=127.0.0.1:29449 network=tcp === CONT TestPreparedQuery_ExecuteCached writer.go:29: 2021-01-29T19:33:03.981Z [INFO] TestPreparedQuery_ExecuteCached: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:03.983Z [INFO] TestPreparedQuery_ExecuteCached: Endpoints down --- PASS: TestPreparedQuery_ExecuteCached (0.80s) === CONT TestPreparedQuery_Execute === RUN TestPreparedQuery_Execute/#00 === CONT TestPreparedQuery_Get/#01 writer.go:29: 2021-01-29T19:33:03.996Z [INFO] TestPreparedQuery_Get/#01: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:03.997Z [INFO] TestPreparedQuery_Get/#01: Endpoints down --- PASS: TestPreparedQuery_Get (1.48s) --- PASS: TestPreparedQuery_Get/#00 (0.68s) --- PASS: TestPreparedQuery_Get/#01 (0.79s) === CONT TestPreparedQuery_List === RUN TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.010Z [WARN] TestPreparedQuery_List/#00: bootstrap = true: do not enable unless necessary === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.012Z [WARN] TestPreparedQuery_Execute/#00: bootstrap = true: do not enable unless necessary === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.012Z [DEBUG] TestPreparedQuery_List/#00.tlsutil: Update: version=1 === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.015Z [DEBUG] TestPreparedQuery_Execute/#00.tlsutil: Update: version=1 === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.015Z [DEBUG] TestPreparedQuery_List/#00.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.017Z [DEBUG] TestPreparedQuery_Execute/#00.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.020Z [INFO] TestPreparedQuery_List/#00.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f7ec8351-21f6-d4e0-bf6f-28d28a73c7d7 Address:127.0.0.1:29471}]" writer.go:29: 2021-01-29T19:33:04.022Z [INFO] TestPreparedQuery_List/#00.server.raft: entering follower state: follower="Node at 127.0.0.1:29471 [Follower]" leader= writer.go:29: 2021-01-29T19:33:04.023Z [INFO] TestPreparedQuery_List/#00.server.serf.wan: serf: EventMemberJoin: Node-f7ec8351-21f6-d4e0-bf6f-28d28a73c7d7.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:04.028Z [INFO] TestPreparedQuery_List/#00.server.serf.lan: serf: EventMemberJoin: Node-f7ec8351-21f6-d4e0-bf6f-28d28a73c7d7 127.0.0.1 writer.go:29: 2021-01-29T19:33:04.044Z [INFO] TestPreparedQuery_List/#00.server: Adding LAN server: server="Node-f7ec8351-21f6-d4e0-bf6f-28d28a73c7d7 (Addr: tcp/127.0.0.1:29471) (DC: dc1)" writer.go:29: 2021-01-29T19:33:04.044Z [INFO] TestPreparedQuery_List/#00.server: Handled event for server in area: event=member-join server=Node-f7ec8351-21f6-d4e0-bf6f-28d28a73c7d7.dc1 area=wan writer.go:29: 2021-01-29T19:33:04.045Z [INFO] TestPreparedQuery_List/#00: Started DNS server: address=127.0.0.1:29466 network=udp writer.go:29: 2021-01-29T19:33:04.046Z [INFO] TestPreparedQuery_List/#00: Started DNS server: address=127.0.0.1:29466 network=tcp === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.048Z [INFO] TestPreparedQuery_Execute/#00.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5262732e-94f3-d69b-a55d-74e327f0bd13 Address:127.0.0.1:29459}]" === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.048Z [INFO] TestPreparedQuery_List/#00: Started HTTP server: address=127.0.0.1:29467 network=tcp === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.049Z [INFO] TestPreparedQuery_Execute/#00.server.raft: entering follower state: follower="Node at 127.0.0.1:29459 [Follower]" leader= === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.049Z [INFO] TestPreparedQuery_List/#00: started state syncer === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.055Z [INFO] TestPreparedQuery_Execute/#00.server.serf.wan: serf: EventMemberJoin: Node-5262732e-94f3-d69b-a55d-74e327f0bd13.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:04.060Z [INFO] TestPreparedQuery_Execute/#00.server.serf.lan: serf: EventMemberJoin: Node-5262732e-94f3-d69b-a55d-74e327f0bd13 127.0.0.1 writer.go:29: 2021-01-29T19:33:04.063Z [INFO] TestPreparedQuery_Execute/#00.server: Adding LAN server: server="Node-5262732e-94f3-d69b-a55d-74e327f0bd13 (Addr: tcp/127.0.0.1:29459) (DC: dc1)" writer.go:29: 2021-01-29T19:33:04.063Z [INFO] TestPreparedQuery_Execute/#00.server: Handled event for server in area: event=member-join server=Node-5262732e-94f3-d69b-a55d-74e327f0bd13.dc1 area=wan writer.go:29: 2021-01-29T19:33:04.064Z [INFO] TestPreparedQuery_Execute/#00: Started DNS server: address=127.0.0.1:29454 network=tcp writer.go:29: 2021-01-29T19:33:04.073Z [INFO] TestPreparedQuery_Execute/#00: Started DNS server: address=127.0.0.1:29454 network=udp writer.go:29: 2021-01-29T19:33:04.075Z [INFO] TestPreparedQuery_Execute/#00: Started HTTP server: address=127.0.0.1:29455 network=tcp writer.go:29: 2021-01-29T19:33:04.075Z [INFO] TestPreparedQuery_Execute/#00: started state syncer === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.081Z [WARN] TestPreparedQuery_List/#00.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:04.082Z [INFO] TestPreparedQuery_List/#00.server.raft: entering candidate state: node="Node at 127.0.0.1:29471 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:04.084Z [DEBUG] TestPreparedQuery_List/#00.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:04.085Z [DEBUG] TestPreparedQuery_List/#00.server.raft: vote granted: from=f7ec8351-21f6-d4e0-bf6f-28d28a73c7d7 term=2 tally=1 writer.go:29: 2021-01-29T19:33:04.087Z [INFO] TestPreparedQuery_List/#00.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:04.088Z [INFO] TestPreparedQuery_List/#00.server.raft: entering leader state: leader="Node at 127.0.0.1:29471 [Leader]" writer.go:29: 2021-01-29T19:33:04.089Z [INFO] TestPreparedQuery_List/#00.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:04.090Z [INFO] TestPreparedQuery_List/#00.server: New leader elected: payload=Node-f7ec8351-21f6-d4e0-bf6f-28d28a73c7d7 writer.go:29: 2021-01-29T19:33:04.091Z [DEBUG] TestPreparedQuery_List/#00.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29471 writer.go:29: 2021-01-29T19:33:04.096Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.099Z [WARN] TestPreparedQuery_Execute/#00.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:04.100Z [INFO] TestPreparedQuery_Execute/#00.server.raft: entering candidate state: node="Node at 127.0.0.1:29459 [Candidate]" term=2 === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.102Z [INFO] TestPreparedQuery_List/#00.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:04.103Z [INFO] TestPreparedQuery_List/#00.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.104Z [DEBUG] TestPreparedQuery_List/#00.server: Skipping self join check for node since the cluster is too small: node=Node-f7ec8351-21f6-d4e0-bf6f-28d28a73c7d7 === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.104Z [DEBUG] TestPreparedQuery_Execute/#00.server.raft: votes: needed=1 === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.105Z [INFO] TestPreparedQuery_List/#00.server: member joined, marking health alive: member=Node-f7ec8351-21f6-d4e0-bf6f-28d28a73c7d7 === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:04.105Z [INFO] TestPreparedQuery_Explain/#00: Waiting for endpoints to shut down === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.105Z [DEBUG] TestPreparedQuery_Execute/#00.server.raft: vote granted: from=5262732e-94f3-d69b-a55d-74e327f0bd13 term=2 tally=1 === CONT TestPreparedQuery_Explain/#00 writer.go:29: 2021-01-29T19:33:04.106Z [INFO] TestPreparedQuery_Explain/#00: Endpoints down === RUN TestPreparedQuery_Explain/#01 === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.107Z [INFO] TestPreparedQuery_Execute/#00.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:04.109Z [INFO] TestPreparedQuery_Execute/#00.server.raft: entering leader state: leader="Node at 127.0.0.1:29459 [Leader]" writer.go:29: 2021-01-29T19:33:04.112Z [INFO] TestPreparedQuery_Execute/#00.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:04.112Z [INFO] TestPreparedQuery_Execute/#00.server: New leader elected: payload=Node-5262732e-94f3-d69b-a55d-74e327f0bd13 writer.go:29: 2021-01-29T19:33:04.114Z [DEBUG] TestPreparedQuery_Execute/#00.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29459 === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.118Z [WARN] TestPreparedQuery_Explain/#01: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:04.119Z [DEBUG] TestPreparedQuery_Explain/#01.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:04.120Z [DEBUG] TestPreparedQuery_Explain/#01.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.123Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.125Z [INFO] TestPreparedQuery_Explain/#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1e0400b7-7592-c3e0-158a-9b0b24611375 Address:127.0.0.1:29465}]" writer.go:29: 2021-01-29T19:33:04.126Z [INFO] TestPreparedQuery_Explain/#01.server.raft: entering follower state: follower="Node at 127.0.0.1:29465 [Follower]" leader= writer.go:29: 2021-01-29T19:33:04.127Z [INFO] TestPreparedQuery_Explain/#01.server.serf.wan: serf: EventMemberJoin: Node-1e0400b7-7592-c3e0-158a-9b0b24611375.dc1 127.0.0.1 === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.131Z [INFO] TestPreparedQuery_Execute/#00.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:04.132Z [INFO] TestPreparedQuery_Execute/#00.leader: started routine: routine="CA root pruning" === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.130Z [INFO] TestPreparedQuery_Explain/#01.server.serf.lan: serf: EventMemberJoin: Node-1e0400b7-7592-c3e0-158a-9b0b24611375 127.0.0.1 === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.133Z [DEBUG] TestPreparedQuery_Execute/#00.server: Skipping self join check for node since the cluster is too small: node=Node-5262732e-94f3-d69b-a55d-74e327f0bd13 writer.go:29: 2021-01-29T19:33:04.132Z [INFO] TestPreparedQuery_Execute/#00: Synced node info writer.go:29: 2021-01-29T19:33:04.135Z [DEBUG] TestPreparedQuery_Execute/#00: Node info in sync writer.go:29: 2021-01-29T19:33:04.134Z [INFO] TestPreparedQuery_Execute/#00.server: member joined, marking health alive: member=Node-5262732e-94f3-d69b-a55d-74e327f0bd13 === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.138Z [INFO] TestPreparedQuery_Explain/#01.server: Adding LAN server: server="Node-1e0400b7-7592-c3e0-158a-9b0b24611375 (Addr: tcp/127.0.0.1:29465) (DC: dc1)" writer.go:29: 2021-01-29T19:33:04.139Z [INFO] TestPreparedQuery_Explain/#01.server: Handled event for server in area: event=member-join server=Node-1e0400b7-7592-c3e0-158a-9b0b24611375.dc1 area=wan writer.go:29: 2021-01-29T19:33:04.139Z [INFO] TestPreparedQuery_Explain/#01: Started DNS server: address=127.0.0.1:29460 network=udp writer.go:29: 2021-01-29T19:33:04.141Z [INFO] TestPreparedQuery_Explain/#01: Started DNS server: address=127.0.0.1:29460 network=tcp writer.go:29: 2021-01-29T19:33:04.152Z [INFO] TestPreparedQuery_Explain/#01: Started HTTP server: address=127.0.0.1:29461 network=tcp writer.go:29: 2021-01-29T19:33:04.154Z [INFO] TestPreparedQuery_Explain/#01: started state syncer === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:04.187Z [INFO] TestServiceManager_PersistService_ConfigFiles: Waiting for endpoints to shut down === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.187Z [WARN] TestPreparedQuery_Explain/#01.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestServiceManager_PersistService_ConfigFiles writer.go:29: 2021-01-29T19:33:04.188Z [INFO] TestServiceManager_PersistService_ConfigFiles: Endpoints down --- PASS: TestServiceManager_PersistService_ConfigFiles (5.83s) === CONT TestPreparedQuery_Create === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.189Z [INFO] TestPreparedQuery_Explain/#01.server.raft: entering candidate state: node="Node at 127.0.0.1:29465 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:04.191Z [DEBUG] TestPreparedQuery_Explain/#01.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:04.193Z [DEBUG] TestPreparedQuery_Explain/#01.server.raft: vote granted: from=1e0400b7-7592-c3e0-158a-9b0b24611375 term=2 tally=1 writer.go:29: 2021-01-29T19:33:04.194Z [INFO] TestPreparedQuery_Explain/#01.server.raft: election won: tally=1 === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.197Z [WARN] TestPreparedQuery_Create: bootstrap = true: do not enable unless necessary === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.196Z [INFO] TestPreparedQuery_Explain/#01.server.raft: entering leader state: leader="Node at 127.0.0.1:29465 [Leader]" === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.198Z [DEBUG] TestPreparedQuery_Create.tlsutil: Update: version=1 === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.198Z [INFO] TestPreparedQuery_Explain/#01.server: cluster leadership acquired === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.199Z [DEBUG] TestPreparedQuery_Create.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.199Z [INFO] TestPreparedQuery_Explain/#01.server: New leader elected: payload=Node-1e0400b7-7592-c3e0-158a-9b0b24611375 === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.204Z [INFO] TestPreparedQuery_Create.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0f3cba05-9f1b-9af7-c31f-fec296d04d77 Address:127.0.0.1:29477}]" === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.203Z [DEBUG] TestPreparedQuery_Explain/#01.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29465 === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.206Z [INFO] TestPreparedQuery_Create.server.raft: entering follower state: follower="Node at 127.0.0.1:29477 [Follower]" leader= writer.go:29: 2021-01-29T19:33:04.206Z [INFO] TestPreparedQuery_Create.server.serf.wan: serf: EventMemberJoin: Node-0f3cba05-9f1b-9af7-c31f-fec296d04d77.dc1 127.0.0.1 === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.210Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.211Z [INFO] TestPreparedQuery_Create.server.serf.lan: serf: EventMemberJoin: Node-0f3cba05-9f1b-9af7-c31f-fec296d04d77 127.0.0.1 writer.go:29: 2021-01-29T19:33:04.215Z [INFO] TestPreparedQuery_Create.server: Handled event for server in area: event=member-join server=Node-0f3cba05-9f1b-9af7-c31f-fec296d04d77.dc1 area=wan writer.go:29: 2021-01-29T19:33:04.216Z [INFO] TestPreparedQuery_Create: Started DNS server: address=127.0.0.1:29472 network=tcp writer.go:29: 2021-01-29T19:33:04.217Z [INFO] TestPreparedQuery_Create: Started DNS server: address=127.0.0.1:29472 network=udp writer.go:29: 2021-01-29T19:33:04.218Z [INFO] TestPreparedQuery_Create: Started HTTP server: address=127.0.0.1:29473 network=tcp writer.go:29: 2021-01-29T19:33:04.219Z [INFO] TestPreparedQuery_Create: started state syncer writer.go:29: 2021-01-29T19:33:04.220Z [INFO] TestPreparedQuery_Create.server: Adding LAN server: server="Node-0f3cba05-9f1b-9af7-c31f-fec296d04d77 (Addr: tcp/127.0.0.1:29477) (DC: dc1)" === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.223Z [INFO] TestPreparedQuery_Explain/#01.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:04.225Z [INFO] TestPreparedQuery_Explain/#01.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.227Z [DEBUG] TestPreparedQuery_Explain/#01.server: Skipping self join check for node since the cluster is too small: node=Node-1e0400b7-7592-c3e0-158a-9b0b24611375 writer.go:29: 2021-01-29T19:33:04.228Z [INFO] TestPreparedQuery_Explain/#01.server: member joined, marking health alive: member=Node-1e0400b7-7592-c3e0-158a-9b0b24611375 === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.272Z [WARN] TestPreparedQuery_Create.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:04.274Z [INFO] TestPreparedQuery_Create.server.raft: entering candidate state: node="Node at 127.0.0.1:29477 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:04.290Z [DEBUG] TestPreparedQuery_Create.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:04.292Z [DEBUG] TestPreparedQuery_Create.server.raft: vote granted: from=0f3cba05-9f1b-9af7-c31f-fec296d04d77 term=2 tally=1 writer.go:29: 2021-01-29T19:33:04.293Z [INFO] TestPreparedQuery_Create.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:04.294Z [INFO] TestPreparedQuery_Create.server.raft: entering leader state: leader="Node at 127.0.0.1:29477 [Leader]" writer.go:29: 2021-01-29T19:33:04.296Z [INFO] TestPreparedQuery_Create.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:04.297Z [INFO] TestPreparedQuery_Create.server: New leader elected: payload=Node-0f3cba05-9f1b-9af7-c31f-fec296d04d77 === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.299Z [WARN] TestPreparedQuery_List/#00.server: endpoint injected; this should only be used for testing === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.299Z [DEBUG] TestPreparedQuery_Create.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29477 === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.301Z [INFO] TestPreparedQuery_List/#00: Requesting shutdown writer.go:29: 2021-01-29T19:33:04.302Z [INFO] TestPreparedQuery_List/#00.server: shutting down server writer.go:29: 2021-01-29T19:33:04.303Z [DEBUG] TestPreparedQuery_List/#00.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.304Z [ERROR] TestPreparedQuery_List/#00.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:04.305Z [WARN] TestPreparedQuery_List/#00.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:04.305Z [DEBUG] TestPreparedQuery_List/#00.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.309Z [WARN] TestPreparedQuery_List/#00.server.serf.wan: serf: Shutdown without a Leave === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.310Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.311Z [INFO] TestPreparedQuery_List/#00.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:04.312Z [INFO] TestPreparedQuery_List/#00: consul server down writer.go:29: 2021-01-29T19:33:04.314Z [INFO] TestPreparedQuery_List/#00: shutdown complete writer.go:29: 2021-01-29T19:33:04.315Z [INFO] TestPreparedQuery_List/#00: Stopping server: protocol=DNS address=127.0.0.1:29466 network=tcp writer.go:29: 2021-01-29T19:33:04.316Z [INFO] TestPreparedQuery_List/#00: Stopping server: protocol=DNS address=127.0.0.1:29466 network=udp writer.go:29: 2021-01-29T19:33:04.317Z [INFO] TestPreparedQuery_List/#00: Stopping server: protocol=HTTP address=127.0.0.1:29467 network=tcp === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.319Z [INFO] TestPreparedQuery_Create.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:04.320Z [INFO] TestPreparedQuery_Create.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.321Z [DEBUG] TestPreparedQuery_Create.server: Skipping self join check for node since the cluster is too small: node=Node-0f3cba05-9f1b-9af7-c31f-fec296d04d77 writer.go:29: 2021-01-29T19:33:04.321Z [INFO] TestPreparedQuery_Create.server: member joined, marking health alive: member=Node-0f3cba05-9f1b-9af7-c31f-fec296d04d77 writer.go:29: 2021-01-29T19:33:04.328Z [DEBUG] TestPreparedQuery_Create: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:04.330Z [INFO] TestPreparedQuery_Create: Synced node info writer.go:29: 2021-01-29T19:33:04.331Z [DEBUG] TestPreparedQuery_Create: Node info in sync === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.357Z [WARN] TestPreparedQuery_Execute/#00.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:04.359Z [INFO] TestPreparedQuery_Execute/#00: Requesting shutdown writer.go:29: 2021-01-29T19:33:04.360Z [INFO] TestPreparedQuery_Execute/#00.server: shutting down server writer.go:29: 2021-01-29T19:33:04.361Z [DEBUG] TestPreparedQuery_Execute/#00.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.361Z [WARN] TestPreparedQuery_Execute/#00.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:04.361Z [DEBUG] TestPreparedQuery_Execute/#00.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.363Z [WARN] TestPreparedQuery_Execute/#00.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:04.365Z [INFO] TestPreparedQuery_Execute/#00.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:04.365Z [INFO] TestPreparedQuery_Execute/#00: consul server down === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.369Z [INFO] TestPreparedQuery_Explain/#01: Requesting shutdown === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.370Z [INFO] TestPreparedQuery_Execute/#00: shutdown complete writer.go:29: 2021-01-29T19:33:04.371Z [INFO] TestPreparedQuery_Execute/#00: Stopping server: protocol=DNS address=127.0.0.1:29454 network=tcp === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.371Z [INFO] TestPreparedQuery_Explain/#01.server: shutting down server === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.372Z [INFO] TestPreparedQuery_Execute/#00: Stopping server: protocol=DNS address=127.0.0.1:29454 network=udp === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.372Z [DEBUG] TestPreparedQuery_Explain/#01.leader: stopping routine: routine="CA root pruning" === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.373Z [INFO] TestPreparedQuery_Execute/#00: Stopping server: protocol=HTTP address=127.0.0.1:29455 network=tcp === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.374Z [WARN] TestPreparedQuery_Explain/#01.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:04.372Z [ERROR] TestPreparedQuery_Explain/#01.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:04.374Z [DEBUG] TestPreparedQuery_Explain/#01.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.376Z [WARN] TestPreparedQuery_Explain/#01.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:04.382Z [INFO] TestPreparedQuery_Explain/#01.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:04.382Z [INFO] TestPreparedQuery_Explain/#01: consul server down writer.go:29: 2021-01-29T19:33:04.384Z [INFO] TestPreparedQuery_Explain/#01: shutdown complete writer.go:29: 2021-01-29T19:33:04.384Z [INFO] TestPreparedQuery_Explain/#01: Stopping server: protocol=DNS address=127.0.0.1:29460 network=tcp writer.go:29: 2021-01-29T19:33:04.385Z [INFO] TestPreparedQuery_Explain/#01: Stopping server: protocol=DNS address=127.0.0.1:29460 network=udp writer.go:29: 2021-01-29T19:33:04.386Z [INFO] TestPreparedQuery_Explain/#01: Stopping server: protocol=HTTP address=127.0.0.1:29461 network=tcp === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:04.555Z [WARN] TestPreparedQuery_Create.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:04.557Z [INFO] TestPreparedQuery_Create: Requesting shutdown writer.go:29: 2021-01-29T19:33:04.558Z [INFO] TestPreparedQuery_Create.server: shutting down server writer.go:29: 2021-01-29T19:33:04.559Z [DEBUG] TestPreparedQuery_Create.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.560Z [WARN] TestPreparedQuery_Create.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:04.560Z [DEBUG] TestPreparedQuery_Create.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.562Z [WARN] TestPreparedQuery_Create.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:04.565Z [INFO] TestPreparedQuery_Create.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:04.565Z [INFO] TestPreparedQuery_Create: consul server down writer.go:29: 2021-01-29T19:33:04.567Z [INFO] TestPreparedQuery_Create: shutdown complete writer.go:29: 2021-01-29T19:33:04.567Z [INFO] TestPreparedQuery_Create: Stopping server: protocol=DNS address=127.0.0.1:29472 network=tcp writer.go:29: 2021-01-29T19:33:04.568Z [INFO] TestPreparedQuery_Create: Stopping server: protocol=DNS address=127.0.0.1:29472 network=udp writer.go:29: 2021-01-29T19:33:04.569Z [INFO] TestPreparedQuery_Create: Stopping server: protocol=HTTP address=127.0.0.1:29473 network=tcp === CONT TestPreparedQuery_List/#00 writer.go:29: 2021-01-29T19:33:04.818Z [INFO] TestPreparedQuery_List/#00: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:04.820Z [INFO] TestPreparedQuery_List/#00: Endpoints down === RUN TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.835Z [WARN] TestPreparedQuery_List/#01: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:04.837Z [DEBUG] TestPreparedQuery_List/#01.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:04.840Z [DEBUG] TestPreparedQuery_List/#01.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:04.848Z [INFO] TestPreparedQuery_List/#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f1f7b02b-0fb0-5ab7-95e1-1eeaf2649bd0 Address:127.0.0.1:29483}]" writer.go:29: 2021-01-29T19:33:04.850Z [INFO] TestPreparedQuery_List/#01.server.raft: entering follower state: follower="Node at 127.0.0.1:29483 [Follower]" leader= writer.go:29: 2021-01-29T19:33:04.851Z [INFO] TestPreparedQuery_List/#01.server.serf.wan: serf: EventMemberJoin: Node-f1f7b02b-0fb0-5ab7-95e1-1eeaf2649bd0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:04.856Z [INFO] TestPreparedQuery_List/#01.server.serf.lan: serf: EventMemberJoin: Node-f1f7b02b-0fb0-5ab7-95e1-1eeaf2649bd0 127.0.0.1 writer.go:29: 2021-01-29T19:33:04.860Z [INFO] TestPreparedQuery_List/#01.server: Adding LAN server: server="Node-f1f7b02b-0fb0-5ab7-95e1-1eeaf2649bd0 (Addr: tcp/127.0.0.1:29483) (DC: dc1)" writer.go:29: 2021-01-29T19:33:04.861Z [INFO] TestPreparedQuery_List/#01.server: Handled event for server in area: event=member-join server=Node-f1f7b02b-0fb0-5ab7-95e1-1eeaf2649bd0.dc1 area=wan writer.go:29: 2021-01-29T19:33:04.862Z [INFO] TestPreparedQuery_List/#01: Started DNS server: address=127.0.0.1:29478 network=tcp writer.go:29: 2021-01-29T19:33:04.863Z [INFO] TestPreparedQuery_List/#01: Started DNS server: address=127.0.0.1:29478 network=udp writer.go:29: 2021-01-29T19:33:04.865Z [INFO] TestPreparedQuery_List/#01: Started HTTP server: address=127.0.0.1:29479 network=tcp writer.go:29: 2021-01-29T19:33:04.870Z [INFO] TestPreparedQuery_List/#01: started state syncer === CONT TestPreparedQuery_Execute/#00 writer.go:29: 2021-01-29T19:33:04.874Z [INFO] TestPreparedQuery_Execute/#00: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:04.875Z [INFO] TestPreparedQuery_Execute/#00: Endpoints down === RUN TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.887Z [WARN] TestPreparedQuery_Execute/#01: bootstrap = true: do not enable unless necessary === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.887Z [INFO] TestPreparedQuery_Explain/#01: Waiting for endpoints to shut down === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.888Z [DEBUG] TestPreparedQuery_Execute/#01.tlsutil: Update: version=1 === CONT TestPreparedQuery_Explain/#01 writer.go:29: 2021-01-29T19:33:04.888Z [INFO] TestPreparedQuery_Explain/#01: Endpoints down === RUN TestPreparedQuery_Explain/#02 === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.890Z [DEBUG] TestPreparedQuery_Execute/#01.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:04.895Z [INFO] TestPreparedQuery_Execute/#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:36851e59-1fab-93ab-b203-01a067c2deca Address:127.0.0.1:29489}]" writer.go:29: 2021-01-29T19:33:04.897Z [INFO] TestPreparedQuery_Execute/#01.server.serf.wan: serf: EventMemberJoin: Node-36851e59-1fab-93ab-b203-01a067c2deca.dc1 127.0.0.1 === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.901Z [WARN] TestPreparedQuery_Explain/#02: bootstrap = true: do not enable unless necessary === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.898Z [INFO] TestPreparedQuery_Execute/#01.server.raft: entering follower state: follower="Node at 127.0.0.1:29489 [Follower]" leader= === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.903Z [DEBUG] TestPreparedQuery_Explain/#02.tlsutil: Update: version=1 === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.900Z [INFO] TestPreparedQuery_Execute/#01.server.serf.lan: serf: EventMemberJoin: Node-36851e59-1fab-93ab-b203-01a067c2deca 127.0.0.1 === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.905Z [DEBUG] TestPreparedQuery_Explain/#02.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.908Z [INFO] TestPreparedQuery_Execute/#01.server: Adding LAN server: server="Node-36851e59-1fab-93ab-b203-01a067c2deca (Addr: tcp/127.0.0.1:29489) (DC: dc1)" writer.go:29: 2021-01-29T19:33:04.910Z [INFO] TestPreparedQuery_Execute/#01.server: Handled event for server in area: event=member-join server=Node-36851e59-1fab-93ab-b203-01a067c2deca.dc1 area=wan === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.910Z [INFO] TestPreparedQuery_Explain/#02.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:08ea097e-50bb-da2e-8d6c-c523ae052b81 Address:127.0.0.1:29495}]" === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.913Z [INFO] TestPreparedQuery_Execute/#01: Started DNS server: address=127.0.0.1:29484 network=tcp === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.913Z [INFO] TestPreparedQuery_Explain/#02.server.raft: entering follower state: follower="Node at 127.0.0.1:29495 [Follower]" leader= writer.go:29: 2021-01-29T19:33:04.914Z [INFO] TestPreparedQuery_Explain/#02.server.serf.wan: serf: EventMemberJoin: Node-08ea097e-50bb-da2e-8d6c-c523ae052b81.dc1 127.0.0.1 === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.915Z [WARN] TestPreparedQuery_List/#01.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:04.917Z [INFO] TestPreparedQuery_List/#01.server.raft: entering candidate state: node="Node at 127.0.0.1:29483 [Candidate]" term=2 === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.919Z [INFO] TestPreparedQuery_Explain/#02.server.serf.lan: serf: EventMemberJoin: Node-08ea097e-50bb-da2e-8d6c-c523ae052b81 127.0.0.1 === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.920Z [INFO] TestPreparedQuery_Execute/#01: Started DNS server: address=127.0.0.1:29484 network=udp === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.924Z [INFO] TestPreparedQuery_Explain/#02.server: Adding LAN server: server="Node-08ea097e-50bb-da2e-8d6c-c523ae052b81 (Addr: tcp/127.0.0.1:29495) (DC: dc1)" === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.922Z [DEBUG] TestPreparedQuery_List/#01.server.raft: votes: needed=1 === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.925Z [INFO] TestPreparedQuery_Explain/#02.server: Handled event for server in area: event=member-join server=Node-08ea097e-50bb-da2e-8d6c-c523ae052b81.dc1 area=wan === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.925Z [DEBUG] TestPreparedQuery_List/#01.server.raft: vote granted: from=f1f7b02b-0fb0-5ab7-95e1-1eeaf2649bd0 term=2 tally=1 === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.928Z [INFO] TestPreparedQuery_Explain/#02: Started DNS server: address=127.0.0.1:29490 network=tcp === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.926Z [INFO] TestPreparedQuery_Execute/#01: Started HTTP server: address=127.0.0.1:29485 network=tcp === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.930Z [INFO] TestPreparedQuery_Explain/#02: Started DNS server: address=127.0.0.1:29490 network=udp writer.go:29: 2021-01-29T19:33:04.932Z [INFO] TestPreparedQuery_Explain/#02: Started HTTP server: address=127.0.0.1:29491 network=tcp === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.929Z [INFO] TestPreparedQuery_List/#01.server.raft: election won: tally=1 === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.933Z [INFO] TestPreparedQuery_Explain/#02: started state syncer === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.930Z [INFO] TestPreparedQuery_Execute/#01: started state syncer === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.933Z [INFO] TestPreparedQuery_List/#01.server.raft: entering leader state: leader="Node at 127.0.0.1:29483 [Leader]" writer.go:29: 2021-01-29T19:33:04.937Z [INFO] TestPreparedQuery_List/#01.server: cluster leadership acquired === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.940Z [WARN] TestPreparedQuery_Execute/#01.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:04.942Z [INFO] TestPreparedQuery_Execute/#01.server.raft: entering candidate state: node="Node at 127.0.0.1:29489 [Candidate]" term=2 === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.938Z [INFO] TestPreparedQuery_List/#01.server: New leader elected: payload=Node-f1f7b02b-0fb0-5ab7-95e1-1eeaf2649bd0 === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.944Z [DEBUG] TestPreparedQuery_Execute/#01.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:04.945Z [DEBUG] TestPreparedQuery_Execute/#01.server.raft: vote granted: from=36851e59-1fab-93ab-b203-01a067c2deca term=2 tally=1 writer.go:29: 2021-01-29T19:33:04.946Z [INFO] TestPreparedQuery_Execute/#01.server.raft: election won: tally=1 === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.939Z [DEBUG] TestPreparedQuery_List/#01.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29483 === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.948Z [INFO] TestPreparedQuery_Execute/#01.server.raft: entering leader state: leader="Node at 127.0.0.1:29489 [Leader]" writer.go:29: 2021-01-29T19:33:04.949Z [INFO] TestPreparedQuery_Execute/#01.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:04.951Z [INFO] TestPreparedQuery_Execute/#01.server: New leader elected: payload=Node-36851e59-1fab-93ab-b203-01a067c2deca === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.953Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.955Z [WARN] TestPreparedQuery_Explain/#02.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:04.956Z [INFO] TestPreparedQuery_Explain/#02.server.raft: entering candidate state: node="Node at 127.0.0.1:29495 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:04.958Z [DEBUG] TestPreparedQuery_Explain/#02.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:04.958Z [DEBUG] TestPreparedQuery_Explain/#02.server.raft: vote granted: from=08ea097e-50bb-da2e-8d6c-c523ae052b81 term=2 tally=1 writer.go:29: 2021-01-29T19:33:04.959Z [INFO] TestPreparedQuery_Explain/#02.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:04.960Z [INFO] TestPreparedQuery_Explain/#02.server.raft: entering leader state: leader="Node at 127.0.0.1:29495 [Leader]" writer.go:29: 2021-01-29T19:33:04.961Z [INFO] TestPreparedQuery_Explain/#02.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:04.961Z [INFO] TestPreparedQuery_Explain/#02.server: New leader elected: payload=Node-08ea097e-50bb-da2e-8d6c-c523ae052b81 === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.953Z [DEBUG] TestPreparedQuery_Execute/#01.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29489 === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.963Z [DEBUG] TestPreparedQuery_Explain/#02.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29495 === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.965Z [INFO] TestPreparedQuery_List/#01.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.968Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.967Z [INFO] TestPreparedQuery_List/#01.leader: started routine: routine="CA root pruning" === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.969Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.970Z [DEBUG] TestPreparedQuery_List/#01.server: Skipping self join check for node since the cluster is too small: node=Node-f1f7b02b-0fb0-5ab7-95e1-1eeaf2649bd0 writer.go:29: 2021-01-29T19:33:04.972Z [INFO] TestPreparedQuery_List/#01.server: member joined, marking health alive: member=Node-f1f7b02b-0fb0-5ab7-95e1-1eeaf2649bd0 === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.975Z [INFO] TestPreparedQuery_Explain/#02.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:04.976Z [INFO] TestPreparedQuery_Explain/#02.leader: started routine: routine="CA root pruning" === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:04.980Z [DEBUG] TestPreparedQuery_List/#01: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:04.983Z [INFO] TestPreparedQuery_List/#01: Synced node info writer.go:29: 2021-01-29T19:33:04.984Z [DEBUG] TestPreparedQuery_List/#01: Node info in sync === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:04.987Z [INFO] TestPreparedQuery_Execute/#01.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:04.988Z [INFO] TestPreparedQuery_Execute/#01.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.989Z [DEBUG] TestPreparedQuery_Execute/#01.server: Skipping self join check for node since the cluster is too small: node=Node-36851e59-1fab-93ab-b203-01a067c2deca writer.go:29: 2021-01-29T19:33:04.991Z [INFO] TestPreparedQuery_Execute/#01.server: member joined, marking health alive: member=Node-36851e59-1fab-93ab-b203-01a067c2deca === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:04.977Z [DEBUG] TestPreparedQuery_Explain/#02.server: Skipping self join check for node since the cluster is too small: node=Node-08ea097e-50bb-da2e-8d6c-c523ae052b81 writer.go:29: 2021-01-29T19:33:04.993Z [INFO] TestPreparedQuery_Explain/#02.server: member joined, marking health alive: member=Node-08ea097e-50bb-da2e-8d6c-c523ae052b81 === CONT TestPreparedQuery_Create writer.go:29: 2021-01-29T19:33:05.078Z [INFO] TestPreparedQuery_Create: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:05.079Z [INFO] TestPreparedQuery_Create: Endpoints down --- PASS: TestPreparedQuery_Create (0.90s) === CONT TestOperator_ServerHealth_Unhealthy === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:05.142Z [DEBUG] TestPreparedQuery_List/#01: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:05.144Z [DEBUG] TestPreparedQuery_List/#01: Node info in sync === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:05.142Z [WARN] TestPreparedQuery_Execute/#01.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:05.146Z [INFO] TestPreparedQuery_Execute/#01: Requesting shutdown writer.go:29: 2021-01-29T19:33:05.150Z [INFO] TestPreparedQuery_Execute/#01.server: shutting down server writer.go:29: 2021-01-29T19:33:05.151Z [DEBUG] TestPreparedQuery_Execute/#01.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.152Z [WARN] TestPreparedQuery_Execute/#01.server.serf.lan: serf: Shutdown without a Leave === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:05.149Z [WARN] TestPreparedQuery_List/#01.server: endpoint injected; this should only be used for testing === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:05.152Z [WARN] TestOperator_ServerHealth_Unhealthy: bootstrap = true: do not enable unless necessary === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:05.153Z [INFO] TestPreparedQuery_List/#01: Requesting shutdown writer.go:29: 2021-01-29T19:33:05.154Z [INFO] TestPreparedQuery_List/#01.server: shutting down server writer.go:29: 2021-01-29T19:33:05.155Z [DEBUG] TestPreparedQuery_List/#01.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.156Z [WARN] TestPreparedQuery_List/#01.server.serf.lan: serf: Shutdown without a Leave === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:05.157Z [ERROR] TestPreparedQuery_Execute/#01.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:05.158Z [DEBUG] TestPreparedQuery_Execute/#01.leader: stopped routine: routine="CA root pruning" === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:05.155Z [DEBUG] TestOperator_ServerHealth_Unhealthy.tlsutil: Update: version=1 === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:05.160Z [DEBUG] TestPreparedQuery_List/#01.leader: stopped routine: routine="CA root pruning" === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:05.166Z [WARN] TestPreparedQuery_Execute/#01.server.serf.wan: serf: Shutdown without a Leave === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:05.162Z [WARN] TestPreparedQuery_List/#01.server.serf.wan: serf: Shutdown without a Leave === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:05.172Z [INFO] TestPreparedQuery_Execute/#01.server.router.manager: shutting down === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:05.173Z [DEBUG] TestOperator_ServerHealth_Unhealthy.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:05.174Z [WARN] TestPreparedQuery_Explain/#02.server: endpoint injected; this should only be used for testing === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:05.175Z [INFO] TestPreparedQuery_List/#01.server.router.manager: shutting down === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:05.178Z [INFO] TestPreparedQuery_Explain/#02: Requesting shutdown === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:05.178Z [INFO] TestPreparedQuery_List/#01: consul server down writer.go:29: 2021-01-29T19:33:05.180Z [INFO] TestPreparedQuery_List/#01: shutdown complete writer.go:29: 2021-01-29T19:33:05.182Z [INFO] TestPreparedQuery_List/#01: Stopping server: protocol=DNS address=127.0.0.1:29478 network=tcp writer.go:29: 2021-01-29T19:33:05.184Z [INFO] TestPreparedQuery_List/#01: Stopping server: protocol=DNS address=127.0.0.1:29478 network=udp === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:05.188Z [INFO] TestPreparedQuery_Execute/#01: consul server down === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:05.179Z [INFO] TestPreparedQuery_Explain/#02.server: shutting down server === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:05.187Z [INFO] TestPreparedQuery_List/#01: Stopping server: protocol=HTTP address=127.0.0.1:29479 network=tcp === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:05.189Z [INFO] TestPreparedQuery_Execute/#01: shutdown complete writer.go:29: 2021-01-29T19:33:05.198Z [INFO] TestPreparedQuery_Execute/#01: Stopping server: protocol=DNS address=127.0.0.1:29484 network=tcp writer.go:29: 2021-01-29T19:33:05.200Z [INFO] TestPreparedQuery_Execute/#01: Stopping server: protocol=DNS address=127.0.0.1:29484 network=udp writer.go:29: 2021-01-29T19:33:05.202Z [INFO] TestPreparedQuery_Execute/#01: Stopping server: protocol=HTTP address=127.0.0.1:29485 network=tcp === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:05.195Z [DEBUG] TestPreparedQuery_Explain/#02.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.224Z [WARN] TestPreparedQuery_Explain/#02.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:05.206Z [ERROR] TestPreparedQuery_Explain/#02.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:05.252Z [DEBUG] TestPreparedQuery_Explain/#02.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.263Z [WARN] TestPreparedQuery_Explain/#02.server.serf.wan: serf: Shutdown without a Leave === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:05.299Z [INFO] TestOperator_ServerHealth_Unhealthy.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:12600cd9-d507-2010-bead-d7f7ec45388a Address:127.0.0.1:29507}]" writer.go:29: 2021-01-29T19:33:05.312Z [INFO] TestOperator_ServerHealth_Unhealthy.server.raft: entering follower state: follower="Node at 127.0.0.1:29507 [Follower]" leader= === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:05.313Z [INFO] TestPreparedQuery_Explain/#02.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:05.314Z [INFO] TestPreparedQuery_Explain/#02: consul server down writer.go:29: 2021-01-29T19:33:05.315Z [INFO] TestPreparedQuery_Explain/#02: shutdown complete writer.go:29: 2021-01-29T19:33:05.317Z [INFO] TestPreparedQuery_Explain/#02: Stopping server: protocol=DNS address=127.0.0.1:29490 network=tcp writer.go:29: 2021-01-29T19:33:05.318Z [INFO] TestPreparedQuery_Explain/#02: Stopping server: protocol=DNS address=127.0.0.1:29490 network=udp writer.go:29: 2021-01-29T19:33:05.320Z [INFO] TestPreparedQuery_Explain/#02: Stopping server: protocol=HTTP address=127.0.0.1:29491 network=tcp === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:05.340Z [INFO] TestOperator_ServerHealth_Unhealthy.server.serf.wan: serf: EventMemberJoin: Node-12600cd9-d507-2010-bead-d7f7ec45388a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:05.344Z [INFO] TestOperator_ServerHealth_Unhealthy.server.serf.lan: serf: EventMemberJoin: Node-12600cd9-d507-2010-bead-d7f7ec45388a 127.0.0.1 writer.go:29: 2021-01-29T19:33:05.350Z [INFO] TestOperator_ServerHealth_Unhealthy: Started DNS server: address=127.0.0.1:29502 network=tcp writer.go:29: 2021-01-29T19:33:05.352Z [INFO] TestOperator_ServerHealth_Unhealthy: Started DNS server: address=127.0.0.1:29502 network=udp writer.go:29: 2021-01-29T19:33:05.350Z [INFO] TestOperator_ServerHealth_Unhealthy.server: Handled event for server in area: event=member-join server=Node-12600cd9-d507-2010-bead-d7f7ec45388a.dc1 area=wan writer.go:29: 2021-01-29T19:33:05.351Z [INFO] TestOperator_ServerHealth_Unhealthy.server: Adding LAN server: server="Node-12600cd9-d507-2010-bead-d7f7ec45388a (Addr: tcp/127.0.0.1:29507) (DC: dc1)" writer.go:29: 2021-01-29T19:33:05.355Z [INFO] TestOperator_ServerHealth_Unhealthy: Started HTTP server: address=127.0.0.1:29503 network=tcp writer.go:29: 2021-01-29T19:33:05.356Z [INFO] TestOperator_ServerHealth_Unhealthy: started state syncer writer.go:29: 2021-01-29T19:33:05.359Z [WARN] TestOperator_ServerHealth_Unhealthy.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:05.360Z [INFO] TestOperator_ServerHealth_Unhealthy.server.raft: entering candidate state: node="Node at 127.0.0.1:29507 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:05.362Z [DEBUG] TestOperator_ServerHealth_Unhealthy.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:05.363Z [DEBUG] TestOperator_ServerHealth_Unhealthy.server.raft: vote granted: from=12600cd9-d507-2010-bead-d7f7ec45388a term=2 tally=1 writer.go:29: 2021-01-29T19:33:05.369Z [INFO] TestOperator_ServerHealth_Unhealthy.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:05.370Z [INFO] TestOperator_ServerHealth_Unhealthy.server.raft: entering leader state: leader="Node at 127.0.0.1:29507 [Leader]" writer.go:29: 2021-01-29T19:33:05.371Z [INFO] TestOperator_ServerHealth_Unhealthy.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:05.372Z [INFO] TestOperator_ServerHealth_Unhealthy.server: New leader elected: payload=Node-12600cd9-d507-2010-bead-d7f7ec45388a writer.go:29: 2021-01-29T19:33:05.381Z [DEBUG] TestOperator_ServerHealth_Unhealthy.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29507 writer.go:29: 2021-01-29T19:33:05.388Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:05.393Z [INFO] TestOperator_ServerHealth_Unhealthy.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:05.394Z [INFO] TestOperator_ServerHealth_Unhealthy.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.395Z [DEBUG] TestOperator_ServerHealth_Unhealthy.server: Skipping self join check for node since the cluster is too small: node=Node-12600cd9-d507-2010-bead-d7f7ec45388a writer.go:29: 2021-01-29T19:33:05.396Z [INFO] TestOperator_ServerHealth_Unhealthy.server: member joined, marking health alive: member=Node-12600cd9-d507-2010-bead-d7f7ec45388a writer.go:29: 2021-01-29T19:33:05.675Z [DEBUG] TestOperator_ServerHealth_Unhealthy: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:05.677Z [INFO] TestOperator_ServerHealth_Unhealthy: Synced node info writer.go:29: 2021-01-29T19:33:05.678Z [DEBUG] TestOperator_ServerHealth_Unhealthy: Node info in sync === CONT TestPreparedQuery_Execute/#01 writer.go:29: 2021-01-29T19:33:05.704Z [INFO] TestPreparedQuery_Execute/#01: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:05.705Z [INFO] TestPreparedQuery_Execute/#01: Endpoints down === RUN TestPreparedQuery_Execute/#02 === CONT TestPreparedQuery_List/#01 writer.go:29: 2021-01-29T19:33:05.711Z [INFO] TestPreparedQuery_List/#01: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:05.712Z [INFO] TestPreparedQuery_List/#01: Endpoints down --- PASS: TestPreparedQuery_List (1.72s) --- PASS: TestPreparedQuery_List/#00 (0.82s) --- PASS: TestPreparedQuery_List/#01 (0.89s) === CONT TestOperator_ServerHealth === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:05.733Z [WARN] TestPreparedQuery_Execute/#02: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:05.734Z [DEBUG] TestPreparedQuery_Execute/#02.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:05.735Z [DEBUG] TestPreparedQuery_Execute/#02.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:05.738Z [INFO] TestPreparedQuery_Execute/#02.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9f597ec9-868d-4e53-4fe6-7a2b8924acba Address:127.0.0.1:29501}]" writer.go:29: 2021-01-29T19:33:05.740Z [INFO] TestPreparedQuery_Execute/#02.server.raft: entering follower state: follower="Node at 127.0.0.1:29501 [Follower]" leader= === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:05.741Z [WARN] TestOperator_ServerHealth: bootstrap = true: do not enable unless necessary === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:05.740Z [INFO] TestPreparedQuery_Execute/#02.server.serf.wan: serf: EventMemberJoin: Node-9f597ec9-868d-4e53-4fe6-7a2b8924acba.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:05.744Z [INFO] TestPreparedQuery_Execute/#02.server.serf.lan: serf: EventMemberJoin: Node-9f597ec9-868d-4e53-4fe6-7a2b8924acba 127.0.0.1 writer.go:29: 2021-01-29T19:33:05.747Z [INFO] TestPreparedQuery_Execute/#02.server: Adding LAN server: server="Node-9f597ec9-868d-4e53-4fe6-7a2b8924acba (Addr: tcp/127.0.0.1:29501) (DC: dc1)" writer.go:29: 2021-01-29T19:33:05.747Z [INFO] TestPreparedQuery_Execute/#02: Started DNS server: address=127.0.0.1:29496 network=udp writer.go:29: 2021-01-29T19:33:05.749Z [INFO] TestPreparedQuery_Execute/#02: Started DNS server: address=127.0.0.1:29496 network=tcp === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:05.745Z [DEBUG] TestOperator_ServerHealth.tlsutil: Update: version=1 === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:05.751Z [INFO] TestPreparedQuery_Execute/#02: Started HTTP server: address=127.0.0.1:29497 network=tcp writer.go:29: 2021-01-29T19:33:05.752Z [INFO] TestPreparedQuery_Execute/#02: started state syncer === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:05.753Z [DEBUG] TestOperator_ServerHealth.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:05.747Z [INFO] TestPreparedQuery_Execute/#02.server: Handled event for server in area: event=member-join server=Node-9f597ec9-868d-4e53-4fe6-7a2b8924acba.dc1 area=wan === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:05.764Z [INFO] TestOperator_ServerHealth.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9f3012b1-e886-cf73-45fc-9e34e0d62744 Address:127.0.0.1:29513}]" writer.go:29: 2021-01-29T19:33:05.765Z [INFO] TestOperator_ServerHealth.server.serf.wan: serf: EventMemberJoin: Node-9f3012b1-e886-cf73-45fc-9e34e0d62744.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:05.765Z [INFO] TestOperator_ServerHealth.server.raft: entering follower state: follower="Node at 127.0.0.1:29513 [Follower]" leader= writer.go:29: 2021-01-29T19:33:05.768Z [INFO] TestOperator_ServerHealth.server.serf.lan: serf: EventMemberJoin: Node-9f3012b1-e886-cf73-45fc-9e34e0d62744 127.0.0.1 writer.go:29: 2021-01-29T19:33:05.771Z [INFO] TestOperator_ServerHealth.server: Handled event for server in area: event=member-join server=Node-9f3012b1-e886-cf73-45fc-9e34e0d62744.dc1 area=wan writer.go:29: 2021-01-29T19:33:05.771Z [INFO] TestOperator_ServerHealth.server: Adding LAN server: server="Node-9f3012b1-e886-cf73-45fc-9e34e0d62744 (Addr: tcp/127.0.0.1:29513) (DC: dc1)" writer.go:29: 2021-01-29T19:33:05.774Z [INFO] TestOperator_ServerHealth: Started DNS server: address=127.0.0.1:29508 network=udp writer.go:29: 2021-01-29T19:33:05.775Z [INFO] TestOperator_ServerHealth: Started DNS server: address=127.0.0.1:29508 network=tcp writer.go:29: 2021-01-29T19:33:05.777Z [INFO] TestOperator_ServerHealth: Started HTTP server: address=127.0.0.1:29509 network=tcp writer.go:29: 2021-01-29T19:33:05.778Z [INFO] TestOperator_ServerHealth: started state syncer === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:05.794Z [WARN] TestPreparedQuery_Execute/#02.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:05.796Z [INFO] TestPreparedQuery_Execute/#02.server.raft: entering candidate state: node="Node at 127.0.0.1:29501 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:05.799Z [DEBUG] TestPreparedQuery_Execute/#02.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:05.800Z [DEBUG] TestPreparedQuery_Execute/#02.server.raft: vote granted: from=9f597ec9-868d-4e53-4fe6-7a2b8924acba term=2 tally=1 writer.go:29: 2021-01-29T19:33:05.801Z [INFO] TestPreparedQuery_Execute/#02.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:05.802Z [INFO] TestPreparedQuery_Execute/#02.server.raft: entering leader state: leader="Node at 127.0.0.1:29501 [Leader]" writer.go:29: 2021-01-29T19:33:05.803Z [INFO] TestPreparedQuery_Execute/#02.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:05.804Z [INFO] TestPreparedQuery_Execute/#02.server: New leader elected: payload=Node-9f597ec9-868d-4e53-4fe6-7a2b8924acba writer.go:29: 2021-01-29T19:33:05.805Z [DEBUG] TestPreparedQuery_Execute/#02.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29501 === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:05.809Z [WARN] TestOperator_ServerHealth.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:05.809Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:05.810Z [INFO] TestOperator_ServerHealth.server.raft: entering candidate state: node="Node at 127.0.0.1:29513 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:05.813Z [DEBUG] TestOperator_ServerHealth.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:05.814Z [DEBUG] TestOperator_ServerHealth.server.raft: vote granted: from=9f3012b1-e886-cf73-45fc-9e34e0d62744 term=2 tally=1 writer.go:29: 2021-01-29T19:33:05.815Z [INFO] TestOperator_ServerHealth.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:05.815Z [INFO] TestOperator_ServerHealth.server.raft: entering leader state: leader="Node at 127.0.0.1:29513 [Leader]" === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:05.817Z [INFO] TestPreparedQuery_Execute/#02.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:05.818Z [INFO] TestOperator_ServerHealth.server: cluster leadership acquired === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:05.818Z [INFO] TestPreparedQuery_Execute/#02.leader: started routine: routine="CA root pruning" === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:05.819Z [INFO] TestOperator_ServerHealth.server: New leader elected: payload=Node-9f3012b1-e886-cf73-45fc-9e34e0d62744 === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:05.819Z [DEBUG] TestPreparedQuery_Execute/#02.server: Skipping self join check for node since the cluster is too small: node=Node-9f597ec9-868d-4e53-4fe6-7a2b8924acba writer.go:29: 2021-01-29T19:33:05.820Z [INFO] TestPreparedQuery_Execute/#02.server: member joined, marking health alive: member=Node-9f597ec9-868d-4e53-4fe6-7a2b8924acba === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:05.821Z [DEBUG] TestOperator_ServerHealth.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29513 writer.go:29: 2021-01-29T19:33:05.826Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPreparedQuery_Explain/#02 writer.go:29: 2021-01-29T19:33:05.828Z [INFO] TestPreparedQuery_Explain/#02: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:05.829Z [INFO] TestPreparedQuery_Explain/#02: Endpoints down --- PASS: TestPreparedQuery_Explain (2.65s) --- PASS: TestPreparedQuery_Explain/#00 (0.93s) --- PASS: TestPreparedQuery_Explain/#01 (0.78s) --- PASS: TestPreparedQuery_Explain/#02 (0.94s) === CONT TestOperator_AutopilotCASConfiguration === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:05.835Z [INFO] TestOperator_ServerHealth.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:05.836Z [INFO] TestOperator_ServerHealth.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.838Z [DEBUG] TestOperator_ServerHealth.server: Skipping self join check for node since the cluster is too small: node=Node-9f3012b1-e886-cf73-45fc-9e34e0d62744 writer.go:29: 2021-01-29T19:33:05.840Z [INFO] TestOperator_ServerHealth.server: member joined, marking health alive: member=Node-9f3012b1-e886-cf73-45fc-9e34e0d62744 === CONT TestOperator_AutopilotCASConfiguration writer.go:29: 2021-01-29T19:33:05.845Z [WARN] TestOperator_AutopilotCASConfiguration: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:05.846Z [DEBUG] TestOperator_AutopilotCASConfiguration.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:05.847Z [DEBUG] TestOperator_AutopilotCASConfiguration.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:05.852Z [INFO] TestOperator_AutopilotCASConfiguration.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7a53a528-8c2e-3a91-6dc9-3216be34f50b Address:127.0.0.1:29519}]" writer.go:29: 2021-01-29T19:33:05.854Z [INFO] TestOperator_AutopilotCASConfiguration.server.raft: entering follower state: follower="Node at 127.0.0.1:29519 [Follower]" leader= writer.go:29: 2021-01-29T19:33:05.858Z [INFO] TestOperator_AutopilotCASConfiguration.server.serf.wan: serf: EventMemberJoin: Node-7a53a528-8c2e-3a91-6dc9-3216be34f50b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:05.861Z [INFO] TestOperator_AutopilotCASConfiguration.server.serf.lan: serf: EventMemberJoin: Node-7a53a528-8c2e-3a91-6dc9-3216be34f50b 127.0.0.1 writer.go:29: 2021-01-29T19:33:05.864Z [INFO] TestOperator_AutopilotCASConfiguration.server: Adding LAN server: server="Node-7a53a528-8c2e-3a91-6dc9-3216be34f50b (Addr: tcp/127.0.0.1:29519) (DC: dc1)" writer.go:29: 2021-01-29T19:33:05.865Z [INFO] TestOperator_AutopilotCASConfiguration.server: Handled event for server in area: event=member-join server=Node-7a53a528-8c2e-3a91-6dc9-3216be34f50b.dc1 area=wan writer.go:29: 2021-01-29T19:33:05.867Z [INFO] TestOperator_AutopilotCASConfiguration: Started DNS server: address=127.0.0.1:29514 network=udp writer.go:29: 2021-01-29T19:33:05.869Z [INFO] TestOperator_AutopilotCASConfiguration: Started DNS server: address=127.0.0.1:29514 network=tcp writer.go:29: 2021-01-29T19:33:05.873Z [INFO] TestOperator_AutopilotCASConfiguration: Started HTTP server: address=127.0.0.1:29515 network=tcp writer.go:29: 2021-01-29T19:33:05.873Z [INFO] TestOperator_AutopilotCASConfiguration: started state syncer writer.go:29: 2021-01-29T19:33:05.896Z [WARN] TestOperator_AutopilotCASConfiguration.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:05.898Z [INFO] TestOperator_AutopilotCASConfiguration.server.raft: entering candidate state: node="Node at 127.0.0.1:29519 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:05.910Z [DEBUG] TestOperator_AutopilotCASConfiguration.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:05.912Z [DEBUG] TestOperator_AutopilotCASConfiguration.server.raft: vote granted: from=7a53a528-8c2e-3a91-6dc9-3216be34f50b term=2 tally=1 writer.go:29: 2021-01-29T19:33:05.914Z [INFO] TestOperator_AutopilotCASConfiguration.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:05.915Z [INFO] TestOperator_AutopilotCASConfiguration.server.raft: entering leader state: leader="Node at 127.0.0.1:29519 [Leader]" writer.go:29: 2021-01-29T19:33:05.916Z [INFO] TestOperator_AutopilotCASConfiguration.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:05.917Z [INFO] TestOperator_AutopilotCASConfiguration.server: New leader elected: payload=Node-7a53a528-8c2e-3a91-6dc9-3216be34f50b writer.go:29: 2021-01-29T19:33:05.921Z [DEBUG] TestOperator_AutopilotCASConfiguration.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29519 writer.go:29: 2021-01-29T19:33:05.928Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:05.933Z [INFO] TestOperator_AutopilotCASConfiguration.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:05.934Z [INFO] TestOperator_AutopilotCASConfiguration.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.935Z [DEBUG] TestOperator_AutopilotCASConfiguration.server: Skipping self join check for node since the cluster is too small: node=Node-7a53a528-8c2e-3a91-6dc9-3216be34f50b writer.go:29: 2021-01-29T19:33:05.936Z [INFO] TestOperator_AutopilotCASConfiguration.server: member joined, marking health alive: member=Node-7a53a528-8c2e-3a91-6dc9-3216be34f50b writer.go:29: 2021-01-29T19:33:05.985Z [INFO] TestOperator_AutopilotCASConfiguration: Requesting shutdown writer.go:29: 2021-01-29T19:33:05.987Z [INFO] TestOperator_AutopilotCASConfiguration.server: shutting down server writer.go:29: 2021-01-29T19:33:05.989Z [DEBUG] TestOperator_AutopilotCASConfiguration.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.991Z [WARN] TestOperator_AutopilotCASConfiguration.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:05.991Z [DEBUG] TestOperator_AutopilotCASConfiguration.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.989Z [ERROR] TestOperator_AutopilotCASConfiguration.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:06.005Z [WARN] TestOperator_AutopilotCASConfiguration.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:06.011Z [INFO] TestOperator_AutopilotCASConfiguration.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:06.011Z [INFO] TestOperator_AutopilotCASConfiguration: consul server down writer.go:29: 2021-01-29T19:33:06.012Z [INFO] TestOperator_AutopilotCASConfiguration: shutdown complete writer.go:29: 2021-01-29T19:33:06.013Z [INFO] TestOperator_AutopilotCASConfiguration: Stopping server: protocol=DNS address=127.0.0.1:29514 network=tcp writer.go:29: 2021-01-29T19:33:06.014Z [INFO] TestOperator_AutopilotCASConfiguration: Stopping server: protocol=DNS address=127.0.0.1:29514 network=udp writer.go:29: 2021-01-29T19:33:06.015Z [INFO] TestOperator_AutopilotCASConfiguration: Stopping server: protocol=HTTP address=127.0.0.1:29515 network=tcp === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:06.063Z [DEBUG] TestPreparedQuery_Execute/#02: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:06.066Z [INFO] TestPreparedQuery_Execute/#02: Synced node info === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:06.144Z [DEBUG] TestOperator_ServerHealth: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:06.144Z [WARN] TestPreparedQuery_Execute/#02.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:06.147Z [INFO] TestPreparedQuery_Execute/#02: Requesting shutdown writer.go:29: 2021-01-29T19:33:06.149Z [INFO] TestPreparedQuery_Execute/#02.server: shutting down server writer.go:29: 2021-01-29T19:33:06.150Z [DEBUG] TestPreparedQuery_Execute/#02.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:06.152Z [WARN] TestPreparedQuery_Execute/#02.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:06.152Z [DEBUG] TestPreparedQuery_Execute/#02.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:06.161Z [WARN] TestPreparedQuery_Execute/#02.server.serf.wan: serf: Shutdown without a Leave === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:06.161Z [INFO] TestOperator_ServerHealth: Synced node info === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:06.163Z [INFO] TestPreparedQuery_Execute/#02.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:06.163Z [INFO] TestPreparedQuery_Execute/#02: consul server down writer.go:29: 2021-01-29T19:33:06.164Z [INFO] TestPreparedQuery_Execute/#02: shutdown complete writer.go:29: 2021-01-29T19:33:06.165Z [INFO] TestPreparedQuery_Execute/#02: Stopping server: protocol=DNS address=127.0.0.1:29496 network=tcp writer.go:29: 2021-01-29T19:33:06.166Z [INFO] TestPreparedQuery_Execute/#02: Stopping server: protocol=DNS address=127.0.0.1:29496 network=udp writer.go:29: 2021-01-29T19:33:06.167Z [INFO] TestPreparedQuery_Execute/#02: Stopping server: protocol=HTTP address=127.0.0.1:29497 network=tcp === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:06.452Z [DEBUG] TestOperator_ServerHealth_Unhealthy: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:06.455Z [DEBUG] TestOperator_ServerHealth_Unhealthy: Node info in sync === CONT TestOperator_AutopilotCASConfiguration writer.go:29: 2021-01-29T19:33:06.516Z [INFO] TestOperator_AutopilotCASConfiguration: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:06.518Z [INFO] TestOperator_AutopilotCASConfiguration: Endpoints down --- PASS: TestOperator_AutopilotCASConfiguration (0.69s) === CONT TestOperator_AutopilotGetConfiguration writer.go:29: 2021-01-29T19:33:06.534Z [WARN] TestOperator_AutopilotGetConfiguration: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:06.545Z [DEBUG] TestOperator_AutopilotGetConfiguration.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:06.551Z [DEBUG] TestOperator_AutopilotGetConfiguration.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:06.561Z [INFO] TestOperator_AutopilotGetConfiguration.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ff51667a-1596-f5e4-c5dd-728ad50e370d Address:127.0.0.1:29531}]" writer.go:29: 2021-01-29T19:33:06.563Z [INFO] TestOperator_AutopilotGetConfiguration.server.raft: entering follower state: follower="Node at 127.0.0.1:29531 [Follower]" leader= writer.go:29: 2021-01-29T19:33:06.564Z [INFO] TestOperator_AutopilotGetConfiguration.server.serf.wan: serf: EventMemberJoin: Node-ff51667a-1596-f5e4-c5dd-728ad50e370d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:06.583Z [INFO] TestOperator_AutopilotGetConfiguration.server.serf.lan: serf: EventMemberJoin: Node-ff51667a-1596-f5e4-c5dd-728ad50e370d 127.0.0.1 writer.go:29: 2021-01-29T19:33:06.585Z [INFO] TestOperator_AutopilotGetConfiguration.server: Adding LAN server: server="Node-ff51667a-1596-f5e4-c5dd-728ad50e370d (Addr: tcp/127.0.0.1:29531) (DC: dc1)" writer.go:29: 2021-01-29T19:33:06.587Z [INFO] TestOperator_AutopilotGetConfiguration.server: Handled event for server in area: event=member-join server=Node-ff51667a-1596-f5e4-c5dd-728ad50e370d.dc1 area=wan writer.go:29: 2021-01-29T19:33:06.590Z [INFO] TestOperator_AutopilotGetConfiguration: Started DNS server: address=127.0.0.1:29526 network=udp writer.go:29: 2021-01-29T19:33:06.591Z [INFO] TestOperator_AutopilotGetConfiguration: Started DNS server: address=127.0.0.1:29526 network=tcp writer.go:29: 2021-01-29T19:33:06.593Z [INFO] TestOperator_AutopilotGetConfiguration: Started HTTP server: address=127.0.0.1:29527 network=tcp writer.go:29: 2021-01-29T19:33:06.594Z [INFO] TestOperator_AutopilotGetConfiguration: started state syncer writer.go:29: 2021-01-29T19:33:06.627Z [WARN] TestOperator_AutopilotGetConfiguration.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:06.628Z [INFO] TestOperator_AutopilotGetConfiguration.server.raft: entering candidate state: node="Node at 127.0.0.1:29531 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:06.630Z [DEBUG] TestOperator_AutopilotGetConfiguration.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:06.631Z [DEBUG] TestOperator_AutopilotGetConfiguration.server.raft: vote granted: from=ff51667a-1596-f5e4-c5dd-728ad50e370d term=2 tally=1 writer.go:29: 2021-01-29T19:33:06.632Z [INFO] TestOperator_AutopilotGetConfiguration.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:06.633Z [INFO] TestOperator_AutopilotGetConfiguration.server.raft: entering leader state: leader="Node at 127.0.0.1:29531 [Leader]" writer.go:29: 2021-01-29T19:33:06.634Z [INFO] TestOperator_AutopilotGetConfiguration.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:06.635Z [INFO] TestOperator_AutopilotGetConfiguration.server: New leader elected: payload=Node-ff51667a-1596-f5e4-c5dd-728ad50e370d writer.go:29: 2021-01-29T19:33:06.636Z [DEBUG] TestOperator_AutopilotGetConfiguration.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29531 writer.go:29: 2021-01-29T19:33:06.643Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:06.649Z [INFO] TestOperator_AutopilotGetConfiguration.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:06.649Z [INFO] TestOperator_AutopilotGetConfiguration.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:06.651Z [DEBUG] TestOperator_AutopilotGetConfiguration.server: Skipping self join check for node since the cluster is too small: node=Node-ff51667a-1596-f5e4-c5dd-728ad50e370d writer.go:29: 2021-01-29T19:33:06.652Z [INFO] TestOperator_AutopilotGetConfiguration.server: member joined, marking health alive: member=Node-ff51667a-1596-f5e4-c5dd-728ad50e370d writer.go:29: 2021-01-29T19:33:06.666Z [INFO] TestOperator_AutopilotGetConfiguration: Requesting shutdown === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:06.668Z [INFO] TestPreparedQuery_Execute/#02: Waiting for endpoints to shut down === CONT TestOperator_AutopilotGetConfiguration writer.go:29: 2021-01-29T19:33:06.668Z [INFO] TestOperator_AutopilotGetConfiguration.server: shutting down server === CONT TestPreparedQuery_Execute/#02 writer.go:29: 2021-01-29T19:33:06.668Z [INFO] TestPreparedQuery_Execute/#02: Endpoints down === RUN TestPreparedQuery_Execute/#03 === CONT TestOperator_AutopilotGetConfiguration writer.go:29: 2021-01-29T19:33:06.668Z [DEBUG] TestOperator_AutopilotGetConfiguration.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:06.670Z [WARN] TestOperator_AutopilotGetConfiguration.server.serf.lan: serf: Shutdown without a Leave === CONT TestPreparedQuery_Execute/#03 writer.go:29: 2021-01-29T19:33:06.680Z [WARN] TestPreparedQuery_Execute/#03: bootstrap = true: do not enable unless necessary === CONT TestOperator_AutopilotGetConfiguration writer.go:29: 2021-01-29T19:33:06.671Z [ERROR] TestOperator_AutopilotGetConfiguration.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:06.671Z [DEBUG] TestOperator_AutopilotGetConfiguration.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:06.673Z [WARN] TestOperator_AutopilotGetConfiguration.server.serf.wan: serf: Shutdown without a Leave === CONT TestPreparedQuery_Execute/#03 writer.go:29: 2021-01-29T19:33:06.681Z [DEBUG] TestPreparedQuery_Execute/#03.tlsutil: Update: version=1 === CONT TestOperator_AutopilotGetConfiguration writer.go:29: 2021-01-29T19:33:06.686Z [INFO] TestOperator_AutopilotGetConfiguration.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:06.686Z [INFO] TestOperator_AutopilotGetConfiguration: consul server down writer.go:29: 2021-01-29T19:33:06.688Z [INFO] TestOperator_AutopilotGetConfiguration: shutdown complete === CONT TestPreparedQuery_Execute/#03 writer.go:29: 2021-01-29T19:33:06.687Z [DEBUG] TestPreparedQuery_Execute/#03.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestOperator_AutopilotGetConfiguration writer.go:29: 2021-01-29T19:33:06.690Z [INFO] TestOperator_AutopilotGetConfiguration: Stopping server: protocol=DNS address=127.0.0.1:29526 network=tcp writer.go:29: 2021-01-29T19:33:06.691Z [INFO] TestOperator_AutopilotGetConfiguration: Stopping server: protocol=DNS address=127.0.0.1:29526 network=udp writer.go:29: 2021-01-29T19:33:06.692Z [INFO] TestOperator_AutopilotGetConfiguration: Stopping server: protocol=HTTP address=127.0.0.1:29527 network=tcp === CONT TestPreparedQuery_Execute/#03 writer.go:29: 2021-01-29T19:33:06.694Z [INFO] TestPreparedQuery_Execute/#03.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:200647d3-8a68-ccbd-c0c3-3f62141990c9 Address:127.0.0.1:29525}]" writer.go:29: 2021-01-29T19:33:06.696Z [INFO] TestPreparedQuery_Execute/#03.server.raft: entering follower state: follower="Node at 127.0.0.1:29525 [Follower]" leader= writer.go:29: 2021-01-29T19:33:06.697Z [INFO] TestPreparedQuery_Execute/#03.server.serf.wan: serf: EventMemberJoin: Node-200647d3-8a68-ccbd-c0c3-3f62141990c9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:06.700Z [INFO] TestPreparedQuery_Execute/#03.server.serf.lan: serf: EventMemberJoin: Node-200647d3-8a68-ccbd-c0c3-3f62141990c9 127.0.0.1 writer.go:29: 2021-01-29T19:33:06.702Z [INFO] TestPreparedQuery_Execute/#03.server: Handled event for server in area: event=member-join server=Node-200647d3-8a68-ccbd-c0c3-3f62141990c9.dc1 area=wan writer.go:29: 2021-01-29T19:33:06.702Z [INFO] TestPreparedQuery_Execute/#03: Started DNS server: address=127.0.0.1:29520 network=udp writer.go:29: 2021-01-29T19:33:06.703Z [INFO] TestPreparedQuery_Execute/#03.server: Adding LAN server: server="Node-200647d3-8a68-ccbd-c0c3-3f62141990c9 (Addr: tcp/127.0.0.1:29525) (DC: dc1)" writer.go:29: 2021-01-29T19:33:06.704Z [INFO] TestPreparedQuery_Execute/#03: Started DNS server: address=127.0.0.1:29520 network=tcp writer.go:29: 2021-01-29T19:33:06.709Z [INFO] TestPreparedQuery_Execute/#03: Started HTTP server: address=127.0.0.1:29521 network=tcp writer.go:29: 2021-01-29T19:33:06.710Z [INFO] TestPreparedQuery_Execute/#03: started state syncer writer.go:29: 2021-01-29T19:33:06.747Z [WARN] TestPreparedQuery_Execute/#03.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:06.748Z [INFO] TestPreparedQuery_Execute/#03.server.raft: entering candidate state: node="Node at 127.0.0.1:29525 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:06.751Z [DEBUG] TestPreparedQuery_Execute/#03.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:06.752Z [DEBUG] TestPreparedQuery_Execute/#03.server.raft: vote granted: from=200647d3-8a68-ccbd-c0c3-3f62141990c9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:06.753Z [INFO] TestPreparedQuery_Execute/#03.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:06.753Z [INFO] TestPreparedQuery_Execute/#03.server.raft: entering leader state: leader="Node at 127.0.0.1:29525 [Leader]" writer.go:29: 2021-01-29T19:33:06.754Z [INFO] TestPreparedQuery_Execute/#03.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:06.755Z [INFO] TestPreparedQuery_Execute/#03.server: New leader elected: payload=Node-200647d3-8a68-ccbd-c0c3-3f62141990c9 writer.go:29: 2021-01-29T19:33:06.758Z [DEBUG] TestPreparedQuery_Execute/#03.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29525 writer.go:29: 2021-01-29T19:33:06.769Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:06.777Z [INFO] TestPreparedQuery_Execute/#03.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:06.779Z [INFO] TestPreparedQuery_Execute/#03.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:06.781Z [DEBUG] TestPreparedQuery_Execute/#03.server: Skipping self join check for node since the cluster is too small: node=Node-200647d3-8a68-ccbd-c0c3-3f62141990c9 writer.go:29: 2021-01-29T19:33:06.782Z [INFO] TestPreparedQuery_Execute/#03: Synced node info writer.go:29: 2021-01-29T19:33:06.783Z [INFO] TestPreparedQuery_Execute/#03.server: member joined, marking health alive: member=Node-200647d3-8a68-ccbd-c0c3-3f62141990c9 writer.go:29: 2021-01-29T19:33:07.066Z [WARN] TestPreparedQuery_Execute/#03.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:07.068Z [INFO] TestPreparedQuery_Execute/#03: Requesting shutdown writer.go:29: 2021-01-29T19:33:07.068Z [INFO] TestPreparedQuery_Execute/#03.server: shutting down server writer.go:29: 2021-01-29T19:33:07.069Z [DEBUG] TestPreparedQuery_Execute/#03.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.070Z [WARN] TestPreparedQuery_Execute/#03.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.071Z [DEBUG] TestPreparedQuery_Execute/#03.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.078Z [WARN] TestPreparedQuery_Execute/#03.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.091Z [INFO] TestPreparedQuery_Execute/#03.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:07.091Z [INFO] TestPreparedQuery_Execute/#03: consul server down writer.go:29: 2021-01-29T19:33:07.093Z [INFO] TestPreparedQuery_Execute/#03: shutdown complete writer.go:29: 2021-01-29T19:33:07.093Z [INFO] TestPreparedQuery_Execute/#03: Stopping server: protocol=DNS address=127.0.0.1:29520 network=tcp writer.go:29: 2021-01-29T19:33:07.094Z [INFO] TestPreparedQuery_Execute/#03: Stopping server: protocol=DNS address=127.0.0.1:29520 network=udp writer.go:29: 2021-01-29T19:33:07.095Z [INFO] TestPreparedQuery_Execute/#03: Stopping server: protocol=HTTP address=127.0.0.1:29521 network=tcp === CONT TestOperator_AutopilotGetConfiguration writer.go:29: 2021-01-29T19:33:07.194Z [INFO] TestOperator_AutopilotGetConfiguration: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:07.195Z [INFO] TestOperator_AutopilotGetConfiguration: Endpoints down --- PASS: TestOperator_AutopilotGetConfiguration (0.67s) === CONT TestOperator_Keyring_LocalOnly writer.go:29: 2021-01-29T19:33:07.204Z [WARN] TestOperator_Keyring_LocalOnly: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:07.222Z [DEBUG] TestOperator_Keyring_LocalOnly.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:07.223Z [DEBUG] TestOperator_Keyring_LocalOnly.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:07.228Z [INFO] TestOperator_Keyring_LocalOnly.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e4794677-3483-26de-bab7-785a36cc6948 Address:127.0.0.1:29537}]" writer.go:29: 2021-01-29T19:33:07.229Z [INFO] TestOperator_Keyring_LocalOnly.server.raft: entering follower state: follower="Node at 127.0.0.1:29537 [Follower]" leader= writer.go:29: 2021-01-29T19:33:07.230Z [INFO] TestOperator_Keyring_LocalOnly.server.serf.wan: serf: EventMemberJoin: Node-e4794677-3483-26de-bab7-785a36cc6948.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:07.234Z [INFO] TestOperator_Keyring_LocalOnly.server.serf.lan: serf: EventMemberJoin: Node-e4794677-3483-26de-bab7-785a36cc6948 127.0.0.1 writer.go:29: 2021-01-29T19:33:07.237Z [INFO] TestOperator_Keyring_LocalOnly.server: Adding LAN server: server="Node-e4794677-3483-26de-bab7-785a36cc6948 (Addr: tcp/127.0.0.1:29537) (DC: dc1)" writer.go:29: 2021-01-29T19:33:07.238Z [INFO] TestOperator_Keyring_LocalOnly: Started DNS server: address=127.0.0.1:29532 network=udp writer.go:29: 2021-01-29T19:33:07.238Z [INFO] TestOperator_Keyring_LocalOnly.server: Handled event for server in area: event=member-join server=Node-e4794677-3483-26de-bab7-785a36cc6948.dc1 area=wan writer.go:29: 2021-01-29T19:33:07.239Z [INFO] TestOperator_Keyring_LocalOnly: Started DNS server: address=127.0.0.1:29532 network=tcp writer.go:29: 2021-01-29T19:33:07.241Z [INFO] TestOperator_Keyring_LocalOnly: Started HTTP server: address=127.0.0.1:29533 network=tcp writer.go:29: 2021-01-29T19:33:07.243Z [INFO] TestOperator_Keyring_LocalOnly: started state syncer writer.go:29: 2021-01-29T19:33:07.285Z [WARN] TestOperator_Keyring_LocalOnly.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:07.286Z [INFO] TestOperator_Keyring_LocalOnly.server.raft: entering candidate state: node="Node at 127.0.0.1:29537 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:07.297Z [DEBUG] TestOperator_Keyring_LocalOnly.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:07.298Z [DEBUG] TestOperator_Keyring_LocalOnly.server.raft: vote granted: from=e4794677-3483-26de-bab7-785a36cc6948 term=2 tally=1 writer.go:29: 2021-01-29T19:33:07.299Z [INFO] TestOperator_Keyring_LocalOnly.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:07.300Z [INFO] TestOperator_Keyring_LocalOnly.server.raft: entering leader state: leader="Node at 127.0.0.1:29537 [Leader]" writer.go:29: 2021-01-29T19:33:07.301Z [INFO] TestOperator_Keyring_LocalOnly.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:07.302Z [INFO] TestOperator_Keyring_LocalOnly.server: New leader elected: payload=Node-e4794677-3483-26de-bab7-785a36cc6948 writer.go:29: 2021-01-29T19:33:07.302Z [DEBUG] TestOperator_Keyring_LocalOnly.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29537 writer.go:29: 2021-01-29T19:33:07.310Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:07.318Z [INFO] TestOperator_Keyring_LocalOnly.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:07.320Z [INFO] TestOperator_Keyring_LocalOnly.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.322Z [DEBUG] TestOperator_Keyring_LocalOnly.server: Skipping self join check for node since the cluster is too small: node=Node-e4794677-3483-26de-bab7-785a36cc6948 writer.go:29: 2021-01-29T19:33:07.324Z [INFO] TestOperator_Keyring_LocalOnly.server: member joined, marking health alive: member=Node-e4794677-3483-26de-bab7-785a36cc6948 === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:07.385Z [DEBUG] TestOperator_ServerHealth_Unhealthy.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:07.421Z [INFO] TestOperator_ServerHealth_Unhealthy: Requesting shutdown writer.go:29: 2021-01-29T19:33:07.424Z [INFO] TestOperator_ServerHealth_Unhealthy.server: shutting down server === CONT TestOperator_Keyring_LocalOnly writer.go:29: 2021-01-29T19:33:07.425Z [INFO] TestOperator_Keyring_LocalOnly.server.serf.wan: serf: Received list-keys query === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:07.425Z [DEBUG] TestOperator_ServerHealth_Unhealthy.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.427Z [WARN] TestOperator_ServerHealth_Unhealthy.server.serf.lan: serf: Shutdown without a Leave === CONT TestOperator_Keyring_LocalOnly writer.go:29: 2021-01-29T19:33:07.428Z [DEBUG] TestOperator_Keyring_LocalOnly.server.serf.wan: serf: messageQueryResponseType: Node-e4794677-3483-26de-bab7-785a36cc6948.dc1 === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:07.429Z [DEBUG] TestOperator_ServerHealth_Unhealthy.leader: stopped routine: routine="CA root pruning" === CONT TestOperator_Keyring_LocalOnly writer.go:29: 2021-01-29T19:33:07.430Z [DEBUG] TestOperator_Keyring_LocalOnly.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:07.432Z [WARN] TestOperator_ServerHealth_Unhealthy.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.435Z [INFO] TestOperator_ServerHealth_Unhealthy.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:07.436Z [INFO] TestOperator_ServerHealth_Unhealthy: consul server down === CONT TestOperator_Keyring_LocalOnly writer.go:29: 2021-01-29T19:33:07.438Z [INFO] TestOperator_Keyring_LocalOnly.server.serf.lan: serf: Received list-keys query === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:07.438Z [INFO] TestOperator_ServerHealth_Unhealthy: shutdown complete writer.go:29: 2021-01-29T19:33:07.439Z [INFO] TestOperator_ServerHealth_Unhealthy: Stopping server: protocol=DNS address=127.0.0.1:29502 network=tcp writer.go:29: 2021-01-29T19:33:07.441Z [INFO] TestOperator_ServerHealth_Unhealthy: Stopping server: protocol=DNS address=127.0.0.1:29502 network=udp writer.go:29: 2021-01-29T19:33:07.443Z [INFO] TestOperator_ServerHealth_Unhealthy: Stopping server: protocol=HTTP address=127.0.0.1:29503 network=tcp === CONT TestOperator_Keyring_LocalOnly writer.go:29: 2021-01-29T19:33:07.443Z [DEBUG] TestOperator_Keyring_LocalOnly.server.serf.lan: serf: messageQueryResponseType: Node-e4794677-3483-26de-bab7-785a36cc6948 writer.go:29: 2021-01-29T19:33:07.451Z [INFO] TestOperator_Keyring_LocalOnly: Requesting shutdown writer.go:29: 2021-01-29T19:33:07.452Z [INFO] TestOperator_Keyring_LocalOnly.server: shutting down server writer.go:29: 2021-01-29T19:33:07.460Z [DEBUG] TestOperator_Keyring_LocalOnly.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.461Z [WARN] TestOperator_Keyring_LocalOnly.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.461Z [ERROR] TestOperator_Keyring_LocalOnly.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:07.461Z [DEBUG] TestOperator_Keyring_LocalOnly.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.467Z [WARN] TestOperator_Keyring_LocalOnly.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.470Z [INFO] TestOperator_Keyring_LocalOnly.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:07.472Z [INFO] TestOperator_Keyring_LocalOnly: consul server down writer.go:29: 2021-01-29T19:33:07.474Z [INFO] TestOperator_Keyring_LocalOnly: shutdown complete writer.go:29: 2021-01-29T19:33:07.475Z [INFO] TestOperator_Keyring_LocalOnly: Stopping server: protocol=DNS address=127.0.0.1:29532 network=tcp writer.go:29: 2021-01-29T19:33:07.476Z [INFO] TestOperator_Keyring_LocalOnly: Stopping server: protocol=DNS address=127.0.0.1:29532 network=udp writer.go:29: 2021-01-29T19:33:07.477Z [INFO] TestOperator_Keyring_LocalOnly: Stopping server: protocol=HTTP address=127.0.0.1:29533 network=tcp === CONT TestPreparedQuery_Execute/#03 writer.go:29: 2021-01-29T19:33:07.597Z [INFO] TestPreparedQuery_Execute/#03: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:07.598Z [INFO] TestPreparedQuery_Execute/#03: Endpoints down === RUN TestPreparedQuery_Execute/#04 writer.go:29: 2021-01-29T19:33:07.631Z [WARN] TestPreparedQuery_Execute/#04: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:07.636Z [DEBUG] TestPreparedQuery_Execute/#04.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:07.638Z [DEBUG] TestPreparedQuery_Execute/#04.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:07.657Z [INFO] TestPreparedQuery_Execute/#04.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ed6165b6-3cb2-2195-73ff-abdf81bbbbb0 Address:127.0.0.1:29543}]" writer.go:29: 2021-01-29T19:33:07.659Z [INFO] TestPreparedQuery_Execute/#04.server.raft: entering follower state: follower="Node at 127.0.0.1:29543 [Follower]" leader= writer.go:29: 2021-01-29T19:33:07.659Z [INFO] TestPreparedQuery_Execute/#04.server.serf.wan: serf: EventMemberJoin: Node-ed6165b6-3cb2-2195-73ff-abdf81bbbbb0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:07.665Z [INFO] TestPreparedQuery_Execute/#04.server.serf.lan: serf: EventMemberJoin: Node-ed6165b6-3cb2-2195-73ff-abdf81bbbbb0 127.0.0.1 writer.go:29: 2021-01-29T19:33:07.669Z [INFO] TestPreparedQuery_Execute/#04.server: Adding LAN server: server="Node-ed6165b6-3cb2-2195-73ff-abdf81bbbbb0 (Addr: tcp/127.0.0.1:29543) (DC: dc1)" writer.go:29: 2021-01-29T19:33:07.669Z [INFO] TestPreparedQuery_Execute/#04.server: Handled event for server in area: event=member-join server=Node-ed6165b6-3cb2-2195-73ff-abdf81bbbbb0.dc1 area=wan writer.go:29: 2021-01-29T19:33:07.670Z [INFO] TestPreparedQuery_Execute/#04: Started DNS server: address=127.0.0.1:29538 network=udp writer.go:29: 2021-01-29T19:33:07.671Z [INFO] TestPreparedQuery_Execute/#04: Started DNS server: address=127.0.0.1:29538 network=tcp writer.go:29: 2021-01-29T19:33:07.673Z [INFO] TestPreparedQuery_Execute/#04: Started HTTP server: address=127.0.0.1:29539 network=tcp writer.go:29: 2021-01-29T19:33:07.674Z [INFO] TestPreparedQuery_Execute/#04: started state syncer writer.go:29: 2021-01-29T19:33:07.716Z [WARN] TestPreparedQuery_Execute/#04.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:07.717Z [INFO] TestPreparedQuery_Execute/#04.server.raft: entering candidate state: node="Node at 127.0.0.1:29543 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:07.720Z [DEBUG] TestPreparedQuery_Execute/#04.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:07.721Z [DEBUG] TestPreparedQuery_Execute/#04.server.raft: vote granted: from=ed6165b6-3cb2-2195-73ff-abdf81bbbbb0 term=2 tally=1 writer.go:29: 2021-01-29T19:33:07.722Z [INFO] TestPreparedQuery_Execute/#04.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:07.723Z [INFO] TestPreparedQuery_Execute/#04.server.raft: entering leader state: leader="Node at 127.0.0.1:29543 [Leader]" writer.go:29: 2021-01-29T19:33:07.725Z [INFO] TestPreparedQuery_Execute/#04.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:07.726Z [INFO] TestPreparedQuery_Execute/#04.server: New leader elected: payload=Node-ed6165b6-3cb2-2195-73ff-abdf81bbbbb0 writer.go:29: 2021-01-29T19:33:07.727Z [DEBUG] TestPreparedQuery_Execute/#04.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29543 writer.go:29: 2021-01-29T19:33:07.735Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:07.747Z [WARN] TestPreparedQuery_Execute/#04.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:07.750Z [INFO] TestPreparedQuery_Execute/#04: Requesting shutdown writer.go:29: 2021-01-29T19:33:07.751Z [INFO] TestPreparedQuery_Execute/#04.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:07.754Z [INFO] TestPreparedQuery_Execute/#04.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.752Z [INFO] TestPreparedQuery_Execute/#04.server: shutting down server writer.go:29: 2021-01-29T19:33:07.757Z [DEBUG] TestPreparedQuery_Execute/#04.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.756Z [DEBUG] TestPreparedQuery_Execute/#04.server: Skipping self join check for node since the cluster is too small: node=Node-ed6165b6-3cb2-2195-73ff-abdf81bbbbb0 writer.go:29: 2021-01-29T19:33:07.757Z [ERROR] TestPreparedQuery_Execute/#04.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:07.759Z [WARN] TestPreparedQuery_Execute/#04.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.759Z [DEBUG] TestPreparedQuery_Execute/#04.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.760Z [INFO] TestPreparedQuery_Execute/#04.server: member joined, marking health alive: member=Node-ed6165b6-3cb2-2195-73ff-abdf81bbbbb0 writer.go:29: 2021-01-29T19:33:07.765Z [WARN] TestPreparedQuery_Execute/#04.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.773Z [INFO] TestPreparedQuery_Execute/#04.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:07.774Z [INFO] TestPreparedQuery_Execute/#04: consul server down writer.go:29: 2021-01-29T19:33:07.777Z [INFO] TestPreparedQuery_Execute/#04: shutdown complete writer.go:29: 2021-01-29T19:33:07.778Z [INFO] TestPreparedQuery_Execute/#04: Stopping server: protocol=DNS address=127.0.0.1:29538 network=tcp writer.go:29: 2021-01-29T19:33:07.779Z [INFO] TestPreparedQuery_Execute/#04: Stopping server: protocol=DNS address=127.0.0.1:29538 network=udp writer.go:29: 2021-01-29T19:33:07.781Z [INFO] TestPreparedQuery_Execute/#04: Stopping server: protocol=HTTP address=127.0.0.1:29539 network=tcp === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:07.828Z [DEBUG] TestOperator_ServerHealth.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:07.856Z [INFO] TestOperator_ServerHealth: Requesting shutdown writer.go:29: 2021-01-29T19:33:07.858Z [INFO] TestOperator_ServerHealth.server: shutting down server writer.go:29: 2021-01-29T19:33:07.860Z [DEBUG] TestOperator_ServerHealth.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.862Z [DEBUG] TestOperator_ServerHealth.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.862Z [WARN] TestOperator_ServerHealth.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.865Z [WARN] TestOperator_ServerHealth.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.867Z [INFO] TestOperator_ServerHealth.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:07.868Z [INFO] TestOperator_ServerHealth: consul server down writer.go:29: 2021-01-29T19:33:07.871Z [INFO] TestOperator_ServerHealth: shutdown complete writer.go:29: 2021-01-29T19:33:07.872Z [INFO] TestOperator_ServerHealth: Stopping server: protocol=DNS address=127.0.0.1:29508 network=tcp writer.go:29: 2021-01-29T19:33:07.874Z [INFO] TestOperator_ServerHealth: Stopping server: protocol=DNS address=127.0.0.1:29508 network=udp writer.go:29: 2021-01-29T19:33:07.875Z [INFO] TestOperator_ServerHealth: Stopping server: protocol=HTTP address=127.0.0.1:29509 network=tcp === CONT TestOperator_ServerHealth_Unhealthy writer.go:29: 2021-01-29T19:33:07.944Z [INFO] TestOperator_ServerHealth_Unhealthy: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:07.946Z [INFO] TestOperator_ServerHealth_Unhealthy: Endpoints down --- PASS: TestOperator_ServerHealth_Unhealthy (2.86s) === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:07.959Z [WARN] TestOperator_Keyring_InvalidRelayFactor: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:07.960Z [DEBUG] TestOperator_Keyring_InvalidRelayFactor.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:07.962Z [DEBUG] TestOperator_Keyring_InvalidRelayFactor.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestOperator_Keyring_LocalOnly writer.go:29: 2021-01-29T19:33:07.982Z [INFO] TestOperator_Keyring_LocalOnly: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:07.983Z [INFO] TestOperator_Keyring_LocalOnly: Endpoints down --- PASS: TestOperator_Keyring_LocalOnly (0.79s) === CONT TestOperator_KeyringUse === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.007Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:78ff9148-57f5-44b3-2dd5-2d7afd4b261e Address:127.0.0.1:29549}]" writer.go:29: 2021-01-29T19:33:08.014Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server.raft: entering follower state: follower="Node at 127.0.0.1:29549 [Follower]" leader= writer.go:29: 2021-01-29T19:33:08.021Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server.serf.wan: serf: EventMemberJoin: Node-78ff9148-57f5-44b3-2dd5-2d7afd4b261e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:08.032Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server.serf.lan: serf: EventMemberJoin: Node-78ff9148-57f5-44b3-2dd5-2d7afd4b261e 127.0.0.1 writer.go:29: 2021-01-29T19:33:08.043Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server: Adding LAN server: server="Node-78ff9148-57f5-44b3-2dd5-2d7afd4b261e (Addr: tcp/127.0.0.1:29549) (DC: dc1)" writer.go:29: 2021-01-29T19:33:08.045Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server: Handled event for server in area: event=member-join server=Node-78ff9148-57f5-44b3-2dd5-2d7afd4b261e.dc1 area=wan writer.go:29: 2021-01-29T19:33:08.057Z [INFO] TestOperator_Keyring_InvalidRelayFactor: Started DNS server: address=127.0.0.1:29544 network=udp writer.go:29: 2021-01-29T19:33:08.059Z [INFO] TestOperator_Keyring_InvalidRelayFactor: Started DNS server: address=127.0.0.1:29544 network=tcp writer.go:29: 2021-01-29T19:33:08.089Z [WARN] TestOperator_Keyring_InvalidRelayFactor.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:08.093Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server.raft: entering candidate state: node="Node at 127.0.0.1:29549 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:08.110Z [INFO] TestOperator_Keyring_InvalidRelayFactor: Started HTTP server: address=127.0.0.1:29545 network=tcp writer.go:29: 2021-01-29T19:33:08.111Z [INFO] TestOperator_Keyring_InvalidRelayFactor: started state syncer writer.go:29: 2021-01-29T19:33:08.112Z [DEBUG] TestOperator_Keyring_InvalidRelayFactor.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:08.114Z [DEBUG] TestOperator_Keyring_InvalidRelayFactor.server.raft: vote granted: from=78ff9148-57f5-44b3-2dd5-2d7afd4b261e term=2 tally=1 writer.go:29: 2021-01-29T19:33:08.115Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:08.116Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server.raft: entering leader state: leader="Node at 127.0.0.1:29549 [Leader]" writer.go:29: 2021-01-29T19:33:08.128Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:08.139Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server: New leader elected: payload=Node-78ff9148-57f5-44b3-2dd5-2d7afd4b261e === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.159Z [WARN] TestOperator_KeyringUse: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:08.161Z [DEBUG] TestOperator_KeyringUse.tlsutil: Update: version=1 === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.172Z [DEBUG] TestOperator_Keyring_InvalidRelayFactor.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29549 === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.180Z [DEBUG] TestOperator_KeyringUse.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.208Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.227Z [INFO] TestOperator_KeyringUse.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4da17b1a-8c23-84fd-7df7-027d4ed94e19 Address:127.0.0.1:29555}]" writer.go:29: 2021-01-29T19:33:08.229Z [INFO] TestOperator_KeyringUse.server.raft: entering follower state: follower="Node at 127.0.0.1:29555 [Follower]" leader= writer.go:29: 2021-01-29T19:33:08.229Z [INFO] TestOperator_KeyringUse.server.serf.wan: serf: EventMemberJoin: Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:08.234Z [INFO] TestOperator_KeyringUse.server.serf.lan: serf: EventMemberJoin: Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19 127.0.0.1 writer.go:29: 2021-01-29T19:33:08.236Z [INFO] TestOperator_KeyringUse.server: Adding LAN server: server="Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19 (Addr: tcp/127.0.0.1:29555) (DC: dc1)" writer.go:29: 2021-01-29T19:33:08.237Z [INFO] TestOperator_KeyringUse: Started DNS server: address=127.0.0.1:29550 network=udp writer.go:29: 2021-01-29T19:33:08.237Z [INFO] TestOperator_KeyringUse.server: Handled event for server in area: event=member-join server=Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19.dc1 area=wan writer.go:29: 2021-01-29T19:33:08.239Z [INFO] TestOperator_KeyringUse: Started DNS server: address=127.0.0.1:29550 network=tcp === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.239Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:08.240Z [INFO] TestOperator_Keyring_InvalidRelayFactor.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:08.241Z [DEBUG] TestOperator_Keyring_InvalidRelayFactor.server: Skipping self join check for node since the cluster is too small: node=Node-78ff9148-57f5-44b3-2dd5-2d7afd4b261e === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.241Z [INFO] TestOperator_KeyringUse: Started HTTP server: address=127.0.0.1:29551 network=tcp writer.go:29: 2021-01-29T19:33:08.242Z [INFO] TestOperator_KeyringUse: started state syncer === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.242Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server: member joined, marking health alive: member=Node-78ff9148-57f5-44b3-2dd5-2d7afd4b261e writer.go:29: 2021-01-29T19:33:08.274Z [INFO] TestOperator_Keyring_InvalidRelayFactor: Requesting shutdown === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.276Z [WARN] TestOperator_KeyringUse.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:08.277Z [INFO] TestOperator_KeyringUse.server.raft: entering candidate state: node="Node at 127.0.0.1:29555 [Candidate]" term=2 === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.276Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server: shutting down server writer.go:29: 2021-01-29T19:33:08.278Z [DEBUG] TestOperator_Keyring_InvalidRelayFactor.leader: stopping routine: routine="CA root pruning" === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.279Z [DEBUG] TestOperator_KeyringUse.server.raft: votes: needed=1 === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.278Z [ERROR] TestOperator_Keyring_InvalidRelayFactor.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.280Z [DEBUG] TestOperator_KeyringUse.server.raft: vote granted: from=4da17b1a-8c23-84fd-7df7-027d4ed94e19 term=2 tally=1 === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.279Z [DEBUG] TestOperator_Keyring_InvalidRelayFactor.leader: stopped routine: routine="CA root pruning" === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.281Z [INFO] TestOperator_KeyringUse.server.raft: election won: tally=1 === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.279Z [WARN] TestOperator_Keyring_InvalidRelayFactor.server.serf.lan: serf: Shutdown without a Leave === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.282Z [INFO] TestOperator_KeyringUse.server.raft: entering leader state: leader="Node at 127.0.0.1:29555 [Leader]" === CONT TestPreparedQuery_Execute/#04 writer.go:29: 2021-01-29T19:33:08.283Z [INFO] TestPreparedQuery_Execute/#04: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:08.284Z [INFO] TestPreparedQuery_Execute/#04: Endpoints down === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.283Z [INFO] TestOperator_KeyringUse.server: cluster leadership acquired === RUN TestPreparedQuery_Execute/#05 === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.284Z [INFO] TestOperator_KeyringUse.server: New leader elected: payload=Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19 === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.286Z [WARN] TestOperator_Keyring_InvalidRelayFactor.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:08.287Z [INFO] TestOperator_Keyring_InvalidRelayFactor.server.router.manager: shutting down === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.288Z [DEBUG] TestOperator_KeyringUse.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29555 === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.287Z [INFO] TestOperator_Keyring_InvalidRelayFactor: consul server down writer.go:29: 2021-01-29T19:33:08.289Z [INFO] TestOperator_Keyring_InvalidRelayFactor: shutdown complete writer.go:29: 2021-01-29T19:33:08.290Z [INFO] TestOperator_Keyring_InvalidRelayFactor: Stopping server: protocol=DNS address=127.0.0.1:29544 network=tcp writer.go:29: 2021-01-29T19:33:08.291Z [INFO] TestOperator_Keyring_InvalidRelayFactor: Stopping server: protocol=DNS address=127.0.0.1:29544 network=udp writer.go:29: 2021-01-29T19:33:08.292Z [INFO] TestOperator_Keyring_InvalidRelayFactor: Stopping server: protocol=HTTP address=127.0.0.1:29545 network=tcp === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.293Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:08.300Z [INFO] TestOperator_KeyringUse.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:08.301Z [INFO] TestOperator_KeyringUse.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:08.302Z [DEBUG] TestOperator_KeyringUse.server: Skipping self join check for node since the cluster is too small: node=Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19 writer.go:29: 2021-01-29T19:33:08.303Z [INFO] TestOperator_KeyringUse.server: member joined, marking health alive: member=Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19 === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.309Z [WARN] TestPreparedQuery_Execute/#05: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:08.312Z [DEBUG] TestPreparedQuery_Execute/#05.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:08.314Z [DEBUG] TestPreparedQuery_Execute/#05.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:08.325Z [INFO] TestPreparedQuery_Execute/#05.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9db14937-369b-f93d-8e63-782571ac5371 Address:127.0.0.1:29567}]" writer.go:29: 2021-01-29T19:33:08.327Z [INFO] TestPreparedQuery_Execute/#05.server.raft: entering follower state: follower="Node at 127.0.0.1:29567 [Follower]" leader= writer.go:29: 2021-01-29T19:33:08.329Z [INFO] TestPreparedQuery_Execute/#05.server.serf.wan: serf: EventMemberJoin: Node-9db14937-369b-f93d-8e63-782571ac5371.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:08.335Z [INFO] TestPreparedQuery_Execute/#05.server.serf.lan: serf: EventMemberJoin: Node-9db14937-369b-f93d-8e63-782571ac5371 127.0.0.1 writer.go:29: 2021-01-29T19:33:08.339Z [INFO] TestPreparedQuery_Execute/#05.server: Handled event for server in area: event=member-join server=Node-9db14937-369b-f93d-8e63-782571ac5371.dc1 area=wan writer.go:29: 2021-01-29T19:33:08.339Z [INFO] TestPreparedQuery_Execute/#05.server: Adding LAN server: server="Node-9db14937-369b-f93d-8e63-782571ac5371 (Addr: tcp/127.0.0.1:29567) (DC: dc1)" writer.go:29: 2021-01-29T19:33:08.346Z [INFO] TestPreparedQuery_Execute/#05: Started DNS server: address=127.0.0.1:29562 network=tcp writer.go:29: 2021-01-29T19:33:08.348Z [INFO] TestPreparedQuery_Execute/#05: Started DNS server: address=127.0.0.1:29562 network=udp writer.go:29: 2021-01-29T19:33:08.351Z [INFO] TestPreparedQuery_Execute/#05: Started HTTP server: address=127.0.0.1:29563 network=tcp writer.go:29: 2021-01-29T19:33:08.353Z [INFO] TestPreparedQuery_Execute/#05: started state syncer === CONT TestOperator_ServerHealth writer.go:29: 2021-01-29T19:33:08.377Z [INFO] TestOperator_ServerHealth: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:08.379Z [INFO] TestOperator_ServerHealth: Endpoints down --- PASS: TestOperator_ServerHealth (2.67s) === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.406Z [WARN] TestOperator_KeyringRemove: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:08.409Z [DEBUG] TestOperator_KeyringRemove.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:08.413Z [DEBUG] TestOperator_KeyringRemove.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:08.420Z [INFO] TestOperator_KeyringRemove.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:34838066-5864-5dc7-8953-b45945408674 Address:127.0.0.1:29561}]" writer.go:29: 2021-01-29T19:33:08.423Z [INFO] TestOperator_KeyringRemove.server.raft: entering follower state: follower="Node at 127.0.0.1:29561 [Follower]" leader= writer.go:29: 2021-01-29T19:33:08.425Z [INFO] TestOperator_KeyringRemove.server.serf.wan: serf: EventMemberJoin: Node-34838066-5864-5dc7-8953-b45945408674.dc1 127.0.0.1 === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.388Z [WARN] TestPreparedQuery_Execute/#05.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:08.432Z [INFO] TestPreparedQuery_Execute/#05.server.raft: entering candidate state: node="Node at 127.0.0.1:29567 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:08.434Z [DEBUG] TestPreparedQuery_Execute/#05.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:08.437Z [DEBUG] TestPreparedQuery_Execute/#05.server.raft: vote granted: from=9db14937-369b-f93d-8e63-782571ac5371 term=2 tally=1 writer.go:29: 2021-01-29T19:33:08.439Z [INFO] TestPreparedQuery_Execute/#05.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:08.459Z [INFO] TestPreparedQuery_Execute/#05.server.raft: entering leader state: leader="Node at 127.0.0.1:29567 [Leader]" === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.465Z [DEBUG] TestOperator_KeyringUse: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.464Z [INFO] TestPreparedQuery_Execute/#05.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:08.467Z [INFO] TestPreparedQuery_Execute/#05.server: New leader elected: payload=Node-9db14937-369b-f93d-8e63-782571ac5371 === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.468Z [INFO] TestOperator_KeyringUse: Synced node info === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.469Z [DEBUG] TestPreparedQuery_Execute/#05.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29567 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.429Z [INFO] TestOperator_KeyringRemove.server.serf.lan: serf: EventMemberJoin: Node-34838066-5864-5dc7-8953-b45945408674 127.0.0.1 === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.474Z [DEBUG] TestOperator_KeyringUse: Node info in sync === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.481Z [INFO] TestOperator_KeyringRemove.server: Handled event for server in area: event=member-join server=Node-34838066-5864-5dc7-8953-b45945408674.dc1 area=wan writer.go:29: 2021-01-29T19:33:08.483Z [INFO] TestOperator_KeyringRemove.server: Adding LAN server: server="Node-34838066-5864-5dc7-8953-b45945408674 (Addr: tcp/127.0.0.1:29561) (DC: dc1)" === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.486Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.484Z [INFO] TestOperator_KeyringRemove: Started DNS server: address=127.0.0.1:29556 network=tcp === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.495Z [INFO] TestPreparedQuery_Execute/#05.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:08.496Z [INFO] TestPreparedQuery_Execute/#05.leader: started routine: routine="CA root pruning" === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.496Z [INFO] TestOperator_KeyringRemove: Started DNS server: address=127.0.0.1:29556 network=udp === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.499Z [DEBUG] TestPreparedQuery_Execute/#05.server: Skipping self join check for node since the cluster is too small: node=Node-9db14937-369b-f93d-8e63-782571ac5371 writer.go:29: 2021-01-29T19:33:08.503Z [INFO] TestPreparedQuery_Execute/#05.server: member joined, marking health alive: member=Node-9db14937-369b-f93d-8e63-782571ac5371 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.484Z [WARN] TestOperator_KeyringRemove.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:08.508Z [INFO] TestOperator_KeyringRemove.server.raft: entering candidate state: node="Node at 127.0.0.1:29561 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:08.502Z [INFO] TestOperator_KeyringRemove: Started HTTP server: address=127.0.0.1:29557 network=tcp writer.go:29: 2021-01-29T19:33:08.521Z [INFO] TestOperator_KeyringRemove: started state syncer writer.go:29: 2021-01-29T19:33:08.515Z [DEBUG] TestOperator_KeyringRemove.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:08.535Z [DEBUG] TestOperator_KeyringRemove.server.raft: vote granted: from=34838066-5864-5dc7-8953-b45945408674 term=2 tally=1 writer.go:29: 2021-01-29T19:33:08.541Z [INFO] TestOperator_KeyringRemove.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:08.547Z [INFO] TestOperator_KeyringRemove.server.raft: entering leader state: leader="Node at 127.0.0.1:29561 [Leader]" writer.go:29: 2021-01-29T19:33:08.552Z [INFO] TestOperator_KeyringRemove.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:08.555Z [INFO] TestOperator_KeyringRemove.server: New leader elected: payload=Node-34838066-5864-5dc7-8953-b45945408674 writer.go:29: 2021-01-29T19:33:08.556Z [DEBUG] TestOperator_KeyringRemove.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29561 writer.go:29: 2021-01-29T19:33:08.571Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:08.585Z [INFO] TestOperator_KeyringRemove.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:08.589Z [INFO] TestOperator_KeyringRemove.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:08.592Z [DEBUG] TestOperator_KeyringRemove.server: Skipping self join check for node since the cluster is too small: node=Node-34838066-5864-5dc7-8953-b45945408674 writer.go:29: 2021-01-29T19:33:08.597Z [INFO] TestOperator_KeyringRemove.server: member joined, marking health alive: member=Node-34838066-5864-5dc7-8953-b45945408674 writer.go:29: 2021-01-29T19:33:08.601Z [DEBUG] TestOperator_KeyringRemove: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:08.604Z [INFO] TestOperator_KeyringRemove: Synced node info === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.649Z [INFO] TestOperator_KeyringUse.server.serf.wan: serf: Received install-key query writer.go:29: 2021-01-29T19:33:08.655Z [DEBUG] TestOperator_KeyringUse.server.serf.wan: serf: messageQueryResponseType: Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19.dc1 writer.go:29: 2021-01-29T19:33:08.666Z [DEBUG] TestOperator_KeyringUse.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.675Z [DEBUG] TestPreparedQuery_Execute/#05: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:08.682Z [INFO] TestPreparedQuery_Execute/#05: Synced node info writer.go:29: 2021-01-29T19:33:08.684Z [DEBUG] TestPreparedQuery_Execute/#05: Node info in sync === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.681Z [INFO] TestOperator_KeyringUse.server.serf.lan: serf: Received install-key query writer.go:29: 2021-01-29T19:33:08.691Z [DEBUG] TestOperator_KeyringUse.server.serf.lan: serf: messageQueryResponseType: Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19 writer.go:29: 2021-01-29T19:33:08.696Z [INFO] TestOperator_KeyringUse.server.serf.wan: serf: Received use-key query writer.go:29: 2021-01-29T19:33:08.699Z [DEBUG] TestOperator_KeyringUse.server.serf.wan: serf: messageQueryResponseType: Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19.dc1 writer.go:29: 2021-01-29T19:33:08.703Z [INFO] TestOperator_KeyringUse.server.serf.lan: serf: Received use-key query writer.go:29: 2021-01-29T19:33:08.706Z [DEBUG] TestOperator_KeyringUse.server.serf.lan: serf: messageQueryResponseType: Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19 writer.go:29: 2021-01-29T19:33:08.710Z [INFO] TestOperator_KeyringUse.server.serf.wan: serf: Received remove-key query writer.go:29: 2021-01-29T19:33:08.712Z [DEBUG] TestOperator_KeyringUse.server.serf.wan: serf: messageQueryResponseType: Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19.dc1 writer.go:29: 2021-01-29T19:33:08.718Z [INFO] TestOperator_KeyringUse.server.serf.lan: serf: Received remove-key query === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.715Z [WARN] TestPreparedQuery_Execute/#05.server: endpoint injected; this should only be used for testing === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.719Z [DEBUG] TestOperator_KeyringUse.server.serf.lan: serf: messageQueryResponseType: Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19 writer.go:29: 2021-01-29T19:33:08.722Z [INFO] TestOperator_KeyringUse.server.serf.wan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:08.725Z [DEBUG] TestOperator_KeyringUse.server.serf.wan: serf: messageQueryResponseType: Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19.dc1 === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.728Z [INFO] TestPreparedQuery_Execute/#05: Requesting shutdown === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.728Z [INFO] TestOperator_KeyringUse.server.serf.lan: serf: Received list-keys query === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.729Z [INFO] TestPreparedQuery_Execute/#05.server: shutting down server writer.go:29: 2021-01-29T19:33:08.731Z [DEBUG] TestPreparedQuery_Execute/#05.leader: stopping routine: routine="CA root pruning" === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.734Z [DEBUG] TestOperator_KeyringUse.server.serf.lan: serf: messageQueryResponseType: Node-4da17b1a-8c23-84fd-7df7-027d4ed94e19 === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.733Z [WARN] TestPreparedQuery_Execute/#05.server.serf.lan: serf: Shutdown without a Leave === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.736Z [INFO] TestOperator_KeyringUse: Requesting shutdown writer.go:29: 2021-01-29T19:33:08.737Z [INFO] TestOperator_KeyringUse.server: shutting down server === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.733Z [DEBUG] TestPreparedQuery_Execute/#05.leader: stopped routine: routine="CA root pruning" === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.739Z [DEBUG] TestOperator_KeyringUse.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:08.740Z [DEBUG] TestOperator_KeyringUse.leader: stopped routine: routine="CA root pruning" === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.738Z [WARN] TestPreparedQuery_Execute/#05.server.serf.wan: serf: Shutdown without a Leave === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.742Z [WARN] TestOperator_KeyringUse.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:08.744Z [WARN] TestOperator_KeyringUse.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:08.747Z [INFO] TestOperator_KeyringUse.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:08.750Z [INFO] TestOperator_KeyringUse: consul server down === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.752Z [INFO] TestPreparedQuery_Execute/#05.server.router.manager: shutting down === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.751Z [INFO] TestOperator_KeyringUse: shutdown complete === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.752Z [INFO] TestPreparedQuery_Execute/#05: consul server down === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.753Z [INFO] TestOperator_KeyringUse: Stopping server: protocol=DNS address=127.0.0.1:29550 network=tcp === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.754Z [INFO] TestPreparedQuery_Execute/#05: shutdown complete === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.755Z [INFO] TestOperator_KeyringUse: Stopping server: protocol=DNS address=127.0.0.1:29550 network=udp === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:08.755Z [INFO] TestPreparedQuery_Execute/#05: Stopping server: protocol=DNS address=127.0.0.1:29562 network=tcp writer.go:29: 2021-01-29T19:33:08.756Z [INFO] TestPreparedQuery_Execute/#05: Stopping server: protocol=DNS address=127.0.0.1:29562 network=udp writer.go:29: 2021-01-29T19:33:08.757Z [INFO] TestPreparedQuery_Execute/#05: Stopping server: protocol=HTTP address=127.0.0.1:29563 network=tcp === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:08.756Z [INFO] TestOperator_KeyringUse: Stopping server: protocol=HTTP address=127.0.0.1:29551 network=tcp === CONT TestOperator_Keyring_InvalidRelayFactor writer.go:29: 2021-01-29T19:33:08.793Z [INFO] TestOperator_Keyring_InvalidRelayFactor: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:08.795Z [INFO] TestOperator_Keyring_InvalidRelayFactor: Endpoints down --- PASS: TestOperator_Keyring_InvalidRelayFactor (0.85s) === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.806Z [WARN] TestOperator_KeyringList: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:08.807Z [DEBUG] TestOperator_KeyringList.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:08.809Z [DEBUG] TestOperator_KeyringList.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.819Z [INFO] TestOperator_KeyringRemove.server.serf.wan: serf: Received install-key query === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.818Z [INFO] TestOperator_KeyringList.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:048c1f49-0014-46ca-87a1-7e1aadb02920 Address:127.0.0.1:29573}]" writer.go:29: 2021-01-29T19:33:08.820Z [INFO] TestOperator_KeyringList.server.raft: entering follower state: follower="Node at 127.0.0.1:29573 [Follower]" leader= === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.822Z [DEBUG] TestOperator_KeyringRemove.server.serf.wan: serf: messageQueryResponseType: Node-34838066-5864-5dc7-8953-b45945408674.dc1 === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.821Z [INFO] TestOperator_KeyringList.server.serf.wan: serf: EventMemberJoin: Node-048c1f49-0014-46ca-87a1-7e1aadb02920.dc1 127.0.0.1 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.824Z [DEBUG] TestOperator_KeyringRemove.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.827Z [INFO] TestOperator_KeyringList.server.serf.lan: serf: EventMemberJoin: Node-048c1f49-0014-46ca-87a1-7e1aadb02920 127.0.0.1 writer.go:29: 2021-01-29T19:33:08.829Z [INFO] TestOperator_KeyringList.server: Handled event for server in area: event=member-join server=Node-048c1f49-0014-46ca-87a1-7e1aadb02920.dc1 area=wan writer.go:29: 2021-01-29T19:33:08.829Z [INFO] TestOperator_KeyringList.server: Adding LAN server: server="Node-048c1f49-0014-46ca-87a1-7e1aadb02920 (Addr: tcp/127.0.0.1:29573) (DC: dc1)" === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.831Z [INFO] TestOperator_KeyringRemove.server.serf.lan: serf: Received install-key query === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.830Z [INFO] TestOperator_KeyringList: Started DNS server: address=127.0.0.1:29568 network=udp writer.go:29: 2021-01-29T19:33:08.832Z [INFO] TestOperator_KeyringList: Started DNS server: address=127.0.0.1:29568 network=tcp writer.go:29: 2021-01-29T19:33:08.833Z [INFO] TestOperator_KeyringList: Started HTTP server: address=127.0.0.1:29569 network=tcp writer.go:29: 2021-01-29T19:33:08.835Z [INFO] TestOperator_KeyringList: started state syncer === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.835Z [DEBUG] TestOperator_KeyringRemove.server.serf.lan: serf: messageQueryResponseType: Node-34838066-5864-5dc7-8953-b45945408674 writer.go:29: 2021-01-29T19:33:08.838Z [INFO] TestOperator_KeyringRemove.server.serf.wan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:08.840Z [DEBUG] TestOperator_KeyringRemove.server.serf.wan: serf: messageQueryResponseType: Node-34838066-5864-5dc7-8953-b45945408674.dc1 writer.go:29: 2021-01-29T19:33:08.842Z [INFO] TestOperator_KeyringRemove.server.serf.lan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:08.844Z [DEBUG] TestOperator_KeyringRemove.server.serf.lan: serf: messageQueryResponseType: Node-34838066-5864-5dc7-8953-b45945408674 writer.go:29: 2021-01-29T19:33:08.847Z [INFO] TestOperator_KeyringRemove.server.serf.wan: serf: Received remove-key query writer.go:29: 2021-01-29T19:33:08.849Z [DEBUG] TestOperator_KeyringRemove.server.serf.wan: serf: messageQueryResponseType: Node-34838066-5864-5dc7-8953-b45945408674.dc1 writer.go:29: 2021-01-29T19:33:08.853Z [INFO] TestOperator_KeyringRemove.server.serf.lan: serf: Received remove-key query === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.858Z [WARN] TestOperator_KeyringList.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:08.859Z [INFO] TestOperator_KeyringList.server.raft: entering candidate state: node="Node at 127.0.0.1:29573 [Candidate]" term=2 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.860Z [DEBUG] TestOperator_KeyringRemove.server.serf.lan: serf: messageQueryResponseType: Node-34838066-5864-5dc7-8953-b45945408674 === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.863Z [DEBUG] TestOperator_KeyringList.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:08.864Z [DEBUG] TestOperator_KeyringList.server.raft: vote granted: from=048c1f49-0014-46ca-87a1-7e1aadb02920 term=2 tally=1 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.864Z [INFO] TestOperator_KeyringRemove.server.serf.wan: serf: Received list-keys query === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.865Z [INFO] TestOperator_KeyringList.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:08.866Z [INFO] TestOperator_KeyringList.server.raft: entering leader state: leader="Node at 127.0.0.1:29573 [Leader]" writer.go:29: 2021-01-29T19:33:08.867Z [INFO] TestOperator_KeyringList.server: cluster leadership acquired === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.866Z [DEBUG] TestOperator_KeyringRemove.server.serf.wan: serf: messageQueryResponseType: Node-34838066-5864-5dc7-8953-b45945408674.dc1 === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.867Z [INFO] TestOperator_KeyringList.server: New leader elected: payload=Node-048c1f49-0014-46ca-87a1-7e1aadb02920 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.870Z [INFO] TestOperator_KeyringRemove.server.serf.lan: serf: Received list-keys query === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.868Z [DEBUG] TestOperator_KeyringList.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29573 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.871Z [DEBUG] TestOperator_KeyringRemove.server.serf.lan: serf: messageQueryResponseType: Node-34838066-5864-5dc7-8953-b45945408674 writer.go:29: 2021-01-29T19:33:08.873Z [INFO] TestOperator_KeyringRemove: Requesting shutdown writer.go:29: 2021-01-29T19:33:08.874Z [INFO] TestOperator_KeyringRemove.server: shutting down server writer.go:29: 2021-01-29T19:33:08.875Z [DEBUG] TestOperator_KeyringRemove.leader: stopping routine: routine="CA root pruning" === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.875Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.876Z [WARN] TestOperator_KeyringRemove.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:08.876Z [DEBUG] TestOperator_KeyringRemove.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:08.878Z [WARN] TestOperator_KeyringRemove.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:08.880Z [INFO] TestOperator_KeyringRemove.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:08.880Z [INFO] TestOperator_KeyringRemove: consul server down writer.go:29: 2021-01-29T19:33:08.882Z [INFO] TestOperator_KeyringRemove: shutdown complete === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.883Z [INFO] TestOperator_KeyringList.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.883Z [INFO] TestOperator_KeyringRemove: Stopping server: protocol=DNS address=127.0.0.1:29556 network=tcp === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.884Z [INFO] TestOperator_KeyringList.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:08.885Z [DEBUG] TestOperator_KeyringList.server: Skipping self join check for node since the cluster is too small: node=Node-048c1f49-0014-46ca-87a1-7e1aadb02920 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.885Z [INFO] TestOperator_KeyringRemove: Stopping server: protocol=DNS address=127.0.0.1:29556 network=udp === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:08.885Z [INFO] TestOperator_KeyringList.server: member joined, marking health alive: member=Node-048c1f49-0014-46ca-87a1-7e1aadb02920 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:08.886Z [INFO] TestOperator_KeyringRemove: Stopping server: protocol=HTTP address=127.0.0.1:29557 network=tcp === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:09.007Z [INFO] TestOperator_KeyringList.server.serf.wan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:09.011Z [DEBUG] TestOperator_KeyringList.server.serf.wan: serf: messageQueryResponseType: Node-048c1f49-0014-46ca-87a1-7e1aadb02920.dc1 writer.go:29: 2021-01-29T19:33:09.013Z [DEBUG] TestOperator_KeyringList.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:09.021Z [INFO] TestOperator_KeyringList.server.serf.lan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:09.027Z [DEBUG] TestOperator_KeyringList.server.serf.lan: serf: messageQueryResponseType: Node-048c1f49-0014-46ca-87a1-7e1aadb02920 writer.go:29: 2021-01-29T19:33:09.030Z [INFO] TestOperator_KeyringList: Requesting shutdown writer.go:29: 2021-01-29T19:33:09.031Z [INFO] TestOperator_KeyringList.server: shutting down server writer.go:29: 2021-01-29T19:33:09.033Z [DEBUG] TestOperator_KeyringList.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:09.034Z [WARN] TestOperator_KeyringList.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:09.033Z [ERROR] TestOperator_KeyringList.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:09.034Z [DEBUG] TestOperator_KeyringList.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:09.036Z [WARN] TestOperator_KeyringList.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:09.042Z [INFO] TestOperator_KeyringList.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:09.043Z [INFO] TestOperator_KeyringList: consul server down writer.go:29: 2021-01-29T19:33:09.046Z [INFO] TestOperator_KeyringList: shutdown complete writer.go:29: 2021-01-29T19:33:09.048Z [INFO] TestOperator_KeyringList: Stopping server: protocol=DNS address=127.0.0.1:29568 network=tcp writer.go:29: 2021-01-29T19:33:09.050Z [INFO] TestOperator_KeyringList: Stopping server: protocol=DNS address=127.0.0.1:29568 network=udp writer.go:29: 2021-01-29T19:33:09.051Z [INFO] TestOperator_KeyringList: Stopping server: protocol=HTTP address=127.0.0.1:29569 network=tcp === CONT TestPreparedQuery_Execute/#05 writer.go:29: 2021-01-29T19:33:09.258Z [INFO] TestPreparedQuery_Execute/#05: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:09.259Z [INFO] TestPreparedQuery_Execute/#05: Endpoints down === RUN TestPreparedQuery_Execute/#06 === CONT TestOperator_KeyringUse writer.go:29: 2021-01-29T19:33:09.261Z [INFO] TestOperator_KeyringUse: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:09.263Z [INFO] TestOperator_KeyringUse: Endpoints down --- PASS: TestOperator_KeyringUse (1.28s) === CONT TestOperator_KeyringInstall === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.286Z [WARN] TestPreparedQuery_Execute/#06: bootstrap = true: do not enable unless necessary === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.289Z [WARN] TestOperator_KeyringInstall: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:09.290Z [DEBUG] TestOperator_KeyringInstall.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:09.291Z [DEBUG] TestOperator_KeyringInstall.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.288Z [DEBUG] TestPreparedQuery_Execute/#06.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:09.295Z [DEBUG] TestPreparedQuery_Execute/#06.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.300Z [INFO] TestOperator_KeyringInstall.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a7053394-701b-1cea-f8f5-5ff50bd36564 Address:127.0.0.1:29579}]" writer.go:29: 2021-01-29T19:33:09.304Z [INFO] TestOperator_KeyringInstall.server.raft: entering follower state: follower="Node at 127.0.0.1:29579 [Follower]" leader= writer.go:29: 2021-01-29T19:33:09.308Z [INFO] TestOperator_KeyringInstall.server.serf.wan: serf: EventMemberJoin: Node-a7053394-701b-1cea-f8f5-5ff50bd36564.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:09.314Z [INFO] TestOperator_KeyringInstall.server.serf.lan: serf: EventMemberJoin: Node-a7053394-701b-1cea-f8f5-5ff50bd36564 127.0.0.1 writer.go:29: 2021-01-29T19:33:09.319Z [INFO] TestOperator_KeyringInstall.server: Handled event for server in area: event=member-join server=Node-a7053394-701b-1cea-f8f5-5ff50bd36564.dc1 area=wan writer.go:29: 2021-01-29T19:33:09.319Z [INFO] TestOperator_KeyringInstall.server: Adding LAN server: server="Node-a7053394-701b-1cea-f8f5-5ff50bd36564 (Addr: tcp/127.0.0.1:29579) (DC: dc1)" === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.321Z [INFO] TestPreparedQuery_Execute/#06.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:af3a25dc-4a66-5383-a18b-bdfee6c38316 Address:127.0.0.1:29585}]" writer.go:29: 2021-01-29T19:33:09.322Z [INFO] TestPreparedQuery_Execute/#06.server.raft: entering follower state: follower="Node at 127.0.0.1:29585 [Follower]" leader= === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.324Z [INFO] TestOperator_KeyringInstall: Started DNS server: address=127.0.0.1:29574 network=udp writer.go:29: 2021-01-29T19:33:09.325Z [INFO] TestOperator_KeyringInstall: Started DNS server: address=127.0.0.1:29574 network=tcp === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.323Z [INFO] TestPreparedQuery_Execute/#06.server.serf.wan: serf: EventMemberJoin: Node-af3a25dc-4a66-5383-a18b-bdfee6c38316.dc1 127.0.0.1 === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.328Z [INFO] TestOperator_KeyringInstall: Started HTTP server: address=127.0.0.1:29575 network=tcp writer.go:29: 2021-01-29T19:33:09.329Z [INFO] TestOperator_KeyringInstall: started state syncer === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.329Z [INFO] TestPreparedQuery_Execute/#06.server.serf.lan: serf: EventMemberJoin: Node-af3a25dc-4a66-5383-a18b-bdfee6c38316 127.0.0.1 === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.342Z [WARN] TestOperator_KeyringInstall.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:09.343Z [INFO] TestOperator_KeyringInstall.server.raft: entering candidate state: node="Node at 127.0.0.1:29579 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:09.345Z [DEBUG] TestOperator_KeyringInstall.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:09.346Z [DEBUG] TestOperator_KeyringInstall.server.raft: vote granted: from=a7053394-701b-1cea-f8f5-5ff50bd36564 term=2 tally=1 writer.go:29: 2021-01-29T19:33:09.348Z [INFO] TestOperator_KeyringInstall.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:09.349Z [INFO] TestOperator_KeyringInstall.server.raft: entering leader state: leader="Node at 127.0.0.1:29579 [Leader]" writer.go:29: 2021-01-29T19:33:09.350Z [INFO] TestOperator_KeyringInstall.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:09.350Z [INFO] TestOperator_KeyringInstall.server: New leader elected: payload=Node-a7053394-701b-1cea-f8f5-5ff50bd36564 === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.356Z [INFO] TestPreparedQuery_Execute/#06.server: Handled event for server in area: event=member-join server=Node-af3a25dc-4a66-5383-a18b-bdfee6c38316.dc1 area=wan writer.go:29: 2021-01-29T19:33:09.357Z [INFO] TestPreparedQuery_Execute/#06: Started DNS server: address=127.0.0.1:29580 network=udp writer.go:29: 2021-01-29T19:33:09.358Z [INFO] TestPreparedQuery_Execute/#06: Started DNS server: address=127.0.0.1:29580 network=tcp writer.go:29: 2021-01-29T19:33:09.358Z [INFO] TestPreparedQuery_Execute/#06.server: Adding LAN server: server="Node-af3a25dc-4a66-5383-a18b-bdfee6c38316 (Addr: tcp/127.0.0.1:29585) (DC: dc1)" writer.go:29: 2021-01-29T19:33:09.360Z [INFO] TestPreparedQuery_Execute/#06: Started HTTP server: address=127.0.0.1:29581 network=tcp writer.go:29: 2021-01-29T19:33:09.362Z [INFO] TestPreparedQuery_Execute/#06: started state syncer === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.361Z [DEBUG] TestOperator_KeyringInstall.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29579 writer.go:29: 2021-01-29T19:33:09.366Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:09.371Z [INFO] TestOperator_KeyringInstall.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:09.372Z [INFO] TestOperator_KeyringInstall.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:09.373Z [DEBUG] TestOperator_KeyringInstall.server: Skipping self join check for node since the cluster is too small: node=Node-a7053394-701b-1cea-f8f5-5ff50bd36564 writer.go:29: 2021-01-29T19:33:09.374Z [INFO] TestOperator_KeyringInstall.server: member joined, marking health alive: member=Node-a7053394-701b-1cea-f8f5-5ff50bd36564 === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.377Z [WARN] TestPreparedQuery_Execute/#06.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:09.378Z [INFO] TestPreparedQuery_Execute/#06.server.raft: entering candidate state: node="Node at 127.0.0.1:29585 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:09.381Z [DEBUG] TestPreparedQuery_Execute/#06.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:09.382Z [DEBUG] TestPreparedQuery_Execute/#06.server.raft: vote granted: from=af3a25dc-4a66-5383-a18b-bdfee6c38316 term=2 tally=1 writer.go:29: 2021-01-29T19:33:09.383Z [INFO] TestPreparedQuery_Execute/#06.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:09.384Z [INFO] TestPreparedQuery_Execute/#06.server.raft: entering leader state: leader="Node at 127.0.0.1:29585 [Leader]" writer.go:29: 2021-01-29T19:33:09.385Z [INFO] TestPreparedQuery_Execute/#06.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:09.386Z [INFO] TestPreparedQuery_Execute/#06.server: New leader elected: payload=Node-af3a25dc-4a66-5383-a18b-bdfee6c38316 writer.go:29: 2021-01-29T19:33:09.386Z [DEBUG] TestPreparedQuery_Execute/#06.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29585 === CONT TestOperator_KeyringRemove writer.go:29: 2021-01-29T19:33:09.387Z [INFO] TestOperator_KeyringRemove: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:09.389Z [INFO] TestOperator_KeyringRemove: Endpoints down --- PASS: TestOperator_KeyringRemove (1.01s) === CONT TestOperator_RaftPeer === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.389Z [DEBUG] TestOperator_KeyringInstall: Skipping remote check since it is managed automatically: check=serfHealth === RUN TestOperator_RaftPeer/#00 === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.391Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.391Z [INFO] TestOperator_KeyringInstall: Synced node info writer.go:29: 2021-01-29T19:33:09.392Z [DEBUG] TestOperator_KeyringInstall: Node info in sync === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.396Z [INFO] TestPreparedQuery_Execute/#06.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:09.397Z [INFO] TestPreparedQuery_Execute/#06.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:09.398Z [DEBUG] TestPreparedQuery_Execute/#06.server: Skipping self join check for node since the cluster is too small: node=Node-af3a25dc-4a66-5383-a18b-bdfee6c38316 writer.go:29: 2021-01-29T19:33:09.399Z [INFO] TestPreparedQuery_Execute/#06.server: member joined, marking health alive: member=Node-af3a25dc-4a66-5383-a18b-bdfee6c38316 === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.401Z [WARN] TestOperator_RaftPeer/#00: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:09.402Z [DEBUG] TestOperator_RaftPeer/#00.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:09.404Z [DEBUG] TestOperator_RaftPeer/#00.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:09.409Z [INFO] TestOperator_RaftPeer/#00.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5dc3823e-f53a-79d5-fad1-829dbaf82d73 Address:127.0.0.1:29591}]" writer.go:29: 2021-01-29T19:33:09.411Z [INFO] TestOperator_RaftPeer/#00.server.raft: entering follower state: follower="Node at 127.0.0.1:29591 [Follower]" leader= writer.go:29: 2021-01-29T19:33:09.412Z [INFO] TestOperator_RaftPeer/#00.server.serf.wan: serf: EventMemberJoin: Node-5dc3823e-f53a-79d5-fad1-829dbaf82d73.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:09.418Z [INFO] TestOperator_RaftPeer/#00.server.serf.lan: serf: EventMemberJoin: Node-5dc3823e-f53a-79d5-fad1-829dbaf82d73 127.0.0.1 writer.go:29: 2021-01-29T19:33:09.421Z [INFO] TestOperator_RaftPeer/#00.server: Handled event for server in area: event=member-join server=Node-5dc3823e-f53a-79d5-fad1-829dbaf82d73.dc1 area=wan writer.go:29: 2021-01-29T19:33:09.421Z [INFO] TestOperator_RaftPeer/#00.server: Adding LAN server: server="Node-5dc3823e-f53a-79d5-fad1-829dbaf82d73 (Addr: tcp/127.0.0.1:29591) (DC: dc1)" writer.go:29: 2021-01-29T19:33:09.422Z [INFO] TestOperator_RaftPeer/#00: Started DNS server: address=127.0.0.1:29586 network=tcp writer.go:29: 2021-01-29T19:33:09.424Z [INFO] TestOperator_RaftPeer/#00: Started DNS server: address=127.0.0.1:29586 network=udp writer.go:29: 2021-01-29T19:33:09.425Z [INFO] TestOperator_RaftPeer/#00: Started HTTP server: address=127.0.0.1:29587 network=tcp writer.go:29: 2021-01-29T19:33:09.426Z [INFO] TestOperator_RaftPeer/#00: started state syncer === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.433Z [INFO] TestOperator_KeyringInstall.server.serf.wan: serf: Received install-key query writer.go:29: 2021-01-29T19:33:09.438Z [DEBUG] TestOperator_KeyringInstall.server.serf.wan: serf: messageQueryResponseType: Node-a7053394-701b-1cea-f8f5-5ff50bd36564.dc1 writer.go:29: 2021-01-29T19:33:09.453Z [DEBUG] TestOperator_KeyringInstall.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:09.458Z [INFO] TestOperator_KeyringInstall.server.serf.lan: serf: Received install-key query writer.go:29: 2021-01-29T19:33:09.462Z [DEBUG] TestOperator_KeyringInstall.server.serf.lan: serf: messageQueryResponseType: Node-a7053394-701b-1cea-f8f5-5ff50bd36564 writer.go:29: 2021-01-29T19:33:09.466Z [INFO] TestOperator_KeyringInstall.server.serf.wan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:09.467Z [DEBUG] TestOperator_KeyringInstall.server.serf.wan: serf: messageQueryResponseType: Node-a7053394-701b-1cea-f8f5-5ff50bd36564.dc1 writer.go:29: 2021-01-29T19:33:09.470Z [INFO] TestOperator_KeyringInstall.server.serf.lan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:09.472Z [DEBUG] TestOperator_KeyringInstall.server.serf.lan: serf: messageQueryResponseType: Node-a7053394-701b-1cea-f8f5-5ff50bd36564 writer.go:29: 2021-01-29T19:33:09.473Z [INFO] TestOperator_KeyringInstall: Requesting shutdown writer.go:29: 2021-01-29T19:33:09.474Z [INFO] TestOperator_KeyringInstall.server: shutting down server writer.go:29: 2021-01-29T19:33:09.475Z [DEBUG] TestOperator_KeyringInstall.leader: stopping routine: routine="CA root pruning" === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.481Z [WARN] TestOperator_RaftPeer/#00.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:09.482Z [INFO] TestOperator_RaftPeer/#00.server.raft: entering candidate state: node="Node at 127.0.0.1:29591 [Candidate]" term=2 === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.482Z [WARN] TestOperator_KeyringInstall.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:09.482Z [DEBUG] TestOperator_KeyringInstall.leader: stopped routine: routine="CA root pruning" === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.484Z [DEBUG] TestOperator_RaftPeer/#00.server.raft: votes: needed=1 === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.484Z [WARN] TestOperator_KeyringInstall.server.serf.wan: serf: Shutdown without a Leave === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.485Z [DEBUG] TestOperator_RaftPeer/#00.server.raft: vote granted: from=5dc3823e-f53a-79d5-fad1-829dbaf82d73 term=2 tally=1 writer.go:29: 2021-01-29T19:33:09.486Z [INFO] TestOperator_RaftPeer/#00.server.raft: election won: tally=1 === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.487Z [INFO] TestOperator_KeyringInstall.server.router.manager: shutting down === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.487Z [INFO] TestOperator_RaftPeer/#00.server.raft: entering leader state: leader="Node at 127.0.0.1:29591 [Leader]" === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.487Z [INFO] TestOperator_KeyringInstall: consul server down === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.489Z [INFO] TestOperator_RaftPeer/#00.server: cluster leadership acquired === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.489Z [INFO] TestOperator_KeyringInstall: shutdown complete === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.489Z [INFO] TestOperator_RaftPeer/#00.server: New leader elected: payload=Node-5dc3823e-f53a-79d5-fad1-829dbaf82d73 === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.490Z [INFO] TestOperator_KeyringInstall: Stopping server: protocol=DNS address=127.0.0.1:29574 network=tcp writer.go:29: 2021-01-29T19:33:09.491Z [INFO] TestOperator_KeyringInstall: Stopping server: protocol=DNS address=127.0.0.1:29574 network=udp writer.go:29: 2021-01-29T19:33:09.492Z [INFO] TestOperator_KeyringInstall: Stopping server: protocol=HTTP address=127.0.0.1:29575 network=tcp === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.490Z [DEBUG] TestOperator_RaftPeer/#00.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29591 writer.go:29: 2021-01-29T19:33:09.498Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:09.524Z [INFO] TestOperator_RaftPeer/#00.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:09.525Z [INFO] TestOperator_RaftPeer/#00.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:09.526Z [DEBUG] TestOperator_RaftPeer/#00.server: Skipping self join check for node since the cluster is too small: node=Node-5dc3823e-f53a-79d5-fad1-829dbaf82d73 writer.go:29: 2021-01-29T19:33:09.527Z [INFO] TestOperator_RaftPeer/#00.server: member joined, marking health alive: member=Node-5dc3823e-f53a-79d5-fad1-829dbaf82d73 === CONT TestOperator_KeyringList writer.go:29: 2021-01-29T19:33:09.552Z [INFO] TestOperator_KeyringList: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:09.553Z [INFO] TestOperator_KeyringList: Endpoints down --- PASS: TestOperator_KeyringList (0.76s) === CONT TestOperator_RaftConfiguration === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.560Z [DEBUG] TestPreparedQuery_Execute/#06: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:09.562Z [INFO] TestPreparedQuery_Execute/#06: Synced node info === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:09.567Z [WARN] TestOperator_RaftConfiguration: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:09.569Z [DEBUG] TestOperator_RaftConfiguration.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:09.571Z [DEBUG] TestOperator_RaftConfiguration.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:09.603Z [INFO] TestOperator_RaftConfiguration.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5bf18591-1b2a-d1a5-fe89-a39ae076bd0b Address:127.0.0.1:29597}]" writer.go:29: 2021-01-29T19:33:09.607Z [INFO] TestOperator_RaftConfiguration.server.raft: entering follower state: follower="Node at 127.0.0.1:29597 [Follower]" leader= writer.go:29: 2021-01-29T19:33:09.638Z [INFO] TestOperator_RaftConfiguration.server.serf.wan: serf: EventMemberJoin: Node-5bf18591-1b2a-d1a5-fe89-a39ae076bd0b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:09.652Z [INFO] TestOperator_RaftConfiguration.server.serf.lan: serf: EventMemberJoin: Node-5bf18591-1b2a-d1a5-fe89-a39ae076bd0b 127.0.0.1 writer.go:29: 2021-01-29T19:33:09.660Z [INFO] TestOperator_RaftConfiguration.server: Handled event for server in area: event=member-join server=Node-5bf18591-1b2a-d1a5-fe89-a39ae076bd0b.dc1 area=wan writer.go:29: 2021-01-29T19:33:09.660Z [INFO] TestOperator_RaftConfiguration.server: Adding LAN server: server="Node-5bf18591-1b2a-d1a5-fe89-a39ae076bd0b (Addr: tcp/127.0.0.1:29597) (DC: dc1)" writer.go:29: 2021-01-29T19:33:09.660Z [WARN] TestOperator_RaftConfiguration.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:09.663Z [INFO] TestOperator_RaftConfiguration.server.raft: entering candidate state: node="Node at 127.0.0.1:29597 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:09.664Z [INFO] TestOperator_RaftConfiguration: Started DNS server: address=127.0.0.1:29592 network=udp writer.go:29: 2021-01-29T19:33:09.665Z [INFO] TestOperator_RaftConfiguration: Started DNS server: address=127.0.0.1:29592 network=tcp writer.go:29: 2021-01-29T19:33:09.668Z [INFO] TestOperator_RaftConfiguration: Started HTTP server: address=127.0.0.1:29593 network=tcp writer.go:29: 2021-01-29T19:33:09.669Z [INFO] TestOperator_RaftConfiguration: started state syncer === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.677Z [WARN] TestPreparedQuery_Execute/#06.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:09.683Z [INFO] TestPreparedQuery_Execute/#06: Requesting shutdown === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:09.685Z [DEBUG] TestOperator_RaftConfiguration.server.raft: votes: needed=1 === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.686Z [INFO] TestPreparedQuery_Execute/#06.server: shutting down server === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:09.689Z [DEBUG] TestOperator_RaftConfiguration.server.raft: vote granted: from=5bf18591-1b2a-d1a5-fe89-a39ae076bd0b term=2 tally=1 === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.690Z [DEBUG] TestPreparedQuery_Execute/#06.leader: stopping routine: routine="CA root pruning" === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:09.693Z [INFO] TestOperator_RaftConfiguration.server.raft: election won: tally=1 === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.693Z [WARN] TestPreparedQuery_Execute/#06.server.serf.lan: serf: Shutdown without a Leave === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:09.696Z [INFO] TestOperator_RaftConfiguration.server.raft: entering leader state: leader="Node at 127.0.0.1:29597 [Leader]" === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.698Z [DEBUG] TestPreparedQuery_Execute/#06.leader: stopped routine: routine="CA root pruning" === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:09.701Z [INFO] TestOperator_RaftConfiguration.server: cluster leadership acquired === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.713Z [WARN] TestPreparedQuery_Execute/#06.server.serf.wan: serf: Shutdown without a Leave === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:09.719Z [INFO] TestOperator_RaftConfiguration.server: New leader elected: payload=Node-5bf18591-1b2a-d1a5-fe89-a39ae076bd0b writer.go:29: 2021-01-29T19:33:09.723Z [INFO] TestOperator_RaftConfiguration: Synced node info writer.go:29: 2021-01-29T19:33:09.730Z [DEBUG] TestOperator_RaftConfiguration.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29597 === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.733Z [INFO] TestPreparedQuery_Execute/#06.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:09.736Z [INFO] TestPreparedQuery_Execute/#06: consul server down === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.742Z [DEBUG] TestOperator_RaftPeer/#00: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.740Z [INFO] TestPreparedQuery_Execute/#06: shutdown complete === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.747Z [INFO] TestOperator_RaftPeer/#00: Synced node info === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.744Z [INFO] TestPreparedQuery_Execute/#06: Stopping server: protocol=DNS address=127.0.0.1:29580 network=tcp === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.748Z [DEBUG] TestOperator_RaftPeer/#00: Node info in sync === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:09.749Z [INFO] TestPreparedQuery_Execute/#06: Stopping server: protocol=DNS address=127.0.0.1:29580 network=udp writer.go:29: 2021-01-29T19:33:09.751Z [INFO] TestPreparedQuery_Execute/#06: Stopping server: protocol=HTTP address=127.0.0.1:29581 network=tcp === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.785Z [INFO] TestOperator_RaftPeer/#00: Requesting shutdown writer.go:29: 2021-01-29T19:33:09.790Z [INFO] TestOperator_RaftPeer/#00.server: shutting down server === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:09.796Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:09.797Z [DEBUG] TestOperator_RaftPeer/#00.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:09.805Z [WARN] TestOperator_RaftPeer/#00.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:09.806Z [DEBUG] TestOperator_RaftPeer/#00.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:09.823Z [WARN] TestOperator_RaftPeer/#00.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:09.826Z [INFO] TestOperator_RaftPeer/#00.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:09.829Z [INFO] TestOperator_RaftPeer/#00: consul server down writer.go:29: 2021-01-29T19:33:09.831Z [INFO] TestOperator_RaftPeer/#00: shutdown complete writer.go:29: 2021-01-29T19:33:09.842Z [INFO] TestOperator_RaftPeer/#00: Stopping server: protocol=DNS address=127.0.0.1:29586 network=tcp writer.go:29: 2021-01-29T19:33:09.843Z [INFO] TestOperator_RaftPeer/#00: Stopping server: protocol=DNS address=127.0.0.1:29586 network=udp writer.go:29: 2021-01-29T19:33:09.844Z [INFO] TestOperator_RaftPeer/#00: Stopping server: protocol=HTTP address=127.0.0.1:29587 network=tcp === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:09.902Z [INFO] TestOperator_RaftConfiguration.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:09.903Z [INFO] TestOperator_RaftConfiguration.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:09.904Z [DEBUG] TestOperator_RaftConfiguration.server: Skipping self join check for node since the cluster is too small: node=Node-5bf18591-1b2a-d1a5-fe89-a39ae076bd0b writer.go:29: 2021-01-29T19:33:09.905Z [INFO] TestOperator_RaftConfiguration.server: member joined, marking health alive: member=Node-5bf18591-1b2a-d1a5-fe89-a39ae076bd0b === CONT TestOperator_KeyringInstall writer.go:29: 2021-01-29T19:33:09.993Z [INFO] TestOperator_KeyringInstall: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:09.994Z [INFO] TestOperator_KeyringInstall: Endpoints down --- PASS: TestOperator_KeyringInstall (0.73s) === CONT TestKVSEndpoint_DELETE_ConflictingFlags === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:09.996Z [INFO] TestOperator_RaftConfiguration: Requesting shutdown writer.go:29: 2021-01-29T19:33:09.998Z [INFO] TestOperator_RaftConfiguration.server: shutting down server writer.go:29: 2021-01-29T19:33:10.000Z [DEBUG] TestOperator_RaftConfiguration.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:10.006Z [WARN] TestOperator_RaftConfiguration.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:10.006Z [DEBUG] TestOperator_RaftConfiguration.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:10.022Z [WARN] TestOperator_RaftConfiguration.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:10.027Z [INFO] TestOperator_RaftConfiguration.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:10.028Z [INFO] TestOperator_RaftConfiguration: consul server down writer.go:29: 2021-01-29T19:33:10.030Z [INFO] TestOperator_RaftConfiguration: shutdown complete === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.011Z [WARN] TestKVSEndpoint_DELETE_ConflictingFlags: bootstrap = true: do not enable unless necessary === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:10.031Z [INFO] TestOperator_RaftConfiguration: Stopping server: protocol=DNS address=127.0.0.1:29592 network=tcp === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.032Z [DEBUG] TestKVSEndpoint_DELETE_ConflictingFlags.tlsutil: Update: version=1 === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:10.033Z [INFO] TestOperator_RaftConfiguration: Stopping server: protocol=DNS address=127.0.0.1:29592 network=udp === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.034Z [DEBUG] TestKVSEndpoint_DELETE_ConflictingFlags.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:10.035Z [INFO] TestOperator_RaftConfiguration: Stopping server: protocol=HTTP address=127.0.0.1:29593 network=tcp === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.044Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6de0225a-e492-c446-ace0-272eb25d62af Address:127.0.0.1:29603}]" writer.go:29: 2021-01-29T19:33:10.046Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server.raft: entering follower state: follower="Node at 127.0.0.1:29603 [Follower]" leader= writer.go:29: 2021-01-29T19:33:10.046Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server.serf.wan: serf: EventMemberJoin: Node-6de0225a-e492-c446-ace0-272eb25d62af.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:10.056Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server.serf.lan: serf: EventMemberJoin: Node-6de0225a-e492-c446-ace0-272eb25d62af 127.0.0.1 writer.go:29: 2021-01-29T19:33:10.060Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server: Adding LAN server: server="Node-6de0225a-e492-c446-ace0-272eb25d62af (Addr: tcp/127.0.0.1:29603) (DC: dc1)" writer.go:29: 2021-01-29T19:33:10.061Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: Started DNS server: address=127.0.0.1:29598 network=udp writer.go:29: 2021-01-29T19:33:10.063Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: Started DNS server: address=127.0.0.1:29598 network=tcp writer.go:29: 2021-01-29T19:33:10.062Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server: Handled event for server in area: event=member-join server=Node-6de0225a-e492-c446-ace0-272eb25d62af.dc1 area=wan writer.go:29: 2021-01-29T19:33:10.065Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: Started HTTP server: address=127.0.0.1:29599 network=tcp writer.go:29: 2021-01-29T19:33:10.068Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: started state syncer writer.go:29: 2021-01-29T19:33:10.088Z [WARN] TestKVSEndpoint_DELETE_ConflictingFlags.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:10.089Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server.raft: entering candidate state: node="Node at 127.0.0.1:29603 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:10.092Z [DEBUG] TestKVSEndpoint_DELETE_ConflictingFlags.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:10.093Z [DEBUG] TestKVSEndpoint_DELETE_ConflictingFlags.server.raft: vote granted: from=6de0225a-e492-c446-ace0-272eb25d62af term=2 tally=1 writer.go:29: 2021-01-29T19:33:10.095Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:10.097Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server.raft: entering leader state: leader="Node at 127.0.0.1:29603 [Leader]" writer.go:29: 2021-01-29T19:33:10.102Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:10.103Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server: New leader elected: payload=Node-6de0225a-e492-c446-ace0-272eb25d62af writer.go:29: 2021-01-29T19:33:10.104Z [DEBUG] TestKVSEndpoint_DELETE_ConflictingFlags.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29603 writer.go:29: 2021-01-29T19:33:10.112Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:10.132Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:10.136Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:10.138Z [DEBUG] TestKVSEndpoint_DELETE_ConflictingFlags.server: Skipping self join check for node since the cluster is too small: node=Node-6de0225a-e492-c446-ace0-272eb25d62af writer.go:29: 2021-01-29T19:33:10.139Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server: member joined, marking health alive: member=Node-6de0225a-e492-c446-ace0-272eb25d62af === CONT TestPreparedQuery_Execute/#06 writer.go:29: 2021-01-29T19:33:10.253Z [INFO] TestPreparedQuery_Execute/#06: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:10.255Z [INFO] TestPreparedQuery_Execute/#06: Endpoints down === RUN TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.274Z [WARN] TestPreparedQuery_Execute/#07: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:10.276Z [DEBUG] TestPreparedQuery_Execute/#07.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:10.279Z [DEBUG] TestPreparedQuery_Execute/#07.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:10.286Z [INFO] TestPreparedQuery_Execute/#07.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5b6ebf59-54b4-206f-08a5-c4be7582c2d9 Address:127.0.0.1:29609}]" writer.go:29: 2021-01-29T19:33:10.288Z [INFO] TestPreparedQuery_Execute/#07.server.raft: entering follower state: follower="Node at 127.0.0.1:29609 [Follower]" leader= writer.go:29: 2021-01-29T19:33:10.288Z [INFO] TestPreparedQuery_Execute/#07.server.serf.wan: serf: EventMemberJoin: Node-5b6ebf59-54b4-206f-08a5-c4be7582c2d9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:10.296Z [INFO] TestPreparedQuery_Execute/#07.server.serf.lan: serf: EventMemberJoin: Node-5b6ebf59-54b4-206f-08a5-c4be7582c2d9 127.0.0.1 writer.go:29: 2021-01-29T19:33:10.305Z [INFO] TestPreparedQuery_Execute/#07.server: Adding LAN server: server="Node-5b6ebf59-54b4-206f-08a5-c4be7582c2d9 (Addr: tcp/127.0.0.1:29609) (DC: dc1)" writer.go:29: 2021-01-29T19:33:10.306Z [INFO] TestPreparedQuery_Execute/#07.server: Handled event for server in area: event=member-join server=Node-5b6ebf59-54b4-206f-08a5-c4be7582c2d9.dc1 area=wan writer.go:29: 2021-01-29T19:33:10.309Z [INFO] TestPreparedQuery_Execute/#07: Started DNS server: address=127.0.0.1:29604 network=udp writer.go:29: 2021-01-29T19:33:10.316Z [INFO] TestPreparedQuery_Execute/#07: Started DNS server: address=127.0.0.1:29604 network=tcp writer.go:29: 2021-01-29T19:33:10.327Z [INFO] TestPreparedQuery_Execute/#07: Started HTTP server: address=127.0.0.1:29605 network=tcp writer.go:29: 2021-01-29T19:33:10.330Z [INFO] TestPreparedQuery_Execute/#07: started state syncer === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.332Z [DEBUG] TestKVSEndpoint_DELETE_ConflictingFlags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:10.337Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: Synced node info === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.340Z [WARN] TestPreparedQuery_Execute/#07.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:10.341Z [INFO] TestPreparedQuery_Execute/#07.server.raft: entering candidate state: node="Node at 127.0.0.1:29609 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:10.343Z [DEBUG] TestPreparedQuery_Execute/#07.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:10.343Z [DEBUG] TestPreparedQuery_Execute/#07.server.raft: vote granted: from=5b6ebf59-54b4-206f-08a5-c4be7582c2d9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:10.344Z [INFO] TestPreparedQuery_Execute/#07.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:10.345Z [INFO] TestPreparedQuery_Execute/#07.server.raft: entering leader state: leader="Node at 127.0.0.1:29609 [Leader]" writer.go:29: 2021-01-29T19:33:10.346Z [INFO] TestPreparedQuery_Execute/#07.server: cluster leadership acquired === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:10.346Z [INFO] TestOperator_RaftPeer/#00: Waiting for endpoints to shut down === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.348Z [INFO] TestPreparedQuery_Execute/#07.server: New leader elected: payload=Node-5b6ebf59-54b4-206f-08a5-c4be7582c2d9 === CONT TestOperator_RaftPeer/#00 writer.go:29: 2021-01-29T19:33:10.347Z [INFO] TestOperator_RaftPeer/#00: Endpoints down === RUN TestOperator_RaftPeer/#01 === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.350Z [DEBUG] TestPreparedQuery_Execute/#07.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29609 === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.358Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: Requesting shutdown writer.go:29: 2021-01-29T19:33:10.359Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server: shutting down server writer.go:29: 2021-01-29T19:33:10.361Z [DEBUG] TestKVSEndpoint_DELETE_ConflictingFlags.leader: stopping routine: routine="CA root pruning" === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.362Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.363Z [WARN] TestKVSEndpoint_DELETE_ConflictingFlags.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:10.363Z [DEBUG] TestKVSEndpoint_DELETE_ConflictingFlags.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:10.366Z [WARN] TestKVSEndpoint_DELETE_ConflictingFlags.server.serf.wan: serf: Shutdown without a Leave === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.368Z [INFO] TestPreparedQuery_Execute/#07.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:10.369Z [INFO] TestPreparedQuery_Execute/#07.leader: started routine: routine="CA root pruning" === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.371Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags.server.router.manager: shutting down === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.370Z [DEBUG] TestPreparedQuery_Execute/#07.server: Skipping self join check for node since the cluster is too small: node=Node-5b6ebf59-54b4-206f-08a5-c4be7582c2d9 === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.373Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: consul server down writer.go:29: 2021-01-29T19:33:10.375Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: shutdown complete === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.374Z [INFO] TestPreparedQuery_Execute/#07.server: member joined, marking health alive: member=Node-5b6ebf59-54b4-206f-08a5-c4be7582c2d9 === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.376Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: Stopping server: protocol=DNS address=127.0.0.1:29598 network=tcp writer.go:29: 2021-01-29T19:33:10.378Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: Stopping server: protocol=DNS address=127.0.0.1:29598 network=udp writer.go:29: 2021-01-29T19:33:10.380Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: Stopping server: protocol=HTTP address=127.0.0.1:29599 network=tcp === CONT TestOperator_RaftPeer/#01 writer.go:29: 2021-01-29T19:33:10.390Z [WARN] TestOperator_RaftPeer/#01: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:10.393Z [DEBUG] TestOperator_RaftPeer/#01.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:10.397Z [DEBUG] TestOperator_RaftPeer/#01.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:10.404Z [INFO] TestOperator_RaftPeer/#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f16044ef-84e1-2555-ca3b-3917f0da6779 Address:127.0.0.1:29615}]" writer.go:29: 2021-01-29T19:33:10.408Z [INFO] TestOperator_RaftPeer/#01.server.raft: entering follower state: follower="Node at 127.0.0.1:29615 [Follower]" leader= writer.go:29: 2021-01-29T19:33:10.414Z [INFO] TestOperator_RaftPeer/#01.server.serf.wan: serf: EventMemberJoin: Node-f16044ef-84e1-2555-ca3b-3917f0da6779.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:10.421Z [INFO] TestOperator_RaftPeer/#01.server.serf.lan: serf: EventMemberJoin: Node-f16044ef-84e1-2555-ca3b-3917f0da6779 127.0.0.1 writer.go:29: 2021-01-29T19:33:10.426Z [INFO] TestOperator_RaftPeer/#01.server: Adding LAN server: server="Node-f16044ef-84e1-2555-ca3b-3917f0da6779 (Addr: tcp/127.0.0.1:29615) (DC: dc1)" writer.go:29: 2021-01-29T19:33:10.428Z [INFO] TestOperator_RaftPeer/#01.server: Handled event for server in area: event=member-join server=Node-f16044ef-84e1-2555-ca3b-3917f0da6779.dc1 area=wan writer.go:29: 2021-01-29T19:33:10.433Z [INFO] TestOperator_RaftPeer/#01: Started DNS server: address=127.0.0.1:29610 network=tcp writer.go:29: 2021-01-29T19:33:10.436Z [INFO] TestOperator_RaftPeer/#01: Started DNS server: address=127.0.0.1:29610 network=udp writer.go:29: 2021-01-29T19:33:10.438Z [INFO] TestOperator_RaftPeer/#01: Started HTTP server: address=127.0.0.1:29611 network=tcp writer.go:29: 2021-01-29T19:33:10.440Z [INFO] TestOperator_RaftPeer/#01: started state syncer writer.go:29: 2021-01-29T19:33:10.467Z [WARN] TestOperator_RaftPeer/#01.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:10.470Z [INFO] TestOperator_RaftPeer/#01.server.raft: entering candidate state: node="Node at 127.0.0.1:29615 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:10.473Z [DEBUG] TestOperator_RaftPeer/#01.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:10.475Z [DEBUG] TestOperator_RaftPeer/#01.server.raft: vote granted: from=f16044ef-84e1-2555-ca3b-3917f0da6779 term=2 tally=1 writer.go:29: 2021-01-29T19:33:10.476Z [INFO] TestOperator_RaftPeer/#01.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:10.479Z [INFO] TestOperator_RaftPeer/#01.server.raft: entering leader state: leader="Node at 127.0.0.1:29615 [Leader]" writer.go:29: 2021-01-29T19:33:10.482Z [INFO] TestOperator_RaftPeer/#01.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:10.482Z [INFO] TestOperator_RaftPeer/#01.server: New leader elected: payload=Node-f16044ef-84e1-2555-ca3b-3917f0da6779 writer.go:29: 2021-01-29T19:33:10.486Z [DEBUG] TestOperator_RaftPeer/#01.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29615 writer.go:29: 2021-01-29T19:33:10.492Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:10.499Z [INFO] TestOperator_RaftPeer/#01.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:10.500Z [INFO] TestOperator_RaftPeer/#01.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:10.502Z [DEBUG] TestOperator_RaftPeer/#01.server: Skipping self join check for node since the cluster is too small: node=Node-f16044ef-84e1-2555-ca3b-3917f0da6779 writer.go:29: 2021-01-29T19:33:10.505Z [INFO] TestOperator_RaftPeer/#01.server: member joined, marking health alive: member=Node-f16044ef-84e1-2555-ca3b-3917f0da6779 === CONT TestOperator_RaftConfiguration writer.go:29: 2021-01-29T19:33:10.538Z [INFO] TestOperator_RaftConfiguration: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:10.539Z [INFO] TestOperator_RaftConfiguration: Endpoints down --- PASS: TestOperator_RaftConfiguration (0.99s) === CONT TestKVSEndpoint_PUT_ConflictingFlags === CONT TestOperator_RaftPeer/#01 writer.go:29: 2021-01-29T19:33:10.550Z [INFO] TestOperator_RaftPeer/#01: Requesting shutdown writer.go:29: 2021-01-29T19:33:10.551Z [INFO] TestOperator_RaftPeer/#01.server: shutting down server === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.554Z [DEBUG] TestPreparedQuery_Execute/#07: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestOperator_RaftPeer/#01 writer.go:29: 2021-01-29T19:33:10.553Z [DEBUG] TestOperator_RaftPeer/#01.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:10.555Z [ERROR] TestOperator_RaftPeer/#01.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:10.556Z [WARN] TestOperator_RaftPeer/#01.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:10.556Z [DEBUG] TestOperator_RaftPeer/#01.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:10.560Z [WARN] TestOperator_RaftPeer/#01.server.serf.wan: serf: Shutdown without a Leave === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.560Z [INFO] TestPreparedQuery_Execute/#07: Synced node info writer.go:29: 2021-01-29T19:33:10.565Z [DEBUG] TestPreparedQuery_Execute/#07: Node info in sync === CONT TestOperator_RaftPeer/#01 writer.go:29: 2021-01-29T19:33:10.565Z [INFO] TestOperator_RaftPeer/#01.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:10.566Z [INFO] TestOperator_RaftPeer/#01: consul server down === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.569Z [WARN] TestKVSEndpoint_PUT_ConflictingFlags: bootstrap = true: do not enable unless necessary === CONT TestOperator_RaftPeer/#01 writer.go:29: 2021-01-29T19:33:10.572Z [INFO] TestOperator_RaftPeer/#01: shutdown complete === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.573Z [DEBUG] TestKVSEndpoint_PUT_ConflictingFlags.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:10.578Z [DEBUG] TestKVSEndpoint_PUT_ConflictingFlags.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestOperator_RaftPeer/#01 writer.go:29: 2021-01-29T19:33:10.576Z [INFO] TestOperator_RaftPeer/#01: Stopping server: protocol=DNS address=127.0.0.1:29610 network=tcp writer.go:29: 2021-01-29T19:33:10.583Z [INFO] TestOperator_RaftPeer/#01: Stopping server: protocol=DNS address=127.0.0.1:29610 network=udp writer.go:29: 2021-01-29T19:33:10.585Z [INFO] TestOperator_RaftPeer/#01: Stopping server: protocol=HTTP address=127.0.0.1:29611 network=tcp === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.592Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2e67aa44-fd66-c2e4-c20b-6cd244d86fb3 Address:127.0.0.1:29621}]" writer.go:29: 2021-01-29T19:33:10.594Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server.raft: entering follower state: follower="Node at 127.0.0.1:29621 [Follower]" leader= writer.go:29: 2021-01-29T19:33:10.594Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server.serf.wan: serf: EventMemberJoin: Node-2e67aa44-fd66-c2e4-c20b-6cd244d86fb3.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:10.600Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server.serf.lan: serf: EventMemberJoin: Node-2e67aa44-fd66-c2e4-c20b-6cd244d86fb3 127.0.0.1 writer.go:29: 2021-01-29T19:33:10.609Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server: Adding LAN server: server="Node-2e67aa44-fd66-c2e4-c20b-6cd244d86fb3 (Addr: tcp/127.0.0.1:29621) (DC: dc1)" writer.go:29: 2021-01-29T19:33:10.612Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server: Handled event for server in area: event=member-join server=Node-2e67aa44-fd66-c2e4-c20b-6cd244d86fb3.dc1 area=wan writer.go:29: 2021-01-29T19:33:10.616Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: Started DNS server: address=127.0.0.1:29616 network=tcp writer.go:29: 2021-01-29T19:33:10.623Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: Started DNS server: address=127.0.0.1:29616 network=udp writer.go:29: 2021-01-29T19:33:10.630Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: Started HTTP server: address=127.0.0.1:29617 network=tcp writer.go:29: 2021-01-29T19:33:10.631Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: started state syncer === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.662Z [INFO] TestPreparedQuery_Execute/#07: Requesting shutdown writer.go:29: 2021-01-29T19:33:10.664Z [INFO] TestPreparedQuery_Execute/#07.server: shutting down server === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.664Z [WARN] TestKVSEndpoint_PUT_ConflictingFlags.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.664Z [DEBUG] TestPreparedQuery_Execute/#07.leader: stopping routine: routine="CA root pruning" === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.665Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server.raft: entering candidate state: node="Node at 127.0.0.1:29621 [Candidate]" term=2 === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.665Z [WARN] TestPreparedQuery_Execute/#07.server.serf.lan: serf: Shutdown without a Leave === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.667Z [DEBUG] TestKVSEndpoint_PUT_ConflictingFlags.server.raft: votes: needed=1 === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.665Z [DEBUG] TestPreparedQuery_Execute/#07.leader: stopped routine: routine="CA root pruning" === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.668Z [DEBUG] TestKVSEndpoint_PUT_ConflictingFlags.server.raft: vote granted: from=2e67aa44-fd66-c2e4-c20b-6cd244d86fb3 term=2 tally=1 writer.go:29: 2021-01-29T19:33:10.668Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server.raft: election won: tally=1 === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.667Z [WARN] TestPreparedQuery_Execute/#07.server.serf.wan: serf: Shutdown without a Leave === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.669Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server.raft: entering leader state: leader="Node at 127.0.0.1:29621 [Leader]" writer.go:29: 2021-01-29T19:33:10.670Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server: cluster leadership acquired === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.671Z [INFO] TestPreparedQuery_Execute/#07.server.router.manager: shutting down === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.671Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server: New leader elected: payload=Node-2e67aa44-fd66-c2e4-c20b-6cd244d86fb3 === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.672Z [INFO] TestPreparedQuery_Execute/#07: consul server down writer.go:29: 2021-01-29T19:33:10.673Z [INFO] TestPreparedQuery_Execute/#07: shutdown complete writer.go:29: 2021-01-29T19:33:10.674Z [INFO] TestPreparedQuery_Execute/#07: Stopping server: protocol=DNS address=127.0.0.1:29604 network=tcp === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.673Z [DEBUG] TestKVSEndpoint_PUT_ConflictingFlags.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29621 === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:10.675Z [INFO] TestPreparedQuery_Execute/#07: Stopping server: protocol=DNS address=127.0.0.1:29604 network=udp writer.go:29: 2021-01-29T19:33:10.676Z [INFO] TestPreparedQuery_Execute/#07: Stopping server: protocol=HTTP address=127.0.0.1:29605 network=tcp === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.680Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:10.687Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:10.690Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:10.692Z [DEBUG] TestKVSEndpoint_PUT_ConflictingFlags.server: Skipping self join check for node since the cluster is too small: node=Node-2e67aa44-fd66-c2e4-c20b-6cd244d86fb3 writer.go:29: 2021-01-29T19:33:10.696Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server: member joined, marking health alive: member=Node-2e67aa44-fd66-c2e4-c20b-6cd244d86fb3 writer.go:29: 2021-01-29T19:33:10.749Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: Requesting shutdown writer.go:29: 2021-01-29T19:33:10.751Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server: shutting down server writer.go:29: 2021-01-29T19:33:10.753Z [DEBUG] TestKVSEndpoint_PUT_ConflictingFlags.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:10.755Z [WARN] TestKVSEndpoint_PUT_ConflictingFlags.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:10.753Z [ERROR] TestKVSEndpoint_PUT_ConflictingFlags.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:10.755Z [DEBUG] TestKVSEndpoint_PUT_ConflictingFlags.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:10.759Z [WARN] TestKVSEndpoint_PUT_ConflictingFlags.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:10.767Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:10.769Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: consul server down writer.go:29: 2021-01-29T19:33:10.771Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: shutdown complete writer.go:29: 2021-01-29T19:33:10.773Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: Stopping server: protocol=DNS address=127.0.0.1:29616 network=tcp writer.go:29: 2021-01-29T19:33:10.775Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: Stopping server: protocol=DNS address=127.0.0.1:29616 network=udp writer.go:29: 2021-01-29T19:33:10.778Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: Stopping server: protocol=HTTP address=127.0.0.1:29617 network=tcp === CONT TestKVSEndpoint_DELETE_ConflictingFlags writer.go:29: 2021-01-29T19:33:10.881Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:10.883Z [INFO] TestKVSEndpoint_DELETE_ConflictingFlags: Endpoints down --- PASS: TestKVSEndpoint_DELETE_ConflictingFlags (0.89s) === CONT TestKVSEndpoint_AcquireRelease writer.go:29: 2021-01-29T19:33:10.892Z [WARN] TestKVSEndpoint_AcquireRelease: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:10.894Z [DEBUG] TestKVSEndpoint_AcquireRelease.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:10.896Z [DEBUG] TestKVSEndpoint_AcquireRelease.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:10.901Z [INFO] TestKVSEndpoint_AcquireRelease.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:986a93f5-c07e-3097-70de-b3d68dc7f367 Address:127.0.0.1:29627}]" writer.go:29: 2021-01-29T19:33:10.904Z [INFO] TestKVSEndpoint_AcquireRelease.server.raft: entering follower state: follower="Node at 127.0.0.1:29627 [Follower]" leader= writer.go:29: 2021-01-29T19:33:10.904Z [INFO] TestKVSEndpoint_AcquireRelease.server.serf.wan: serf: EventMemberJoin: Node-986a93f5-c07e-3097-70de-b3d68dc7f367.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:10.908Z [INFO] TestKVSEndpoint_AcquireRelease.server.serf.lan: serf: EventMemberJoin: Node-986a93f5-c07e-3097-70de-b3d68dc7f367 127.0.0.1 writer.go:29: 2021-01-29T19:33:10.911Z [INFO] TestKVSEndpoint_AcquireRelease: Started DNS server: address=127.0.0.1:29622 network=udp writer.go:29: 2021-01-29T19:33:10.913Z [INFO] TestKVSEndpoint_AcquireRelease: Started DNS server: address=127.0.0.1:29622 network=tcp writer.go:29: 2021-01-29T19:33:10.912Z [INFO] TestKVSEndpoint_AcquireRelease.server: Adding LAN server: server="Node-986a93f5-c07e-3097-70de-b3d68dc7f367 (Addr: tcp/127.0.0.1:29627) (DC: dc1)" writer.go:29: 2021-01-29T19:33:10.912Z [INFO] TestKVSEndpoint_AcquireRelease.server: Handled event for server in area: event=member-join server=Node-986a93f5-c07e-3097-70de-b3d68dc7f367.dc1 area=wan writer.go:29: 2021-01-29T19:33:10.917Z [INFO] TestKVSEndpoint_AcquireRelease: Started HTTP server: address=127.0.0.1:29623 network=tcp writer.go:29: 2021-01-29T19:33:10.918Z [INFO] TestKVSEndpoint_AcquireRelease: started state syncer writer.go:29: 2021-01-29T19:33:10.972Z [WARN] TestKVSEndpoint_AcquireRelease.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:10.974Z [INFO] TestKVSEndpoint_AcquireRelease.server.raft: entering candidate state: node="Node at 127.0.0.1:29627 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:10.977Z [DEBUG] TestKVSEndpoint_AcquireRelease.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:10.979Z [DEBUG] TestKVSEndpoint_AcquireRelease.server.raft: vote granted: from=986a93f5-c07e-3097-70de-b3d68dc7f367 term=2 tally=1 writer.go:29: 2021-01-29T19:33:10.981Z [INFO] TestKVSEndpoint_AcquireRelease.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:10.983Z [INFO] TestKVSEndpoint_AcquireRelease.server.raft: entering leader state: leader="Node at 127.0.0.1:29627 [Leader]" writer.go:29: 2021-01-29T19:33:10.985Z [INFO] TestKVSEndpoint_AcquireRelease.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:10.987Z [INFO] TestKVSEndpoint_AcquireRelease.server: New leader elected: payload=Node-986a93f5-c07e-3097-70de-b3d68dc7f367 writer.go:29: 2021-01-29T19:33:10.992Z [DEBUG] TestKVSEndpoint_AcquireRelease.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29627 writer.go:29: 2021-01-29T19:33:10.998Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:11.004Z [INFO] TestKVSEndpoint_AcquireRelease.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:11.006Z [INFO] TestKVSEndpoint_AcquireRelease.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.008Z [DEBUG] TestKVSEndpoint_AcquireRelease.server: Skipping self join check for node since the cluster is too small: node=Node-986a93f5-c07e-3097-70de-b3d68dc7f367 writer.go:29: 2021-01-29T19:33:11.009Z [INFO] TestKVSEndpoint_AcquireRelease.server: member joined, marking health alive: member=Node-986a93f5-c07e-3097-70de-b3d68dc7f367 writer.go:29: 2021-01-29T19:33:11.034Z [DEBUG] TestKVSEndpoint_AcquireRelease: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:11.036Z [INFO] TestKVSEndpoint_AcquireRelease: Synced node info writer.go:29: 2021-01-29T19:33:11.068Z [INFO] TestKVSEndpoint_AcquireRelease: Requesting shutdown writer.go:29: 2021-01-29T19:33:11.069Z [INFO] TestKVSEndpoint_AcquireRelease.server: shutting down server writer.go:29: 2021-01-29T19:33:11.069Z [DEBUG] TestKVSEndpoint_AcquireRelease.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.070Z [WARN] TestKVSEndpoint_AcquireRelease.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:11.070Z [DEBUG] TestKVSEndpoint_AcquireRelease.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.072Z [WARN] TestKVSEndpoint_AcquireRelease.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:11.075Z [INFO] TestKVSEndpoint_AcquireRelease.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:11.075Z [INFO] TestKVSEndpoint_AcquireRelease: consul server down writer.go:29: 2021-01-29T19:33:11.076Z [INFO] TestKVSEndpoint_AcquireRelease: shutdown complete writer.go:29: 2021-01-29T19:33:11.077Z [INFO] TestKVSEndpoint_AcquireRelease: Stopping server: protocol=DNS address=127.0.0.1:29622 network=tcp writer.go:29: 2021-01-29T19:33:11.078Z [INFO] TestKVSEndpoint_AcquireRelease: Stopping server: protocol=DNS address=127.0.0.1:29622 network=udp writer.go:29: 2021-01-29T19:33:11.079Z [INFO] TestKVSEndpoint_AcquireRelease: Stopping server: protocol=HTTP address=127.0.0.1:29623 network=tcp === CONT TestOperator_RaftPeer/#01 writer.go:29: 2021-01-29T19:33:11.086Z [INFO] TestOperator_RaftPeer/#01: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:11.088Z [INFO] TestOperator_RaftPeer/#01: Endpoints down --- PASS: TestOperator_RaftPeer (1.70s) --- PASS: TestOperator_RaftPeer/#00 (0.96s) --- PASS: TestOperator_RaftPeer/#01 (0.74s) === CONT TestKVSEndpoint_CAS writer.go:29: 2021-01-29T19:33:11.097Z [WARN] TestKVSEndpoint_CAS: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:11.098Z [DEBUG] TestKVSEndpoint_CAS.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:11.099Z [DEBUG] TestKVSEndpoint_CAS.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:11.104Z [INFO] TestKVSEndpoint_CAS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c870c863-e55f-cc3a-61e9-7bdbdc97066a Address:127.0.0.1:29633}]" writer.go:29: 2021-01-29T19:33:11.107Z [INFO] TestKVSEndpoint_CAS.server.raft: entering follower state: follower="Node at 127.0.0.1:29633 [Follower]" leader= writer.go:29: 2021-01-29T19:33:11.113Z [INFO] TestKVSEndpoint_CAS.server.serf.wan: serf: EventMemberJoin: Node-c870c863-e55f-cc3a-61e9-7bdbdc97066a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:11.120Z [INFO] TestKVSEndpoint_CAS.server.serf.lan: serf: EventMemberJoin: Node-c870c863-e55f-cc3a-61e9-7bdbdc97066a 127.0.0.1 writer.go:29: 2021-01-29T19:33:11.125Z [INFO] TestKVSEndpoint_CAS.server: Adding LAN server: server="Node-c870c863-e55f-cc3a-61e9-7bdbdc97066a (Addr: tcp/127.0.0.1:29633) (DC: dc1)" writer.go:29: 2021-01-29T19:33:11.127Z [INFO] TestKVSEndpoint_CAS.server: Handled event for server in area: event=member-join server=Node-c870c863-e55f-cc3a-61e9-7bdbdc97066a.dc1 area=wan writer.go:29: 2021-01-29T19:33:11.131Z [INFO] TestKVSEndpoint_CAS: Started DNS server: address=127.0.0.1:29628 network=tcp writer.go:29: 2021-01-29T19:33:11.134Z [INFO] TestKVSEndpoint_CAS: Started DNS server: address=127.0.0.1:29628 network=udp writer.go:29: 2021-01-29T19:33:11.136Z [INFO] TestKVSEndpoint_CAS: Started HTTP server: address=127.0.0.1:29629 network=tcp writer.go:29: 2021-01-29T19:33:11.139Z [INFO] TestKVSEndpoint_CAS: started state syncer writer.go:29: 2021-01-29T19:33:11.153Z [WARN] TestKVSEndpoint_CAS.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:11.155Z [INFO] TestKVSEndpoint_CAS.server.raft: entering candidate state: node="Node at 127.0.0.1:29633 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:11.157Z [DEBUG] TestKVSEndpoint_CAS.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:11.159Z [DEBUG] TestKVSEndpoint_CAS.server.raft: vote granted: from=c870c863-e55f-cc3a-61e9-7bdbdc97066a term=2 tally=1 writer.go:29: 2021-01-29T19:33:11.161Z [INFO] TestKVSEndpoint_CAS.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:11.162Z [INFO] TestKVSEndpoint_CAS.server.raft: entering leader state: leader="Node at 127.0.0.1:29633 [Leader]" writer.go:29: 2021-01-29T19:33:11.164Z [INFO] TestKVSEndpoint_CAS.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:11.166Z [INFO] TestKVSEndpoint_CAS.server: New leader elected: payload=Node-c870c863-e55f-cc3a-61e9-7bdbdc97066a writer.go:29: 2021-01-29T19:33:11.169Z [DEBUG] TestKVSEndpoint_CAS.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29633 === CONT TestPreparedQuery_Execute/#07 writer.go:29: 2021-01-29T19:33:11.178Z [INFO] TestPreparedQuery_Execute/#07: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:11.179Z [INFO] TestPreparedQuery_Execute/#07: Endpoints down --- PASS: TestPreparedQuery_Execute (7.20s) --- PASS: TestPreparedQuery_Execute/#00 (0.89s) --- PASS: TestPreparedQuery_Execute/#01 (0.83s) --- PASS: TestPreparedQuery_Execute/#02 (0.96s) --- PASS: TestPreparedQuery_Execute/#03 (0.93s) --- PASS: TestPreparedQuery_Execute/#04 (0.68s) --- PASS: TestPreparedQuery_Execute/#05 (0.97s) --- PASS: TestPreparedQuery_Execute/#06 (1.00s) --- PASS: TestPreparedQuery_Execute/#07 (0.92s) === CONT TestIntentionsCreate_good === CONT TestKVSEndpoint_CAS writer.go:29: 2021-01-29T19:33:11.176Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:11.197Z [INFO] TestKVSEndpoint_CAS.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:11.199Z [INFO] TestKVSEndpoint_CAS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.201Z [DEBUG] TestKVSEndpoint_CAS.server: Skipping self join check for node since the cluster is too small: node=Node-c870c863-e55f-cc3a-61e9-7bdbdc97066a writer.go:29: 2021-01-29T19:33:11.202Z [INFO] TestKVSEndpoint_CAS.server: member joined, marking health alive: member=Node-c870c863-e55f-cc3a-61e9-7bdbdc97066a === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:11.206Z [WARN] TestIntentionsCreate_good: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:11.208Z [DEBUG] TestIntentionsCreate_good.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:11.210Z [DEBUG] TestIntentionsCreate_good.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:11.217Z [INFO] TestIntentionsCreate_good.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:61604b19-bbf3-fd46-4842-d3813163f0a1 Address:127.0.0.1:29639}]" writer.go:29: 2021-01-29T19:33:11.219Z [INFO] TestIntentionsCreate_good.server.raft: entering follower state: follower="Node at 127.0.0.1:29639 [Follower]" leader= writer.go:29: 2021-01-29T19:33:11.232Z [INFO] TestIntentionsCreate_good.server.serf.wan: serf: EventMemberJoin: Node-61604b19-bbf3-fd46-4842-d3813163f0a1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:11.234Z [INFO] TestIntentionsCreate_good.server.serf.lan: serf: EventMemberJoin: Node-61604b19-bbf3-fd46-4842-d3813163f0a1 127.0.0.1 writer.go:29: 2021-01-29T19:33:11.237Z [INFO] TestIntentionsCreate_good.server: Adding LAN server: server="Node-61604b19-bbf3-fd46-4842-d3813163f0a1 (Addr: tcp/127.0.0.1:29639) (DC: dc1)" writer.go:29: 2021-01-29T19:33:11.238Z [INFO] TestIntentionsCreate_good.server: Handled event for server in area: event=member-join server=Node-61604b19-bbf3-fd46-4842-d3813163f0a1.dc1 area=wan writer.go:29: 2021-01-29T19:33:11.237Z [INFO] TestIntentionsCreate_good: Started DNS server: address=127.0.0.1:29634 network=udp writer.go:29: 2021-01-29T19:33:11.241Z [INFO] TestIntentionsCreate_good: Started DNS server: address=127.0.0.1:29634 network=tcp writer.go:29: 2021-01-29T19:33:11.244Z [INFO] TestIntentionsCreate_good: Started HTTP server: address=127.0.0.1:29635 network=tcp writer.go:29: 2021-01-29T19:33:11.245Z [INFO] TestIntentionsCreate_good: started state syncer writer.go:29: 2021-01-29T19:33:11.260Z [WARN] TestIntentionsCreate_good.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:11.262Z [INFO] TestIntentionsCreate_good.server.raft: entering candidate state: node="Node at 127.0.0.1:29639 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:11.264Z [DEBUG] TestIntentionsCreate_good.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:11.266Z [DEBUG] TestIntentionsCreate_good.server.raft: vote granted: from=61604b19-bbf3-fd46-4842-d3813163f0a1 term=2 tally=1 writer.go:29: 2021-01-29T19:33:11.268Z [INFO] TestIntentionsCreate_good.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:11.269Z [INFO] TestIntentionsCreate_good.server.raft: entering leader state: leader="Node at 127.0.0.1:29639 [Leader]" writer.go:29: 2021-01-29T19:33:11.271Z [INFO] TestIntentionsCreate_good.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:11.272Z [INFO] TestIntentionsCreate_good.server: New leader elected: payload=Node-61604b19-bbf3-fd46-4842-d3813163f0a1 writer.go:29: 2021-01-29T19:33:11.277Z [DEBUG] TestIntentionsCreate_good.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29639 === CONT TestKVSEndpoint_PUT_ConflictingFlags writer.go:29: 2021-01-29T19:33:11.280Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:11.283Z [INFO] TestKVSEndpoint_PUT_ConflictingFlags: Endpoints down === CONT TestKVSEndpoint_DELETE_CAS --- PASS: TestKVSEndpoint_PUT_ConflictingFlags (0.75s) === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:11.301Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.307Z [WARN] TestKVSEndpoint_DELETE_CAS: bootstrap = true: do not enable unless necessary === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:11.310Z [INFO] TestIntentionsCreate_good.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.310Z [DEBUG] TestKVSEndpoint_DELETE_CAS.tlsutil: Update: version=1 === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:11.311Z [INFO] TestIntentionsCreate_good.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.312Z [DEBUG] TestIntentionsCreate_good.server: Skipping self join check for node since the cluster is too small: node=Node-61604b19-bbf3-fd46-4842-d3813163f0a1 writer.go:29: 2021-01-29T19:33:11.313Z [INFO] TestIntentionsCreate_good.server: member joined, marking health alive: member=Node-61604b19-bbf3-fd46-4842-d3813163f0a1 === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.312Z [DEBUG] TestKVSEndpoint_DELETE_CAS.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:11.317Z [INFO] TestKVSEndpoint_DELETE_CAS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9df213a7-b2d4-e7ef-5fab-0695d084fd8d Address:127.0.0.1:29645}]" writer.go:29: 2021-01-29T19:33:11.320Z [INFO] TestKVSEndpoint_DELETE_CAS.server.raft: entering follower state: follower="Node at 127.0.0.1:29645 [Follower]" leader= writer.go:29: 2021-01-29T19:33:11.322Z [INFO] TestKVSEndpoint_DELETE_CAS.server.serf.wan: serf: EventMemberJoin: Node-9df213a7-b2d4-e7ef-5fab-0695d084fd8d.dc1 127.0.0.1 === CONT TestKVSEndpoint_CAS writer.go:29: 2021-01-29T19:33:11.327Z [DEBUG] TestKVSEndpoint_CAS: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.336Z [INFO] TestKVSEndpoint_DELETE_CAS.server.serf.lan: serf: EventMemberJoin: Node-9df213a7-b2d4-e7ef-5fab-0695d084fd8d 127.0.0.1 === CONT TestKVSEndpoint_CAS writer.go:29: 2021-01-29T19:33:11.340Z [INFO] TestKVSEndpoint_CAS: Synced node info writer.go:29: 2021-01-29T19:33:11.341Z [DEBUG] TestKVSEndpoint_CAS: Node info in sync === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.344Z [INFO] TestKVSEndpoint_DELETE_CAS: Started DNS server: address=127.0.0.1:29640 network=udp writer.go:29: 2021-01-29T19:33:11.347Z [INFO] TestKVSEndpoint_DELETE_CAS: Started DNS server: address=127.0.0.1:29640 network=tcp writer.go:29: 2021-01-29T19:33:11.350Z [INFO] TestKVSEndpoint_DELETE_CAS.server: Adding LAN server: server="Node-9df213a7-b2d4-e7ef-5fab-0695d084fd8d (Addr: tcp/127.0.0.1:29645) (DC: dc1)" writer.go:29: 2021-01-29T19:33:11.354Z [INFO] TestKVSEndpoint_DELETE_CAS: Started HTTP server: address=127.0.0.1:29641 network=tcp writer.go:29: 2021-01-29T19:33:11.355Z [INFO] TestKVSEndpoint_DELETE_CAS: started state syncer === CONT TestKVSEndpoint_CAS writer.go:29: 2021-01-29T19:33:11.358Z [INFO] TestKVSEndpoint_CAS: Requesting shutdown writer.go:29: 2021-01-29T19:33:11.359Z [INFO] TestKVSEndpoint_CAS.server: shutting down server writer.go:29: 2021-01-29T19:33:11.359Z [DEBUG] TestKVSEndpoint_CAS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.360Z [WARN] TestKVSEndpoint_CAS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:11.362Z [DEBUG] TestKVSEndpoint_CAS.leader: stopped routine: routine="CA root pruning" === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.372Z [INFO] TestKVSEndpoint_DELETE_CAS.server: Handled event for server in area: event=member-join server=Node-9df213a7-b2d4-e7ef-5fab-0695d084fd8d.dc1 area=wan === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:11.376Z [DEBUG] TestIntentionsCreate_good: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestKVSEndpoint_CAS writer.go:29: 2021-01-29T19:33:11.381Z [WARN] TestKVSEndpoint_CAS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:11.384Z [INFO] TestKVSEndpoint_CAS.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:11.386Z [INFO] TestKVSEndpoint_CAS: consul server down writer.go:29: 2021-01-29T19:33:11.387Z [INFO] TestKVSEndpoint_CAS: shutdown complete writer.go:29: 2021-01-29T19:33:11.387Z [INFO] TestKVSEndpoint_CAS: Stopping server: protocol=DNS address=127.0.0.1:29628 network=tcp writer.go:29: 2021-01-29T19:33:11.388Z [INFO] TestKVSEndpoint_CAS: Stopping server: protocol=DNS address=127.0.0.1:29628 network=udp writer.go:29: 2021-01-29T19:33:11.389Z [INFO] TestKVSEndpoint_CAS: Stopping server: protocol=HTTP address=127.0.0.1:29629 network=tcp === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.390Z [WARN] TestKVSEndpoint_DELETE_CAS.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:11.391Z [INFO] TestKVSEndpoint_DELETE_CAS.server.raft: entering candidate state: node="Node at 127.0.0.1:29645 [Candidate]" term=2 === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:11.401Z [INFO] TestIntentionsCreate_good: Synced node info === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.407Z [DEBUG] TestKVSEndpoint_DELETE_CAS.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:11.410Z [DEBUG] TestKVSEndpoint_DELETE_CAS.server.raft: vote granted: from=9df213a7-b2d4-e7ef-5fab-0695d084fd8d term=2 tally=1 writer.go:29: 2021-01-29T19:33:11.413Z [INFO] TestKVSEndpoint_DELETE_CAS.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:11.418Z [INFO] TestKVSEndpoint_DELETE_CAS.server.raft: entering leader state: leader="Node at 127.0.0.1:29645 [Leader]" writer.go:29: 2021-01-29T19:33:11.421Z [INFO] TestKVSEndpoint_DELETE_CAS.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:11.423Z [INFO] TestKVSEndpoint_DELETE_CAS.server: New leader elected: payload=Node-9df213a7-b2d4-e7ef-5fab-0695d084fd8d writer.go:29: 2021-01-29T19:33:11.425Z [DEBUG] TestKVSEndpoint_DELETE_CAS.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29645 === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:11.446Z [DEBUG] TestIntentionsCreate_good: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:11.450Z [DEBUG] TestIntentionsCreate_good: Node info in sync writer.go:29: 2021-01-29T19:33:11.451Z [DEBUG] TestIntentionsCreate_good: Node info in sync === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.475Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:11.505Z [INFO] TestIntentionsCreate_good: Requesting shutdown writer.go:29: 2021-01-29T19:33:11.507Z [INFO] TestIntentionsCreate_good.server: shutting down server writer.go:29: 2021-01-29T19:33:11.508Z [DEBUG] TestIntentionsCreate_good.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.510Z [WARN] TestIntentionsCreate_good.server.serf.lan: serf: Shutdown without a Leave === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.509Z [INFO] TestKVSEndpoint_DELETE_CAS.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:11.512Z [DEBUG] TestIntentionsCreate_good.leader: stopped routine: routine="CA root pruning" === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.512Z [INFO] TestKVSEndpoint_DELETE_CAS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.517Z [DEBUG] TestKVSEndpoint_DELETE_CAS.server: Skipping self join check for node since the cluster is too small: node=Node-9df213a7-b2d4-e7ef-5fab-0695d084fd8d writer.go:29: 2021-01-29T19:33:11.520Z [INFO] TestKVSEndpoint_DELETE_CAS.server: member joined, marking health alive: member=Node-9df213a7-b2d4-e7ef-5fab-0695d084fd8d === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:11.532Z [WARN] TestIntentionsCreate_good.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:11.540Z [INFO] TestIntentionsCreate_good.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:11.542Z [INFO] TestIntentionsCreate_good: consul server down writer.go:29: 2021-01-29T19:33:11.554Z [INFO] TestIntentionsCreate_good: shutdown complete writer.go:29: 2021-01-29T19:33:11.555Z [INFO] TestIntentionsCreate_good: Stopping server: protocol=DNS address=127.0.0.1:29634 network=tcp writer.go:29: 2021-01-29T19:33:11.557Z [INFO] TestIntentionsCreate_good: Stopping server: protocol=DNS address=127.0.0.1:29634 network=udp writer.go:29: 2021-01-29T19:33:11.558Z [INFO] TestIntentionsCreate_good: Stopping server: protocol=HTTP address=127.0.0.1:29635 network=tcp === CONT TestKVSEndpoint_AcquireRelease writer.go:29: 2021-01-29T19:33:11.582Z [INFO] TestKVSEndpoint_AcquireRelease: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:11.583Z [INFO] TestKVSEndpoint_AcquireRelease: Endpoints down --- PASS: TestKVSEndpoint_AcquireRelease (0.70s) === CONT TestKVSEndpoint_Recurse writer.go:29: 2021-01-29T19:33:11.662Z [WARN] TestKVSEndpoint_Recurse: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:11.663Z [DEBUG] TestKVSEndpoint_Recurse.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:11.664Z [DEBUG] TestKVSEndpoint_Recurse.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:11.669Z [INFO] TestKVSEndpoint_Recurse.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a2ab82ea-a147-2c0f-8f41-2cca5a4c501e Address:127.0.0.1:29651}]" === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.673Z [INFO] TestKVSEndpoint_DELETE_CAS: Requesting shutdown writer.go:29: 2021-01-29T19:33:11.674Z [INFO] TestKVSEndpoint_DELETE_CAS.server: shutting down server writer.go:29: 2021-01-29T19:33:11.675Z [DEBUG] TestKVSEndpoint_DELETE_CAS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.677Z [WARN] TestKVSEndpoint_DELETE_CAS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:11.676Z [ERROR] TestKVSEndpoint_DELETE_CAS.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestKVSEndpoint_Recurse writer.go:29: 2021-01-29T19:33:11.672Z [INFO] TestKVSEndpoint_Recurse.server.raft: entering follower state: follower="Node at 127.0.0.1:29651 [Follower]" leader= writer.go:29: 2021-01-29T19:33:11.672Z [INFO] TestKVSEndpoint_Recurse.server.serf.wan: serf: EventMemberJoin: Node-a2ab82ea-a147-2c0f-8f41-2cca5a4c501e.dc1 127.0.0.1 === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.678Z [DEBUG] TestKVSEndpoint_DELETE_CAS.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.682Z [WARN] TestKVSEndpoint_DELETE_CAS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:11.689Z [INFO] TestKVSEndpoint_DELETE_CAS.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:11.691Z [INFO] TestKVSEndpoint_DELETE_CAS: consul server down writer.go:29: 2021-01-29T19:33:11.692Z [INFO] TestKVSEndpoint_DELETE_CAS: shutdown complete === CONT TestKVSEndpoint_Recurse writer.go:29: 2021-01-29T19:33:11.692Z [INFO] TestKVSEndpoint_Recurse.server.serf.lan: serf: EventMemberJoin: Node-a2ab82ea-a147-2c0f-8f41-2cca5a4c501e 127.0.0.1 === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:11.693Z [INFO] TestKVSEndpoint_DELETE_CAS: Stopping server: protocol=DNS address=127.0.0.1:29640 network=tcp writer.go:29: 2021-01-29T19:33:11.695Z [INFO] TestKVSEndpoint_DELETE_CAS: Stopping server: protocol=DNS address=127.0.0.1:29640 network=udp writer.go:29: 2021-01-29T19:33:11.696Z [INFO] TestKVSEndpoint_DELETE_CAS: Stopping server: protocol=HTTP address=127.0.0.1:29641 network=tcp === CONT TestKVSEndpoint_Recurse writer.go:29: 2021-01-29T19:33:11.697Z [INFO] TestKVSEndpoint_Recurse.server: Adding LAN server: server="Node-a2ab82ea-a147-2c0f-8f41-2cca5a4c501e (Addr: tcp/127.0.0.1:29651) (DC: dc1)" writer.go:29: 2021-01-29T19:33:11.698Z [INFO] TestKVSEndpoint_Recurse.server: Handled event for server in area: event=member-join server=Node-a2ab82ea-a147-2c0f-8f41-2cca5a4c501e.dc1 area=wan writer.go:29: 2021-01-29T19:33:11.700Z [INFO] TestKVSEndpoint_Recurse: Started DNS server: address=127.0.0.1:29646 network=udp writer.go:29: 2021-01-29T19:33:11.701Z [INFO] TestKVSEndpoint_Recurse: Started DNS server: address=127.0.0.1:29646 network=tcp writer.go:29: 2021-01-29T19:33:11.703Z [INFO] TestKVSEndpoint_Recurse: Started HTTP server: address=127.0.0.1:29647 network=tcp writer.go:29: 2021-01-29T19:33:11.704Z [INFO] TestKVSEndpoint_Recurse: started state syncer writer.go:29: 2021-01-29T19:33:11.718Z [WARN] TestKVSEndpoint_Recurse.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:11.720Z [INFO] TestKVSEndpoint_Recurse.server.raft: entering candidate state: node="Node at 127.0.0.1:29651 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:11.721Z [DEBUG] TestKVSEndpoint_Recurse.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:11.722Z [DEBUG] TestKVSEndpoint_Recurse.server.raft: vote granted: from=a2ab82ea-a147-2c0f-8f41-2cca5a4c501e term=2 tally=1 writer.go:29: 2021-01-29T19:33:11.723Z [INFO] TestKVSEndpoint_Recurse.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:11.724Z [INFO] TestKVSEndpoint_Recurse.server.raft: entering leader state: leader="Node at 127.0.0.1:29651 [Leader]" writer.go:29: 2021-01-29T19:33:11.725Z [INFO] TestKVSEndpoint_Recurse.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:11.727Z [INFO] TestKVSEndpoint_Recurse.server: New leader elected: payload=Node-a2ab82ea-a147-2c0f-8f41-2cca5a4c501e writer.go:29: 2021-01-29T19:33:11.732Z [DEBUG] TestKVSEndpoint_Recurse.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29651 writer.go:29: 2021-01-29T19:33:11.741Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:11.747Z [INFO] TestKVSEndpoint_Recurse.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:11.748Z [INFO] TestKVSEndpoint_Recurse.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:11.749Z [DEBUG] TestKVSEndpoint_Recurse.server: Skipping self join check for node since the cluster is too small: node=Node-a2ab82ea-a147-2c0f-8f41-2cca5a4c501e writer.go:29: 2021-01-29T19:33:11.750Z [INFO] TestKVSEndpoint_Recurse.server: member joined, marking health alive: member=Node-a2ab82ea-a147-2c0f-8f41-2cca5a4c501e === CONT TestKVSEndpoint_CAS writer.go:29: 2021-01-29T19:33:11.891Z [INFO] TestKVSEndpoint_CAS: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:11.893Z [INFO] TestKVSEndpoint_CAS: Endpoints down --- PASS: TestKVSEndpoint_CAS (0.80s) === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:11.906Z [WARN] TestKVSEndpoint_PUT_GET_DELETE: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:11.908Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:11.911Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:11.916Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ea444f0a-da40-78e2-e15c-2b23384bc12d Address:127.0.0.1:29657}]" writer.go:29: 2021-01-29T19:33:11.919Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server.raft: entering follower state: follower="Node at 127.0.0.1:29657 [Follower]" leader= writer.go:29: 2021-01-29T19:33:11.923Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server.serf.wan: serf: EventMemberJoin: Node-ea444f0a-da40-78e2-e15c-2b23384bc12d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:11.927Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server.serf.lan: serf: EventMemberJoin: Node-ea444f0a-da40-78e2-e15c-2b23384bc12d 127.0.0.1 writer.go:29: 2021-01-29T19:33:11.933Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server: Handled event for server in area: event=member-join server=Node-ea444f0a-da40-78e2-e15c-2b23384bc12d.dc1 area=wan writer.go:29: 2021-01-29T19:33:11.933Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server: Adding LAN server: server="Node-ea444f0a-da40-78e2-e15c-2b23384bc12d (Addr: tcp/127.0.0.1:29657) (DC: dc1)" === CONT TestKVSEndpoint_Recurse writer.go:29: 2021-01-29T19:33:11.939Z [DEBUG] TestKVSEndpoint_Recurse: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:11.935Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: Started DNS server: address=127.0.0.1:29652 network=tcp === CONT TestKVSEndpoint_Recurse writer.go:29: 2021-01-29T19:33:11.942Z [INFO] TestKVSEndpoint_Recurse: Synced node info === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:11.941Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: Started DNS server: address=127.0.0.1:29652 network=udp writer.go:29: 2021-01-29T19:33:11.949Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: Started HTTP server: address=127.0.0.1:29653 network=tcp writer.go:29: 2021-01-29T19:33:11.950Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: started state syncer writer.go:29: 2021-01-29T19:33:11.990Z [WARN] TestKVSEndpoint_PUT_GET_DELETE.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:11.991Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server.raft: entering candidate state: node="Node at 127.0.0.1:29657 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:11.993Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:11.995Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE.server.raft: vote granted: from=ea444f0a-da40-78e2-e15c-2b23384bc12d term=2 tally=1 writer.go:29: 2021-01-29T19:33:11.996Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:11.997Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server.raft: entering leader state: leader="Node at 127.0.0.1:29657 [Leader]" writer.go:29: 2021-01-29T19:33:11.999Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:11.999Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server: New leader elected: payload=Node-ea444f0a-da40-78e2-e15c-2b23384bc12d writer.go:29: 2021-01-29T19:33:12.003Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29657 writer.go:29: 2021-01-29T19:33:12.007Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:12.012Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:12.014Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.015Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE.server: Skipping self join check for node since the cluster is too small: node=Node-ea444f0a-da40-78e2-e15c-2b23384bc12d writer.go:29: 2021-01-29T19:33:12.017Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server: member joined, marking health alive: member=Node-ea444f0a-da40-78e2-e15c-2b23384bc12d === CONT TestKVSEndpoint_Recurse writer.go:29: 2021-01-29T19:33:12.038Z [INFO] TestKVSEndpoint_Recurse: Requesting shutdown writer.go:29: 2021-01-29T19:33:12.039Z [INFO] TestKVSEndpoint_Recurse.server: shutting down server writer.go:29: 2021-01-29T19:33:12.040Z [DEBUG] TestKVSEndpoint_Recurse.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.042Z [WARN] TestKVSEndpoint_Recurse.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:12.043Z [DEBUG] TestKVSEndpoint_Recurse.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.047Z [WARN] TestKVSEndpoint_Recurse.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:12.050Z [INFO] TestKVSEndpoint_Recurse.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:12.050Z [INFO] TestKVSEndpoint_Recurse: consul server down writer.go:29: 2021-01-29T19:33:12.054Z [INFO] TestKVSEndpoint_Recurse: shutdown complete writer.go:29: 2021-01-29T19:33:12.055Z [INFO] TestKVSEndpoint_Recurse: Stopping server: protocol=DNS address=127.0.0.1:29646 network=tcp writer.go:29: 2021-01-29T19:33:12.057Z [INFO] TestKVSEndpoint_Recurse: Stopping server: protocol=DNS address=127.0.0.1:29646 network=udp === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:12.060Z [INFO] TestIntentionsCreate_good: Waiting for endpoints to shut down === CONT TestKVSEndpoint_Recurse writer.go:29: 2021-01-29T19:33:12.058Z [INFO] TestKVSEndpoint_Recurse: Stopping server: protocol=HTTP address=127.0.0.1:29647 network=tcp === CONT TestIntentionsCreate_good writer.go:29: 2021-01-29T19:33:12.061Z [INFO] TestIntentionsCreate_good: Endpoints down --- PASS: TestIntentionsCreate_good (0.88s) === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.070Z [WARN] TestAgentKeyring_ACL: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:12.071Z [WARN] TestAgentKeyring_ACL: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:12.075Z [DEBUG] TestAgentKeyring_ACL.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:12.079Z [DEBUG] TestAgentKeyring_ACL.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.092Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:12.094Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: Synced node info === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.095Z [INFO] TestAgentKeyring_ACL.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7bef7f2a-97f2-ebe6-de99-f981a6820feb Address:127.0.0.1:29663}]" === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.095Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE: Node info in sync === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.096Z [INFO] TestAgentKeyring_ACL.server.raft: entering follower state: follower="Node at 127.0.0.1:29663 [Follower]" leader= writer.go:29: 2021-01-29T19:33:12.097Z [INFO] TestAgentKeyring_ACL.server.serf.wan: serf: EventMemberJoin: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:12.100Z [INFO] TestAgentKeyring_ACL.server.serf.lan: serf: EventMemberJoin: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb 127.0.0.1 writer.go:29: 2021-01-29T19:33:12.105Z [INFO] TestAgentKeyring_ACL.server: Handled event for server in area: event=member-join server=Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb.dc1 area=wan writer.go:29: 2021-01-29T19:33:12.105Z [INFO] TestAgentKeyring_ACL.server: Adding LAN server: server="Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb (Addr: tcp/127.0.0.1:29663) (DC: dc1)" writer.go:29: 2021-01-29T19:33:12.107Z [INFO] TestAgentKeyring_ACL: Started DNS server: address=127.0.0.1:29658 network=udp writer.go:29: 2021-01-29T19:33:12.109Z [INFO] TestAgentKeyring_ACL: Started DNS server: address=127.0.0.1:29658 network=tcp writer.go:29: 2021-01-29T19:33:12.111Z [INFO] TestAgentKeyring_ACL: Started HTTP server: address=127.0.0.1:29659 network=tcp writer.go:29: 2021-01-29T19:33:12.113Z [INFO] TestAgentKeyring_ACL: started state syncer writer.go:29: 2021-01-29T19:33:12.154Z [WARN] TestAgentKeyring_ACL.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:12.155Z [INFO] TestAgentKeyring_ACL.server.raft: entering candidate state: node="Node at 127.0.0.1:29663 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:12.156Z [DEBUG] TestAgentKeyring_ACL.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:33:12.158Z [DEBUG] TestAgentKeyring_ACL.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:12.159Z [DEBUG] TestAgentKeyring_ACL.server.raft: vote granted: from=7bef7f2a-97f2-ebe6-de99-f981a6820feb term=2 tally=1 writer.go:29: 2021-01-29T19:33:12.161Z [INFO] TestAgentKeyring_ACL.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:12.163Z [INFO] TestAgentKeyring_ACL.server.raft: entering leader state: leader="Node at 127.0.0.1:29663 [Leader]" writer.go:29: 2021-01-29T19:33:12.165Z [INFO] TestAgentKeyring_ACL.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:12.165Z [INFO] TestAgentKeyring_ACL.server: New leader elected: payload=Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb writer.go:29: 2021-01-29T19:33:12.172Z [INFO] TestAgentKeyring_ACL.server: initializing acls writer.go:29: 2021-01-29T19:33:12.178Z [INFO] TestAgentKeyring_ACL.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:12.190Z [WARN] TestAgentKeyring_ACL.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:33:12.192Z [INFO] TestAgentKeyring_ACL.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:33:12.194Z [INFO] TestAgentKeyring_ACL.server: Created ACL anonymous token from configuration === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:12.198Z [INFO] TestKVSEndpoint_DELETE_CAS: Waiting for endpoints to shut down === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.195Z [INFO] TestAgentKeyring_ACL.leader: started routine: routine="legacy ACL token upgrade" === CONT TestKVSEndpoint_DELETE_CAS writer.go:29: 2021-01-29T19:33:12.199Z [INFO] TestKVSEndpoint_DELETE_CAS: Endpoints down --- PASS: TestKVSEndpoint_DELETE_CAS (0.91s) === CONT TestAgent_InitKeyring --- PASS: TestAgent_InitKeyring (0.00s) === CONT TestAgent_InmemKeyrings === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.200Z [INFO] TestAgentKeyring_ACL.leader: started routine: routine="acl token reaping" === RUN TestAgent_InmemKeyrings/no_keys === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.199Z [INFO] TestAgentKeyring_ACL.server.serf.wan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:12.201Z [INFO] TestAgentKeyring_ACL.server.serf.lan: serf: EventMemberUpdate: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb writer.go:29: 2021-01-29T19:33:12.204Z [INFO] TestAgentKeyring_ACL.server.serf.wan: serf: EventMemberUpdate: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb.dc1 writer.go:29: 2021-01-29T19:33:12.204Z [INFO] TestAgentKeyring_ACL.server: Updating LAN server: server="Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb (Addr: tcp/127.0.0.1:29663) (DC: dc1)" writer.go:29: 2021-01-29T19:33:12.206Z [INFO] TestAgentKeyring_ACL.server: Handled event for server in area: event=member-update server=Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb.dc1 area=wan === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.209Z [WARN] TestAgent_InmemKeyrings/no_keys: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:12.210Z [DEBUG] TestAgent_InmemKeyrings/no_keys.tlsutil: Update: version=1 === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.225Z [DEBUG] TestAgentKeyring_ACL.server.serf.wan: serf: messageQueryResponseType: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb.dc1 writer.go:29: 2021-01-29T19:33:12.230Z [DEBUG] TestAgentKeyring_ACL.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:12.233Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.242Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.240Z [INFO] TestAgentKeyring_ACL.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.242Z [DEBUG] TestAgent_InmemKeyrings/no_keys.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.243Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE: Node info in sync === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.243Z [INFO] TestAgentKeyring_ACL.leader: started routine: routine="CA root pruning" === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.250Z [INFO] TestAgent_InmemKeyrings/no_keys.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9cbae4ed-99a3-d3c4-9611-ce54169f2e2a Address:127.0.0.1:29669}]" === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.253Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: Requesting shutdown === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.253Z [INFO] TestAgent_InmemKeyrings/no_keys.server.serf.wan: serf: EventMemberJoin: Node-9cbae4ed-99a3-d3c4-9611-ce54169f2e2a.dc1 127.0.0.1 === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.254Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server: shutting down server writer.go:29: 2021-01-29T19:33:12.256Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.258Z [WARN] TestKVSEndpoint_PUT_GET_DELETE.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.258Z [INFO] TestAgent_InmemKeyrings/no_keys.server.serf.lan: serf: EventMemberJoin: Node-9cbae4ed-99a3-d3c4-9611-ce54169f2e2a 127.0.0.1 === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.260Z [DEBUG] TestKVSEndpoint_PUT_GET_DELETE.leader: stopped routine: routine="CA root pruning" === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.248Z [DEBUG] TestAgentKeyring_ACL.server: Skipping self join check for node since the cluster is too small: node=Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb writer.go:29: 2021-01-29T19:33:12.265Z [INFO] TestAgentKeyring_ACL.server: member joined, marking health alive: member=Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.273Z [INFO] TestAgent_InmemKeyrings/no_keys: Started DNS server: address=127.0.0.1:29664 network=udp === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.275Z [WARN] TestKVSEndpoint_PUT_GET_DELETE.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.269Z [INFO] TestAgentKeyring_ACL.server.serf.lan: serf: Received list-keys query === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.279Z [INFO] TestAgent_InmemKeyrings/no_keys: Started DNS server: address=127.0.0.1:29664 network=tcp === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.284Z [DEBUG] TestAgentKeyring_ACL.server.serf.lan: serf: messageQueryResponseType: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.280Z [INFO] TestAgent_InmemKeyrings/no_keys.server: Handled event for server in area: event=member-join server=Node-9cbae4ed-99a3-d3c4-9611-ce54169f2e2a.dc1 area=wan === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.288Z [DEBUG] TestAgentKeyring_ACL.server: Skipping self join check for node since the cluster is too small: node=Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.281Z [INFO] TestAgent_InmemKeyrings/no_keys.server: Adding LAN server: server="Node-9cbae4ed-99a3-d3c4-9611-ce54169f2e2a (Addr: tcp/127.0.0.1:29669) (DC: dc1)" writer.go:29: 2021-01-29T19:33:12.285Z [INFO] TestAgent_InmemKeyrings/no_keys.server.raft: entering follower state: follower="Node at 127.0.0.1:29669 [Follower]" leader= === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.296Z [INFO] TestKVSEndpoint_PUT_GET_DELETE.server.router.manager: shutting down === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.294Z [INFO] TestAgent_InmemKeyrings/no_keys: Started HTTP server: address=127.0.0.1:29665 network=tcp === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.299Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: consul server down === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.314Z [INFO] TestAgent_InmemKeyrings/no_keys: started state syncer === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.314Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: shutdown complete === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.295Z [INFO] TestAgentKeyring_ACL.server.serf.wan: serf: Received install-key query === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.315Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: Stopping server: protocol=DNS address=127.0.0.1:29652 network=tcp writer.go:29: 2021-01-29T19:33:12.317Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: Stopping server: protocol=DNS address=127.0.0.1:29652 network=udp writer.go:29: 2021-01-29T19:33:12.320Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: Stopping server: protocol=HTTP address=127.0.0.1:29653 network=tcp === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.320Z [DEBUG] TestAgentKeyring_ACL.server.serf.wan: serf: messageQueryResponseType: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb.dc1 writer.go:29: 2021-01-29T19:33:12.327Z [INFO] TestAgentKeyring_ACL.server.serf.lan: serf: Received install-key query writer.go:29: 2021-01-29T19:33:12.329Z [DEBUG] TestAgentKeyring_ACL.server.serf.lan: serf: messageQueryResponseType: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb writer.go:29: 2021-01-29T19:33:12.331Z [INFO] TestAgentKeyring_ACL.server.serf.wan: serf: Received use-key query writer.go:29: 2021-01-29T19:33:12.333Z [DEBUG] TestAgentKeyring_ACL.server.serf.wan: serf: messageQueryResponseType: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb.dc1 writer.go:29: 2021-01-29T19:33:12.335Z [INFO] TestAgentKeyring_ACL.server.serf.lan: serf: Received use-key query writer.go:29: 2021-01-29T19:33:12.336Z [DEBUG] TestAgentKeyring_ACL.server.serf.lan: serf: messageQueryResponseType: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb writer.go:29: 2021-01-29T19:33:12.338Z [INFO] TestAgentKeyring_ACL.server.serf.wan: serf: Received remove-key query writer.go:29: 2021-01-29T19:33:12.350Z [DEBUG] TestAgentKeyring_ACL.server.serf.wan: serf: messageQueryResponseType: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb.dc1 writer.go:29: 2021-01-29T19:33:12.352Z [INFO] TestAgentKeyring_ACL.server.serf.lan: serf: Received remove-key query writer.go:29: 2021-01-29T19:33:12.354Z [DEBUG] TestAgentKeyring_ACL.server.serf.lan: serf: messageQueryResponseType: Node-7bef7f2a-97f2-ebe6-de99-f981a6820feb writer.go:29: 2021-01-29T19:33:12.355Z [INFO] TestAgentKeyring_ACL: Requesting shutdown writer.go:29: 2021-01-29T19:33:12.356Z [INFO] TestAgentKeyring_ACL.server: shutting down server writer.go:29: 2021-01-29T19:33:12.357Z [DEBUG] TestAgentKeyring_ACL.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:12.358Z [DEBUG] TestAgentKeyring_ACL.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:12.359Z [DEBUG] TestAgentKeyring_ACL.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.357Z [ERROR] TestAgentKeyring_ACL.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:12.359Z [DEBUG] TestAgentKeyring_ACL.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:12.359Z [DEBUG] TestAgentKeyring_ACL.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:12.360Z [DEBUG] TestAgentKeyring_ACL.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.360Z [WARN] TestAgentKeyring_ACL.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.365Z [WARN] TestAgent_InmemKeyrings/no_keys.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.368Z [WARN] TestAgentKeyring_ACL.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.371Z [INFO] TestAgent_InmemKeyrings/no_keys.server.raft: entering candidate state: node="Node at 127.0.0.1:29669 [Candidate]" term=2 === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.375Z [INFO] TestAgentKeyring_ACL.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:12.376Z [INFO] TestAgentKeyring_ACL: consul server down writer.go:29: 2021-01-29T19:33:12.377Z [INFO] TestAgentKeyring_ACL: shutdown complete writer.go:29: 2021-01-29T19:33:12.378Z [INFO] TestAgentKeyring_ACL: Stopping server: protocol=DNS address=127.0.0.1:29658 network=tcp writer.go:29: 2021-01-29T19:33:12.379Z [INFO] TestAgentKeyring_ACL: Stopping server: protocol=DNS address=127.0.0.1:29658 network=udp writer.go:29: 2021-01-29T19:33:12.380Z [INFO] TestAgentKeyring_ACL: Stopping server: protocol=HTTP address=127.0.0.1:29659 network=tcp === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.379Z [DEBUG] TestAgent_InmemKeyrings/no_keys.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:12.383Z [DEBUG] TestAgent_InmemKeyrings/no_keys.server.raft: vote granted: from=9cbae4ed-99a3-d3c4-9611-ce54169f2e2a term=2 tally=1 writer.go:29: 2021-01-29T19:33:12.384Z [INFO] TestAgent_InmemKeyrings/no_keys.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:12.386Z [INFO] TestAgent_InmemKeyrings/no_keys.server.raft: entering leader state: leader="Node at 127.0.0.1:29669 [Leader]" writer.go:29: 2021-01-29T19:33:12.388Z [INFO] TestAgent_InmemKeyrings/no_keys.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:12.389Z [INFO] TestAgent_InmemKeyrings/no_keys.server: New leader elected: payload=Node-9cbae4ed-99a3-d3c4-9611-ce54169f2e2a writer.go:29: 2021-01-29T19:33:12.392Z [DEBUG] TestAgent_InmemKeyrings/no_keys.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29669 writer.go:29: 2021-01-29T19:33:12.398Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:12.404Z [INFO] TestAgent_InmemKeyrings/no_keys.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:12.406Z [INFO] TestAgent_InmemKeyrings/no_keys.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.407Z [DEBUG] TestAgent_InmemKeyrings/no_keys.server: Skipping self join check for node since the cluster is too small: node=Node-9cbae4ed-99a3-d3c4-9611-ce54169f2e2a writer.go:29: 2021-01-29T19:33:12.409Z [INFO] TestAgent_InmemKeyrings/no_keys.server: member joined, marking health alive: member=Node-9cbae4ed-99a3-d3c4-9611-ce54169f2e2a === CONT TestKVSEndpoint_Recurse writer.go:29: 2021-01-29T19:33:12.562Z [INFO] TestKVSEndpoint_Recurse: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:12.563Z [INFO] TestKVSEndpoint_Recurse: Endpoints down --- PASS: TestKVSEndpoint_Recurse (0.98s) === CONT TestAgent_LoadKeyrings === RUN TestAgent_LoadKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.582Z [WARN] TestAgent_LoadKeyrings/no_keys: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:12.584Z [DEBUG] TestAgent_LoadKeyrings/no_keys.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:12.585Z [DEBUG] TestAgent_LoadKeyrings/no_keys.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:12.590Z [INFO] TestAgent_LoadKeyrings/no_keys.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:99416b59-c1aa-9392-223c-16139bc2e255 Address:127.0.0.1:29675}]" writer.go:29: 2021-01-29T19:33:12.591Z [INFO] TestAgent_LoadKeyrings/no_keys.server.raft: entering follower state: follower="Node at 127.0.0.1:29675 [Follower]" leader= writer.go:29: 2021-01-29T19:33:12.594Z [INFO] TestAgent_LoadKeyrings/no_keys.server.serf.wan: serf: EventMemberJoin: Node-99416b59-c1aa-9392-223c-16139bc2e255.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:12.597Z [INFO] TestAgent_LoadKeyrings/no_keys.server.serf.lan: serf: EventMemberJoin: Node-99416b59-c1aa-9392-223c-16139bc2e255 127.0.0.1 writer.go:29: 2021-01-29T19:33:12.599Z [INFO] TestAgent_LoadKeyrings/no_keys.server: Handled event for server in area: event=member-join server=Node-99416b59-c1aa-9392-223c-16139bc2e255.dc1 area=wan writer.go:29: 2021-01-29T19:33:12.600Z [INFO] TestAgent_LoadKeyrings/no_keys.server: Adding LAN server: server="Node-99416b59-c1aa-9392-223c-16139bc2e255 (Addr: tcp/127.0.0.1:29675) (DC: dc1)" writer.go:29: 2021-01-29T19:33:12.603Z [INFO] TestAgent_LoadKeyrings/no_keys: Started DNS server: address=127.0.0.1:29670 network=tcp writer.go:29: 2021-01-29T19:33:12.605Z [INFO] TestAgent_LoadKeyrings/no_keys: Started DNS server: address=127.0.0.1:29670 network=udp writer.go:29: 2021-01-29T19:33:12.606Z [INFO] TestAgent_LoadKeyrings/no_keys: Started HTTP server: address=127.0.0.1:29671 network=tcp writer.go:29: 2021-01-29T19:33:12.607Z [INFO] TestAgent_LoadKeyrings/no_keys: started state syncer writer.go:29: 2021-01-29T19:33:12.647Z [WARN] TestAgent_LoadKeyrings/no_keys.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:12.648Z [INFO] TestAgent_LoadKeyrings/no_keys.server.raft: entering candidate state: node="Node at 127.0.0.1:29675 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:12.650Z [DEBUG] TestAgent_LoadKeyrings/no_keys.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:12.651Z [DEBUG] TestAgent_LoadKeyrings/no_keys.server.raft: vote granted: from=99416b59-c1aa-9392-223c-16139bc2e255 term=2 tally=1 writer.go:29: 2021-01-29T19:33:12.652Z [INFO] TestAgent_LoadKeyrings/no_keys.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:12.653Z [INFO] TestAgent_LoadKeyrings/no_keys.server.raft: entering leader state: leader="Node at 127.0.0.1:29675 [Leader]" writer.go:29: 2021-01-29T19:33:12.654Z [INFO] TestAgent_LoadKeyrings/no_keys.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:12.654Z [INFO] TestAgent_LoadKeyrings/no_keys.server: New leader elected: payload=Node-99416b59-c1aa-9392-223c-16139bc2e255 writer.go:29: 2021-01-29T19:33:12.657Z [DEBUG] TestAgent_LoadKeyrings/no_keys.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29675 writer.go:29: 2021-01-29T19:33:12.662Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:12.667Z [INFO] TestAgent_LoadKeyrings/no_keys.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:12.667Z [INFO] TestAgent_LoadKeyrings/no_keys.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.668Z [DEBUG] TestAgent_LoadKeyrings/no_keys.server: Skipping self join check for node since the cluster is too small: node=Node-99416b59-c1aa-9392-223c-16139bc2e255 writer.go:29: 2021-01-29T19:33:12.669Z [INFO] TestAgent_LoadKeyrings/no_keys.server: member joined, marking health alive: member=Node-99416b59-c1aa-9392-223c-16139bc2e255 === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.669Z [INFO] TestAgent_InmemKeyrings/no_keys: Requesting shutdown writer.go:29: 2021-01-29T19:33:12.670Z [INFO] TestAgent_InmemKeyrings/no_keys.server: shutting down server writer.go:29: 2021-01-29T19:33:12.671Z [DEBUG] TestAgent_InmemKeyrings/no_keys.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.672Z [WARN] TestAgent_InmemKeyrings/no_keys.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:12.671Z [ERROR] TestAgent_InmemKeyrings/no_keys.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:12.672Z [DEBUG] TestAgent_InmemKeyrings/no_keys.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.676Z [WARN] TestAgent_InmemKeyrings/no_keys.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:12.679Z [INFO] TestAgent_InmemKeyrings/no_keys.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:12.679Z [INFO] TestAgent_InmemKeyrings/no_keys: consul server down writer.go:29: 2021-01-29T19:33:12.681Z [INFO] TestAgent_InmemKeyrings/no_keys: shutdown complete writer.go:29: 2021-01-29T19:33:12.682Z [INFO] TestAgent_InmemKeyrings/no_keys: Stopping server: protocol=DNS address=127.0.0.1:29664 network=tcp writer.go:29: 2021-01-29T19:33:12.683Z [INFO] TestAgent_InmemKeyrings/no_keys: Stopping server: protocol=DNS address=127.0.0.1:29664 network=udp writer.go:29: 2021-01-29T19:33:12.684Z [INFO] TestAgent_InmemKeyrings/no_keys: Stopping server: protocol=HTTP address=127.0.0.1:29665 network=tcp === CONT TestKVSEndpoint_PUT_GET_DELETE writer.go:29: 2021-01-29T19:33:12.822Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:12.824Z [INFO] TestKVSEndpoint_PUT_GET_DELETE: Endpoints down --- PASS: TestKVSEndpoint_PUT_GET_DELETE (0.93s) === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:12.852Z [WARN] TestIntentionsSpecificDelete_good: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:12.854Z [DEBUG] TestIntentionsSpecificDelete_good.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:12.857Z [DEBUG] TestIntentionsSpecificDelete_good.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_LoadKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.862Z [INFO] TestAgent_LoadKeyrings/no_keys: Requesting shutdown writer.go:29: 2021-01-29T19:33:12.864Z [DEBUG] TestAgent_LoadKeyrings/no_keys: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:12.865Z [INFO] TestAgent_LoadKeyrings/no_keys.server: shutting down server writer.go:29: 2021-01-29T19:33:12.868Z [DEBUG] TestAgent_LoadKeyrings/no_keys.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.870Z [WARN] TestAgent_LoadKeyrings/no_keys.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:12.872Z [DEBUG] TestAgent_LoadKeyrings/no_keys.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.874Z [WARN] TestAgent_LoadKeyrings/no_keys.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:12.877Z [INFO] TestAgent_LoadKeyrings/no_keys: Synced node info writer.go:29: 2021-01-29T19:33:12.879Z [DEBUG] TestAgent_LoadKeyrings/no_keys: Node info in sync === CONT TestAgentKeyring_ACL writer.go:29: 2021-01-29T19:33:12.882Z [INFO] TestAgentKeyring_ACL: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:12.884Z [INFO] TestAgentKeyring_ACL: Endpoints down --- PASS: TestAgentKeyring_ACL (0.82s) === CONT TestIntentionsSpecificUpdate_good === CONT TestAgent_LoadKeyrings/no_keys writer.go:29: 2021-01-29T19:33:12.881Z [INFO] TestAgent_LoadKeyrings/no_keys.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:12.882Z [INFO] TestAgent_LoadKeyrings/no_keys: consul server down writer.go:29: 2021-01-29T19:33:12.895Z [INFO] TestAgent_LoadKeyrings/no_keys: shutdown complete writer.go:29: 2021-01-29T19:33:12.896Z [INFO] TestAgent_LoadKeyrings/no_keys: Stopping server: protocol=DNS address=127.0.0.1:29670 network=tcp writer.go:29: 2021-01-29T19:33:12.899Z [INFO] TestAgent_LoadKeyrings/no_keys: Stopping server: protocol=DNS address=127.0.0.1:29670 network=udp writer.go:29: 2021-01-29T19:33:12.900Z [INFO] TestAgent_LoadKeyrings/no_keys: Stopping server: protocol=HTTP address=127.0.0.1:29671 network=tcp === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:12.902Z [INFO] TestIntentionsSpecificDelete_good.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5f5457e8-b7a9-e92b-80ae-0cd0de1cc67d Address:127.0.0.1:29681}]" writer.go:29: 2021-01-29T19:33:12.906Z [INFO] TestIntentionsSpecificDelete_good.server.serf.wan: serf: EventMemberJoin: Node-5f5457e8-b7a9-e92b-80ae-0cd0de1cc67d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:12.906Z [INFO] TestIntentionsSpecificDelete_good.server.raft: entering follower state: follower="Node at 127.0.0.1:29681 [Follower]" leader= === CONT TestIntentionsSpecificUpdate_good writer.go:29: 2021-01-29T19:33:12.920Z [WARN] TestIntentionsSpecificUpdate_good: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:12.922Z [DEBUG] TestIntentionsSpecificUpdate_good.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:12.925Z [DEBUG] TestIntentionsSpecificUpdate_good.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:12.924Z [INFO] TestIntentionsSpecificDelete_good.server.serf.lan: serf: EventMemberJoin: Node-5f5457e8-b7a9-e92b-80ae-0cd0de1cc67d 127.0.0.1 writer.go:29: 2021-01-29T19:33:12.929Z [INFO] TestIntentionsSpecificDelete_good.server: Adding LAN server: server="Node-5f5457e8-b7a9-e92b-80ae-0cd0de1cc67d (Addr: tcp/127.0.0.1:29681) (DC: dc1)" writer.go:29: 2021-01-29T19:33:12.931Z [INFO] TestIntentionsSpecificDelete_good: Started DNS server: address=127.0.0.1:29676 network=udp writer.go:29: 2021-01-29T19:33:12.931Z [INFO] TestIntentionsSpecificDelete_good.server: Handled event for server in area: event=member-join server=Node-5f5457e8-b7a9-e92b-80ae-0cd0de1cc67d.dc1 area=wan writer.go:29: 2021-01-29T19:33:12.934Z [INFO] TestIntentionsSpecificDelete_good: Started DNS server: address=127.0.0.1:29676 network=tcp writer.go:29: 2021-01-29T19:33:12.938Z [INFO] TestIntentionsSpecificDelete_good: Started HTTP server: address=127.0.0.1:29677 network=tcp writer.go:29: 2021-01-29T19:33:12.940Z [INFO] TestIntentionsSpecificDelete_good: started state syncer === CONT TestIntentionsSpecificUpdate_good writer.go:29: 2021-01-29T19:33:12.946Z [INFO] TestIntentionsSpecificUpdate_good.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:00f5c10f-c2f7-c7fc-6327-eb6475d68e3f Address:127.0.0.1:29687}]" writer.go:29: 2021-01-29T19:33:12.948Z [INFO] TestIntentionsSpecificUpdate_good.server.raft: entering follower state: follower="Node at 127.0.0.1:29687 [Follower]" leader= writer.go:29: 2021-01-29T19:33:12.949Z [INFO] TestIntentionsSpecificUpdate_good.server.serf.wan: serf: EventMemberJoin: Node-00f5c10f-c2f7-c7fc-6327-eb6475d68e3f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:12.955Z [INFO] TestIntentionsSpecificUpdate_good.server.serf.lan: serf: EventMemberJoin: Node-00f5c10f-c2f7-c7fc-6327-eb6475d68e3f 127.0.0.1 writer.go:29: 2021-01-29T19:33:12.960Z [INFO] TestIntentionsSpecificUpdate_good.server: Adding LAN server: server="Node-00f5c10f-c2f7-c7fc-6327-eb6475d68e3f (Addr: tcp/127.0.0.1:29687) (DC: dc1)" writer.go:29: 2021-01-29T19:33:12.961Z [INFO] TestIntentionsSpecificUpdate_good.server: Handled event for server in area: event=member-join server=Node-00f5c10f-c2f7-c7fc-6327-eb6475d68e3f.dc1 area=wan writer.go:29: 2021-01-29T19:33:12.963Z [INFO] TestIntentionsSpecificUpdate_good: Started DNS server: address=127.0.0.1:29682 network=udp === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:12.964Z [WARN] TestIntentionsSpecificDelete_good.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestIntentionsSpecificUpdate_good writer.go:29: 2021-01-29T19:33:12.964Z [INFO] TestIntentionsSpecificUpdate_good: Started DNS server: address=127.0.0.1:29682 network=tcp === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:12.965Z [INFO] TestIntentionsSpecificDelete_good.server.raft: entering candidate state: node="Node at 127.0.0.1:29681 [Candidate]" term=2 === CONT TestIntentionsSpecificUpdate_good writer.go:29: 2021-01-29T19:33:12.967Z [INFO] TestIntentionsSpecificUpdate_good: Started HTTP server: address=127.0.0.1:29683 network=tcp writer.go:29: 2021-01-29T19:33:12.969Z [INFO] TestIntentionsSpecificUpdate_good: started state syncer === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:12.972Z [DEBUG] TestIntentionsSpecificDelete_good.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:12.974Z [DEBUG] TestIntentionsSpecificDelete_good.server.raft: vote granted: from=5f5457e8-b7a9-e92b-80ae-0cd0de1cc67d term=2 tally=1 writer.go:29: 2021-01-29T19:33:12.976Z [INFO] TestIntentionsSpecificDelete_good.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:12.977Z [INFO] TestIntentionsSpecificDelete_good.server.raft: entering leader state: leader="Node at 127.0.0.1:29681 [Leader]" writer.go:29: 2021-01-29T19:33:12.979Z [INFO] TestIntentionsSpecificDelete_good.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:12.980Z [INFO] TestIntentionsSpecificDelete_good.server: New leader elected: payload=Node-5f5457e8-b7a9-e92b-80ae-0cd0de1cc67d writer.go:29: 2021-01-29T19:33:12.981Z [DEBUG] TestIntentionsSpecificDelete_good.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29681 writer.go:29: 2021-01-29T19:33:12.987Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:12.993Z [INFO] TestIntentionsSpecificDelete_good.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:12.994Z [INFO] TestIntentionsSpecificDelete_good.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.996Z [DEBUG] TestIntentionsSpecificDelete_good.server: Skipping self join check for node since the cluster is too small: node=Node-5f5457e8-b7a9-e92b-80ae-0cd0de1cc67d writer.go:29: 2021-01-29T19:33:12.997Z [INFO] TestIntentionsSpecificDelete_good.server: member joined, marking health alive: member=Node-5f5457e8-b7a9-e92b-80ae-0cd0de1cc67d === CONT TestIntentionsSpecificUpdate_good writer.go:29: 2021-01-29T19:33:13.005Z [WARN] TestIntentionsSpecificUpdate_good.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:13.006Z [INFO] TestIntentionsSpecificUpdate_good.server.raft: entering candidate state: node="Node at 127.0.0.1:29687 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:13.009Z [DEBUG] TestIntentionsSpecificUpdate_good.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:13.011Z [DEBUG] TestIntentionsSpecificUpdate_good.server.raft: vote granted: from=00f5c10f-c2f7-c7fc-6327-eb6475d68e3f term=2 tally=1 writer.go:29: 2021-01-29T19:33:13.012Z [INFO] TestIntentionsSpecificUpdate_good.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:13.014Z [INFO] TestIntentionsSpecificUpdate_good.server.raft: entering leader state: leader="Node at 127.0.0.1:29687 [Leader]" writer.go:29: 2021-01-29T19:33:13.015Z [INFO] TestIntentionsSpecificUpdate_good.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:13.018Z [INFO] TestIntentionsSpecificUpdate_good.server: New leader elected: payload=Node-00f5c10f-c2f7-c7fc-6327-eb6475d68e3f writer.go:29: 2021-01-29T19:33:13.046Z [DEBUG] TestIntentionsSpecificUpdate_good.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29687 writer.go:29: 2021-01-29T19:33:13.054Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:13.062Z [INFO] TestIntentionsSpecificUpdate_good.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:13.063Z [INFO] TestIntentionsSpecificUpdate_good.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.065Z [DEBUG] TestIntentionsSpecificUpdate_good.server: Skipping self join check for node since the cluster is too small: node=Node-00f5c10f-c2f7-c7fc-6327-eb6475d68e3f writer.go:29: 2021-01-29T19:33:13.066Z [INFO] TestIntentionsSpecificUpdate_good.server: member joined, marking health alive: member=Node-00f5c10f-c2f7-c7fc-6327-eb6475d68e3f writer.go:29: 2021-01-29T19:33:13.119Z [INFO] TestIntentionsSpecificUpdate_good: Requesting shutdown writer.go:29: 2021-01-29T19:33:13.120Z [INFO] TestIntentionsSpecificUpdate_good.server: shutting down server writer.go:29: 2021-01-29T19:33:13.122Z [DEBUG] TestIntentionsSpecificUpdate_good.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.123Z [WARN] TestIntentionsSpecificUpdate_good.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:13.123Z [ERROR] TestIntentionsSpecificUpdate_good.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:13.124Z [DEBUG] TestIntentionsSpecificUpdate_good.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.127Z [WARN] TestIntentionsSpecificUpdate_good.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:13.132Z [INFO] TestIntentionsSpecificUpdate_good.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:13.133Z [INFO] TestIntentionsSpecificUpdate_good: consul server down writer.go:29: 2021-01-29T19:33:13.135Z [INFO] TestIntentionsSpecificUpdate_good: shutdown complete writer.go:29: 2021-01-29T19:33:13.137Z [INFO] TestIntentionsSpecificUpdate_good: Stopping server: protocol=DNS address=127.0.0.1:29682 network=tcp writer.go:29: 2021-01-29T19:33:13.138Z [INFO] TestIntentionsSpecificUpdate_good: Stopping server: protocol=DNS address=127.0.0.1:29682 network=udp writer.go:29: 2021-01-29T19:33:13.139Z [INFO] TestIntentionsSpecificUpdate_good: Stopping server: protocol=HTTP address=127.0.0.1:29683 network=tcp === CONT TestAgent_InmemKeyrings/no_keys writer.go:29: 2021-01-29T19:33:13.185Z [INFO] TestAgent_InmemKeyrings/no_keys: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:13.186Z [INFO] TestAgent_InmemKeyrings/no_keys: Endpoints down === RUN TestAgent_InmemKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.195Z [WARN] TestAgent_InmemKeyrings/server_with_keys: bootstrap = true: do not enable unless necessary === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:13.211Z [INFO] TestIntentionsSpecificDelete_good: Requesting shutdown writer.go:29: 2021-01-29T19:33:13.212Z [INFO] TestIntentionsSpecificDelete_good.server: shutting down server === CONT TestAgent_InmemKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.212Z [DEBUG] TestAgent_InmemKeyrings/server_with_keys.tlsutil: Update: version=1 === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:13.213Z [DEBUG] TestIntentionsSpecificDelete_good.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_InmemKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.213Z [DEBUG] TestAgent_InmemKeyrings/server_with_keys.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:13.213Z [ERROR] TestIntentionsSpecificDelete_good.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:13.214Z [WARN] TestIntentionsSpecificDelete_good.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:13.214Z [DEBUG] TestIntentionsSpecificDelete_good.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.217Z [WARN] TestIntentionsSpecificDelete_good.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:13.223Z [INFO] TestIntentionsSpecificDelete_good.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:13.223Z [INFO] TestIntentionsSpecificDelete_good: consul server down writer.go:29: 2021-01-29T19:33:13.226Z [INFO] TestIntentionsSpecificDelete_good: shutdown complete === CONT TestAgent_InmemKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.226Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:27581d2f-911b-88ef-6fcf-577b03207699 Address:127.0.0.1:29693}]" === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:13.227Z [INFO] TestIntentionsSpecificDelete_good: Stopping server: protocol=DNS address=127.0.0.1:29676 network=tcp === CONT TestAgent_InmemKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.228Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server.raft: entering follower state: follower="Node at 127.0.0.1:29693 [Follower]" leader= === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:13.229Z [INFO] TestIntentionsSpecificDelete_good: Stopping server: protocol=DNS address=127.0.0.1:29676 network=udp writer.go:29: 2021-01-29T19:33:13.231Z [INFO] TestIntentionsSpecificDelete_good: Stopping server: protocol=HTTP address=127.0.0.1:29677 network=tcp === CONT TestAgent_InmemKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.232Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server.serf.wan: serf: EventMemberJoin: Node-27581d2f-911b-88ef-6fcf-577b03207699.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:13.238Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server.serf.lan: serf: EventMemberJoin: Node-27581d2f-911b-88ef-6fcf-577b03207699 127.0.0.1 writer.go:29: 2021-01-29T19:33:13.256Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server: Handled event for server in area: event=member-join server=Node-27581d2f-911b-88ef-6fcf-577b03207699.dc1 area=wan writer.go:29: 2021-01-29T19:33:13.257Z [INFO] TestAgent_InmemKeyrings/server_with_keys: Started DNS server: address=127.0.0.1:29688 network=tcp writer.go:29: 2021-01-29T19:33:13.259Z [INFO] TestAgent_InmemKeyrings/server_with_keys: Started DNS server: address=127.0.0.1:29688 network=udp writer.go:29: 2021-01-29T19:33:13.258Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server: Adding LAN server: server="Node-27581d2f-911b-88ef-6fcf-577b03207699 (Addr: tcp/127.0.0.1:29693) (DC: dc1)" writer.go:29: 2021-01-29T19:33:13.262Z [INFO] TestAgent_InmemKeyrings/server_with_keys: Started HTTP server: address=127.0.0.1:29689 network=tcp writer.go:29: 2021-01-29T19:33:13.264Z [INFO] TestAgent_InmemKeyrings/server_with_keys: started state syncer writer.go:29: 2021-01-29T19:33:13.296Z [WARN] TestAgent_InmemKeyrings/server_with_keys.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:13.297Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server.raft: entering candidate state: node="Node at 127.0.0.1:29693 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:13.299Z [DEBUG] TestAgent_InmemKeyrings/server_with_keys.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:13.300Z [DEBUG] TestAgent_InmemKeyrings/server_with_keys.server.raft: vote granted: from=27581d2f-911b-88ef-6fcf-577b03207699 term=2 tally=1 writer.go:29: 2021-01-29T19:33:13.301Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:13.302Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server.raft: entering leader state: leader="Node at 127.0.0.1:29693 [Leader]" writer.go:29: 2021-01-29T19:33:13.303Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:13.305Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server: New leader elected: payload=Node-27581d2f-911b-88ef-6fcf-577b03207699 writer.go:29: 2021-01-29T19:33:13.310Z [DEBUG] TestAgent_InmemKeyrings/server_with_keys.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29693 writer.go:29: 2021-01-29T19:33:13.316Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:13.321Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:13.322Z [INFO] TestAgent_InmemKeyrings/server_with_keys.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.324Z [DEBUG] TestAgent_InmemKeyrings/server_with_keys.server: Skipping self join check for node since the cluster is too small: node=Node-27581d2f-911b-88ef-6fcf-577b03207699 writer.go:29: 2021-01-29T19:33:13.325Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server: member joined, marking health alive: member=Node-27581d2f-911b-88ef-6fcf-577b03207699 writer.go:29: 2021-01-29T19:33:13.344Z [INFO] TestAgent_InmemKeyrings/server_with_keys: Requesting shutdown writer.go:29: 2021-01-29T19:33:13.346Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server: shutting down server writer.go:29: 2021-01-29T19:33:13.347Z [DEBUG] TestAgent_InmemKeyrings/server_with_keys.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.349Z [WARN] TestAgent_InmemKeyrings/server_with_keys.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:13.347Z [ERROR] TestAgent_InmemKeyrings/server_with_keys.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:13.349Z [DEBUG] TestAgent_InmemKeyrings/server_with_keys.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.352Z [WARN] TestAgent_InmemKeyrings/server_with_keys.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:13.365Z [INFO] TestAgent_InmemKeyrings/server_with_keys.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:13.365Z [INFO] TestAgent_InmemKeyrings/server_with_keys: consul server down writer.go:29: 2021-01-29T19:33:13.369Z [INFO] TestAgent_InmemKeyrings/server_with_keys: shutdown complete writer.go:29: 2021-01-29T19:33:13.370Z [INFO] TestAgent_InmemKeyrings/server_with_keys: Stopping server: protocol=DNS address=127.0.0.1:29688 network=tcp writer.go:29: 2021-01-29T19:33:13.372Z [INFO] TestAgent_InmemKeyrings/server_with_keys: Stopping server: protocol=DNS address=127.0.0.1:29688 network=udp writer.go:29: 2021-01-29T19:33:13.373Z [INFO] TestAgent_InmemKeyrings/server_with_keys: Stopping server: protocol=HTTP address=127.0.0.1:29689 network=tcp === CONT TestAgent_LoadKeyrings/no_keys writer.go:29: 2021-01-29T19:33:13.402Z [INFO] TestAgent_LoadKeyrings/no_keys: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:13.404Z [INFO] TestAgent_LoadKeyrings/no_keys: Endpoints down === RUN TestAgent_LoadKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.423Z [WARN] TestAgent_LoadKeyrings/server_with_keys: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:13.425Z [DEBUG] TestAgent_LoadKeyrings/server_with_keys.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:13.429Z [DEBUG] TestAgent_LoadKeyrings/server_with_keys.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:13.439Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:af010a45-100a-410d-cd82-6103c5c1d047 Address:127.0.0.1:29699}]" writer.go:29: 2021-01-29T19:33:13.442Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server.raft: entering follower state: follower="Node at 127.0.0.1:29699 [Follower]" leader= writer.go:29: 2021-01-29T19:33:13.447Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server.serf.wan: serf: EventMemberJoin: Node-af010a45-100a-410d-cd82-6103c5c1d047.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:13.456Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server.serf.lan: serf: EventMemberJoin: Node-af010a45-100a-410d-cd82-6103c5c1d047 127.0.0.1 writer.go:29: 2021-01-29T19:33:13.465Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server: Handled event for server in area: event=member-join server=Node-af010a45-100a-410d-cd82-6103c5c1d047.dc1 area=wan writer.go:29: 2021-01-29T19:33:13.465Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server: Adding LAN server: server="Node-af010a45-100a-410d-cd82-6103c5c1d047 (Addr: tcp/127.0.0.1:29699) (DC: dc1)" writer.go:29: 2021-01-29T19:33:13.471Z [INFO] TestAgent_LoadKeyrings/server_with_keys: Started DNS server: address=127.0.0.1:29694 network=tcp writer.go:29: 2021-01-29T19:33:13.474Z [INFO] TestAgent_LoadKeyrings/server_with_keys: Started DNS server: address=127.0.0.1:29694 network=udp writer.go:29: 2021-01-29T19:33:13.476Z [INFO] TestAgent_LoadKeyrings/server_with_keys: Started HTTP server: address=127.0.0.1:29695 network=tcp writer.go:29: 2021-01-29T19:33:13.477Z [INFO] TestAgent_LoadKeyrings/server_with_keys: started state syncer writer.go:29: 2021-01-29T19:33:13.494Z [WARN] TestAgent_LoadKeyrings/server_with_keys.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:13.495Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server.raft: entering candidate state: node="Node at 127.0.0.1:29699 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:13.497Z [DEBUG] TestAgent_LoadKeyrings/server_with_keys.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:13.498Z [DEBUG] TestAgent_LoadKeyrings/server_with_keys.server.raft: vote granted: from=af010a45-100a-410d-cd82-6103c5c1d047 term=2 tally=1 writer.go:29: 2021-01-29T19:33:13.499Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:13.500Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server.raft: entering leader state: leader="Node at 127.0.0.1:29699 [Leader]" writer.go:29: 2021-01-29T19:33:13.501Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:13.502Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server: New leader elected: payload=Node-af010a45-100a-410d-cd82-6103c5c1d047 writer.go:29: 2021-01-29T19:33:13.503Z [DEBUG] TestAgent_LoadKeyrings/server_with_keys.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29699 writer.go:29: 2021-01-29T19:33:13.503Z [INFO] TestAgent_LoadKeyrings/server_with_keys: Synced node info writer.go:29: 2021-01-29T19:33:13.513Z [DEBUG] TestAgent_LoadKeyrings/server_with_keys: Node info in sync writer.go:29: 2021-01-29T19:33:13.514Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:13.562Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:13.563Z [INFO] TestAgent_LoadKeyrings/server_with_keys.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.564Z [DEBUG] TestAgent_LoadKeyrings/server_with_keys.server: Skipping self join check for node since the cluster is too small: node=Node-af010a45-100a-410d-cd82-6103c5c1d047 writer.go:29: 2021-01-29T19:33:13.565Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server: member joined, marking health alive: member=Node-af010a45-100a-410d-cd82-6103c5c1d047 === CONT TestIntentionsSpecificUpdate_good writer.go:29: 2021-01-29T19:33:13.642Z [INFO] TestIntentionsSpecificUpdate_good: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:13.645Z [INFO] TestIntentionsSpecificUpdate_good: Endpoints down --- PASS: TestIntentionsSpecificUpdate_good (0.76s) === CONT TestIntentionsSpecificGet_invalidId === CONT TestAgent_LoadKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.724Z [INFO] TestAgent_LoadKeyrings/server_with_keys: Requesting shutdown writer.go:29: 2021-01-29T19:33:13.728Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server: shutting down server writer.go:29: 2021-01-29T19:33:13.731Z [DEBUG] TestAgent_LoadKeyrings/server_with_keys.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.735Z [WARN] TestAgent_LoadKeyrings/server_with_keys.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:13.740Z [DEBUG] TestAgent_LoadKeyrings/server_with_keys.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.742Z [WARN] TestAgent_LoadKeyrings/server_with_keys.server.serf.wan: serf: Shutdown without a Leave === CONT TestIntentionsSpecificDelete_good writer.go:29: 2021-01-29T19:33:13.743Z [INFO] TestIntentionsSpecificDelete_good: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:13.746Z [INFO] TestIntentionsSpecificDelete_good: Endpoints down --- PASS: TestIntentionsSpecificDelete_good (0.92s) === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.749Z [WARN] TestIntentionsSpecificGet_invalidId: bootstrap = true: do not enable unless necessary === CONT TestIntentionsSpecificGet_good === CONT TestAgent_LoadKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.752Z [INFO] TestAgent_LoadKeyrings/server_with_keys.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:13.755Z [INFO] TestAgent_LoadKeyrings/server_with_keys: consul server down writer.go:29: 2021-01-29T19:33:13.757Z [INFO] TestAgent_LoadKeyrings/server_with_keys: shutdown complete writer.go:29: 2021-01-29T19:33:13.758Z [INFO] TestAgent_LoadKeyrings/server_with_keys: Stopping server: protocol=DNS address=127.0.0.1:29694 network=tcp === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.762Z [DEBUG] TestIntentionsSpecificGet_invalidId.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:13.766Z [DEBUG] TestIntentionsSpecificGet_invalidId.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_LoadKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.760Z [INFO] TestAgent_LoadKeyrings/server_with_keys: Stopping server: protocol=DNS address=127.0.0.1:29694 network=udp writer.go:29: 2021-01-29T19:33:13.772Z [INFO] TestAgent_LoadKeyrings/server_with_keys: Stopping server: protocol=HTTP address=127.0.0.1:29695 network=tcp === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.881Z [INFO] TestIntentionsSpecificGet_invalidId.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c7094bdc-645e-9431-5a31-05443c259f43 Address:127.0.0.1:29711}]" writer.go:29: 2021-01-29T19:33:13.887Z [INFO] TestIntentionsSpecificGet_invalidId.server.raft: entering follower state: follower="Node at 127.0.0.1:29711 [Follower]" leader= === CONT TestAgent_InmemKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:13.889Z [INFO] TestAgent_InmemKeyrings/server_with_keys: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:13.891Z [INFO] TestAgent_InmemKeyrings/server_with_keys: Endpoints down === RUN TestAgent_InmemKeyrings/client_with_keys === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.891Z [INFO] TestIntentionsSpecificGet_invalidId.server.serf.wan: serf: EventMemberJoin: Node-c7094bdc-645e-9431-5a31-05443c259f43.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:13.897Z [INFO] TestIntentionsSpecificGet_invalidId.server.serf.lan: serf: EventMemberJoin: Node-c7094bdc-645e-9431-5a31-05443c259f43 127.0.0.1 === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:13.908Z [WARN] TestIntentionsSpecificGet_good: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:13.910Z [DEBUG] TestIntentionsSpecificGet_good.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:13.913Z [DEBUG] TestIntentionsSpecificGet_good.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.915Z [DEBUG] TestAgent_InmemKeyrings/client_with_keys.tlsutil: Update: version=1 === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.916Z [INFO] TestIntentionsSpecificGet_invalidId.server: Adding LAN server: server="Node-c7094bdc-645e-9431-5a31-05443c259f43 (Addr: tcp/127.0.0.1:29711) (DC: dc1)" writer.go:29: 2021-01-29T19:33:13.917Z [INFO] TestIntentionsSpecificGet_invalidId.server: Handled event for server in area: event=member-join server=Node-c7094bdc-645e-9431-5a31-05443c259f43.dc1 area=wan === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.918Z [INFO] TestAgent_InmemKeyrings/client_with_keys.client.serf.lan: serf: EventMemberJoin: Node-d56fa349-a7a2-0b0d-e719-ccce2e98cf65 127.0.0.1 === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.920Z [INFO] TestIntentionsSpecificGet_invalidId: Started DNS server: address=127.0.0.1:29706 network=udp === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.924Z [INFO] TestAgent_InmemKeyrings/client_with_keys: Started DNS server: address=127.0.0.1:29712 network=tcp === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.924Z [INFO] TestIntentionsSpecificGet_invalidId: Started DNS server: address=127.0.0.1:29706 network=tcp writer.go:29: 2021-01-29T19:33:13.928Z [INFO] TestIntentionsSpecificGet_invalidId: Started HTTP server: address=127.0.0.1:29707 network=tcp writer.go:29: 2021-01-29T19:33:13.929Z [INFO] TestIntentionsSpecificGet_invalidId: started state syncer === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.931Z [INFO] TestAgent_InmemKeyrings/client_with_keys: Started DNS server: address=127.0.0.1:29712 network=udp writer.go:29: 2021-01-29T19:33:13.933Z [INFO] TestAgent_InmemKeyrings/client_with_keys: Started HTTP server: address=127.0.0.1:29713 network=tcp === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:13.933Z [INFO] TestIntentionsSpecificGet_good.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:53da9a74-ec86-bcab-e974-dfc78ce768fc Address:127.0.0.1:29705}]" === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.935Z [INFO] TestAgent_InmemKeyrings/client_with_keys: started state syncer writer.go:29: 2021-01-29T19:33:13.937Z [INFO] TestAgent_InmemKeyrings/client_with_keys: Requesting shutdown writer.go:29: 2021-01-29T19:33:13.939Z [WARN] TestAgent_InmemKeyrings/client_with_keys.client.manager: No servers available writer.go:29: 2021-01-29T19:33:13.941Z [ERROR] TestAgent_InmemKeyrings/client_with_keys.anti_entropy: failed to sync remote state: error="No known Consul servers" === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:13.936Z [INFO] TestIntentionsSpecificGet_good.server.serf.wan: serf: EventMemberJoin: Node-53da9a74-ec86-bcab-e974-dfc78ce768fc.dc1 127.0.0.1 === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.939Z [INFO] TestAgent_InmemKeyrings/client_with_keys.client: shutting down client writer.go:29: 2021-01-29T19:33:13.944Z [WARN] TestAgent_InmemKeyrings/client_with_keys.client.serf.lan: serf: Shutdown without a Leave === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:13.937Z [INFO] TestIntentionsSpecificGet_good.server.raft: entering follower state: follower="Node at 127.0.0.1:29705 [Follower]" leader= === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.947Z [INFO] TestAgent_InmemKeyrings/client_with_keys.client.manager: shutting down === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:13.950Z [INFO] TestIntentionsSpecificGet_good.server.serf.lan: serf: EventMemberJoin: Node-53da9a74-ec86-bcab-e974-dfc78ce768fc 127.0.0.1 === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.953Z [INFO] TestAgent_InmemKeyrings/client_with_keys: consul client down === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.953Z [WARN] TestIntentionsSpecificGet_invalidId.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.954Z [INFO] TestAgent_InmemKeyrings/client_with_keys: shutdown complete === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.954Z [INFO] TestIntentionsSpecificGet_invalidId.server.raft: entering candidate state: node="Node at 127.0.0.1:29711 [Candidate]" term=2 === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.955Z [INFO] TestAgent_InmemKeyrings/client_with_keys: Stopping server: protocol=DNS address=127.0.0.1:29712 network=tcp writer.go:29: 2021-01-29T19:33:13.957Z [INFO] TestAgent_InmemKeyrings/client_with_keys: Stopping server: protocol=DNS address=127.0.0.1:29712 network=udp === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.957Z [DEBUG] TestIntentionsSpecificGet_invalidId.server.raft: votes: needed=1 === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:13.958Z [INFO] TestAgent_InmemKeyrings/client_with_keys: Stopping server: protocol=HTTP address=127.0.0.1:29713 network=tcp === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:13.959Z [INFO] TestIntentionsSpecificGet_good.server: Handled event for server in area: event=member-join server=Node-53da9a74-ec86-bcab-e974-dfc78ce768fc.dc1 area=wan === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.959Z [DEBUG] TestIntentionsSpecificGet_invalidId.server.raft: vote granted: from=c7094bdc-645e-9431-5a31-05443c259f43 term=2 tally=1 writer.go:29: 2021-01-29T19:33:13.960Z [INFO] TestIntentionsSpecificGet_invalidId.server.raft: election won: tally=1 === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:13.961Z [INFO] TestIntentionsSpecificGet_good.server: Adding LAN server: server="Node-53da9a74-ec86-bcab-e974-dfc78ce768fc (Addr: tcp/127.0.0.1:29705) (DC: dc1)" === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.962Z [INFO] TestIntentionsSpecificGet_invalidId.server.raft: entering leader state: leader="Node at 127.0.0.1:29711 [Leader]" === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:13.963Z [INFO] TestIntentionsSpecificGet_good: Started DNS server: address=127.0.0.1:29700 network=tcp writer.go:29: 2021-01-29T19:33:13.964Z [INFO] TestIntentionsSpecificGet_good: Started DNS server: address=127.0.0.1:29700 network=udp === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.963Z [INFO] TestIntentionsSpecificGet_invalidId.server: cluster leadership acquired === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:13.965Z [INFO] TestIntentionsSpecificGet_good: Started HTTP server: address=127.0.0.1:29701 network=tcp writer.go:29: 2021-01-29T19:33:13.967Z [INFO] TestIntentionsSpecificGet_good: started state syncer === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:13.964Z [INFO] TestIntentionsSpecificGet_invalidId.server: New leader elected: payload=Node-c7094bdc-645e-9431-5a31-05443c259f43 writer.go:29: 2021-01-29T19:33:13.965Z [DEBUG] TestIntentionsSpecificGet_invalidId.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29711 writer.go:29: 2021-01-29T19:33:13.975Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:13.982Z [INFO] TestIntentionsSpecificGet_invalidId.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:13.984Z [INFO] TestIntentionsSpecificGet_invalidId.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:13.985Z [DEBUG] TestIntentionsSpecificGet_invalidId.server: Skipping self join check for node since the cluster is too small: node=Node-c7094bdc-645e-9431-5a31-05443c259f43 writer.go:29: 2021-01-29T19:33:13.987Z [INFO] TestIntentionsSpecificGet_invalidId.server: member joined, marking health alive: member=Node-c7094bdc-645e-9431-5a31-05443c259f43 === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:14.015Z [WARN] TestIntentionsSpecificGet_good.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:14.017Z [INFO] TestIntentionsSpecificGet_good.server.raft: entering candidate state: node="Node at 127.0.0.1:29705 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:14.019Z [DEBUG] TestIntentionsSpecificGet_good.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:14.021Z [DEBUG] TestIntentionsSpecificGet_good.server.raft: vote granted: from=53da9a74-ec86-bcab-e974-dfc78ce768fc term=2 tally=1 writer.go:29: 2021-01-29T19:33:14.022Z [INFO] TestIntentionsSpecificGet_good.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:14.024Z [INFO] TestIntentionsSpecificGet_good.server.raft: entering leader state: leader="Node at 127.0.0.1:29705 [Leader]" writer.go:29: 2021-01-29T19:33:14.027Z [INFO] TestIntentionsSpecificGet_good.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:14.029Z [INFO] TestIntentionsSpecificGet_good.server: New leader elected: payload=Node-53da9a74-ec86-bcab-e974-dfc78ce768fc === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:14.032Z [DEBUG] TestIntentionsSpecificGet_invalidId: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:14.032Z [DEBUG] TestIntentionsSpecificGet_good.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29705 === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:14.035Z [INFO] TestIntentionsSpecificGet_invalidId: Synced node info writer.go:29: 2021-01-29T19:33:14.036Z [DEBUG] TestIntentionsSpecificGet_invalidId: Node info in sync === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:14.037Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:14.042Z [INFO] TestIntentionsSpecificGet_good.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:14.042Z [INFO] TestIntentionsSpecificGet_good.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.043Z [DEBUG] TestIntentionsSpecificGet_good.server: Skipping self join check for node since the cluster is too small: node=Node-53da9a74-ec86-bcab-e974-dfc78ce768fc writer.go:29: 2021-01-29T19:33:14.044Z [INFO] TestIntentionsSpecificGet_good.server: member joined, marking health alive: member=Node-53da9a74-ec86-bcab-e974-dfc78ce768fc === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:14.131Z [INFO] TestIntentionsSpecificGet_invalidId: Requesting shutdown writer.go:29: 2021-01-29T19:33:14.132Z [INFO] TestIntentionsSpecificGet_invalidId.server: shutting down server writer.go:29: 2021-01-29T19:33:14.133Z [DEBUG] TestIntentionsSpecificGet_invalidId.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.134Z [WARN] TestIntentionsSpecificGet_invalidId.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:14.134Z [DEBUG] TestIntentionsSpecificGet_invalidId.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.142Z [WARN] TestIntentionsSpecificGet_invalidId.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:14.145Z [INFO] TestIntentionsSpecificGet_invalidId.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:14.146Z [INFO] TestIntentionsSpecificGet_invalidId: consul server down writer.go:29: 2021-01-29T19:33:14.148Z [INFO] TestIntentionsSpecificGet_invalidId: shutdown complete writer.go:29: 2021-01-29T19:33:14.150Z [INFO] TestIntentionsSpecificGet_invalidId: Stopping server: protocol=DNS address=127.0.0.1:29706 network=tcp writer.go:29: 2021-01-29T19:33:14.152Z [INFO] TestIntentionsSpecificGet_invalidId: Stopping server: protocol=DNS address=127.0.0.1:29706 network=udp writer.go:29: 2021-01-29T19:33:14.155Z [INFO] TestIntentionsSpecificGet_invalidId: Stopping server: protocol=HTTP address=127.0.0.1:29707 network=tcp === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:14.231Z [INFO] TestIntentionsSpecificGet_good: Requesting shutdown writer.go:29: 2021-01-29T19:33:14.231Z [DEBUG] TestIntentionsSpecificGet_good: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:14.232Z [INFO] TestIntentionsSpecificGet_good.server: shutting down server writer.go:29: 2021-01-29T19:33:14.235Z [DEBUG] TestIntentionsSpecificGet_good.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.237Z [WARN] TestIntentionsSpecificGet_good.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:14.237Z [DEBUG] TestIntentionsSpecificGet_good.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.240Z [INFO] TestIntentionsSpecificGet_good: Synced node info writer.go:29: 2021-01-29T19:33:14.240Z [WARN] TestIntentionsSpecificGet_good.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:14.245Z [INFO] TestIntentionsSpecificGet_good.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:14.245Z [INFO] TestIntentionsSpecificGet_good: consul server down writer.go:29: 2021-01-29T19:33:14.248Z [INFO] TestIntentionsSpecificGet_good: shutdown complete writer.go:29: 2021-01-29T19:33:14.250Z [INFO] TestIntentionsSpecificGet_good: Stopping server: protocol=DNS address=127.0.0.1:29700 network=tcp writer.go:29: 2021-01-29T19:33:14.252Z [INFO] TestIntentionsSpecificGet_good: Stopping server: protocol=DNS address=127.0.0.1:29700 network=udp writer.go:29: 2021-01-29T19:33:14.254Z [INFO] TestIntentionsSpecificGet_good: Stopping server: protocol=HTTP address=127.0.0.1:29701 network=tcp === CONT TestAgent_LoadKeyrings/server_with_keys writer.go:29: 2021-01-29T19:33:14.274Z [INFO] TestAgent_LoadKeyrings/server_with_keys: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:14.275Z [INFO] TestAgent_LoadKeyrings/server_with_keys: Endpoints down === RUN TestAgent_LoadKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:14.291Z [DEBUG] TestAgent_LoadKeyrings/client_with_keys.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:14.296Z [INFO] TestAgent_LoadKeyrings/client_with_keys.client.serf.lan: serf: EventMemberJoin: Node-5ea567af-33f1-d462-c4c5-3b3951ce8489 127.0.0.1 writer.go:29: 2021-01-29T19:33:14.301Z [INFO] TestAgent_LoadKeyrings/client_with_keys: Started DNS server: address=127.0.0.1:29718 network=tcp writer.go:29: 2021-01-29T19:33:14.303Z [INFO] TestAgent_LoadKeyrings/client_with_keys: Started DNS server: address=127.0.0.1:29718 network=udp writer.go:29: 2021-01-29T19:33:14.304Z [INFO] TestAgent_LoadKeyrings/client_with_keys: Started HTTP server: address=127.0.0.1:29719 network=tcp writer.go:29: 2021-01-29T19:33:14.305Z [INFO] TestAgent_LoadKeyrings/client_with_keys: started state syncer writer.go:29: 2021-01-29T19:33:14.307Z [INFO] TestAgent_LoadKeyrings/client_with_keys: Requesting shutdown writer.go:29: 2021-01-29T19:33:14.308Z [INFO] TestAgent_LoadKeyrings/client_with_keys.client: shutting down client writer.go:29: 2021-01-29T19:33:14.308Z [WARN] TestAgent_LoadKeyrings/client_with_keys.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:14.310Z [WARN] TestAgent_LoadKeyrings/client_with_keys.client.manager: No servers available writer.go:29: 2021-01-29T19:33:14.312Z [ERROR] TestAgent_LoadKeyrings/client_with_keys.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:33:14.312Z [INFO] TestAgent_LoadKeyrings/client_with_keys.client.manager: shutting down writer.go:29: 2021-01-29T19:33:14.316Z [INFO] TestAgent_LoadKeyrings/client_with_keys: consul client down writer.go:29: 2021-01-29T19:33:14.317Z [INFO] TestAgent_LoadKeyrings/client_with_keys: shutdown complete writer.go:29: 2021-01-29T19:33:14.319Z [INFO] TestAgent_LoadKeyrings/client_with_keys: Stopping server: protocol=DNS address=127.0.0.1:29718 network=tcp writer.go:29: 2021-01-29T19:33:14.321Z [INFO] TestAgent_LoadKeyrings/client_with_keys: Stopping server: protocol=DNS address=127.0.0.1:29718 network=udp writer.go:29: 2021-01-29T19:33:14.322Z [INFO] TestAgent_LoadKeyrings/client_with_keys: Stopping server: protocol=HTTP address=127.0.0.1:29719 network=tcp === CONT TestAgent_InmemKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:14.460Z [INFO] TestAgent_InmemKeyrings/client_with_keys: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:14.462Z [INFO] TestAgent_InmemKeyrings/client_with_keys: Endpoints down === RUN TestAgent_InmemKeyrings/ignore_files writer.go:29: 2021-01-29T19:33:14.477Z [WARN] TestAgent_InmemKeyrings/ignore_files: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:14.478Z [DEBUG] TestAgent_InmemKeyrings/ignore_files.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:14.495Z [DEBUG] TestAgent_InmemKeyrings/ignore_files.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:14.509Z [INFO] TestAgent_InmemKeyrings/ignore_files.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:34016d04-adb4-5b6d-b916-6b7fe5254251 Address:127.0.0.1:29729}]" writer.go:29: 2021-01-29T19:33:14.512Z [INFO] TestAgent_InmemKeyrings/ignore_files.server.raft: entering follower state: follower="Node at 127.0.0.1:29729 [Follower]" leader= writer.go:29: 2021-01-29T19:33:14.523Z [INFO] TestAgent_InmemKeyrings/ignore_files.server.serf.wan: serf: EventMemberJoin: Node-34016d04-adb4-5b6d-b916-6b7fe5254251.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:14.531Z [INFO] TestAgent_InmemKeyrings/ignore_files.server.serf.lan: serf: EventMemberJoin: Node-34016d04-adb4-5b6d-b916-6b7fe5254251 127.0.0.1 writer.go:29: 2021-01-29T19:33:14.536Z [INFO] TestAgent_InmemKeyrings/ignore_files.server: Handled event for server in area: event=member-join server=Node-34016d04-adb4-5b6d-b916-6b7fe5254251.dc1 area=wan writer.go:29: 2021-01-29T19:33:14.536Z [INFO] TestAgent_InmemKeyrings/ignore_files.server: Adding LAN server: server="Node-34016d04-adb4-5b6d-b916-6b7fe5254251 (Addr: tcp/127.0.0.1:29729) (DC: dc1)" writer.go:29: 2021-01-29T19:33:14.537Z [INFO] TestAgent_InmemKeyrings/ignore_files: Started DNS server: address=127.0.0.1:29724 network=udp writer.go:29: 2021-01-29T19:33:14.539Z [INFO] TestAgent_InmemKeyrings/ignore_files: Started DNS server: address=127.0.0.1:29724 network=tcp writer.go:29: 2021-01-29T19:33:14.540Z [INFO] TestAgent_InmemKeyrings/ignore_files: Started HTTP server: address=127.0.0.1:29725 network=tcp writer.go:29: 2021-01-29T19:33:14.542Z [INFO] TestAgent_InmemKeyrings/ignore_files: started state syncer writer.go:29: 2021-01-29T19:33:14.569Z [WARN] TestAgent_InmemKeyrings/ignore_files.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:14.571Z [INFO] TestAgent_InmemKeyrings/ignore_files.server.raft: entering candidate state: node="Node at 127.0.0.1:29729 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:14.575Z [DEBUG] TestAgent_InmemKeyrings/ignore_files.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:14.576Z [DEBUG] TestAgent_InmemKeyrings/ignore_files.server.raft: vote granted: from=34016d04-adb4-5b6d-b916-6b7fe5254251 term=2 tally=1 writer.go:29: 2021-01-29T19:33:14.577Z [INFO] TestAgent_InmemKeyrings/ignore_files.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:14.578Z [INFO] TestAgent_InmemKeyrings/ignore_files.server.raft: entering leader state: leader="Node at 127.0.0.1:29729 [Leader]" writer.go:29: 2021-01-29T19:33:14.580Z [INFO] TestAgent_InmemKeyrings/ignore_files.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:14.581Z [INFO] TestAgent_InmemKeyrings/ignore_files.server: New leader elected: payload=Node-34016d04-adb4-5b6d-b916-6b7fe5254251 writer.go:29: 2021-01-29T19:33:14.587Z [DEBUG] TestAgent_InmemKeyrings/ignore_files.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29729 writer.go:29: 2021-01-29T19:33:14.597Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:14.613Z [INFO] TestAgent_InmemKeyrings/ignore_files.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:14.615Z [INFO] TestAgent_InmemKeyrings/ignore_files.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.618Z [DEBUG] TestAgent_InmemKeyrings/ignore_files.server: Skipping self join check for node since the cluster is too small: node=Node-34016d04-adb4-5b6d-b916-6b7fe5254251 writer.go:29: 2021-01-29T19:33:14.620Z [INFO] TestAgent_InmemKeyrings/ignore_files.server: member joined, marking health alive: member=Node-34016d04-adb4-5b6d-b916-6b7fe5254251 === CONT TestIntentionsSpecificGet_invalidId writer.go:29: 2021-01-29T19:33:14.656Z [INFO] TestIntentionsSpecificGet_invalidId: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:14.658Z [INFO] TestIntentionsSpecificGet_invalidId: Endpoints down --- PASS: TestIntentionsSpecificGet_invalidId (1.01s) === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:14.680Z [WARN] TestIntentionsCreate_noBody: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:14.681Z [DEBUG] TestIntentionsCreate_noBody.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:14.682Z [DEBUG] TestIntentionsCreate_noBody.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:14.686Z [INFO] TestIntentionsCreate_noBody.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:29b080b2-acca-350b-ad4a-c39c9bde476a Address:127.0.0.1:29735}]" writer.go:29: 2021-01-29T19:33:14.688Z [INFO] TestIntentionsCreate_noBody.server.raft: entering follower state: follower="Node at 127.0.0.1:29735 [Follower]" leader= writer.go:29: 2021-01-29T19:33:14.688Z [INFO] TestIntentionsCreate_noBody.server.serf.wan: serf: EventMemberJoin: Node-29b080b2-acca-350b-ad4a-c39c9bde476a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:14.697Z [INFO] TestIntentionsCreate_noBody.server.serf.lan: serf: EventMemberJoin: Node-29b080b2-acca-350b-ad4a-c39c9bde476a 127.0.0.1 writer.go:29: 2021-01-29T19:33:14.700Z [INFO] TestIntentionsCreate_noBody.server: Handled event for server in area: event=member-join server=Node-29b080b2-acca-350b-ad4a-c39c9bde476a.dc1 area=wan writer.go:29: 2021-01-29T19:33:14.701Z [INFO] TestIntentionsCreate_noBody.server: Adding LAN server: server="Node-29b080b2-acca-350b-ad4a-c39c9bde476a (Addr: tcp/127.0.0.1:29735) (DC: dc1)" writer.go:29: 2021-01-29T19:33:14.701Z [INFO] TestIntentionsCreate_noBody: Started DNS server: address=127.0.0.1:29730 network=udp writer.go:29: 2021-01-29T19:33:14.709Z [INFO] TestIntentionsCreate_noBody: Started DNS server: address=127.0.0.1:29730 network=tcp writer.go:29: 2021-01-29T19:33:14.710Z [INFO] TestIntentionsCreate_noBody: Started HTTP server: address=127.0.0.1:29731 network=tcp writer.go:29: 2021-01-29T19:33:14.712Z [INFO] TestIntentionsCreate_noBody: started state syncer writer.go:29: 2021-01-29T19:33:14.745Z [WARN] TestIntentionsCreate_noBody.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:14.746Z [INFO] TestIntentionsCreate_noBody.server.raft: entering candidate state: node="Node at 127.0.0.1:29735 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:14.749Z [DEBUG] TestIntentionsCreate_noBody.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:14.749Z [DEBUG] TestIntentionsCreate_noBody.server.raft: vote granted: from=29b080b2-acca-350b-ad4a-c39c9bde476a term=2 tally=1 writer.go:29: 2021-01-29T19:33:14.750Z [INFO] TestIntentionsCreate_noBody.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:14.751Z [INFO] TestIntentionsCreate_noBody.server.raft: entering leader state: leader="Node at 127.0.0.1:29735 [Leader]" writer.go:29: 2021-01-29T19:33:14.752Z [INFO] TestIntentionsCreate_noBody.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:14.752Z [INFO] TestIntentionsCreate_noBody.server: New leader elected: payload=Node-29b080b2-acca-350b-ad4a-c39c9bde476a === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:14.756Z [INFO] TestIntentionsSpecificGet_good: Waiting for endpoints to shut down === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:14.753Z [DEBUG] TestIntentionsCreate_noBody.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29735 === CONT TestIntentionsSpecificGet_good writer.go:29: 2021-01-29T19:33:14.757Z [INFO] TestIntentionsSpecificGet_good: Endpoints down --- PASS: TestIntentionsSpecificGet_good (1.01s) === CONT TestIntentionsCheck_noDestination === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:14.762Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:14.770Z [INFO] TestIntentionsCreate_noBody.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:14.773Z [INFO] TestIntentionsCreate_noBody.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.775Z [DEBUG] TestIntentionsCreate_noBody.server: Skipping self join check for node since the cluster is too small: node=Node-29b080b2-acca-350b-ad4a-c39c9bde476a writer.go:29: 2021-01-29T19:33:14.777Z [INFO] TestIntentionsCreate_noBody.server: member joined, marking health alive: member=Node-29b080b2-acca-350b-ad4a-c39c9bde476a === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.784Z [WARN] TestIntentionsCheck_noDestination: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:14.785Z [DEBUG] TestIntentionsCheck_noDestination.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:14.786Z [DEBUG] TestIntentionsCheck_noDestination.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_InmemKeyrings/ignore_files writer.go:29: 2021-01-29T19:33:14.799Z [DEBUG] TestAgent_InmemKeyrings/ignore_files: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:14.801Z [INFO] TestAgent_InmemKeyrings/ignore_files: Synced node info === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.807Z [INFO] TestIntentionsCheck_noDestination.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6aac4e40-52dc-a95d-f7f1-30352dc88ae9 Address:127.0.0.1:29747}]" writer.go:29: 2021-01-29T19:33:14.809Z [INFO] TestIntentionsCheck_noDestination.server.raft: entering follower state: follower="Node at 127.0.0.1:29747 [Follower]" leader= writer.go:29: 2021-01-29T19:33:14.810Z [INFO] TestIntentionsCheck_noDestination.server.serf.wan: serf: EventMemberJoin: Node-6aac4e40-52dc-a95d-f7f1-30352dc88ae9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:14.813Z [INFO] TestIntentionsCheck_noDestination.server.serf.lan: serf: EventMemberJoin: Node-6aac4e40-52dc-a95d-f7f1-30352dc88ae9 127.0.0.1 writer.go:29: 2021-01-29T19:33:14.816Z [INFO] TestIntentionsCheck_noDestination: Started DNS server: address=127.0.0.1:29742 network=udp === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:14.815Z [INFO] TestIntentionsCreate_noBody: Requesting shutdown === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.816Z [INFO] TestIntentionsCheck_noDestination.server: Adding LAN server: server="Node-6aac4e40-52dc-a95d-f7f1-30352dc88ae9 (Addr: tcp/127.0.0.1:29747) (DC: dc1)" writer.go:29: 2021-01-29T19:33:14.817Z [INFO] TestIntentionsCheck_noDestination.server: Handled event for server in area: event=member-join server=Node-6aac4e40-52dc-a95d-f7f1-30352dc88ae9.dc1 area=wan === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:14.818Z [INFO] TestIntentionsCreate_noBody.server: shutting down server === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.819Z [INFO] TestIntentionsCheck_noDestination: Started DNS server: address=127.0.0.1:29742 network=tcp === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:14.819Z [DEBUG] TestIntentionsCreate_noBody.leader: stopping routine: routine="CA root pruning" === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.820Z [INFO] TestIntentionsCheck_noDestination: Started HTTP server: address=127.0.0.1:29743 network=tcp writer.go:29: 2021-01-29T19:33:14.821Z [INFO] TestIntentionsCheck_noDestination: started state syncer === CONT TestAgent_LoadKeyrings/client_with_keys writer.go:29: 2021-01-29T19:33:14.824Z [INFO] TestAgent_LoadKeyrings/client_with_keys: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:14.825Z [INFO] TestAgent_LoadKeyrings/client_with_keys: Endpoints down --- PASS: TestAgent_LoadKeyrings (2.26s) --- PASS: TestAgent_LoadKeyrings/no_keys (0.84s) --- PASS: TestAgent_LoadKeyrings/server_with_keys (0.87s) --- PASS: TestAgent_LoadKeyrings/client_with_keys (0.55s) === CONT TestIntentionsCheck_noSource === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:14.821Z [WARN] TestIntentionsCreate_noBody.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:14.820Z [ERROR] TestIntentionsCreate_noBody.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_InmemKeyrings/ignore_files writer.go:29: 2021-01-29T19:33:14.833Z [INFO] TestAgent_InmemKeyrings/ignore_files: Requesting shutdown === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:14.835Z [WARN] TestIntentionsCheck_noSource: bootstrap = true: do not enable unless necessary === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:14.821Z [DEBUG] TestIntentionsCreate_noBody.leader: stopped routine: routine="CA root pruning" === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:14.836Z [DEBUG] TestIntentionsCheck_noSource.tlsutil: Update: version=1 === CONT TestAgent_InmemKeyrings/ignore_files writer.go:29: 2021-01-29T19:33:14.836Z [INFO] TestAgent_InmemKeyrings/ignore_files.server: shutting down server === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:14.835Z [WARN] TestIntentionsCreate_noBody.server.serf.wan: serf: Shutdown without a Leave === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:14.837Z [DEBUG] TestIntentionsCheck_noSource.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_InmemKeyrings/ignore_files writer.go:29: 2021-01-29T19:33:14.838Z [DEBUG] TestAgent_InmemKeyrings/ignore_files.leader: stopping routine: routine="CA root pruning" === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:14.840Z [INFO] TestIntentionsCreate_noBody.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:14.840Z [INFO] TestIntentionsCreate_noBody: consul server down writer.go:29: 2021-01-29T19:33:14.843Z [INFO] TestIntentionsCreate_noBody: shutdown complete writer.go:29: 2021-01-29T19:33:14.843Z [INFO] TestIntentionsCreate_noBody: Stopping server: protocol=DNS address=127.0.0.1:29730 network=tcp writer.go:29: 2021-01-29T19:33:14.845Z [INFO] TestIntentionsCreate_noBody: Stopping server: protocol=DNS address=127.0.0.1:29730 network=udp writer.go:29: 2021-01-29T19:33:14.845Z [INFO] TestIntentionsCreate_noBody: Stopping server: protocol=HTTP address=127.0.0.1:29731 network=tcp === CONT TestAgent_InmemKeyrings/ignore_files writer.go:29: 2021-01-29T19:33:14.840Z [WARN] TestAgent_InmemKeyrings/ignore_files.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:14.840Z [DEBUG] TestAgent_InmemKeyrings/ignore_files.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.850Z [WARN] TestAgent_InmemKeyrings/ignore_files.server.serf.wan: serf: Shutdown without a Leave === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:14.843Z [INFO] TestIntentionsCheck_noSource.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4e518567-9ec2-95f8-6b68-3925a90e1742 Address:127.0.0.1:29759}]" === CONT TestAgent_InmemKeyrings/ignore_files writer.go:29: 2021-01-29T19:33:14.852Z [INFO] TestAgent_InmemKeyrings/ignore_files.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:14.852Z [INFO] TestAgent_InmemKeyrings/ignore_files: consul server down writer.go:29: 2021-01-29T19:33:14.854Z [INFO] TestAgent_InmemKeyrings/ignore_files: shutdown complete === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.855Z [WARN] TestIntentionsCheck_noDestination.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:14.853Z [INFO] TestIntentionsCheck_noSource.server.raft: entering follower state: follower="Node at 127.0.0.1:29759 [Follower]" leader= === CONT TestAgent_InmemKeyrings/ignore_files writer.go:29: 2021-01-29T19:33:14.856Z [INFO] TestAgent_InmemKeyrings/ignore_files: Stopping server: protocol=DNS address=127.0.0.1:29724 network=tcp === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.857Z [INFO] TestIntentionsCheck_noDestination.server.raft: entering candidate state: node="Node at 127.0.0.1:29747 [Candidate]" term=2 === CONT TestAgent_InmemKeyrings/ignore_files writer.go:29: 2021-01-29T19:33:14.858Z [INFO] TestAgent_InmemKeyrings/ignore_files: Stopping server: protocol=DNS address=127.0.0.1:29724 network=udp writer.go:29: 2021-01-29T19:33:14.860Z [INFO] TestAgent_InmemKeyrings/ignore_files: Stopping server: protocol=HTTP address=127.0.0.1:29725 network=tcp === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.861Z [DEBUG] TestIntentionsCheck_noDestination.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:14.862Z [DEBUG] TestIntentionsCheck_noDestination.server.raft: vote granted: from=6aac4e40-52dc-a95d-f7f1-30352dc88ae9 term=2 tally=1 === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:14.855Z [INFO] TestIntentionsCheck_noSource.server.serf.wan: serf: EventMemberJoin: Node-4e518567-9ec2-95f8-6b68-3925a90e1742.dc1 127.0.0.1 === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.864Z [INFO] TestIntentionsCheck_noDestination.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:14.866Z [INFO] TestIntentionsCheck_noDestination.server.raft: entering leader state: leader="Node at 127.0.0.1:29747 [Leader]" writer.go:29: 2021-01-29T19:33:14.868Z [INFO] TestIntentionsCheck_noDestination.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:14.869Z [INFO] TestIntentionsCheck_noDestination.server: New leader elected: payload=Node-6aac4e40-52dc-a95d-f7f1-30352dc88ae9 writer.go:29: 2021-01-29T19:33:14.869Z [DEBUG] TestIntentionsCheck_noDestination.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29747 writer.go:29: 2021-01-29T19:33:14.873Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:14.869Z [INFO] TestIntentionsCheck_noSource.server.serf.lan: serf: EventMemberJoin: Node-4e518567-9ec2-95f8-6b68-3925a90e1742 127.0.0.1 writer.go:29: 2021-01-29T19:33:14.877Z [INFO] TestIntentionsCheck_noSource.server: Handled event for server in area: event=member-join server=Node-4e518567-9ec2-95f8-6b68-3925a90e1742.dc1 area=wan writer.go:29: 2021-01-29T19:33:14.878Z [INFO] TestIntentionsCheck_noSource.server: Adding LAN server: server="Node-4e518567-9ec2-95f8-6b68-3925a90e1742 (Addr: tcp/127.0.0.1:29759) (DC: dc1)" writer.go:29: 2021-01-29T19:33:14.880Z [INFO] TestIntentionsCheck_noSource: Started DNS server: address=127.0.0.1:29754 network=tcp writer.go:29: 2021-01-29T19:33:14.881Z [INFO] TestIntentionsCheck_noSource: Started DNS server: address=127.0.0.1:29754 network=udp === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.884Z [INFO] TestIntentionsCheck_noDestination.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:14.883Z [INFO] TestIntentionsCheck_noSource: Started HTTP server: address=127.0.0.1:29755 network=tcp === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.885Z [INFO] TestIntentionsCheck_noDestination.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.886Z [DEBUG] TestIntentionsCheck_noDestination.server: Skipping self join check for node since the cluster is too small: node=Node-6aac4e40-52dc-a95d-f7f1-30352dc88ae9 writer.go:29: 2021-01-29T19:33:14.887Z [INFO] TestIntentionsCheck_noDestination.server: member joined, marking health alive: member=Node-6aac4e40-52dc-a95d-f7f1-30352dc88ae9 === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:14.886Z [INFO] TestIntentionsCheck_noSource: started state syncer === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:14.891Z [INFO] TestIntentionsCheck_noDestination: Requesting shutdown writer.go:29: 2021-01-29T19:33:14.893Z [INFO] TestIntentionsCheck_noDestination.server: shutting down server writer.go:29: 2021-01-29T19:33:14.895Z [DEBUG] TestIntentionsCheck_noDestination.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.895Z [ERROR] TestIntentionsCheck_noDestination.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:14.898Z [WARN] TestIntentionsCheck_noDestination.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:14.898Z [DEBUG] TestIntentionsCheck_noDestination.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.901Z [WARN] TestIntentionsCheck_noDestination.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:14.909Z [INFO] TestIntentionsCheck_noDestination.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:14.909Z [INFO] TestIntentionsCheck_noDestination: consul server down writer.go:29: 2021-01-29T19:33:14.911Z [INFO] TestIntentionsCheck_noDestination: shutdown complete writer.go:29: 2021-01-29T19:33:14.912Z [INFO] TestIntentionsCheck_noDestination: Stopping server: protocol=DNS address=127.0.0.1:29742 network=tcp writer.go:29: 2021-01-29T19:33:14.913Z [INFO] TestIntentionsCheck_noDestination: Stopping server: protocol=DNS address=127.0.0.1:29742 network=udp writer.go:29: 2021-01-29T19:33:14.914Z [INFO] TestIntentionsCheck_noDestination: Stopping server: protocol=HTTP address=127.0.0.1:29743 network=tcp === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:14.922Z [WARN] TestIntentionsCheck_noSource.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:14.923Z [INFO] TestIntentionsCheck_noSource.server.raft: entering candidate state: node="Node at 127.0.0.1:29759 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:14.924Z [DEBUG] TestIntentionsCheck_noSource.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:14.925Z [DEBUG] TestIntentionsCheck_noSource.server.raft: vote granted: from=4e518567-9ec2-95f8-6b68-3925a90e1742 term=2 tally=1 writer.go:29: 2021-01-29T19:33:14.926Z [INFO] TestIntentionsCheck_noSource.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:14.927Z [INFO] TestIntentionsCheck_noSource.server.raft: entering leader state: leader="Node at 127.0.0.1:29759 [Leader]" writer.go:29: 2021-01-29T19:33:14.928Z [INFO] TestIntentionsCheck_noSource.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:14.929Z [INFO] TestIntentionsCheck_noSource.server: New leader elected: payload=Node-4e518567-9ec2-95f8-6b68-3925a90e1742 writer.go:29: 2021-01-29T19:33:14.931Z [DEBUG] TestIntentionsCheck_noSource.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29759 writer.go:29: 2021-01-29T19:33:14.936Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:14.939Z [INFO] TestIntentionsCheck_noSource: Synced node info writer.go:29: 2021-01-29T19:33:14.941Z [DEBUG] TestIntentionsCheck_noSource: Node info in sync writer.go:29: 2021-01-29T19:33:14.943Z [INFO] TestIntentionsCheck_noSource.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:14.946Z [INFO] TestIntentionsCheck_noSource.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:14.948Z [DEBUG] TestIntentionsCheck_noSource.server: Skipping self join check for node since the cluster is too small: node=Node-4e518567-9ec2-95f8-6b68-3925a90e1742 writer.go:29: 2021-01-29T19:33:14.959Z [INFO] TestIntentionsCheck_noSource.server: member joined, marking health alive: member=Node-4e518567-9ec2-95f8-6b68-3925a90e1742 writer.go:29: 2021-01-29T19:33:15.188Z [INFO] TestIntentionsCheck_noSource: Requesting shutdown writer.go:29: 2021-01-29T19:33:15.190Z [INFO] TestIntentionsCheck_noSource.server: shutting down server writer.go:29: 2021-01-29T19:33:15.193Z [DEBUG] TestIntentionsCheck_noSource.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:15.195Z [WARN] TestIntentionsCheck_noSource.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:15.195Z [DEBUG] TestIntentionsCheck_noSource.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:15.199Z [WARN] TestIntentionsCheck_noSource.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:15.201Z [INFO] TestIntentionsCheck_noSource.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:15.202Z [INFO] TestIntentionsCheck_noSource: consul server down writer.go:29: 2021-01-29T19:33:15.203Z [INFO] TestIntentionsCheck_noSource: shutdown complete writer.go:29: 2021-01-29T19:33:15.203Z [INFO] TestIntentionsCheck_noSource: Stopping server: protocol=DNS address=127.0.0.1:29754 network=tcp writer.go:29: 2021-01-29T19:33:15.204Z [INFO] TestIntentionsCheck_noSource: Stopping server: protocol=DNS address=127.0.0.1:29754 network=udp writer.go:29: 2021-01-29T19:33:15.207Z [INFO] TestIntentionsCheck_noSource: Stopping server: protocol=HTTP address=127.0.0.1:29755 network=tcp === CONT TestIntentionsCreate_noBody writer.go:29: 2021-01-29T19:33:15.347Z [INFO] TestIntentionsCreate_noBody: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:15.349Z [INFO] TestIntentionsCreate_noBody: Endpoints down --- PASS: TestIntentionsCreate_noBody (0.69s) === CONT TestIntentionsCheck_basic === CONT TestAgent_InmemKeyrings/ignore_files writer.go:29: 2021-01-29T19:33:15.361Z [INFO] TestAgent_InmemKeyrings/ignore_files: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:15.362Z [INFO] TestAgent_InmemKeyrings/ignore_files: Endpoints down --- PASS: TestAgent_InmemKeyrings (3.16s) --- PASS: TestAgent_InmemKeyrings/no_keys (0.99s) --- PASS: TestAgent_InmemKeyrings/server_with_keys (0.71s) --- PASS: TestAgent_InmemKeyrings/client_with_keys (0.57s) --- PASS: TestAgent_InmemKeyrings/ignore_files (0.90s) === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.381Z [WARN] TestIntentionsMatch_noName: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:15.386Z [DEBUG] TestIntentionsMatch_noName.tlsutil: Update: version=1 === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.386Z [WARN] TestIntentionsCheck_basic: bootstrap = true: do not enable unless necessary === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.387Z [DEBUG] TestIntentionsMatch_noName.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.388Z [DEBUG] TestIntentionsCheck_basic.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:15.391Z [DEBUG] TestIntentionsCheck_basic.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.394Z [INFO] TestIntentionsMatch_noName.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:76adfd09-c313-cd68-f918-5f7f532758c8 Address:127.0.0.1:29753}]" writer.go:29: 2021-01-29T19:33:15.397Z [INFO] TestIntentionsMatch_noName.server.raft: entering follower state: follower="Node at 127.0.0.1:29753 [Follower]" leader= === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.401Z [INFO] TestIntentionsCheck_basic.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c01c66a3-64d3-baeb-d6a3-1db29a5f9887 Address:127.0.0.1:29741}]" writer.go:29: 2021-01-29T19:33:15.405Z [INFO] TestIntentionsCheck_basic.server.raft: entering follower state: follower="Node at 127.0.0.1:29741 [Follower]" leader= writer.go:29: 2021-01-29T19:33:15.405Z [INFO] TestIntentionsCheck_basic.server.serf.wan: serf: EventMemberJoin: Node-c01c66a3-64d3-baeb-d6a3-1db29a5f9887.dc1 127.0.0.1 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.397Z [INFO] TestIntentionsMatch_noName.server.serf.wan: serf: EventMemberJoin: Node-76adfd09-c313-cd68-f918-5f7f532758c8.dc1 127.0.0.1 === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:15.414Z [INFO] TestIntentionsCheck_noDestination: Waiting for endpoints to shut down === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.412Z [INFO] TestIntentionsCheck_basic.server.serf.lan: serf: EventMemberJoin: Node-c01c66a3-64d3-baeb-d6a3-1db29a5f9887 127.0.0.1 === CONT TestIntentionsCheck_noDestination writer.go:29: 2021-01-29T19:33:15.415Z [INFO] TestIntentionsCheck_noDestination: Endpoints down --- PASS: TestIntentionsCheck_noDestination (0.66s) === CONT TestIntentionsMatch_byInvalid === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.420Z [INFO] TestIntentionsCheck_basic: Started DNS server: address=127.0.0.1:29736 network=udp === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.418Z [INFO] TestIntentionsMatch_noName.server.serf.lan: serf: EventMemberJoin: Node-76adfd09-c313-cd68-f918-5f7f532758c8 127.0.0.1 === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.421Z [INFO] TestIntentionsCheck_basic.server: Adding LAN server: server="Node-c01c66a3-64d3-baeb-d6a3-1db29a5f9887 (Addr: tcp/127.0.0.1:29741) (DC: dc1)" writer.go:29: 2021-01-29T19:33:15.423Z [INFO] TestIntentionsCheck_basic: Started DNS server: address=127.0.0.1:29736 network=tcp === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.423Z [INFO] TestIntentionsMatch_noName: Started DNS server: address=127.0.0.1:29748 network=udp writer.go:29: 2021-01-29T19:33:15.424Z [INFO] TestIntentionsMatch_noName.server: Adding LAN server: server="Node-76adfd09-c313-cd68-f918-5f7f532758c8 (Addr: tcp/127.0.0.1:29753) (DC: dc1)" === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.422Z [INFO] TestIntentionsCheck_basic.server: Handled event for server in area: event=member-join server=Node-c01c66a3-64d3-baeb-d6a3-1db29a5f9887.dc1 area=wan === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.426Z [WARN] TestIntentionsMatch_byInvalid: bootstrap = true: do not enable unless necessary === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.424Z [INFO] TestIntentionsMatch_noName.server: Handled event for server in area: event=member-join server=Node-76adfd09-c313-cd68-f918-5f7f532758c8.dc1 area=wan === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.425Z [INFO] TestIntentionsCheck_basic: Started HTTP server: address=127.0.0.1:29737 network=tcp === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.427Z [DEBUG] TestIntentionsMatch_byInvalid.tlsutil: Update: version=1 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.426Z [INFO] TestIntentionsMatch_noName: Started DNS server: address=127.0.0.1:29748 network=tcp === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.427Z [INFO] TestIntentionsCheck_basic: started state syncer === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.428Z [DEBUG] TestIntentionsMatch_byInvalid.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.428Z [INFO] TestIntentionsMatch_noName: Started HTTP server: address=127.0.0.1:29749 network=tcp writer.go:29: 2021-01-29T19:33:15.431Z [INFO] TestIntentionsMatch_noName: started state syncer === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.445Z [INFO] TestIntentionsMatch_byInvalid.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:72aaef9e-bd94-b23a-bc69-66e032666172 Address:127.0.0.1:29765}]" === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.442Z [WARN] TestIntentionsMatch_noName.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:15.453Z [INFO] TestIntentionsMatch_noName.server.raft: entering candidate state: node="Node at 127.0.0.1:29753 [Candidate]" term=2 === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.455Z [WARN] TestIntentionsCheck_basic.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.454Z [INFO] TestIntentionsMatch_byInvalid.server.raft: entering follower state: follower="Node at 127.0.0.1:29765 [Follower]" leader= === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.456Z [INFO] TestIntentionsCheck_basic.server.raft: entering candidate state: node="Node at 127.0.0.1:29741 [Candidate]" term=2 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.458Z [DEBUG] TestIntentionsMatch_noName.server.raft: votes: needed=1 === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.459Z [DEBUG] TestIntentionsCheck_basic.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:15.461Z [DEBUG] TestIntentionsCheck_basic.server.raft: vote granted: from=c01c66a3-64d3-baeb-d6a3-1db29a5f9887 term=2 tally=1 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.461Z [DEBUG] TestIntentionsMatch_noName.server.raft: vote granted: from=76adfd09-c313-cd68-f918-5f7f532758c8 term=2 tally=1 writer.go:29: 2021-01-29T19:33:15.463Z [INFO] TestIntentionsMatch_noName.server.raft: election won: tally=1 === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.463Z [INFO] TestIntentionsCheck_basic.server.raft: election won: tally=1 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.465Z [INFO] TestIntentionsMatch_noName.server.raft: entering leader state: leader="Node at 127.0.0.1:29753 [Leader]" === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.466Z [INFO] TestIntentionsCheck_basic.server.raft: entering leader state: leader="Node at 127.0.0.1:29741 [Leader]" writer.go:29: 2021-01-29T19:33:15.468Z [INFO] TestIntentionsCheck_basic.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:15.471Z [INFO] TestIntentionsCheck_basic.server: New leader elected: payload=Node-c01c66a3-64d3-baeb-d6a3-1db29a5f9887 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.467Z [INFO] TestIntentionsMatch_noName.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:15.475Z [DEBUG] TestIntentionsMatch_noName.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29753 writer.go:29: 2021-01-29T19:33:15.476Z [INFO] TestIntentionsMatch_noName.server: New leader elected: payload=Node-76adfd09-c313-cd68-f918-5f7f532758c8 === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.468Z [INFO] TestIntentionsMatch_byInvalid.server.serf.wan: serf: EventMemberJoin: Node-72aaef9e-bd94-b23a-bc69-66e032666172.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:15.483Z [INFO] TestIntentionsMatch_byInvalid.server.serf.lan: serf: EventMemberJoin: Node-72aaef9e-bd94-b23a-bc69-66e032666172 127.0.0.1 writer.go:29: 2021-01-29T19:33:15.489Z [INFO] TestIntentionsMatch_byInvalid: Started DNS server: address=127.0.0.1:29760 network=udp === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.480Z [DEBUG] TestIntentionsCheck_basic.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29741 === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.498Z [INFO] TestIntentionsMatch_byInvalid.server: Adding LAN server: server="Node-72aaef9e-bd94-b23a-bc69-66e032666172 (Addr: tcp/127.0.0.1:29765) (DC: dc1)" === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.503Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.501Z [INFO] TestIntentionsMatch_byInvalid.server: Handled event for server in area: event=member-join server=Node-72aaef9e-bd94-b23a-bc69-66e032666172.dc1 area=wan === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.506Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.512Z [INFO] TestIntentionsMatch_byInvalid: Started DNS server: address=127.0.0.1:29760 network=tcp writer.go:29: 2021-01-29T19:33:15.517Z [INFO] TestIntentionsMatch_byInvalid: Started HTTP server: address=127.0.0.1:29761 network=tcp writer.go:29: 2021-01-29T19:33:15.518Z [INFO] TestIntentionsMatch_byInvalid: started state syncer writer.go:29: 2021-01-29T19:33:15.522Z [WARN] TestIntentionsMatch_byInvalid.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:15.523Z [INFO] TestIntentionsMatch_byInvalid.server.raft: entering candidate state: node="Node at 127.0.0.1:29765 [Candidate]" term=2 === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.530Z [INFO] TestIntentionsCheck_basic.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.530Z [DEBUG] TestIntentionsMatch_byInvalid.server.raft: votes: needed=1 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.527Z [INFO] TestIntentionsMatch_noName.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.532Z [INFO] TestIntentionsCheck_basic.leader: started routine: routine="CA root pruning" === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.533Z [DEBUG] TestIntentionsMatch_byInvalid.server.raft: vote granted: from=72aaef9e-bd94-b23a-bc69-66e032666172 term=2 tally=1 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.535Z [INFO] TestIntentionsMatch_noName.leader: started routine: routine="CA root pruning" === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.536Z [INFO] TestIntentionsMatch_byInvalid.server.raft: election won: tally=1 === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.539Z [DEBUG] TestIntentionsMatch_noName.server: Skipping self join check for node since the cluster is too small: node=Node-76adfd09-c313-cd68-f918-5f7f532758c8 === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.544Z [INFO] TestIntentionsMatch_byInvalid.server.raft: entering leader state: leader="Node at 127.0.0.1:29765 [Leader]" writer.go:29: 2021-01-29T19:33:15.549Z [INFO] TestIntentionsMatch_byInvalid.server: cluster leadership acquired === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.546Z [INFO] TestIntentionsMatch_noName.server: member joined, marking health alive: member=Node-76adfd09-c313-cd68-f918-5f7f532758c8 === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.535Z [DEBUG] TestIntentionsCheck_basic.server: Skipping self join check for node since the cluster is too small: node=Node-c01c66a3-64d3-baeb-d6a3-1db29a5f9887 writer.go:29: 2021-01-29T19:33:15.553Z [INFO] TestIntentionsCheck_basic.server: member joined, marking health alive: member=Node-c01c66a3-64d3-baeb-d6a3-1db29a5f9887 === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.555Z [INFO] TestIntentionsMatch_byInvalid.server: New leader elected: payload=Node-72aaef9e-bd94-b23a-bc69-66e032666172 writer.go:29: 2021-01-29T19:33:15.559Z [DEBUG] TestIntentionsMatch_byInvalid.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29765 writer.go:29: 2021-01-29T19:33:15.573Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:15.600Z [INFO] TestIntentionsMatch_byInvalid.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:15.601Z [INFO] TestIntentionsMatch_byInvalid.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:15.602Z [DEBUG] TestIntentionsMatch_byInvalid.server: Skipping self join check for node since the cluster is too small: node=Node-72aaef9e-bd94-b23a-bc69-66e032666172 writer.go:29: 2021-01-29T19:33:15.603Z [INFO] TestIntentionsMatch_byInvalid.server: member joined, marking health alive: member=Node-72aaef9e-bd94-b23a-bc69-66e032666172 === CONT TestIntentionsCheck_noSource writer.go:29: 2021-01-29T19:33:15.710Z [INFO] TestIntentionsCheck_noSource: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:15.711Z [INFO] TestIntentionsCheck_noSource: Endpoints down --- PASS: TestIntentionsCheck_noSource (0.89s) === CONT TestIntentionsMatch_noBy === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.712Z [INFO] TestIntentionsMatch_noName: Requesting shutdown writer.go:29: 2021-01-29T19:33:15.715Z [INFO] TestIntentionsMatch_noName.server: shutting down server writer.go:29: 2021-01-29T19:33:15.716Z [DEBUG] TestIntentionsMatch_noName.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:15.717Z [WARN] TestIntentionsMatch_noName.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:15.718Z [ERROR] TestIntentionsMatch_noName.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.719Z [DEBUG] TestIntentionsCheck_basic: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:15.723Z [INFO] TestIntentionsCheck_basic: Synced node info === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.718Z [DEBUG] TestIntentionsMatch_noName.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:15.727Z [WARN] TestIntentionsMatch_noName.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:15.759Z [INFO] TestIntentionsMatch_noName.server.router.manager: shutting down === CONT TestIntentionsMatch_noBy writer.go:29: 2021-01-29T19:33:15.760Z [WARN] TestIntentionsMatch_noBy: bootstrap = true: do not enable unless necessary === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:15.761Z [INFO] TestIntentionsMatch_noName: consul server down writer.go:29: 2021-01-29T19:33:15.761Z [INFO] TestIntentionsMatch_noName: shutdown complete writer.go:29: 2021-01-29T19:33:15.762Z [INFO] TestIntentionsMatch_noName: Stopping server: protocol=DNS address=127.0.0.1:29748 network=tcp writer.go:29: 2021-01-29T19:33:15.763Z [INFO] TestIntentionsMatch_noName: Stopping server: protocol=DNS address=127.0.0.1:29748 network=udp writer.go:29: 2021-01-29T19:33:15.764Z [INFO] TestIntentionsMatch_noName: Stopping server: protocol=HTTP address=127.0.0.1:29749 network=tcp === CONT TestIntentionsMatch_noBy writer.go:29: 2021-01-29T19:33:15.772Z [DEBUG] TestIntentionsMatch_noBy.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:15.775Z [DEBUG] TestIntentionsMatch_noBy.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:15.790Z [INFO] TestIntentionsMatch_noBy.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7c97fe8a-a3b2-147d-0471-09b307a4f11c Address:127.0.0.1:29771}]" writer.go:29: 2021-01-29T19:33:15.793Z [INFO] TestIntentionsMatch_noBy.server.raft: entering follower state: follower="Node at 127.0.0.1:29771 [Follower]" leader= writer.go:29: 2021-01-29T19:33:15.795Z [INFO] TestIntentionsMatch_noBy.server.serf.wan: serf: EventMemberJoin: Node-7c97fe8a-a3b2-147d-0471-09b307a4f11c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:15.799Z [INFO] TestIntentionsMatch_noBy.server.serf.lan: serf: EventMemberJoin: Node-7c97fe8a-a3b2-147d-0471-09b307a4f11c 127.0.0.1 writer.go:29: 2021-01-29T19:33:15.801Z [INFO] TestIntentionsMatch_noBy.server: Handled event for server in area: event=member-join server=Node-7c97fe8a-a3b2-147d-0471-09b307a4f11c.dc1 area=wan writer.go:29: 2021-01-29T19:33:15.801Z [INFO] TestIntentionsMatch_noBy.server: Adding LAN server: server="Node-7c97fe8a-a3b2-147d-0471-09b307a4f11c (Addr: tcp/127.0.0.1:29771) (DC: dc1)" writer.go:29: 2021-01-29T19:33:15.803Z [INFO] TestIntentionsMatch_noBy: Started DNS server: address=127.0.0.1:29766 network=tcp writer.go:29: 2021-01-29T19:33:15.805Z [INFO] TestIntentionsMatch_noBy: Started DNS server: address=127.0.0.1:29766 network=udp writer.go:29: 2021-01-29T19:33:15.806Z [INFO] TestIntentionsMatch_noBy: Started HTTP server: address=127.0.0.1:29767 network=tcp writer.go:29: 2021-01-29T19:33:15.807Z [INFO] TestIntentionsMatch_noBy: started state syncer writer.go:29: 2021-01-29T19:33:15.856Z [WARN] TestIntentionsMatch_noBy.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:15.858Z [INFO] TestIntentionsMatch_noBy.server.raft: entering candidate state: node="Node at 127.0.0.1:29771 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:15.860Z [DEBUG] TestIntentionsMatch_noBy.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:15.862Z [DEBUG] TestIntentionsMatch_noBy.server.raft: vote granted: from=7c97fe8a-a3b2-147d-0471-09b307a4f11c term=2 tally=1 writer.go:29: 2021-01-29T19:33:15.863Z [INFO] TestIntentionsMatch_noBy.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:15.864Z [INFO] TestIntentionsMatch_noBy.server.raft: entering leader state: leader="Node at 127.0.0.1:29771 [Leader]" === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.864Z [DEBUG] TestIntentionsMatch_byInvalid: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestIntentionsMatch_noBy writer.go:29: 2021-01-29T19:33:15.866Z [INFO] TestIntentionsMatch_noBy.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:15.866Z [INFO] TestIntentionsMatch_noBy.server: New leader elected: payload=Node-7c97fe8a-a3b2-147d-0471-09b307a4f11c === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.868Z [INFO] TestIntentionsMatch_byInvalid: Synced node info === CONT TestIntentionsMatch_noBy writer.go:29: 2021-01-29T19:33:15.871Z [DEBUG] TestIntentionsMatch_noBy.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29771 writer.go:29: 2021-01-29T19:33:15.875Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:15.881Z [INFO] TestIntentionsMatch_noBy.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:15.882Z [INFO] TestIntentionsMatch_noBy.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:15.883Z [DEBUG] TestIntentionsMatch_noBy.server: Skipping self join check for node since the cluster is too small: node=Node-7c97fe8a-a3b2-147d-0471-09b307a4f11c writer.go:29: 2021-01-29T19:33:15.885Z [INFO] TestIntentionsMatch_noBy.server: member joined, marking health alive: member=Node-7c97fe8a-a3b2-147d-0471-09b307a4f11c === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.893Z [INFO] TestIntentionsCheck_basic: Requesting shutdown === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.920Z [INFO] TestIntentionsMatch_byInvalid: Requesting shutdown writer.go:29: 2021-01-29T19:33:15.921Z [INFO] TestIntentionsMatch_byInvalid.server: shutting down server writer.go:29: 2021-01-29T19:33:15.922Z [DEBUG] TestIntentionsMatch_byInvalid.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:15.923Z [WARN] TestIntentionsMatch_byInvalid.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:15.924Z [DEBUG] TestIntentionsMatch_byInvalid.leader: stopped routine: routine="CA root pruning" === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.894Z [INFO] TestIntentionsCheck_basic.server: shutting down server writer.go:29: 2021-01-29T19:33:15.927Z [DEBUG] TestIntentionsCheck_basic.leader: stopping routine: routine="CA root pruning" === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.931Z [WARN] TestIntentionsMatch_byInvalid.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:15.934Z [INFO] TestIntentionsMatch_byInvalid.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:15.934Z [INFO] TestIntentionsMatch_byInvalid: consul server down === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.928Z [WARN] TestIntentionsCheck_basic.server.serf.lan: serf: Shutdown without a Leave === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.936Z [INFO] TestIntentionsMatch_byInvalid: shutdown complete === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.928Z [DEBUG] TestIntentionsCheck_basic.leader: stopped routine: routine="CA root pruning" === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.937Z [INFO] TestIntentionsMatch_byInvalid: Stopping server: protocol=DNS address=127.0.0.1:29760 network=tcp === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.937Z [WARN] TestIntentionsCheck_basic.server.serf.wan: serf: Shutdown without a Leave === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:15.938Z [INFO] TestIntentionsMatch_byInvalid: Stopping server: protocol=DNS address=127.0.0.1:29760 network=udp writer.go:29: 2021-01-29T19:33:15.939Z [INFO] TestIntentionsMatch_byInvalid: Stopping server: protocol=HTTP address=127.0.0.1:29761 network=tcp === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:15.939Z [INFO] TestIntentionsCheck_basic.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:15.939Z [INFO] TestIntentionsCheck_basic: consul server down writer.go:29: 2021-01-29T19:33:15.942Z [INFO] TestIntentionsCheck_basic: shutdown complete writer.go:29: 2021-01-29T19:33:15.944Z [INFO] TestIntentionsCheck_basic: Stopping server: protocol=DNS address=127.0.0.1:29736 network=tcp writer.go:29: 2021-01-29T19:33:15.945Z [INFO] TestIntentionsCheck_basic: Stopping server: protocol=DNS address=127.0.0.1:29736 network=udp writer.go:29: 2021-01-29T19:33:15.947Z [INFO] TestIntentionsCheck_basic: Stopping server: protocol=HTTP address=127.0.0.1:29737 network=tcp === CONT TestIntentionsMatch_noBy writer.go:29: 2021-01-29T19:33:16.101Z [INFO] TestIntentionsMatch_noBy: Requesting shutdown writer.go:29: 2021-01-29T19:33:16.104Z [INFO] TestIntentionsMatch_noBy.server: shutting down server writer.go:29: 2021-01-29T19:33:16.105Z [DEBUG] TestIntentionsMatch_noBy.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:16.107Z [WARN] TestIntentionsMatch_noBy.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:16.109Z [ERROR] TestIntentionsMatch_noBy.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:16.110Z [DEBUG] TestIntentionsMatch_noBy.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:16.110Z [WARN] TestIntentionsMatch_noBy.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:16.124Z [INFO] TestIntentionsMatch_noBy.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:16.124Z [INFO] TestIntentionsMatch_noBy: consul server down writer.go:29: 2021-01-29T19:33:16.127Z [INFO] TestIntentionsMatch_noBy: shutdown complete writer.go:29: 2021-01-29T19:33:16.128Z [INFO] TestIntentionsMatch_noBy: Stopping server: protocol=DNS address=127.0.0.1:29766 network=tcp writer.go:29: 2021-01-29T19:33:16.130Z [INFO] TestIntentionsMatch_noBy: Stopping server: protocol=DNS address=127.0.0.1:29766 network=udp writer.go:29: 2021-01-29T19:33:16.131Z [INFO] TestIntentionsMatch_noBy: Stopping server: protocol=HTTP address=127.0.0.1:29767 network=tcp === CONT TestIntentionsMatch_noName writer.go:29: 2021-01-29T19:33:16.265Z [INFO] TestIntentionsMatch_noName: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:16.267Z [INFO] TestIntentionsMatch_noName: Endpoints down --- PASS: TestIntentionsMatch_noName (0.91s) === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.290Z [WARN] TestIntentionsMatch_basic: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:16.292Z [DEBUG] TestIntentionsMatch_basic.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:16.294Z [DEBUG] TestIntentionsMatch_basic.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:16.308Z [INFO] TestIntentionsMatch_basic.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9977d29f-11b3-c245-4cce-0fe48ae84c8d Address:127.0.0.1:29777}]" writer.go:29: 2021-01-29T19:33:16.311Z [INFO] TestIntentionsMatch_basic.server.raft: entering follower state: follower="Node at 127.0.0.1:29777 [Follower]" leader= writer.go:29: 2021-01-29T19:33:16.312Z [INFO] TestIntentionsMatch_basic.server.serf.wan: serf: EventMemberJoin: Node-9977d29f-11b3-c245-4cce-0fe48ae84c8d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:16.317Z [INFO] TestIntentionsMatch_basic.server.serf.lan: serf: EventMemberJoin: Node-9977d29f-11b3-c245-4cce-0fe48ae84c8d 127.0.0.1 writer.go:29: 2021-01-29T19:33:16.320Z [INFO] TestIntentionsMatch_basic.server: Adding LAN server: server="Node-9977d29f-11b3-c245-4cce-0fe48ae84c8d (Addr: tcp/127.0.0.1:29777) (DC: dc1)" writer.go:29: 2021-01-29T19:33:16.321Z [INFO] TestIntentionsMatch_basic.server: Handled event for server in area: event=member-join server=Node-9977d29f-11b3-c245-4cce-0fe48ae84c8d.dc1 area=wan writer.go:29: 2021-01-29T19:33:16.323Z [INFO] TestIntentionsMatch_basic: Started DNS server: address=127.0.0.1:29772 network=tcp writer.go:29: 2021-01-29T19:33:16.328Z [INFO] TestIntentionsMatch_basic: Started DNS server: address=127.0.0.1:29772 network=udp writer.go:29: 2021-01-29T19:33:16.330Z [INFO] TestIntentionsMatch_basic: Started HTTP server: address=127.0.0.1:29773 network=tcp writer.go:29: 2021-01-29T19:33:16.331Z [INFO] TestIntentionsMatch_basic: started state syncer writer.go:29: 2021-01-29T19:33:16.350Z [WARN] TestIntentionsMatch_basic.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:16.351Z [INFO] TestIntentionsMatch_basic.server.raft: entering candidate state: node="Node at 127.0.0.1:29777 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:16.353Z [DEBUG] TestIntentionsMatch_basic.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:16.354Z [DEBUG] TestIntentionsMatch_basic.server.raft: vote granted: from=9977d29f-11b3-c245-4cce-0fe48ae84c8d term=2 tally=1 writer.go:29: 2021-01-29T19:33:16.356Z [INFO] TestIntentionsMatch_basic.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:16.357Z [INFO] TestIntentionsMatch_basic.server.raft: entering leader state: leader="Node at 127.0.0.1:29777 [Leader]" writer.go:29: 2021-01-29T19:33:16.358Z [INFO] TestIntentionsMatch_basic.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:16.359Z [INFO] TestIntentionsMatch_basic.server: New leader elected: payload=Node-9977d29f-11b3-c245-4cce-0fe48ae84c8d writer.go:29: 2021-01-29T19:33:16.360Z [DEBUG] TestIntentionsMatch_basic.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29777 writer.go:29: 2021-01-29T19:33:16.370Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:16.383Z [INFO] TestIntentionsMatch_basic.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:16.386Z [INFO] TestIntentionsMatch_basic.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:16.390Z [DEBUG] TestIntentionsMatch_basic.server: Skipping self join check for node since the cluster is too small: node=Node-9977d29f-11b3-c245-4cce-0fe48ae84c8d writer.go:29: 2021-01-29T19:33:16.394Z [INFO] TestIntentionsMatch_basic.server: member joined, marking health alive: member=Node-9977d29f-11b3-c245-4cce-0fe48ae84c8d === CONT TestIntentionsMatch_byInvalid writer.go:29: 2021-01-29T19:33:16.440Z [INFO] TestIntentionsMatch_byInvalid: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:16.441Z [INFO] TestIntentionsMatch_byInvalid: Endpoints down --- PASS: TestIntentionsMatch_byInvalid (1.03s) === CONT TestIntentionsList_values === CONT TestIntentionsCheck_basic writer.go:29: 2021-01-29T19:33:16.448Z [INFO] TestIntentionsCheck_basic: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:16.449Z [INFO] TestIntentionsCheck_basic: Endpoints down --- PASS: TestIntentionsCheck_basic (1.10s) === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.459Z [WARN] TestIntentionsList_empty: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:16.460Z [DEBUG] TestIntentionsList_empty.tlsutil: Update: version=1 === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.462Z [WARN] TestIntentionsList_values: bootstrap = true: do not enable unless necessary === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.463Z [DEBUG] TestIntentionsList_empty.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:16.468Z [INFO] TestIntentionsList_empty.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6555c6de-88b7-2fb5-4257-c1b1f793c811 Address:127.0.0.1:29789}]" writer.go:29: 2021-01-29T19:33:16.470Z [INFO] TestIntentionsList_empty.server.raft: entering follower state: follower="Node at 127.0.0.1:29789 [Follower]" leader= writer.go:29: 2021-01-29T19:33:16.471Z [INFO] TestIntentionsList_empty.server.serf.wan: serf: EventMemberJoin: Node-6555c6de-88b7-2fb5-4257-c1b1f793c811.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:16.474Z [INFO] TestIntentionsList_empty.server.serf.lan: serf: EventMemberJoin: Node-6555c6de-88b7-2fb5-4257-c1b1f793c811 127.0.0.1 writer.go:29: 2021-01-29T19:33:16.476Z [INFO] TestIntentionsList_empty.server: Handled event for server in area: event=member-join server=Node-6555c6de-88b7-2fb5-4257-c1b1f793c811.dc1 area=wan writer.go:29: 2021-01-29T19:33:16.476Z [INFO] TestIntentionsList_empty.server: Adding LAN server: server="Node-6555c6de-88b7-2fb5-4257-c1b1f793c811 (Addr: tcp/127.0.0.1:29789) (DC: dc1)" === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.464Z [DEBUG] TestIntentionsList_values.tlsutil: Update: version=1 === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.476Z [INFO] TestIntentionsList_empty: Started DNS server: address=127.0.0.1:29784 network=udp writer.go:29: 2021-01-29T19:33:16.479Z [INFO] TestIntentionsList_empty: Started DNS server: address=127.0.0.1:29784 network=tcp === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.479Z [DEBUG] TestIntentionsList_values.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.483Z [INFO] TestIntentionsList_empty: Started HTTP server: address=127.0.0.1:29785 network=tcp writer.go:29: 2021-01-29T19:33:16.485Z [INFO] TestIntentionsList_empty: started state syncer === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.489Z [INFO] TestIntentionsList_values.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e4eb69fe-f6aa-d903-4df6-a24bf00a5fe4 Address:127.0.0.1:29783}]" writer.go:29: 2021-01-29T19:33:16.492Z [INFO] TestIntentionsList_values.server.raft: entering follower state: follower="Node at 127.0.0.1:29783 [Follower]" leader= writer.go:29: 2021-01-29T19:33:16.493Z [INFO] TestIntentionsList_values.server.serf.wan: serf: EventMemberJoin: Node-e4eb69fe-f6aa-d903-4df6-a24bf00a5fe4.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:16.496Z [INFO] TestIntentionsList_values.server.serf.lan: serf: EventMemberJoin: Node-e4eb69fe-f6aa-d903-4df6-a24bf00a5fe4 127.0.0.1 writer.go:29: 2021-01-29T19:33:16.497Z [INFO] TestIntentionsList_values.server: Handled event for server in area: event=member-join server=Node-e4eb69fe-f6aa-d903-4df6-a24bf00a5fe4.dc1 area=wan writer.go:29: 2021-01-29T19:33:16.498Z [INFO] TestIntentionsList_values.server: Adding LAN server: server="Node-e4eb69fe-f6aa-d903-4df6-a24bf00a5fe4 (Addr: tcp/127.0.0.1:29783) (DC: dc1)" writer.go:29: 2021-01-29T19:33:16.498Z [INFO] TestIntentionsList_values: Started DNS server: address=127.0.0.1:29778 network=udp writer.go:29: 2021-01-29T19:33:16.501Z [INFO] TestIntentionsList_values: Started DNS server: address=127.0.0.1:29778 network=tcp writer.go:29: 2021-01-29T19:33:16.504Z [INFO] TestIntentionsList_values: Started HTTP server: address=127.0.0.1:29779 network=tcp writer.go:29: 2021-01-29T19:33:16.505Z [INFO] TestIntentionsList_values: started state syncer === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.511Z [WARN] TestIntentionsList_empty.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:16.513Z [INFO] TestIntentionsList_empty.server.raft: entering candidate state: node="Node at 127.0.0.1:29789 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:16.522Z [DEBUG] TestIntentionsList_empty.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:16.523Z [DEBUG] TestIntentionsList_empty.server.raft: vote granted: from=6555c6de-88b7-2fb5-4257-c1b1f793c811 term=2 tally=1 writer.go:29: 2021-01-29T19:33:16.524Z [INFO] TestIntentionsList_empty.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:16.532Z [INFO] TestIntentionsList_empty.server.raft: entering leader state: leader="Node at 127.0.0.1:29789 [Leader]" writer.go:29: 2021-01-29T19:33:16.535Z [INFO] TestIntentionsList_empty.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:16.536Z [INFO] TestIntentionsList_empty.server: New leader elected: payload=Node-6555c6de-88b7-2fb5-4257-c1b1f793c811 writer.go:29: 2021-01-29T19:33:16.536Z [DEBUG] TestIntentionsList_empty.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29789 writer.go:29: 2021-01-29T19:33:16.543Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.546Z [INFO] TestIntentionsMatch_basic: Requesting shutdown === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.550Z [INFO] TestIntentionsList_empty.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.547Z [INFO] TestIntentionsMatch_basic.server: shutting down server writer.go:29: 2021-01-29T19:33:16.552Z [DEBUG] TestIntentionsMatch_basic.leader: stopping routine: routine="CA root pruning" === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.552Z [INFO] TestIntentionsList_empty.leader: started routine: routine="CA root pruning" === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.553Z [WARN] TestIntentionsMatch_basic.server.serf.lan: serf: Shutdown without a Leave === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.553Z [DEBUG] TestIntentionsList_empty.server: Skipping self join check for node since the cluster is too small: node=Node-6555c6de-88b7-2fb5-4257-c1b1f793c811 === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.553Z [DEBUG] TestIntentionsMatch_basic.leader: stopped routine: routine="CA root pruning" === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.555Z [WARN] TestIntentionsList_values.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:16.555Z [INFO] TestIntentionsList_values.server.raft: entering candidate state: node="Node at 127.0.0.1:29783 [Candidate]" term=2 === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.552Z [ERROR] TestIntentionsMatch_basic.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.555Z [INFO] TestIntentionsList_empty.server: member joined, marking health alive: member=Node-6555c6de-88b7-2fb5-4257-c1b1f793c811 === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.557Z [WARN] TestIntentionsMatch_basic.server.serf.wan: serf: Shutdown without a Leave === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.557Z [DEBUG] TestIntentionsList_values.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:16.558Z [DEBUG] TestIntentionsList_values.server.raft: vote granted: from=e4eb69fe-f6aa-d903-4df6-a24bf00a5fe4 term=2 tally=1 === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.558Z [INFO] TestIntentionsMatch_basic.server.router.manager: shutting down === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.559Z [INFO] TestIntentionsList_values.server.raft: election won: tally=1 === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.559Z [INFO] TestIntentionsMatch_basic: consul server down === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.560Z [INFO] TestIntentionsList_values.server.raft: entering leader state: leader="Node at 127.0.0.1:29783 [Leader]" === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.560Z [INFO] TestIntentionsMatch_basic: shutdown complete writer.go:29: 2021-01-29T19:33:16.561Z [INFO] TestIntentionsMatch_basic: Stopping server: protocol=DNS address=127.0.0.1:29772 network=tcp === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.561Z [INFO] TestIntentionsList_values.server: cluster leadership acquired === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.562Z [INFO] TestIntentionsMatch_basic: Stopping server: protocol=DNS address=127.0.0.1:29772 network=udp === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.561Z [INFO] TestIntentionsList_values.server: New leader elected: payload=Node-e4eb69fe-f6aa-d903-4df6-a24bf00a5fe4 === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:16.563Z [INFO] TestIntentionsMatch_basic: Stopping server: protocol=HTTP address=127.0.0.1:29773 network=tcp === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.562Z [DEBUG] TestIntentionsList_values.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29783 writer.go:29: 2021-01-29T19:33:16.573Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:16.585Z [INFO] TestIntentionsList_values.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:16.587Z [INFO] TestIntentionsList_values.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:16.588Z [DEBUG] TestIntentionsList_values.server: Skipping self join check for node since the cluster is too small: node=Node-e4eb69fe-f6aa-d903-4df6-a24bf00a5fe4 writer.go:29: 2021-01-29T19:33:16.590Z [INFO] TestIntentionsList_values.server: member joined, marking health alive: member=Node-e4eb69fe-f6aa-d903-4df6-a24bf00a5fe4 === CONT TestIntentionsMatch_noBy writer.go:29: 2021-01-29T19:33:16.641Z [INFO] TestIntentionsMatch_noBy: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:16.643Z [INFO] TestIntentionsMatch_noBy: Endpoints down --- PASS: TestIntentionsMatch_noBy (0.93s) === CONT TestRPC_HTTPSMaxConnsPerClient === RUN TestRPC_HTTPSMaxConnsPerClient/HTTP writer.go:29: 2021-01-29T19:33:16.693Z [WARN] TestRPC_HTTPSMaxConnsPerClient/HTTP: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:16.694Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTP.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:16.696Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTP.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:16.713Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8aa31f9d-e50d-1ede-c8a2-2777cda8cd7c Address:127.0.0.1:29801}]" writer.go:29: 2021-01-29T19:33:16.716Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.raft: entering follower state: follower="Node at 127.0.0.1:29801 [Follower]" leader= writer.go:29: 2021-01-29T19:33:16.722Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.serf.wan: serf: EventMemberJoin: Node-8aa31f9d-e50d-1ede-c8a2-2777cda8cd7c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:16.729Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.serf.lan: serf: EventMemberJoin: Node-8aa31f9d-e50d-1ede-c8a2-2777cda8cd7c 127.0.0.1 writer.go:29: 2021-01-29T19:33:16.731Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: Started DNS server: address=127.0.0.1:29796 network=udp writer.go:29: 2021-01-29T19:33:16.732Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server: Adding LAN server: server="Node-8aa31f9d-e50d-1ede-c8a2-2777cda8cd7c (Addr: tcp/127.0.0.1:29801) (DC: dc1)" writer.go:29: 2021-01-29T19:33:16.733Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server: Handled event for server in area: event=member-join server=Node-8aa31f9d-e50d-1ede-c8a2-2777cda8cd7c.dc1 area=wan writer.go:29: 2021-01-29T19:33:16.735Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: Started DNS server: address=127.0.0.1:29796 network=tcp writer.go:29: 2021-01-29T19:33:16.741Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: Started HTTP server: address=127.0.0.1:29797 network=tcp writer.go:29: 2021-01-29T19:33:16.743Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: started state syncer writer.go:29: 2021-01-29T19:33:16.760Z [WARN] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:16.761Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.raft: entering candidate state: node="Node at 127.0.0.1:29801 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:16.762Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.raft: votes: needed=1 === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.762Z [DEBUG] TestIntentionsList_empty: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestRPC_HTTPSMaxConnsPerClient/HTTP writer.go:29: 2021-01-29T19:33:16.763Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.raft: vote granted: from=8aa31f9d-e50d-1ede-c8a2-2777cda8cd7c term=2 tally=1 writer.go:29: 2021-01-29T19:33:16.764Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:16.765Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.raft: entering leader state: leader="Node at 127.0.0.1:29801 [Leader]" === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.764Z [INFO] TestIntentionsList_empty: Synced node info === CONT TestRPC_HTTPSMaxConnsPerClient/HTTP writer.go:29: 2021-01-29T19:33:16.766Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:16.766Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server: New leader elected: payload=Node-8aa31f9d-e50d-1ede-c8a2-2777cda8cd7c === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.766Z [DEBUG] TestIntentionsList_empty: Node info in sync === CONT TestRPC_HTTPSMaxConnsPerClient/HTTP writer.go:29: 2021-01-29T19:33:16.768Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTP.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29801 writer.go:29: 2021-01-29T19:33:16.773Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:16.788Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:16.789Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:16.791Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTP.server: Skipping self join check for node since the cluster is too small: node=Node-8aa31f9d-e50d-1ede-c8a2-2777cda8cd7c writer.go:29: 2021-01-29T19:33:16.792Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server: member joined, marking health alive: member=Node-8aa31f9d-e50d-1ede-c8a2-2777cda8cd7c === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:16.819Z [DEBUG] TestIntentionsList_values: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:16.823Z [INFO] TestIntentionsList_values: Synced node info writer.go:29: 2021-01-29T19:33:16.835Z [INFO] TestIntentionsList_values: Requesting shutdown writer.go:29: 2021-01-29T19:33:16.841Z [INFO] TestIntentionsList_values.server: shutting down server writer.go:29: 2021-01-29T19:33:16.843Z [DEBUG] TestIntentionsList_values.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:16.844Z [WARN] TestIntentionsList_values.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:16.844Z [DEBUG] TestIntentionsList_values.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:16.852Z [WARN] TestIntentionsList_values.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:16.857Z [INFO] TestIntentionsList_values.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:16.857Z [INFO] TestIntentionsList_values: consul server down writer.go:29: 2021-01-29T19:33:16.859Z [INFO] TestIntentionsList_values: shutdown complete writer.go:29: 2021-01-29T19:33:16.859Z [INFO] TestIntentionsList_values: Stopping server: protocol=DNS address=127.0.0.1:29778 network=tcp writer.go:29: 2021-01-29T19:33:16.860Z [INFO] TestIntentionsList_values: Stopping server: protocol=DNS address=127.0.0.1:29778 network=udp writer.go:29: 2021-01-29T19:33:16.862Z [INFO] TestIntentionsList_values: Stopping server: protocol=HTTP address=127.0.0.1:29779 network=tcp === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:16.871Z [INFO] TestIntentionsList_empty: Requesting shutdown writer.go:29: 2021-01-29T19:33:16.873Z [INFO] TestIntentionsList_empty.server: shutting down server writer.go:29: 2021-01-29T19:33:16.874Z [DEBUG] TestIntentionsList_empty.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:16.876Z [WARN] TestIntentionsList_empty.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:16.876Z [DEBUG] TestIntentionsList_empty.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:16.880Z [WARN] TestIntentionsList_empty.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:16.882Z [INFO] TestIntentionsList_empty.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:16.882Z [INFO] TestIntentionsList_empty: consul server down writer.go:29: 2021-01-29T19:33:16.885Z [INFO] TestIntentionsList_empty: shutdown complete writer.go:29: 2021-01-29T19:33:16.886Z [INFO] TestIntentionsList_empty: Stopping server: protocol=DNS address=127.0.0.1:29784 network=tcp writer.go:29: 2021-01-29T19:33:16.888Z [INFO] TestIntentionsList_empty: Stopping server: protocol=DNS address=127.0.0.1:29784 network=udp writer.go:29: 2021-01-29T19:33:16.891Z [INFO] TestIntentionsList_empty: Stopping server: protocol=HTTP address=127.0.0.1:29785 network=tcp === CONT TestRPC_HTTPSMaxConnsPerClient/HTTP writer.go:29: 2021-01-29T19:33:16.989Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTP.tlsutil: Update: version=2 writer.go:29: 2021-01-29T19:33:16.992Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: Requesting shutdown writer.go:29: 2021-01-29T19:33:16.994Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server: shutting down server writer.go:29: 2021-01-29T19:33:16.995Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTP.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:16.997Z [WARN] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:16.996Z [ERROR] TestRPC_HTTPSMaxConnsPerClient/HTTP.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:16.997Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTP.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.003Z [WARN] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:17.013Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:17.013Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: consul server down writer.go:29: 2021-01-29T19:33:17.015Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: shutdown complete writer.go:29: 2021-01-29T19:33:17.017Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: Stopping server: protocol=DNS address=127.0.0.1:29796 network=tcp writer.go:29: 2021-01-29T19:33:17.018Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: Stopping server: protocol=DNS address=127.0.0.1:29796 network=udp writer.go:29: 2021-01-29T19:33:17.021Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: Stopping server: protocol=HTTP address=127.0.0.1:29797 network=tcp === CONT TestIntentionsMatch_basic writer.go:29: 2021-01-29T19:33:17.064Z [INFO] TestIntentionsMatch_basic: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:17.065Z [INFO] TestIntentionsMatch_basic: Endpoints down --- PASS: TestIntentionsMatch_basic (0.80s) === CONT TestHTTPServer_HandshakeTimeout writer.go:29: 2021-01-29T19:33:17.087Z [WARN] TestHTTPServer_HandshakeTimeout: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:17.090Z [DEBUG] TestHTTPServer_HandshakeTimeout.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:17.094Z [DEBUG] TestHTTPServer_HandshakeTimeout.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:17.103Z [INFO] TestHTTPServer_HandshakeTimeout.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9e6d51a1-e08a-c1c4-6e4b-160e6114c12a Address:127.0.0.1:29795}]" writer.go:29: 2021-01-29T19:33:17.111Z [INFO] TestHTTPServer_HandshakeTimeout.server.raft: entering follower state: follower="Node at 127.0.0.1:29795 [Follower]" leader= writer.go:29: 2021-01-29T19:33:17.119Z [INFO] TestHTTPServer_HandshakeTimeout.server.serf.wan: serf: EventMemberJoin: Node-9e6d51a1-e08a-c1c4-6e4b-160e6114c12a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:17.145Z [INFO] TestHTTPServer_HandshakeTimeout.server.serf.lan: serf: EventMemberJoin: Node-9e6d51a1-e08a-c1c4-6e4b-160e6114c12a 127.0.0.1 writer.go:29: 2021-01-29T19:33:17.164Z [INFO] TestHTTPServer_HandshakeTimeout.server: Adding LAN server: server="Node-9e6d51a1-e08a-c1c4-6e4b-160e6114c12a (Addr: tcp/127.0.0.1:29795) (DC: dc1)" writer.go:29: 2021-01-29T19:33:17.165Z [INFO] TestHTTPServer_HandshakeTimeout.server: Handled event for server in area: event=member-join server=Node-9e6d51a1-e08a-c1c4-6e4b-160e6114c12a.dc1 area=wan writer.go:29: 2021-01-29T19:33:17.167Z [INFO] TestHTTPServer_HandshakeTimeout: Started DNS server: address=127.0.0.1:29790 network=tcp writer.go:29: 2021-01-29T19:33:17.168Z [INFO] TestHTTPServer_HandshakeTimeout: Started DNS server: address=127.0.0.1:29790 network=udp writer.go:29: 2021-01-29T19:33:17.169Z [DEBUG] TestHTTPServer_HandshakeTimeout.tlsutil: IncomingHTTPSConfig: version=1 writer.go:29: 2021-01-29T19:33:17.170Z [WARN] TestHTTPServer_HandshakeTimeout.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:17.172Z [INFO] TestHTTPServer_HandshakeTimeout.server.raft: entering candidate state: node="Node at 127.0.0.1:29795 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:17.172Z [INFO] TestHTTPServer_HandshakeTimeout: Started HTTPS server: address=127.0.0.1:29792 network=tcp writer.go:29: 2021-01-29T19:33:17.176Z [INFO] TestHTTPServer_HandshakeTimeout: started state syncer writer.go:29: 2021-01-29T19:33:17.192Z [DEBUG] TestHTTPServer_HandshakeTimeout.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:17.194Z [DEBUG] TestHTTPServer_HandshakeTimeout.server.raft: vote granted: from=9e6d51a1-e08a-c1c4-6e4b-160e6114c12a term=2 tally=1 writer.go:29: 2021-01-29T19:33:17.195Z [INFO] TestHTTPServer_HandshakeTimeout.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:17.196Z [INFO] TestHTTPServer_HandshakeTimeout.server.raft: entering leader state: leader="Node at 127.0.0.1:29795 [Leader]" writer.go:29: 2021-01-29T19:33:17.198Z [INFO] TestHTTPServer_HandshakeTimeout.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:17.198Z [INFO] TestHTTPServer_HandshakeTimeout.server: New leader elected: payload=Node-9e6d51a1-e08a-c1c4-6e4b-160e6114c12a writer.go:29: 2021-01-29T19:33:17.199Z [DEBUG] TestHTTPServer_HandshakeTimeout.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29795 writer.go:29: 2021-01-29T19:33:17.233Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:17.237Z [INFO] TestHTTPServer_HandshakeTimeout.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:17.238Z [INFO] TestHTTPServer_HandshakeTimeout.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.239Z [DEBUG] TestHTTPServer_HandshakeTimeout.server: Skipping self join check for node since the cluster is too small: node=Node-9e6d51a1-e08a-c1c4-6e4b-160e6114c12a writer.go:29: 2021-01-29T19:33:17.240Z [INFO] TestHTTPServer_HandshakeTimeout.server: member joined, marking health alive: member=Node-9e6d51a1-e08a-c1c4-6e4b-160e6114c12a 2021/01/29 19:33:17 http: TLS handshake error from 127.0.0.1:48950: read tcp 127.0.0.1:29792->127.0.0.1:48950: i/o timeout writer.go:29: 2021-01-29T19:33:17.345Z [INFO] TestHTTPServer_HandshakeTimeout: Requesting shutdown writer.go:29: 2021-01-29T19:33:17.347Z [INFO] TestHTTPServer_HandshakeTimeout.server: shutting down server writer.go:29: 2021-01-29T19:33:17.347Z [DEBUG] TestHTTPServer_HandshakeTimeout.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.348Z [WARN] TestHTTPServer_HandshakeTimeout.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:17.349Z [DEBUG] TestHTTPServer_HandshakeTimeout.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.349Z [ERROR] TestHTTPServer_HandshakeTimeout.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:17.350Z [WARN] TestHTTPServer_HandshakeTimeout.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:17.354Z [INFO] TestHTTPServer_HandshakeTimeout.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:17.354Z [INFO] TestHTTPServer_HandshakeTimeout: consul server down writer.go:29: 2021-01-29T19:33:17.356Z [INFO] TestHTTPServer_HandshakeTimeout: shutdown complete writer.go:29: 2021-01-29T19:33:17.357Z [INFO] TestHTTPServer_HandshakeTimeout: Stopping server: protocol=DNS address=127.0.0.1:29790 network=tcp writer.go:29: 2021-01-29T19:33:17.358Z [INFO] TestHTTPServer_HandshakeTimeout: Stopping server: protocol=DNS address=127.0.0.1:29790 network=udp writer.go:29: 2021-01-29T19:33:17.359Z [INFO] TestHTTPServer_HandshakeTimeout: Stopping server: protocol=HTTPS address=127.0.0.1:29792 network=tcp === CONT TestIntentionsList_values writer.go:29: 2021-01-29T19:33:17.363Z [INFO] TestIntentionsList_values: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:17.364Z [INFO] TestIntentionsList_values: Endpoints down --- PASS: TestIntentionsList_values (0.92s) === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.391Z [WARN] TestEnableWebUI: bootstrap = true: do not enable unless necessary === CONT TestIntentionsList_empty writer.go:29: 2021-01-29T19:33:17.393Z [INFO] TestIntentionsList_empty: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:17.394Z [INFO] TestIntentionsList_empty: Endpoints down --- PASS: TestIntentionsList_empty (0.94s) === CONT TestACLResolution === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.393Z [DEBUG] TestEnableWebUI.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:17.395Z [DEBUG] TestEnableWebUI.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:17.400Z [INFO] TestEnableWebUI.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ae6b9f77-d209-37e4-796b-b57a1e93e351 Address:127.0.0.1:29813}]" writer.go:29: 2021-01-29T19:33:17.403Z [INFO] TestEnableWebUI.server.raft: entering follower state: follower="Node at 127.0.0.1:29813 [Follower]" leader= writer.go:29: 2021-01-29T19:33:17.410Z [INFO] TestEnableWebUI.server.serf.wan: serf: EventMemberJoin: Node-ae6b9f77-d209-37e4-796b-b57a1e93e351.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:17.418Z [INFO] TestEnableWebUI.server.serf.lan: serf: EventMemberJoin: Node-ae6b9f77-d209-37e4-796b-b57a1e93e351 127.0.0.1 writer.go:29: 2021-01-29T19:33:17.429Z [INFO] TestEnableWebUI.server: Adding LAN server: server="Node-ae6b9f77-d209-37e4-796b-b57a1e93e351 (Addr: tcp/127.0.0.1:29813) (DC: dc1)" writer.go:29: 2021-01-29T19:33:17.430Z [INFO] TestEnableWebUI.server: Handled event for server in area: event=member-join server=Node-ae6b9f77-d209-37e4-796b-b57a1e93e351.dc1 area=wan === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.414Z [WARN] TestACLResolution: bootstrap = true: do not enable unless necessary === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.432Z [INFO] TestEnableWebUI: Started DNS server: address=127.0.0.1:29808 network=tcp writer.go:29: 2021-01-29T19:33:17.434Z [INFO] TestEnableWebUI: Started DNS server: address=127.0.0.1:29808 network=udp writer.go:29: 2021-01-29T19:33:17.436Z [INFO] TestEnableWebUI: Started HTTP server: address=127.0.0.1:29809 network=tcp writer.go:29: 2021-01-29T19:33:17.436Z [INFO] TestEnableWebUI: started state syncer === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.432Z [DEBUG] TestACLResolution.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:17.439Z [DEBUG] TestACLResolution.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.449Z [WARN] TestEnableWebUI.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:17.452Z [INFO] TestEnableWebUI.server.raft: entering candidate state: node="Node at 127.0.0.1:29813 [Candidate]" term=2 === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.457Z [INFO] TestACLResolution.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6cf86ded-4037-81d8-e243-4a448bc9bc87 Address:127.0.0.1:29807}]" === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.457Z [DEBUG] TestEnableWebUI.server.raft: votes: needed=1 === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.460Z [INFO] TestACLResolution.server.raft: entering follower state: follower="Node at 127.0.0.1:29807 [Follower]" leader= === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.459Z [DEBUG] TestEnableWebUI.server.raft: vote granted: from=ae6b9f77-d209-37e4-796b-b57a1e93e351 term=2 tally=1 writer.go:29: 2021-01-29T19:33:17.464Z [INFO] TestEnableWebUI.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:17.466Z [INFO] TestEnableWebUI.server.raft: entering leader state: leader="Node at 127.0.0.1:29813 [Leader]" writer.go:29: 2021-01-29T19:33:17.467Z [INFO] TestEnableWebUI.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:17.468Z [INFO] TestEnableWebUI.server: New leader elected: payload=Node-ae6b9f77-d209-37e4-796b-b57a1e93e351 === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.465Z [INFO] TestACLResolution.server.serf.wan: serf: EventMemberJoin: Node-6cf86ded-4037-81d8-e243-4a448bc9bc87.dc1 127.0.0.1 === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.472Z [DEBUG] TestEnableWebUI.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29813 writer.go:29: 2021-01-29T19:33:17.480Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.476Z [INFO] TestACLResolution.server.serf.lan: serf: EventMemberJoin: Node-6cf86ded-4037-81d8-e243-4a448bc9bc87 127.0.0.1 writer.go:29: 2021-01-29T19:33:17.484Z [INFO] TestACLResolution.server: Adding LAN server: server="Node-6cf86ded-4037-81d8-e243-4a448bc9bc87 (Addr: tcp/127.0.0.1:29807) (DC: dc1)" writer.go:29: 2021-01-29T19:33:17.486Z [INFO] TestACLResolution.server: Handled event for server in area: event=member-join server=Node-6cf86ded-4037-81d8-e243-4a448bc9bc87.dc1 area=wan writer.go:29: 2021-01-29T19:33:17.488Z [INFO] TestACLResolution: Started DNS server: address=127.0.0.1:29802 network=udp writer.go:29: 2021-01-29T19:33:17.491Z [INFO] TestACLResolution: Started DNS server: address=127.0.0.1:29802 network=tcp === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.492Z [INFO] TestEnableWebUI.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:17.494Z [INFO] TestEnableWebUI.leader: started routine: routine="CA root pruning" === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.497Z [INFO] TestACLResolution: Started HTTP server: address=127.0.0.1:29803 network=tcp writer.go:29: 2021-01-29T19:33:17.499Z [INFO] TestACLResolution: started state syncer === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.496Z [DEBUG] TestEnableWebUI.server: Skipping self join check for node since the cluster is too small: node=Node-ae6b9f77-d209-37e4-796b-b57a1e93e351 writer.go:29: 2021-01-29T19:33:17.502Z [INFO] TestEnableWebUI.server: member joined, marking health alive: member=Node-ae6b9f77-d209-37e4-796b-b57a1e93e351 === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.518Z [WARN] TestACLResolution.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:17.519Z [INFO] TestACLResolution.server.raft: entering candidate state: node="Node at 127.0.0.1:29807 [Candidate]" term=2 === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.522Z [INFO] TestEnableWebUI: Requesting shutdown writer.go:29: 2021-01-29T19:33:17.523Z [INFO] TestEnableWebUI.server: shutting down server === CONT TestRPC_HTTPSMaxConnsPerClient/HTTP writer.go:29: 2021-01-29T19:33:17.528Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:17.530Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTP: Endpoints down === RUN TestRPC_HTTPSMaxConnsPerClient/HTTPS === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.526Z [DEBUG] TestACLResolution.server.raft: votes: needed=1 === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.525Z [DEBUG] TestEnableWebUI.leader: stopping routine: routine="CA root pruning" === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.534Z [DEBUG] TestACLResolution.server.raft: vote granted: from=6cf86ded-4037-81d8-e243-4a448bc9bc87 term=2 tally=1 === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.535Z [WARN] TestEnableWebUI.server.serf.lan: serf: Shutdown without a Leave === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.539Z [INFO] TestACLResolution.server.raft: election won: tally=1 === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.527Z [ERROR] TestEnableWebUI.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.544Z [INFO] TestACLResolution.server.raft: entering leader state: leader="Node at 127.0.0.1:29807 [Leader]" === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.535Z [DEBUG] TestEnableWebUI.leader: stopped routine: routine="CA root pruning" === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.550Z [INFO] TestACLResolution.server: cluster leadership acquired === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.550Z [WARN] TestEnableWebUI.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:17.560Z [INFO] TestEnableWebUI.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:17.561Z [INFO] TestEnableWebUI: consul server down === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.551Z [INFO] TestACLResolution.server: New leader elected: payload=Node-6cf86ded-4037-81d8-e243-4a448bc9bc87 === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:17.562Z [INFO] TestEnableWebUI: shutdown complete writer.go:29: 2021-01-29T19:33:17.563Z [INFO] TestEnableWebUI: Stopping server: protocol=DNS address=127.0.0.1:29808 network=tcp writer.go:29: 2021-01-29T19:33:17.565Z [INFO] TestEnableWebUI: Stopping server: protocol=DNS address=127.0.0.1:29808 network=udp writer.go:29: 2021-01-29T19:33:17.566Z [INFO] TestEnableWebUI: Stopping server: protocol=HTTP address=127.0.0.1:29809 network=tcp === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.552Z [DEBUG] TestACLResolution.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29807 === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:17.585Z [WARN] TestRPC_HTTPSMaxConnsPerClient/HTTPS: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:17.588Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:17.591Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.594Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:17.599Z [INFO] TestACLResolution.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:17.601Z [INFO] TestACLResolution.leader: started routine: routine="CA root pruning" === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:17.602Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:54e71b53-053c-d446-78e4-43c1322256ce Address:127.0.0.1:29825}]" === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.604Z [DEBUG] TestACLResolution.server: Skipping self join check for node since the cluster is too small: node=Node-6cf86ded-4037-81d8-e243-4a448bc9bc87 === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:17.606Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.raft: entering follower state: follower="Node at 127.0.0.1:29825 [Follower]" leader= === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.608Z [INFO] TestACLResolution.server: member joined, marking health alive: member=Node-6cf86ded-4037-81d8-e243-4a448bc9bc87 === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:17.609Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.serf.wan: serf: EventMemberJoin: Node-54e71b53-053c-d446-78e4-43c1322256ce.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:17.616Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.serf.lan: serf: EventMemberJoin: Node-54e71b53-053c-d446-78e4-43c1322256ce 127.0.0.1 writer.go:29: 2021-01-29T19:33:17.619Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server: Adding LAN server: server="Node-54e71b53-053c-d446-78e4-43c1322256ce (Addr: tcp/127.0.0.1:29825) (DC: dc1)" writer.go:29: 2021-01-29T19:33:17.621Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server: Handled event for server in area: event=member-join server=Node-54e71b53-053c-d446-78e4-43c1322256ce.dc1 area=wan writer.go:29: 2021-01-29T19:33:17.622Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Started DNS server: address=127.0.0.1:29820 network=udp writer.go:29: 2021-01-29T19:33:17.624Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Started DNS server: address=127.0.0.1:29820 network=tcp writer.go:29: 2021-01-29T19:33:17.626Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS.tlsutil: IncomingHTTPSConfig: version=1 writer.go:29: 2021-01-29T19:33:17.630Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Started HTTPS server: address=127.0.0.1:29822 network=tcp writer.go:29: 2021-01-29T19:33:17.632Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: started state syncer writer.go:29: 2021-01-29T19:33:17.671Z [WARN] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:17.673Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.raft: entering candidate state: node="Node at 127.0.0.1:29825 [Candidate]" term=2 === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.676Z [INFO] TestACLResolution: Requesting shutdown writer.go:29: 2021-01-29T19:33:17.679Z [INFO] TestACLResolution.server: shutting down server writer.go:29: 2021-01-29T19:33:17.680Z [DEBUG] TestACLResolution.leader: stopping routine: routine="CA root pruning" === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:17.676Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.raft: votes: needed=1 === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.681Z [WARN] TestACLResolution.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:17.680Z [ERROR] TestACLResolution.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:17.681Z [DEBUG] TestACLResolution.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.684Z [WARN] TestACLResolution.server.serf.wan: serf: Shutdown without a Leave === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:17.681Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.raft: vote granted: from=54e71b53-053c-d446-78e4-43c1322256ce term=2 tally=1 writer.go:29: 2021-01-29T19:33:17.691Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:17.693Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.raft: entering leader state: leader="Node at 127.0.0.1:29825 [Leader]" writer.go:29: 2021-01-29T19:33:17.695Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:17.695Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server: New leader elected: payload=Node-54e71b53-053c-d446-78e4-43c1322256ce === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.699Z [INFO] TestACLResolution.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:17.700Z [INFO] TestACLResolution: consul server down writer.go:29: 2021-01-29T19:33:17.701Z [INFO] TestACLResolution: shutdown complete === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:17.700Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29825 === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:17.702Z [INFO] TestACLResolution: Stopping server: protocol=DNS address=127.0.0.1:29802 network=tcp writer.go:29: 2021-01-29T19:33:17.703Z [INFO] TestACLResolution: Stopping server: protocol=DNS address=127.0.0.1:29802 network=udp writer.go:29: 2021-01-29T19:33:17.704Z [INFO] TestACLResolution: Stopping server: protocol=HTTP address=127.0.0.1:29803 network=tcp === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:17.705Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:17.712Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:17.715Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.718Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server: Skipping self join check for node since the cluster is too small: node=Node-54e71b53-053c-d446-78e4-43c1322256ce writer.go:29: 2021-01-29T19:33:17.721Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server: member joined, marking health alive: member=Node-54e71b53-053c-d446-78e4-43c1322256ce === CONT TestHTTPServer_HandshakeTimeout writer.go:29: 2021-01-29T19:33:17.860Z [INFO] TestHTTPServer_HandshakeTimeout: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:17.862Z [INFO] TestHTTPServer_HandshakeTimeout: Endpoints down --- PASS: TestHTTPServer_HandshakeTimeout (0.80s) === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:17.943Z [WARN] TestParseConsistency_Invalid: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:17.947Z [DEBUG] TestParseConsistency_Invalid.tlsutil: Update: version=1 === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:17.952Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:17.958Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Synced node info === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:17.969Z [DEBUG] TestParseConsistency_Invalid.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:17.998Z [INFO] TestParseConsistency_Invalid.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d32ccaa1-7292-4699-8be4-017a6bec5a07 Address:127.0.0.1:29819}]" writer.go:29: 2021-01-29T19:33:18.003Z [INFO] TestParseConsistency_Invalid.server.serf.wan: serf: EventMemberJoin: Node-d32ccaa1-7292-4699-8be4-017a6bec5a07.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.008Z [INFO] TestParseConsistency_Invalid.server.serf.lan: serf: EventMemberJoin: Node-d32ccaa1-7292-4699-8be4-017a6bec5a07 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.011Z [INFO] TestParseConsistency_Invalid.server.raft: entering follower state: follower="Node at 127.0.0.1:29819 [Follower]" leader= 2021/01/29 19:33:18 http: TLS handshake error from 127.0.0.1:51852: read tcp 127.0.0.1:29822->127.0.0.1:51852: use of closed network connection writer.go:29: 2021-01-29T19:33:18.016Z [INFO] TestParseConsistency_Invalid.server: Adding LAN server: server="Node-d32ccaa1-7292-4699-8be4-017a6bec5a07 (Addr: tcp/127.0.0.1:29819) (DC: dc1)" writer.go:29: 2021-01-29T19:33:18.017Z [INFO] TestParseConsistency_Invalid.server: Handled event for server in area: event=member-join server=Node-d32ccaa1-7292-4699-8be4-017a6bec5a07.dc1 area=wan writer.go:29: 2021-01-29T19:33:18.021Z [INFO] TestParseConsistency_Invalid: Started DNS server: address=127.0.0.1:29814 network=udp writer.go:29: 2021-01-29T19:33:18.024Z [INFO] TestParseConsistency_Invalid: Started DNS server: address=127.0.0.1:29814 network=tcp === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:18.024Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS.tlsutil: Update: version=2 === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:18.030Z [INFO] TestParseConsistency_Invalid: Started HTTP server: address=127.0.0.1:29815 network=tcp writer.go:29: 2021-01-29T19:33:18.032Z [INFO] TestParseConsistency_Invalid: started state syncer 2021/01/29 19:33:18 http: TLS handshake error from 127.0.0.1:51848: EOF 2021/01/29 19:33:18 http: TLS handshake error from 127.0.0.1:51850: EOF 2021/01/29 19:33:18 http: TLS handshake error from 127.0.0.1:51854: EOF === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:18.036Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Requesting shutdown writer.go:29: 2021-01-29T19:33:18.038Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server: shutting down server writer.go:29: 2021-01-29T19:33:18.039Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.041Z [WARN] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:18.045Z [DEBUG] TestRPC_HTTPSMaxConnsPerClient/HTTPS.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.053Z [WARN] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:18.064Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:18.066Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: consul server down writer.go:29: 2021-01-29T19:33:18.067Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: shutdown complete writer.go:29: 2021-01-29T19:33:18.069Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Stopping server: protocol=DNS address=127.0.0.1:29820 network=tcp writer.go:29: 2021-01-29T19:33:18.071Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Stopping server: protocol=DNS address=127.0.0.1:29820 network=udp writer.go:29: 2021-01-29T19:33:18.073Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Stopping server: protocol=HTTPS address=127.0.0.1:29822 network=tcp === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:18.069Z [INFO] TestEnableWebUI: Waiting for endpoints to shut down === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:18.074Z [WARN] TestParseConsistency_Invalid.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestEnableWebUI writer.go:29: 2021-01-29T19:33:18.075Z [INFO] TestEnableWebUI: Endpoints down --- PASS: TestEnableWebUI (0.75s) === CONT TestParseConsistency === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:18.111Z [INFO] TestParseConsistency_Invalid.server.raft: entering candidate state: node="Node at 127.0.0.1:29819 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:18.129Z [DEBUG] TestParseConsistency_Invalid.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:18.130Z [DEBUG] TestParseConsistency_Invalid.server.raft: vote granted: from=d32ccaa1-7292-4699-8be4-017a6bec5a07 term=2 tally=1 writer.go:29: 2021-01-29T19:33:18.131Z [INFO] TestParseConsistency_Invalid.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:18.132Z [INFO] TestParseConsistency_Invalid.server.raft: entering leader state: leader="Node at 127.0.0.1:29819 [Leader]" writer.go:29: 2021-01-29T19:33:18.133Z [INFO] TestParseConsistency_Invalid.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:18.135Z [INFO] TestParseConsistency_Invalid.server: New leader elected: payload=Node-d32ccaa1-7292-4699-8be4-017a6bec5a07 writer.go:29: 2021-01-29T19:33:18.140Z [DEBUG] TestParseConsistency_Invalid.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29819 === CONT TestACLResolution writer.go:29: 2021-01-29T19:33:18.217Z [INFO] TestACLResolution: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:18.223Z [INFO] TestACLResolution: Endpoints down --- PASS: TestACLResolution (0.83s) === CONT TestParseWait_InvalidIndex --- PASS: TestParseWait_InvalidIndex (0.00s) === CONT TestParseWait_InvalidTime --- PASS: TestParseWait_InvalidTime (0.00s) === CONT TestPProfHandlers_ACLs === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:18.225Z [INFO] TestParseConsistency_Invalid: Synced node info writer.go:29: 2021-01-29T19:33:18.243Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.249Z [WARN] TestParseConsistency: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:18.250Z [DEBUG] TestParseConsistency.tlsutil: Update: version=1 === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:18.252Z [INFO] TestParseConsistency_Invalid.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:18.253Z [INFO] TestParseConsistency_Invalid.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.254Z [DEBUG] TestParseConsistency_Invalid.server: Skipping self join check for node since the cluster is too small: node=Node-d32ccaa1-7292-4699-8be4-017a6bec5a07 === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.254Z [DEBUG] TestParseConsistency.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:18.255Z [INFO] TestParseConsistency_Invalid.server: member joined, marking health alive: member=Node-d32ccaa1-7292-4699-8be4-017a6bec5a07 === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.262Z [WARN] TestPProfHandlers_ACLs: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:18.267Z [WARN] TestPProfHandlers_ACLs: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:18.271Z [DEBUG] TestPProfHandlers_ACLs.tlsutil: Update: version=1 === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.274Z [INFO] TestParseConsistency.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1f068ccc-39f1-c640-814c-945e50c4886f Address:127.0.0.1:29837}]" === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.274Z [DEBUG] TestPProfHandlers_ACLs.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.275Z [INFO] TestParseConsistency.server.raft: entering follower state: follower="Node at 127.0.0.1:29837 [Follower]" leader= writer.go:29: 2021-01-29T19:33:18.276Z [INFO] TestParseConsistency.server.serf.wan: serf: EventMemberJoin: Node-1f068ccc-39f1-c640-814c-945e50c4886f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.279Z [INFO] TestParseConsistency.server.serf.lan: serf: EventMemberJoin: Node-1f068ccc-39f1-c640-814c-945e50c4886f 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.282Z [INFO] TestParseConsistency.server: Adding LAN server: server="Node-1f068ccc-39f1-c640-814c-945e50c4886f (Addr: tcp/127.0.0.1:29837) (DC: dc1)" writer.go:29: 2021-01-29T19:33:18.282Z [INFO] TestParseConsistency.server: Handled event for server in area: event=member-join server=Node-1f068ccc-39f1-c640-814c-945e50c4886f.dc1 area=wan === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.284Z [INFO] TestPProfHandlers_ACLs.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e2b4a64c-274c-047f-e3f4-8b310733dd72 Address:127.0.0.1:29831}]" === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.284Z [INFO] TestParseConsistency: Started DNS server: address=127.0.0.1:29832 network=tcp writer.go:29: 2021-01-29T19:33:18.286Z [INFO] TestParseConsistency: Started DNS server: address=127.0.0.1:29832 network=udp === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.286Z [INFO] TestPProfHandlers_ACLs.server.raft: entering follower state: follower="Node at 127.0.0.1:29831 [Follower]" leader= === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.289Z [INFO] TestParseConsistency: Started HTTP server: address=127.0.0.1:29833 network=tcp writer.go:29: 2021-01-29T19:33:18.290Z [INFO] TestParseConsistency: started state syncer === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.287Z [INFO] TestPProfHandlers_ACLs.server.serf.wan: serf: EventMemberJoin: Node-e2b4a64c-274c-047f-e3f4-8b310733dd72.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.299Z [INFO] TestPProfHandlers_ACLs.server.serf.lan: serf: EventMemberJoin: Node-e2b4a64c-274c-047f-e3f4-8b310733dd72 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.302Z [INFO] TestPProfHandlers_ACLs.server: Adding LAN server: server="Node-e2b4a64c-274c-047f-e3f4-8b310733dd72 (Addr: tcp/127.0.0.1:29831) (DC: dc1)" writer.go:29: 2021-01-29T19:33:18.303Z [INFO] TestPProfHandlers_ACLs.server: Handled event for server in area: event=member-join server=Node-e2b4a64c-274c-047f-e3f4-8b310733dd72.dc1 area=wan writer.go:29: 2021-01-29T19:33:18.305Z [INFO] TestPProfHandlers_ACLs: Started DNS server: address=127.0.0.1:29826 network=tcp writer.go:29: 2021-01-29T19:33:18.307Z [INFO] TestPProfHandlers_ACLs: Started DNS server: address=127.0.0.1:29826 network=udp writer.go:29: 2021-01-29T19:33:18.310Z [INFO] TestPProfHandlers_ACLs: Started HTTP server: address=127.0.0.1:29827 network=tcp writer.go:29: 2021-01-29T19:33:18.312Z [INFO] TestPProfHandlers_ACLs: started state syncer === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.342Z [WARN] TestParseConsistency.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:18.344Z [INFO] TestParseConsistency.server.raft: entering candidate state: node="Node at 127.0.0.1:29837 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:18.347Z [DEBUG] TestParseConsistency.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:18.347Z [DEBUG] TestParseConsistency.server.raft: vote granted: from=1f068ccc-39f1-c640-814c-945e50c4886f term=2 tally=1 === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.348Z [WARN] TestPProfHandlers_ACLs.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.348Z [INFO] TestParseConsistency.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:18.349Z [INFO] TestParseConsistency.server.raft: entering leader state: leader="Node at 127.0.0.1:29837 [Leader]" === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.349Z [INFO] TestPProfHandlers_ACLs.server.raft: entering candidate state: node="Node at 127.0.0.1:29831 [Candidate]" term=2 === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.351Z [INFO] TestParseConsistency.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:18.351Z [INFO] TestParseConsistency.server: New leader elected: payload=Node-1f068ccc-39f1-c640-814c-945e50c4886f === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.352Z [DEBUG] TestPProfHandlers_ACLs.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:18.353Z [DEBUG] TestPProfHandlers_ACLs.server.raft: vote granted: from=e2b4a64c-274c-047f-e3f4-8b310733dd72 term=2 tally=1 === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.352Z [DEBUG] TestParseConsistency.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29837 === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.355Z [INFO] TestPProfHandlers_ACLs.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:18.354Z [DEBUG] TestPProfHandlers_ACLs.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.358Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.357Z [INFO] TestPProfHandlers_ACLs.server.raft: entering leader state: leader="Node at 127.0.0.1:29831 [Leader]" writer.go:29: 2021-01-29T19:33:18.362Z [INFO] TestPProfHandlers_ACLs.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:18.363Z [INFO] TestPProfHandlers_ACLs.server: New leader elected: payload=Node-e2b4a64c-274c-047f-e3f4-8b310733dd72 writer.go:29: 2021-01-29T19:33:18.364Z [INFO] TestPProfHandlers_ACLs.server: initializing acls writer.go:29: 2021-01-29T19:33:18.366Z [INFO] TestPProfHandlers_ACLs.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:18.367Z [WARN] TestPProfHandlers_ACLs.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:33:18.369Z [INFO] TestPProfHandlers_ACLs.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:33:18.370Z [INFO] TestPProfHandlers_ACLs.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:18.371Z [INFO] TestPProfHandlers_ACLs.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:18.372Z [INFO] TestPProfHandlers_ACLs.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:18.373Z [INFO] TestPProfHandlers_ACLs.server.serf.lan: serf: EventMemberUpdate: Node-e2b4a64c-274c-047f-e3f4-8b310733dd72 writer.go:29: 2021-01-29T19:33:18.374Z [INFO] TestPProfHandlers_ACLs.server.serf.wan: serf: EventMemberUpdate: Node-e2b4a64c-274c-047f-e3f4-8b310733dd72.dc1 === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.375Z [INFO] TestParseConsistency.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.374Z [INFO] TestPProfHandlers_ACLs.server: Updating LAN server: server="Node-e2b4a64c-274c-047f-e3f4-8b310733dd72 (Addr: tcp/127.0.0.1:29831) (DC: dc1)" === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.377Z [INFO] TestParseConsistency.leader: started routine: routine="CA root pruning" === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.376Z [INFO] TestPProfHandlers_ACLs.server: Handled event for server in area: event=member-update server=Node-e2b4a64c-274c-047f-e3f4-8b310733dd72.dc1 area=wan === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.379Z [DEBUG] TestParseConsistency.server: Skipping self join check for node since the cluster is too small: node=Node-1f068ccc-39f1-c640-814c-945e50c4886f writer.go:29: 2021-01-29T19:33:18.380Z [INFO] TestParseConsistency.server: member joined, marking health alive: member=Node-1f068ccc-39f1-c640-814c-945e50c4886f === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:18.389Z [DEBUG] TestParseConsistency_Invalid: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:18.391Z [DEBUG] TestParseConsistency_Invalid: Node info in sync writer.go:29: 2021-01-29T19:33:18.393Z [DEBUG] TestParseConsistency_Invalid: Node info in sync === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.397Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:18.400Z [INFO] TestPProfHandlers_ACLs.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:18.401Z [INFO] TestPProfHandlers_ACLs.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.402Z [DEBUG] TestPProfHandlers_ACLs.server: Skipping self join check for node since the cluster is too small: node=Node-e2b4a64c-274c-047f-e3f4-8b310733dd72 writer.go:29: 2021-01-29T19:33:18.400Z [ERROR] TestPProfHandlers_ACLs.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:18.403Z [INFO] TestParseConsistency_Invalid: Requesting shutdown === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.403Z [INFO] TestPProfHandlers_ACLs.server: member joined, marking health alive: member=Node-e2b4a64c-274c-047f-e3f4-8b310733dd72 === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:18.404Z [INFO] TestParseConsistency_Invalid.server: shutting down server writer.go:29: 2021-01-29T19:33:18.406Z [DEBUG] TestParseConsistency_Invalid.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.407Z [WARN] TestParseConsistency_Invalid.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:18.407Z [DEBUG] TestParseConsistency_Invalid.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.410Z [WARN] TestParseConsistency_Invalid.server.serf.wan: serf: Shutdown without a Leave === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.412Z [DEBUG] TestPProfHandlers_ACLs.server: Skipping self join check for node since the cluster is too small: node=Node-e2b4a64c-274c-047f-e3f4-8b310733dd72 === CONT TestParseConsistency_Invalid writer.go:29: 2021-01-29T19:33:18.413Z [INFO] TestParseConsistency_Invalid.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:18.414Z [INFO] TestParseConsistency_Invalid: consul server down writer.go:29: 2021-01-29T19:33:18.417Z [INFO] TestParseConsistency_Invalid: shutdown complete writer.go:29: 2021-01-29T19:33:18.418Z [INFO] TestParseConsistency_Invalid: Stopping server: protocol=DNS address=127.0.0.1:29814 network=tcp writer.go:29: 2021-01-29T19:33:18.437Z [INFO] TestParseConsistency_Invalid: Stopping server: protocol=DNS address=127.0.0.1:29814 network=udp writer.go:29: 2021-01-29T19:33:18.440Z [INFO] TestParseConsistency_Invalid: Stopping server: protocol=HTTP address=127.0.0.1:29815 network=tcp writer.go:29: 2021-01-29T19:33:18.441Z [INFO] TestParseConsistency_Invalid: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:18.445Z [INFO] TestParseConsistency_Invalid: Endpoints down --- PASS: TestParseConsistency_Invalid (0.58s) === CONT TestPProfHandlers_DisableDebugNoACLs writer.go:29: 2021-01-29T19:33:18.480Z [WARN] TestPProfHandlers_DisableDebugNoACLs: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:18.481Z [DEBUG] TestPProfHandlers_DisableDebugNoACLs.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:18.482Z [DEBUG] TestPProfHandlers_DisableDebugNoACLs.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:18.486Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:dd309b35-76e9-616a-fcfd-5475dc2b5c55 Address:127.0.0.1:29843}]" writer.go:29: 2021-01-29T19:33:18.488Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server.raft: entering follower state: follower="Node at 127.0.0.1:29843 [Follower]" leader= writer.go:29: 2021-01-29T19:33:18.493Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server.serf.wan: serf: EventMemberJoin: Node-dd309b35-76e9-616a-fcfd-5475dc2b5c55.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.500Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server.serf.lan: serf: EventMemberJoin: Node-dd309b35-76e9-616a-fcfd-5475dc2b5c55 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.504Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server: Adding LAN server: server="Node-dd309b35-76e9-616a-fcfd-5475dc2b5c55 (Addr: tcp/127.0.0.1:29843) (DC: dc1)" writer.go:29: 2021-01-29T19:33:18.504Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server: Handled event for server in area: event=member-join server=Node-dd309b35-76e9-616a-fcfd-5475dc2b5c55.dc1 area=wan writer.go:29: 2021-01-29T19:33:18.507Z [INFO] TestPProfHandlers_DisableDebugNoACLs: Started DNS server: address=127.0.0.1:29838 network=tcp writer.go:29: 2021-01-29T19:33:18.510Z [INFO] TestPProfHandlers_DisableDebugNoACLs: Started DNS server: address=127.0.0.1:29838 network=udp writer.go:29: 2021-01-29T19:33:18.515Z [INFO] TestPProfHandlers_DisableDebugNoACLs: Started HTTP server: address=127.0.0.1:29839 network=tcp writer.go:29: 2021-01-29T19:33:18.517Z [INFO] TestPProfHandlers_DisableDebugNoACLs: started state syncer === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:18.526Z [INFO] TestParseConsistency: Requesting shutdown writer.go:29: 2021-01-29T19:33:18.527Z [INFO] TestParseConsistency.server: shutting down server writer.go:29: 2021-01-29T19:33:18.528Z [DEBUG] TestParseConsistency.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.529Z [WARN] TestParseConsistency.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:18.529Z [DEBUG] TestParseConsistency.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.528Z [ERROR] TestParseConsistency.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:18.537Z [WARN] TestParseConsistency.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:18.539Z [INFO] TestParseConsistency.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:18.540Z [INFO] TestParseConsistency: consul server down writer.go:29: 2021-01-29T19:33:18.543Z [INFO] TestParseConsistency: shutdown complete writer.go:29: 2021-01-29T19:33:18.544Z [INFO] TestParseConsistency: Stopping server: protocol=DNS address=127.0.0.1:29832 network=tcp writer.go:29: 2021-01-29T19:33:18.546Z [INFO] TestParseConsistency: Stopping server: protocol=DNS address=127.0.0.1:29832 network=udp writer.go:29: 2021-01-29T19:33:18.547Z [INFO] TestParseConsistency: Stopping server: protocol=HTTP address=127.0.0.1:29833 network=tcp === CONT TestPProfHandlers_DisableDebugNoACLs writer.go:29: 2021-01-29T19:33:18.556Z [WARN] TestPProfHandlers_DisableDebugNoACLs.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:18.558Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server.raft: entering candidate state: node="Node at 127.0.0.1:29843 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:18.562Z [DEBUG] TestPProfHandlers_DisableDebugNoACLs.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:18.563Z [DEBUG] TestPProfHandlers_DisableDebugNoACLs.server.raft: vote granted: from=dd309b35-76e9-616a-fcfd-5475dc2b5c55 term=2 tally=1 writer.go:29: 2021-01-29T19:33:18.565Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:18.567Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server.raft: entering leader state: leader="Node at 127.0.0.1:29843 [Leader]" writer.go:29: 2021-01-29T19:33:18.569Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:18.570Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server: New leader elected: payload=Node-dd309b35-76e9-616a-fcfd-5475dc2b5c55 === CONT TestRPC_HTTPSMaxConnsPerClient/HTTPS writer.go:29: 2021-01-29T19:33:18.574Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:18.576Z [INFO] TestRPC_HTTPSMaxConnsPerClient/HTTPS: Endpoints down --- PASS: TestRPC_HTTPSMaxConnsPerClient (1.93s) --- PASS: TestRPC_HTTPSMaxConnsPerClient/HTTP (0.88s) --- PASS: TestRPC_HTTPSMaxConnsPerClient/HTTPS (1.05s) === CONT TestPProfHandlers_EnableDebug === CONT TestPProfHandlers_DisableDebugNoACLs writer.go:29: 2021-01-29T19:33:18.574Z [DEBUG] TestPProfHandlers_DisableDebugNoACLs.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29843 writer.go:29: 2021-01-29T19:33:18.594Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:18.605Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestPProfHandlers_EnableDebug writer.go:29: 2021-01-29T19:33:18.606Z [WARN] TestPProfHandlers_EnableDebug: bootstrap = true: do not enable unless necessary === CONT TestPProfHandlers_DisableDebugNoACLs writer.go:29: 2021-01-29T19:33:18.607Z [INFO] TestPProfHandlers_DisableDebugNoACLs.leader: started routine: routine="CA root pruning" === CONT TestPProfHandlers_EnableDebug writer.go:29: 2021-01-29T19:33:18.609Z [DEBUG] TestPProfHandlers_EnableDebug.tlsutil: Update: version=1 === CONT TestPProfHandlers_DisableDebugNoACLs writer.go:29: 2021-01-29T19:33:18.610Z [DEBUG] TestPProfHandlers_DisableDebugNoACLs.server: Skipping self join check for node since the cluster is too small: node=Node-dd309b35-76e9-616a-fcfd-5475dc2b5c55 === CONT TestPProfHandlers_EnableDebug writer.go:29: 2021-01-29T19:33:18.611Z [DEBUG] TestPProfHandlers_EnableDebug.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPProfHandlers_DisableDebugNoACLs writer.go:29: 2021-01-29T19:33:18.611Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server: member joined, marking health alive: member=Node-dd309b35-76e9-616a-fcfd-5475dc2b5c55 === CONT TestPProfHandlers_EnableDebug writer.go:29: 2021-01-29T19:33:18.615Z [INFO] TestPProfHandlers_EnableDebug.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:decccb1d-6da3-0399-3f41-4fc48fa269ae Address:127.0.0.1:29849}]" writer.go:29: 2021-01-29T19:33:18.617Z [INFO] TestPProfHandlers_EnableDebug.server.raft: entering follower state: follower="Node at 127.0.0.1:29849 [Follower]" leader= writer.go:29: 2021-01-29T19:33:18.618Z [INFO] TestPProfHandlers_EnableDebug.server.serf.wan: serf: EventMemberJoin: Node-decccb1d-6da3-0399-3f41-4fc48fa269ae.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.622Z [INFO] TestPProfHandlers_EnableDebug.server.serf.lan: serf: EventMemberJoin: Node-decccb1d-6da3-0399-3f41-4fc48fa269ae 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.624Z [INFO] TestPProfHandlers_EnableDebug.server: Handled event for server in area: event=member-join server=Node-decccb1d-6da3-0399-3f41-4fc48fa269ae.dc1 area=wan writer.go:29: 2021-01-29T19:33:18.624Z [INFO] TestPProfHandlers_EnableDebug.server: Adding LAN server: server="Node-decccb1d-6da3-0399-3f41-4fc48fa269ae (Addr: tcp/127.0.0.1:29849) (DC: dc1)" writer.go:29: 2021-01-29T19:33:18.625Z [INFO] TestPProfHandlers_EnableDebug: Started DNS server: address=127.0.0.1:29844 network=udp writer.go:29: 2021-01-29T19:33:18.627Z [INFO] TestPProfHandlers_EnableDebug: Started DNS server: address=127.0.0.1:29844 network=tcp writer.go:29: 2021-01-29T19:33:18.629Z [INFO] TestPProfHandlers_EnableDebug: Started HTTP server: address=127.0.0.1:29845 network=tcp writer.go:29: 2021-01-29T19:33:18.630Z [INFO] TestPProfHandlers_EnableDebug: started state syncer writer.go:29: 2021-01-29T19:33:18.677Z [WARN] TestPProfHandlers_EnableDebug.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:18.684Z [INFO] TestPProfHandlers_EnableDebug.server.raft: entering candidate state: node="Node at 127.0.0.1:29849 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:18.687Z [DEBUG] TestPProfHandlers_EnableDebug.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:18.688Z [DEBUG] TestPProfHandlers_EnableDebug.server.raft: vote granted: from=decccb1d-6da3-0399-3f41-4fc48fa269ae term=2 tally=1 writer.go:29: 2021-01-29T19:33:18.690Z [INFO] TestPProfHandlers_EnableDebug.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:18.691Z [INFO] TestPProfHandlers_EnableDebug.server.raft: entering leader state: leader="Node at 127.0.0.1:29849 [Leader]" writer.go:29: 2021-01-29T19:33:18.692Z [INFO] TestPProfHandlers_EnableDebug.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:18.693Z [INFO] TestPProfHandlers_EnableDebug.server: New leader elected: payload=Node-decccb1d-6da3-0399-3f41-4fc48fa269ae writer.go:29: 2021-01-29T19:33:18.695Z [DEBUG] TestPProfHandlers_EnableDebug.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29849 writer.go:29: 2021-01-29T19:33:18.702Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:18.707Z [INFO] TestPProfHandlers_EnableDebug.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:18.708Z [INFO] TestPProfHandlers_EnableDebug.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.709Z [DEBUG] TestPProfHandlers_EnableDebug.server: Skipping self join check for node since the cluster is too small: node=Node-decccb1d-6da3-0399-3f41-4fc48fa269ae writer.go:29: 2021-01-29T19:33:18.710Z [INFO] TestPProfHandlers_EnableDebug.server: member joined, marking health alive: member=Node-decccb1d-6da3-0399-3f41-4fc48fa269ae === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:18.741Z [DEBUG] TestPProfHandlers_ACLs.acl: dropping node from result due to ACLs: node=Node-e2b4a64c-274c-047f-e3f4-8b310733dd72 writer.go:29: 2021-01-29T19:33:18.745Z [DEBUG] TestPProfHandlers_ACLs.acl: dropping node from result due to ACLs: node=Node-e2b4a64c-274c-047f-e3f4-8b310733dd72 === RUN TestPProfHandlers_ACLs/case_0_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:200,_token:"master",_endpoint:"/debug/pprof/heap",_nilResponse:false}) === CONT TestPProfHandlers_DisableDebugNoACLs writer.go:29: 2021-01-29T19:33:18.770Z [INFO] TestPProfHandlers_DisableDebugNoACLs: Requesting shutdown writer.go:29: 2021-01-29T19:33:18.772Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server: shutting down server writer.go:29: 2021-01-29T19:33:18.773Z [DEBUG] TestPProfHandlers_DisableDebugNoACLs.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.773Z [ERROR] TestPProfHandlers_DisableDebugNoACLs.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:18.774Z [DEBUG] TestPProfHandlers_DisableDebugNoACLs.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.774Z [WARN] TestPProfHandlers_DisableDebugNoACLs.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:18.777Z [WARN] TestPProfHandlers_DisableDebugNoACLs.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:18.779Z [INFO] TestPProfHandlers_DisableDebugNoACLs.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:18.779Z [INFO] TestPProfHandlers_DisableDebugNoACLs: consul server down writer.go:29: 2021-01-29T19:33:18.781Z [INFO] TestPProfHandlers_DisableDebugNoACLs: shutdown complete writer.go:29: 2021-01-29T19:33:18.783Z [INFO] TestPProfHandlers_DisableDebugNoACLs: Stopping server: protocol=DNS address=127.0.0.1:29838 network=tcp writer.go:29: 2021-01-29T19:33:18.784Z [INFO] TestPProfHandlers_DisableDebugNoACLs: Stopping server: protocol=DNS address=127.0.0.1:29838 network=udp writer.go:29: 2021-01-29T19:33:18.785Z [INFO] TestPProfHandlers_DisableDebugNoACLs: Stopping server: protocol=HTTP address=127.0.0.1:29839 network=tcp === CONT TestPProfHandlers_EnableDebug writer.go:29: 2021-01-29T19:33:19.023Z [DEBUG] TestPProfHandlers_EnableDebug: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:19.025Z [INFO] TestPProfHandlers_EnableDebug: Synced node info writer.go:29: 2021-01-29T19:33:19.026Z [DEBUG] TestPProfHandlers_EnableDebug: Node info in sync === CONT TestParseConsistency writer.go:29: 2021-01-29T19:33:19.050Z [INFO] TestParseConsistency: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:19.051Z [INFO] TestParseConsistency: Endpoints down --- PASS: TestParseConsistency (0.94s) === CONT TestParseWait --- PASS: TestParseWait (0.00s) === CONT TestParseSource writer.go:29: 2021-01-29T19:33:19.061Z [WARN] TestParseSource: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:19.072Z [DEBUG] TestParseSource.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:19.074Z [DEBUG] TestParseSource.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:19.077Z [INFO] TestParseSource.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2d62a138-9337-22a5-3372-600e0985690e Address:127.0.0.1:29855}]" writer.go:29: 2021-01-29T19:33:19.079Z [INFO] TestParseSource.server.raft: entering follower state: follower="Node at 127.0.0.1:29855 [Follower]" leader= writer.go:29: 2021-01-29T19:33:19.079Z [INFO] TestParseSource.server.serf.wan: serf: EventMemberJoin: Node-2d62a138-9337-22a5-3372-600e0985690e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:19.085Z [INFO] TestParseSource.server.serf.lan: serf: EventMemberJoin: Node-2d62a138-9337-22a5-3372-600e0985690e 127.0.0.1 writer.go:29: 2021-01-29T19:33:19.088Z [INFO] TestParseSource.server: Adding LAN server: server="Node-2d62a138-9337-22a5-3372-600e0985690e (Addr: tcp/127.0.0.1:29855) (DC: dc1)" writer.go:29: 2021-01-29T19:33:19.089Z [INFO] TestParseSource.server: Handled event for server in area: event=member-join server=Node-2d62a138-9337-22a5-3372-600e0985690e.dc1 area=wan writer.go:29: 2021-01-29T19:33:19.092Z [INFO] TestParseSource: Started DNS server: address=127.0.0.1:29850 network=tcp writer.go:29: 2021-01-29T19:33:19.094Z [INFO] TestParseSource: Started DNS server: address=127.0.0.1:29850 network=udp writer.go:29: 2021-01-29T19:33:19.095Z [INFO] TestParseSource: Started HTTP server: address=127.0.0.1:29851 network=tcp writer.go:29: 2021-01-29T19:33:19.096Z [INFO] TestParseSource: started state syncer === RUN TestPProfHandlers_ACLs/case_1_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:403,_token:"agent",_endpoint:"/debug/pprof/heap",_nilResponse:true}) === RUN TestPProfHandlers_ACLs/case_2_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:403,_token:"agent",_endpoint:"/debug/pprof/",_nilResponse:true}) === RUN TestPProfHandlers_ACLs/case_3_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:403,_token:"",_endpoint:"/debug/pprof/",_nilResponse:true}) === RUN TestPProfHandlers_ACLs/case_4_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:200,_token:"master",_endpoint:"/debug/pprof/heap",_nilResponse:false}) === CONT TestParseSource writer.go:29: 2021-01-29T19:33:19.142Z [WARN] TestParseSource.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:19.144Z [INFO] TestParseSource.server.raft: entering candidate state: node="Node at 127.0.0.1:29855 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:19.146Z [DEBUG] TestParseSource.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:19.148Z [DEBUG] TestParseSource.server.raft: vote granted: from=2d62a138-9337-22a5-3372-600e0985690e term=2 tally=1 writer.go:29: 2021-01-29T19:33:19.149Z [INFO] TestParseSource.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:19.151Z [INFO] TestParseSource.server.raft: entering leader state: leader="Node at 127.0.0.1:29855 [Leader]" writer.go:29: 2021-01-29T19:33:19.152Z [INFO] TestParseSource.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:19.153Z [INFO] TestParseSource.server: New leader elected: payload=Node-2d62a138-9337-22a5-3372-600e0985690e writer.go:29: 2021-01-29T19:33:19.154Z [DEBUG] TestParseSource.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29855 writer.go:29: 2021-01-29T19:33:19.158Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:19.164Z [INFO] TestParseSource.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:19.166Z [INFO] TestParseSource.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.168Z [DEBUG] TestParseSource.server: Skipping self join check for node since the cluster is too small: node=Node-2d62a138-9337-22a5-3372-600e0985690e writer.go:29: 2021-01-29T19:33:19.170Z [INFO] TestParseSource.server: member joined, marking health alive: member=Node-2d62a138-9337-22a5-3372-600e0985690e === CONT TestPProfHandlers_DisableDebugNoACLs writer.go:29: 2021-01-29T19:33:19.287Z [INFO] TestPProfHandlers_DisableDebugNoACLs: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:19.288Z [INFO] TestPProfHandlers_DisableDebugNoACLs: Endpoints down --- PASS: TestPProfHandlers_DisableDebugNoACLs (0.84s) === CONT TestPrettyPrintBare writer.go:29: 2021-01-29T19:33:19.307Z [WARN] TestPrettyPrintBare: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:19.309Z [DEBUG] TestPrettyPrintBare.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:19.313Z [DEBUG] TestPrettyPrintBare.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:19.330Z [INFO] TestPrettyPrintBare.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2ebe9a73-8044-60c1-5755-1b5571820db4 Address:127.0.0.1:29861}]" writer.go:29: 2021-01-29T19:33:19.338Z [INFO] TestPrettyPrintBare.server.raft: entering follower state: follower="Node at 127.0.0.1:29861 [Follower]" leader= writer.go:29: 2021-01-29T19:33:19.340Z [INFO] TestPrettyPrintBare.server.serf.wan: serf: EventMemberJoin: Node-2ebe9a73-8044-60c1-5755-1b5571820db4.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:19.350Z [INFO] TestPrettyPrintBare.server.serf.lan: serf: EventMemberJoin: Node-2ebe9a73-8044-60c1-5755-1b5571820db4 127.0.0.1 writer.go:29: 2021-01-29T19:33:19.357Z [INFO] TestPrettyPrintBare.server: Adding LAN server: server="Node-2ebe9a73-8044-60c1-5755-1b5571820db4 (Addr: tcp/127.0.0.1:29861) (DC: dc1)" writer.go:29: 2021-01-29T19:33:19.357Z [INFO] TestPrettyPrintBare.server: Handled event for server in area: event=member-join server=Node-2ebe9a73-8044-60c1-5755-1b5571820db4.dc1 area=wan writer.go:29: 2021-01-29T19:33:19.359Z [INFO] TestPrettyPrintBare: Started DNS server: address=127.0.0.1:29856 network=udp writer.go:29: 2021-01-29T19:33:19.362Z [INFO] TestPrettyPrintBare: Started DNS server: address=127.0.0.1:29856 network=tcp writer.go:29: 2021-01-29T19:33:19.365Z [INFO] TestPrettyPrintBare: Started HTTP server: address=127.0.0.1:29857 network=tcp writer.go:29: 2021-01-29T19:33:19.366Z [INFO] TestPrettyPrintBare: started state syncer writer.go:29: 2021-01-29T19:33:19.378Z [WARN] TestPrettyPrintBare.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:19.380Z [INFO] TestPrettyPrintBare.server.raft: entering candidate state: node="Node at 127.0.0.1:29861 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:19.383Z [DEBUG] TestPrettyPrintBare.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:19.385Z [DEBUG] TestPrettyPrintBare.server.raft: vote granted: from=2ebe9a73-8044-60c1-5755-1b5571820db4 term=2 tally=1 writer.go:29: 2021-01-29T19:33:19.387Z [INFO] TestPrettyPrintBare.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:19.389Z [INFO] TestPrettyPrintBare.server.raft: entering leader state: leader="Node at 127.0.0.1:29861 [Leader]" writer.go:29: 2021-01-29T19:33:19.391Z [INFO] TestPrettyPrintBare.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:19.392Z [INFO] TestPrettyPrintBare.server: New leader elected: payload=Node-2ebe9a73-8044-60c1-5755-1b5571820db4 writer.go:29: 2021-01-29T19:33:19.393Z [DEBUG] TestPrettyPrintBare.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29861 writer.go:29: 2021-01-29T19:33:19.397Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:19.403Z [INFO] TestPrettyPrintBare.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:19.404Z [INFO] TestPrettyPrintBare.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.404Z [DEBUG] TestPrettyPrintBare.server: Skipping self join check for node since the cluster is too small: node=Node-2ebe9a73-8044-60c1-5755-1b5571820db4 writer.go:29: 2021-01-29T19:33:19.405Z [INFO] TestPrettyPrintBare.server: member joined, marking health alive: member=Node-2ebe9a73-8044-60c1-5755-1b5571820db4 === RUN TestPProfHandlers_ACLs/case_5_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:403,_token:"towel",_endpoint:"/debug/pprof/heap",_nilResponse:true}) === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:19.435Z [INFO] TestPProfHandlers_ACLs: Requesting shutdown writer.go:29: 2021-01-29T19:33:19.436Z [INFO] TestPProfHandlers_ACLs.server: shutting down server writer.go:29: 2021-01-29T19:33:19.437Z [DEBUG] TestPProfHandlers_ACLs.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:19.439Z [DEBUG] TestPProfHandlers_ACLs.leader: stopping routine: routine="acl token reaping" === CONT TestParseSource writer.go:29: 2021-01-29T19:33:19.440Z [DEBUG] TestParseSource: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:19.441Z [DEBUG] TestPProfHandlers_ACLs.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.442Z [WARN] TestPProfHandlers_ACLs.server.serf.lan: serf: Shutdown without a Leave === CONT TestParseSource writer.go:29: 2021-01-29T19:33:19.445Z [INFO] TestParseSource: Synced node info === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:19.440Z [DEBUG] TestPProfHandlers_ACLs.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestParseSource writer.go:29: 2021-01-29T19:33:19.446Z [DEBUG] TestParseSource: Node info in sync === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:19.441Z [DEBUG] TestPProfHandlers_ACLs.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:19.442Z [DEBUG] TestPProfHandlers_ACLs.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.445Z [WARN] TestPProfHandlers_ACLs.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:19.453Z [INFO] TestPProfHandlers_ACLs.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:19.454Z [INFO] TestPProfHandlers_ACLs: consul server down writer.go:29: 2021-01-29T19:33:19.456Z [INFO] TestPProfHandlers_ACLs: shutdown complete writer.go:29: 2021-01-29T19:33:19.458Z [INFO] TestPProfHandlers_ACLs: Stopping server: protocol=DNS address=127.0.0.1:29826 network=tcp writer.go:29: 2021-01-29T19:33:19.460Z [INFO] TestPProfHandlers_ACLs: Stopping server: protocol=DNS address=127.0.0.1:29826 network=udp writer.go:29: 2021-01-29T19:33:19.463Z [INFO] TestPProfHandlers_ACLs: Stopping server: protocol=HTTP address=127.0.0.1:29827 network=tcp === CONT TestParseSource writer.go:29: 2021-01-29T19:33:19.469Z [INFO] TestParseSource: Requesting shutdown writer.go:29: 2021-01-29T19:33:19.470Z [INFO] TestParseSource.server: shutting down server writer.go:29: 2021-01-29T19:33:19.471Z [DEBUG] TestParseSource.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.472Z [WARN] TestParseSource.server.serf.lan: serf: Shutdown without a Leave === CONT TestPrettyPrintBare writer.go:29: 2021-01-29T19:33:19.472Z [DEBUG] TestPrettyPrintBare.http: Request finished: method=GET url=/v1/kv/key?pretty from= latency=1.67021ms === CONT TestParseSource writer.go:29: 2021-01-29T19:33:19.472Z [DEBUG] TestParseSource.leader: stopped routine: routine="CA root pruning" === CONT TestPrettyPrintBare writer.go:29: 2021-01-29T19:33:19.473Z [INFO] TestPrettyPrintBare: Requesting shutdown writer.go:29: 2021-01-29T19:33:19.475Z [INFO] TestPrettyPrintBare.server: shutting down server writer.go:29: 2021-01-29T19:33:19.476Z [DEBUG] TestPrettyPrintBare.leader: stopping routine: routine="CA root pruning" === CONT TestParseSource writer.go:29: 2021-01-29T19:33:19.475Z [WARN] TestParseSource.server.serf.wan: serf: Shutdown without a Leave === CONT TestPrettyPrintBare writer.go:29: 2021-01-29T19:33:19.478Z [WARN] TestPrettyPrintBare.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:19.477Z [ERROR] TestPrettyPrintBare.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:19.480Z [DEBUG] TestPrettyPrintBare.leader: stopped routine: routine="CA root pruning" === CONT TestParseSource writer.go:29: 2021-01-29T19:33:19.481Z [INFO] TestParseSource.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:19.482Z [INFO] TestParseSource: consul server down writer.go:29: 2021-01-29T19:33:19.486Z [INFO] TestParseSource: shutdown complete writer.go:29: 2021-01-29T19:33:19.487Z [INFO] TestParseSource: Stopping server: protocol=DNS address=127.0.0.1:29850 network=tcp === CONT TestPrettyPrintBare writer.go:29: 2021-01-29T19:33:19.481Z [WARN] TestPrettyPrintBare.server.serf.wan: serf: Shutdown without a Leave === CONT TestParseSource writer.go:29: 2021-01-29T19:33:19.488Z [INFO] TestParseSource: Stopping server: protocol=DNS address=127.0.0.1:29850 network=udp writer.go:29: 2021-01-29T19:33:19.494Z [INFO] TestParseSource: Stopping server: protocol=HTTP address=127.0.0.1:29851 network=tcp === CONT TestPrettyPrintBare writer.go:29: 2021-01-29T19:33:19.497Z [INFO] TestPrettyPrintBare.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:19.498Z [INFO] TestPrettyPrintBare: consul server down writer.go:29: 2021-01-29T19:33:19.500Z [INFO] TestPrettyPrintBare: shutdown complete writer.go:29: 2021-01-29T19:33:19.503Z [INFO] TestPrettyPrintBare: Stopping server: protocol=DNS address=127.0.0.1:29856 network=tcp writer.go:29: 2021-01-29T19:33:19.505Z [INFO] TestPrettyPrintBare: Stopping server: protocol=DNS address=127.0.0.1:29856 network=udp writer.go:29: 2021-01-29T19:33:19.508Z [INFO] TestPrettyPrintBare: Stopping server: protocol=HTTP address=127.0.0.1:29857 network=tcp === CONT TestPProfHandlers_EnableDebug writer.go:29: 2021-01-29T19:33:19.849Z [INFO] TestPProfHandlers_EnableDebug: Requesting shutdown writer.go:29: 2021-01-29T19:33:19.852Z [INFO] TestPProfHandlers_EnableDebug.server: shutting down server writer.go:29: 2021-01-29T19:33:19.853Z [DEBUG] TestPProfHandlers_EnableDebug.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.855Z [WARN] TestPProfHandlers_EnableDebug.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:19.855Z [DEBUG] TestPProfHandlers_EnableDebug.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.858Z [WARN] TestPProfHandlers_EnableDebug.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:19.864Z [INFO] TestPProfHandlers_EnableDebug.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:19.865Z [INFO] TestPProfHandlers_EnableDebug: consul server down writer.go:29: 2021-01-29T19:33:19.867Z [INFO] TestPProfHandlers_EnableDebug: shutdown complete writer.go:29: 2021-01-29T19:33:19.868Z [INFO] TestPProfHandlers_EnableDebug: Stopping server: protocol=DNS address=127.0.0.1:29844 network=tcp writer.go:29: 2021-01-29T19:33:19.870Z [INFO] TestPProfHandlers_EnableDebug: Stopping server: protocol=DNS address=127.0.0.1:29844 network=udp writer.go:29: 2021-01-29T19:33:19.871Z [INFO] TestPProfHandlers_EnableDebug: Stopping server: protocol=HTTP address=127.0.0.1:29845 network=tcp === CONT TestPProfHandlers_ACLs writer.go:29: 2021-01-29T19:33:19.964Z [INFO] TestPProfHandlers_ACLs: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:19.965Z [INFO] TestPProfHandlers_ACLs: Endpoints down --- PASS: TestPProfHandlers_ACLs (1.74s) --- PASS: TestPProfHandlers_ACLs/case_0_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:200,_token:"master",_endpoint:"/debug/pprof/heap",_nilResponse:false}) (0.37s) --- PASS: TestPProfHandlers_ACLs/case_1_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:403,_token:"agent",_endpoint:"/debug/pprof/heap",_nilResponse:true}) (0.00s) --- PASS: TestPProfHandlers_ACLs/case_2_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:403,_token:"agent",_endpoint:"/debug/pprof/",_nilResponse:true}) (0.00s) --- PASS: TestPProfHandlers_ACLs/case_3_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:403,_token:"",_endpoint:"/debug/pprof/",_nilResponse:true}) (0.00s) --- PASS: TestPProfHandlers_ACLs/case_4_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:200,_token:"master",_endpoint:"/debug/pprof/heap",_nilResponse:false}) (0.31s) --- PASS: TestPProfHandlers_ACLs/case_5_(struct_{_code_int;_token_string;_endpoint_string;_nilResponse_bool_}{code:403,_token:"towel",_endpoint:"/debug/pprof/heap",_nilResponse:true}) (0.00s) === CONT TestPrettyPrint === CONT TestParseSource writer.go:29: 2021-01-29T19:33:20.005Z [INFO] TestParseSource: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:20.007Z [INFO] TestParseSource: Endpoints down --- PASS: TestParseSource (0.96s) === CONT TestHTTP_wrap_obfuscateLog === CONT TestPrettyPrintBare writer.go:29: 2021-01-29T19:33:20.026Z [INFO] TestPrettyPrintBare: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:20.038Z [INFO] TestPrettyPrintBare: Endpoints down --- PASS: TestPrettyPrintBare (0.76s) === CONT TestContentTypeIsJSON === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.069Z [WARN] TestPrettyPrint: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:20.074Z [DEBUG] TestPrettyPrint.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:20.078Z [DEBUG] TestPrettyPrint.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.142Z [WARN] TestContentTypeIsJSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:20.144Z [DEBUG] TestContentTypeIsJSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:20.161Z [DEBUG] TestContentTypeIsJSON.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:20.189Z [INFO] TestContentTypeIsJSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f5ce4a51-911b-d251-96f9-83bd9e80daf2 Address:127.0.0.1:29879}]" === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.190Z [INFO] TestPrettyPrint.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c39fd998-463c-d952-d1f8-2116d3740d2c Address:127.0.0.1:29867}]" writer.go:29: 2021-01-29T19:33:20.194Z [INFO] TestPrettyPrint.server.raft: entering follower state: follower="Node at 127.0.0.1:29867 [Follower]" leader= === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.197Z [INFO] TestContentTypeIsJSON.server.raft: entering follower state: follower="Node at 127.0.0.1:29879 [Follower]" leader= writer.go:29: 2021-01-29T19:33:20.212Z [INFO] TestContentTypeIsJSON.server.serf.wan: serf: EventMemberJoin: Node-f5ce4a51-911b-d251-96f9-83bd9e80daf2.dc1 127.0.0.1 === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.216Z [INFO] TestPrettyPrint.server.serf.wan: serf: EventMemberJoin: Node-c39fd998-463c-d952-d1f8-2116d3740d2c.dc1 127.0.0.1 === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.264Z [WARN] TestContentTypeIsJSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:20.264Z [INFO] TestContentTypeIsJSON.server.raft: entering candidate state: node="Node at 127.0.0.1:29879 [Candidate]" term=2 === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.262Z [WARN] TestPrettyPrint.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.266Z [INFO] TestContentTypeIsJSON.server.serf.lan: serf: EventMemberJoin: Node-f5ce4a51-911b-d251-96f9-83bd9e80daf2 127.0.0.1 === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.267Z [INFO] TestPrettyPrint.server.raft: entering candidate state: node="Node at 127.0.0.1:29867 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:20.265Z [INFO] TestPrettyPrint.server.serf.lan: serf: EventMemberJoin: Node-c39fd998-463c-d952-d1f8-2116d3740d2c 127.0.0.1 === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.303Z [INFO] TestContentTypeIsJSON.server: Adding LAN server: server="Node-f5ce4a51-911b-d251-96f9-83bd9e80daf2 (Addr: tcp/127.0.0.1:29879) (DC: dc1)" === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.303Z [INFO] TestPrettyPrint.server: Adding LAN server: server="Node-c39fd998-463c-d952-d1f8-2116d3740d2c (Addr: tcp/127.0.0.1:29867) (DC: dc1)" === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.304Z [INFO] TestContentTypeIsJSON: Started DNS server: address=127.0.0.1:29874 network=udp writer.go:29: 2021-01-29T19:33:20.305Z [WARN] TestContentTypeIsJSON.server.raft: unable to get address for sever, using fallback address: id=f5ce4a51-911b-d251-96f9-83bd9e80daf2 fallback=127.0.0.1:29879 error="Could not find address for server id f5ce4a51-911b-d251-96f9-83bd9e80daf2" === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.304Z [INFO] TestPrettyPrint: Started DNS server: address=127.0.0.1:29862 network=udp === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.310Z [INFO] TestContentTypeIsJSON.server: Handled event for server in area: event=member-join server=Node-f5ce4a51-911b-d251-96f9-83bd9e80daf2.dc1 area=wan === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.304Z [WARN] TestPrettyPrint.server.raft: unable to get address for sever, using fallback address: id=c39fd998-463c-d952-d1f8-2116d3740d2c fallback=127.0.0.1:29867 error="Could not find address for server id c39fd998-463c-d952-d1f8-2116d3740d2c" === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.311Z [INFO] TestContentTypeIsJSON: Started DNS server: address=127.0.0.1:29874 network=tcp === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.319Z [INFO] TestPrettyPrint.server: Handled event for server in area: event=member-join server=Node-c39fd998-463c-d952-d1f8-2116d3740d2c.dc1 area=wan writer.go:29: 2021-01-29T19:33:20.321Z [INFO] TestPrettyPrint: Started DNS server: address=127.0.0.1:29862 network=tcp writer.go:29: 2021-01-29T19:33:20.323Z [INFO] TestPrettyPrint: Started HTTP server: address=127.0.0.1:29863 network=tcp writer.go:29: 2021-01-29T19:33:20.324Z [INFO] TestPrettyPrint: started state syncer === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.324Z [INFO] TestContentTypeIsJSON: Started HTTP server: address=127.0.0.1:29875 network=tcp writer.go:29: 2021-01-29T19:33:20.325Z [INFO] TestContentTypeIsJSON: started state syncer === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.326Z [DEBUG] TestPrettyPrint.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:20.327Z [DEBUG] TestPrettyPrint.server.raft: vote granted: from=c39fd998-463c-d952-d1f8-2116d3740d2c term=2 tally=1 writer.go:29: 2021-01-29T19:33:20.328Z [INFO] TestPrettyPrint.server.raft: election won: tally=1 === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.326Z [DEBUG] TestContentTypeIsJSON.server.raft: votes: needed=1 === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.328Z [INFO] TestPrettyPrint.server.raft: entering leader state: leader="Node at 127.0.0.1:29867 [Leader]" writer.go:29: 2021-01-29T19:33:20.330Z [INFO] TestPrettyPrint.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:20.331Z [INFO] TestPrettyPrint.server: New leader elected: payload=Node-c39fd998-463c-d952-d1f8-2116d3740d2c === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.329Z [DEBUG] TestContentTypeIsJSON.server.raft: vote granted: from=f5ce4a51-911b-d251-96f9-83bd9e80daf2 term=2 tally=1 === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.332Z [DEBUG] TestPrettyPrint.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29867 writer.go:29: 2021-01-29T19:33:20.337Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.333Z [INFO] TestContentTypeIsJSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:20.340Z [INFO] TestContentTypeIsJSON.server.raft: entering leader state: leader="Node at 127.0.0.1:29879 [Leader]" writer.go:29: 2021-01-29T19:33:20.348Z [INFO] TestContentTypeIsJSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:20.349Z [INFO] TestContentTypeIsJSON.server: New leader elected: payload=Node-f5ce4a51-911b-d251-96f9-83bd9e80daf2 === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.354Z [INFO] TestPrettyPrint.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:20.355Z [INFO] TestPrettyPrint.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:20.357Z [DEBUG] TestPrettyPrint.server: Skipping self join check for node since the cluster is too small: node=Node-c39fd998-463c-d952-d1f8-2116d3740d2c writer.go:29: 2021-01-29T19:33:20.358Z [INFO] TestPrettyPrint.server: member joined, marking health alive: member=Node-c39fd998-463c-d952-d1f8-2116d3740d2c === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.359Z [DEBUG] TestContentTypeIsJSON.http: Request finished: method=GET url=/v1/kv/key from= latency=24.729µs writer.go:29: 2021-01-29T19:33:20.362Z [INFO] TestContentTypeIsJSON: Requesting shutdown writer.go:29: 2021-01-29T19:33:20.363Z [INFO] TestContentTypeIsJSON.server: shutting down server writer.go:29: 2021-01-29T19:33:20.364Z [WARN] TestContentTypeIsJSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:20.365Z [ERROR] TestContentTypeIsJSON.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:20.367Z [WARN] TestContentTypeIsJSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:20.368Z [DEBUG] TestContentTypeIsJSON.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29879 writer.go:29: 2021-01-29T19:33:20.371Z [ERROR] TestContentTypeIsJSON.server.autopilot: failed to initialize config: error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:20.373Z [ERROR] TestContentTypeIsJSON.server: failed to establish leadership: error="raft is already shutdown" === CONT TestPProfHandlers_EnableDebug writer.go:29: 2021-01-29T19:33:20.373Z [INFO] TestPProfHandlers_EnableDebug: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:20.374Z [INFO] TestPProfHandlers_EnableDebug: Endpoints down --- PASS: TestPProfHandlers_EnableDebug (1.80s) === CONT TestUIResponseHeaders === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.369Z [INFO] TestContentTypeIsJSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:20.370Z [INFO] TestContentTypeIsJSON: consul server down writer.go:29: 2021-01-29T19:33:20.377Z [INFO] TestContentTypeIsJSON: shutdown complete writer.go:29: 2021-01-29T19:33:20.378Z [INFO] TestContentTypeIsJSON: Stopping server: protocol=DNS address=127.0.0.1:29874 network=tcp writer.go:29: 2021-01-29T19:33:20.380Z [INFO] TestContentTypeIsJSON: Stopping server: protocol=DNS address=127.0.0.1:29874 network=udp writer.go:29: 2021-01-29T19:33:20.381Z [INFO] TestContentTypeIsJSON: Stopping server: protocol=HTTP address=127.0.0.1:29875 network=tcp === CONT TestUIResponseHeaders writer.go:29: 2021-01-29T19:33:20.384Z [WARN] TestUIResponseHeaders: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:20.385Z [DEBUG] TestUIResponseHeaders.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:20.386Z [DEBUG] TestUIResponseHeaders.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:20.390Z [INFO] TestUIResponseHeaders.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0b5f92ad-491b-4438-ec2f-386d6a875208 Address:127.0.0.1:29885}]" writer.go:29: 2021-01-29T19:33:20.391Z [INFO] TestUIResponseHeaders.server.raft: entering follower state: follower="Node at 127.0.0.1:29885 [Follower]" leader= writer.go:29: 2021-01-29T19:33:20.394Z [INFO] TestUIResponseHeaders.server.serf.wan: serf: EventMemberJoin: Node-0b5f92ad-491b-4438-ec2f-386d6a875208.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:20.396Z [INFO] TestUIResponseHeaders.server.serf.lan: serf: EventMemberJoin: Node-0b5f92ad-491b-4438-ec2f-386d6a875208 127.0.0.1 writer.go:29: 2021-01-29T19:33:20.399Z [INFO] TestUIResponseHeaders: Started DNS server: address=127.0.0.1:29880 network=udp writer.go:29: 2021-01-29T19:33:20.399Z [INFO] TestUIResponseHeaders.server: Adding LAN server: server="Node-0b5f92ad-491b-4438-ec2f-386d6a875208 (Addr: tcp/127.0.0.1:29885) (DC: dc1)" writer.go:29: 2021-01-29T19:33:20.400Z [INFO] TestUIResponseHeaders: Started DNS server: address=127.0.0.1:29880 network=tcp writer.go:29: 2021-01-29T19:33:20.400Z [INFO] TestUIResponseHeaders.server: Handled event for server in area: event=member-join server=Node-0b5f92ad-491b-4438-ec2f-386d6a875208.dc1 area=wan writer.go:29: 2021-01-29T19:33:20.403Z [INFO] TestUIResponseHeaders: Started HTTP server: address=127.0.0.1:29881 network=tcp writer.go:29: 2021-01-29T19:33:20.411Z [INFO] TestUIResponseHeaders: started state syncer writer.go:29: 2021-01-29T19:33:20.446Z [WARN] TestUIResponseHeaders.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:20.447Z [INFO] TestUIResponseHeaders.server.raft: entering candidate state: node="Node at 127.0.0.1:29885 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:20.460Z [DEBUG] TestUIResponseHeaders.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:20.461Z [DEBUG] TestUIResponseHeaders.server.raft: vote granted: from=0b5f92ad-491b-4438-ec2f-386d6a875208 term=2 tally=1 writer.go:29: 2021-01-29T19:33:20.462Z [INFO] TestUIResponseHeaders.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:20.464Z [INFO] TestUIResponseHeaders.server.raft: entering leader state: leader="Node at 127.0.0.1:29885 [Leader]" writer.go:29: 2021-01-29T19:33:20.466Z [INFO] TestUIResponseHeaders.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:20.467Z [INFO] TestUIResponseHeaders.server: New leader elected: payload=Node-0b5f92ad-491b-4438-ec2f-386d6a875208 writer.go:29: 2021-01-29T19:33:20.473Z [DEBUG] TestUIResponseHeaders.http: Request finished: method=GET url=/ui from= latency=1.508µs writer.go:29: 2021-01-29T19:33:20.474Z [INFO] TestUIResponseHeaders: Requesting shutdown writer.go:29: 2021-01-29T19:33:20.475Z [INFO] TestUIResponseHeaders.server: shutting down server writer.go:29: 2021-01-29T19:33:20.477Z [WARN] TestUIResponseHeaders.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:20.477Z [ERROR] TestUIResponseHeaders.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:20.479Z [DEBUG] TestUIResponseHeaders.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29885 writer.go:29: 2021-01-29T19:33:20.479Z [WARN] TestUIResponseHeaders.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:20.484Z [INFO] TestUIResponseHeaders.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:20.484Z [ERROR] TestUIResponseHeaders.server: failed to establish leadership: error="leadership lost while committing log" writer.go:29: 2021-01-29T19:33:20.484Z [INFO] TestUIResponseHeaders: consul server down writer.go:29: 2021-01-29T19:33:20.488Z [INFO] TestUIResponseHeaders: shutdown complete writer.go:29: 2021-01-29T19:33:20.489Z [INFO] TestUIResponseHeaders: Stopping server: protocol=DNS address=127.0.0.1:29880 network=tcp writer.go:29: 2021-01-29T19:33:20.488Z [ERROR] TestUIResponseHeaders.server: failed to transfer leadership attempt, will retry: attempt=0 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:20.492Z [ERROR] TestUIResponseHeaders.server: failed to transfer leadership attempt, will retry: attempt=1 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:20.493Z [ERROR] TestUIResponseHeaders.server: failed to transfer leadership attempt, will retry: attempt=2 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:20.494Z [ERROR] TestUIResponseHeaders.server: failed to transfer leadership: error="failed to transfer leadership in 3 attempts" writer.go:29: 2021-01-29T19:33:20.490Z [INFO] TestUIResponseHeaders: Stopping server: protocol=DNS address=127.0.0.1:29880 network=udp writer.go:29: 2021-01-29T19:33:20.498Z [INFO] TestUIResponseHeaders: Stopping server: protocol=HTTP address=127.0.0.1:29881 network=tcp === RUN TestHTTP_wrap_obfuscateLog//some/url?token=secret1&token=secret2 === RUN TestHTTP_wrap_obfuscateLog//v1/acl/clone/secret1 === RUN TestHTTP_wrap_obfuscateLog//v1/acl/clone/secret1?token=secret2 === RUN TestHTTP_wrap_obfuscateLog//v1/acl/destroy/secret1 === RUN TestHTTP_wrap_obfuscateLog//v1/acl/destroy/secret1?token=secret2 === RUN TestHTTP_wrap_obfuscateLog//v1/acl/info/secret1 === RUN TestHTTP_wrap_obfuscateLog//v1/acl/info/secret1?token=secret2 === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:20.624Z [DEBUG] TestPrettyPrint.http: Request finished: method=GET url=/v1/kv/key?pretty=1 from= latency=51.728µs writer.go:29: 2021-01-29T19:33:20.632Z [INFO] TestPrettyPrint: Requesting shutdown writer.go:29: 2021-01-29T19:33:20.636Z [INFO] TestPrettyPrint.server: shutting down server writer.go:29: 2021-01-29T19:33:20.640Z [DEBUG] TestPrettyPrint.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:20.640Z [ERROR] TestPrettyPrint.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:20.642Z [WARN] TestPrettyPrint.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:20.642Z [DEBUG] TestPrettyPrint.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:20.646Z [WARN] TestPrettyPrint.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:20.670Z [INFO] TestPrettyPrint.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:20.671Z [INFO] TestPrettyPrint: consul server down writer.go:29: 2021-01-29T19:33:20.674Z [INFO] TestPrettyPrint: shutdown complete writer.go:29: 2021-01-29T19:33:20.676Z [INFO] TestPrettyPrint: Stopping server: protocol=DNS address=127.0.0.1:29862 network=tcp writer.go:29: 2021-01-29T19:33:20.678Z [INFO] TestPrettyPrint: Stopping server: protocol=DNS address=127.0.0.1:29862 network=udp writer.go:29: 2021-01-29T19:33:20.679Z [INFO] TestPrettyPrint: Stopping server: protocol=HTTP address=127.0.0.1:29863 network=tcp === CONT TestContentTypeIsJSON writer.go:29: 2021-01-29T19:33:20.883Z [INFO] TestContentTypeIsJSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:20.884Z [INFO] TestContentTypeIsJSON: Endpoints down --- PASS: TestContentTypeIsJSON (0.84s) === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:20.896Z [WARN] TestHTTPAPIResponseHeaders: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:20.898Z [DEBUG] TestHTTPAPIResponseHeaders.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:20.899Z [DEBUG] TestHTTPAPIResponseHeaders.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:20.904Z [INFO] TestHTTPAPIResponseHeaders.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:22f1b740-8042-110f-a0fc-f9ea9c325487 Address:127.0.0.1:29891}]" writer.go:29: 2021-01-29T19:33:20.907Z [INFO] TestHTTPAPIResponseHeaders.server.raft: entering follower state: follower="Node at 127.0.0.1:29891 [Follower]" leader= writer.go:29: 2021-01-29T19:33:20.909Z [INFO] TestHTTPAPIResponseHeaders.server.serf.wan: serf: EventMemberJoin: Node-22f1b740-8042-110f-a0fc-f9ea9c325487.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:20.912Z [INFO] TestHTTPAPIResponseHeaders.server.serf.lan: serf: EventMemberJoin: Node-22f1b740-8042-110f-a0fc-f9ea9c325487 127.0.0.1 writer.go:29: 2021-01-29T19:33:20.914Z [INFO] TestHTTPAPIResponseHeaders: Started DNS server: address=127.0.0.1:29886 network=udp writer.go:29: 2021-01-29T19:33:20.916Z [INFO] TestHTTPAPIResponseHeaders.server: Handled event for server in area: event=member-join server=Node-22f1b740-8042-110f-a0fc-f9ea9c325487.dc1 area=wan writer.go:29: 2021-01-29T19:33:20.915Z [INFO] TestHTTPAPIResponseHeaders.server: Adding LAN server: server="Node-22f1b740-8042-110f-a0fc-f9ea9c325487 (Addr: tcp/127.0.0.1:29891) (DC: dc1)" writer.go:29: 2021-01-29T19:33:20.918Z [INFO] TestHTTPAPIResponseHeaders: Started DNS server: address=127.0.0.1:29886 network=tcp writer.go:29: 2021-01-29T19:33:20.920Z [INFO] TestHTTPAPIResponseHeaders: Started HTTP server: address=127.0.0.1:29887 network=tcp writer.go:29: 2021-01-29T19:33:20.921Z [INFO] TestHTTPAPIResponseHeaders: started state syncer writer.go:29: 2021-01-29T19:33:20.957Z [WARN] TestHTTPAPIResponseHeaders.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:20.958Z [INFO] TestHTTPAPIResponseHeaders.server.raft: entering candidate state: node="Node at 127.0.0.1:29891 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:20.960Z [DEBUG] TestHTTPAPIResponseHeaders.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:20.961Z [DEBUG] TestHTTPAPIResponseHeaders.server.raft: vote granted: from=22f1b740-8042-110f-a0fc-f9ea9c325487 term=2 tally=1 writer.go:29: 2021-01-29T19:33:20.962Z [INFO] TestHTTPAPIResponseHeaders.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:20.963Z [INFO] TestHTTPAPIResponseHeaders.server.raft: entering leader state: leader="Node at 127.0.0.1:29891 [Leader]" writer.go:29: 2021-01-29T19:33:20.964Z [INFO] TestHTTPAPIResponseHeaders.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:20.965Z [INFO] TestHTTPAPIResponseHeaders.server: New leader elected: payload=Node-22f1b740-8042-110f-a0fc-f9ea9c325487 writer.go:29: 2021-01-29T19:33:20.965Z [DEBUG] TestHTTPAPIResponseHeaders.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29891 writer.go:29: 2021-01-29T19:33:20.970Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:20.975Z [INFO] TestHTTPAPIResponseHeaders.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:20.976Z [INFO] TestHTTPAPIResponseHeaders.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:20.977Z [DEBUG] TestHTTPAPIResponseHeaders.server: Skipping self join check for node since the cluster is too small: node=Node-22f1b740-8042-110f-a0fc-f9ea9c325487 writer.go:29: 2021-01-29T19:33:20.977Z [INFO] TestHTTPAPIResponseHeaders.server: member joined, marking health alive: member=Node-22f1b740-8042-110f-a0fc-f9ea9c325487 === CONT TestUIResponseHeaders writer.go:29: 2021-01-29T19:33:20.999Z [INFO] TestUIResponseHeaders: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:21.000Z [INFO] TestUIResponseHeaders: Endpoints down --- PASS: TestUIResponseHeaders (0.63s) === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.009Z [WARN] TestHTTPAPI_TranslateAddrHeader: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:21.010Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:21.011Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:21.015Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:186b73e3-50f5-120b-60c7-c7e7c0e78a12 Address:127.0.0.1:29897}]" writer.go:29: 2021-01-29T19:33:21.017Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.raft: entering follower state: follower="Node at 127.0.0.1:29897 [Follower]" leader= writer.go:29: 2021-01-29T19:33:21.017Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.serf.wan: serf: EventMemberJoin: Node-186b73e3-50f5-120b-60c7-c7e7c0e78a12.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.023Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.serf.lan: serf: EventMemberJoin: Node-186b73e3-50f5-120b-60c7-c7e7c0e78a12 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.027Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: Adding LAN server: server="Node-186b73e3-50f5-120b-60c7-c7e7c0e78a12 (Addr: tcp/127.0.0.1:29897) (DC: dc1)" writer.go:29: 2021-01-29T19:33:21.030Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: Handled event for server in area: event=member-join server=Node-186b73e3-50f5-120b-60c7-c7e7c0e78a12.dc1 area=wan writer.go:29: 2021-01-29T19:33:21.031Z [INFO] TestHTTPAPI_TranslateAddrHeader: Started DNS server: address=127.0.0.1:29892 network=udp writer.go:29: 2021-01-29T19:33:21.033Z [INFO] TestHTTPAPI_TranslateAddrHeader: Started DNS server: address=127.0.0.1:29892 network=tcp writer.go:29: 2021-01-29T19:33:21.034Z [INFO] TestHTTPAPI_TranslateAddrHeader: Started HTTP server: address=127.0.0.1:29893 network=tcp writer.go:29: 2021-01-29T19:33:21.035Z [INFO] TestHTTPAPI_TranslateAddrHeader: started state syncer --- PASS: TestHTTP_wrap_obfuscateLog (1.05s) --- PASS: TestHTTP_wrap_obfuscateLog//some/url?token=secret1&token=secret2 (0.00s) --- PASS: TestHTTP_wrap_obfuscateLog//v1/acl/clone/secret1 (0.00s) --- PASS: TestHTTP_wrap_obfuscateLog//v1/acl/clone/secret1?token=secret2 (0.00s) --- PASS: TestHTTP_wrap_obfuscateLog//v1/acl/destroy/secret1 (0.00s) --- PASS: TestHTTP_wrap_obfuscateLog//v1/acl/destroy/secret1?token=secret2 (0.00s) --- PASS: TestHTTP_wrap_obfuscateLog//v1/acl/info/secret1 (0.00s) --- PASS: TestHTTP_wrap_obfuscateLog//v1/acl/info/secret1?token=secret2 (0.00s) === CONT TestHTTPAPI_BlockEndpoints === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.063Z [WARN] TestHTTPAPI_TranslateAddrHeader.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:21.064Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.raft: entering candidate state: node="Node at 127.0.0.1:29897 [Candidate]" term=2 === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.065Z [WARN] TestHTTPAPI_BlockEndpoints: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:21.066Z [DEBUG] TestHTTPAPI_BlockEndpoints.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:21.067Z [DEBUG] TestHTTPAPI_BlockEndpoints.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.067Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:21.070Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.server.raft: vote granted: from=186b73e3-50f5-120b-60c7-c7e7c0e78a12 term=2 tally=1 writer.go:29: 2021-01-29T19:33:21.071Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.raft: election won: tally=1 === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.072Z [DEBUG] TestHTTPAPIResponseHeaders.http: Request finished: method=GET url=/v1/agent/self from= latency=1.832µs writer.go:29: 2021-01-29T19:33:21.074Z [INFO] TestHTTPAPIResponseHeaders: Requesting shutdown === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.073Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.raft: entering leader state: leader="Node at 127.0.0.1:29897 [Leader]" === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.075Z [INFO] TestHTTPAPIResponseHeaders.server: shutting down server writer.go:29: 2021-01-29T19:33:21.076Z [DEBUG] TestHTTPAPIResponseHeaders.leader: stopping routine: routine="CA root pruning" === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.075Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: cluster leadership acquired === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.076Z [ERROR] TestHTTPAPIResponseHeaders.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.076Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: New leader elected: payload=Node-186b73e3-50f5-120b-60c7-c7e7c0e78a12 === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.076Z [WARN] TestHTTPAPIResponseHeaders.server.serf.lan: serf: Shutdown without a Leave === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.076Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29897 === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.076Z [DEBUG] TestHTTPAPIResponseHeaders.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.082Z [WARN] TestHTTPAPIResponseHeaders.server.serf.wan: serf: Shutdown without a Leave === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.083Z [INFO] TestHTTPAPI_BlockEndpoints.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4c12c1ec-880a-4935-46b6-c594f63b7f2b Address:127.0.0.1:29903}]" writer.go:29: 2021-01-29T19:33:21.084Z [INFO] TestHTTPAPI_BlockEndpoints.server.raft: entering follower state: follower="Node at 127.0.0.1:29903 [Follower]" leader= === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.084Z [INFO] TestHTTPAPIResponseHeaders.server.router.manager: shutting down === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.086Z [INFO] TestHTTPAPI_BlockEndpoints.server.serf.wan: serf: EventMemberJoin: Node-4c12c1ec-880a-4935-46b6-c594f63b7f2b.dc1 127.0.0.1 === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.087Z [INFO] TestHTTPAPIResponseHeaders: consul server down === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.088Z [INFO] TestHTTPAPI_BlockEndpoints.server.serf.lan: serf: EventMemberJoin: Node-4c12c1ec-880a-4935-46b6-c594f63b7f2b 127.0.0.1 === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.089Z [INFO] TestHTTPAPIResponseHeaders: shutdown complete === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.091Z [INFO] TestHTTPAPI_BlockEndpoints.server: Adding LAN server: server="Node-4c12c1ec-880a-4935-46b6-c594f63b7f2b (Addr: tcp/127.0.0.1:29903) (DC: dc1)" writer.go:29: 2021-01-29T19:33:21.092Z [INFO] TestHTTPAPI_BlockEndpoints.server: Handled event for server in area: event=member-join server=Node-4c12c1ec-880a-4935-46b6-c594f63b7f2b.dc1 area=wan writer.go:29: 2021-01-29T19:33:21.091Z [INFO] TestHTTPAPI_BlockEndpoints: Started DNS server: address=127.0.0.1:29898 network=udp === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.091Z [INFO] TestHTTPAPIResponseHeaders: Stopping server: protocol=DNS address=127.0.0.1:29886 network=tcp === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.094Z [INFO] TestHTTPAPI_BlockEndpoints: Started DNS server: address=127.0.0.1:29898 network=tcp === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.094Z [INFO] TestHTTPAPIResponseHeaders: Stopping server: protocol=DNS address=127.0.0.1:29886 network=udp === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.095Z [INFO] TestHTTPAPI_BlockEndpoints: Started HTTP server: address=127.0.0.1:29899 network=tcp writer.go:29: 2021-01-29T19:33:21.096Z [INFO] TestHTTPAPI_BlockEndpoints: started state syncer === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.096Z [INFO] TestHTTPAPIResponseHeaders: Stopping server: protocol=HTTP address=127.0.0.1:29887 network=tcp === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.097Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:21.107Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:21.109Z [INFO] TestHTTPAPI_TranslateAddrHeader.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.111Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.server: Skipping self join check for node since the cluster is too small: node=Node-186b73e3-50f5-120b-60c7-c7e7c0e78a12 writer.go:29: 2021-01-29T19:33:21.112Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: member joined, marking health alive: member=Node-186b73e3-50f5-120b-60c7-c7e7c0e78a12 === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.124Z [WARN] TestHTTPAPI_BlockEndpoints.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:21.125Z [INFO] TestHTTPAPI_BlockEndpoints.server.raft: entering candidate state: node="Node at 127.0.0.1:29903 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:21.127Z [DEBUG] TestHTTPAPI_BlockEndpoints.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:21.128Z [DEBUG] TestHTTPAPI_BlockEndpoints.server.raft: vote granted: from=4c12c1ec-880a-4935-46b6-c594f63b7f2b term=2 tally=1 writer.go:29: 2021-01-29T19:33:21.129Z [INFO] TestHTTPAPI_BlockEndpoints.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:21.130Z [INFO] TestHTTPAPI_BlockEndpoints.server.raft: entering leader state: leader="Node at 127.0.0.1:29903 [Leader]" writer.go:29: 2021-01-29T19:33:21.131Z [INFO] TestHTTPAPI_BlockEndpoints.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:21.132Z [INFO] TestHTTPAPI_BlockEndpoints.server: New leader elected: payload=Node-4c12c1ec-880a-4935-46b6-c594f63b7f2b writer.go:29: 2021-01-29T19:33:21.134Z [DEBUG] TestHTTPAPI_BlockEndpoints.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29903 writer.go:29: 2021-01-29T19:33:21.156Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:21.164Z [INFO] TestHTTPAPI_BlockEndpoints.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:21.166Z [INFO] TestHTTPAPI_BlockEndpoints.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.168Z [DEBUG] TestHTTPAPI_BlockEndpoints.server: Skipping self join check for node since the cluster is too small: node=Node-4c12c1ec-880a-4935-46b6-c594f63b7f2b writer.go:29: 2021-01-29T19:33:21.170Z [INFO] TestHTTPAPI_BlockEndpoints.server: member joined, marking health alive: member=Node-4c12c1ec-880a-4935-46b6-c594f63b7f2b === CONT TestPrettyPrint writer.go:29: 2021-01-29T19:33:21.180Z [INFO] TestPrettyPrint: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:21.182Z [INFO] TestPrettyPrint: Endpoints down --- PASS: TestPrettyPrint (1.22s) === CONT TestSetMeta --- PASS: TestSetMeta (0.00s) === CONT TestSetLastContact === RUN TestSetLastContact/neg === RUN TestSetLastContact/zero === RUN TestSetLastContact/pos === RUN TestSetLastContact/pos_ms_only --- PASS: TestSetLastContact (0.01s) --- PASS: TestSetLastContact/neg (0.00s) --- PASS: TestSetLastContact/zero (0.00s) --- PASS: TestSetLastContact/pos (0.00s) --- PASS: TestSetLastContact/pos_ms_only (0.00s) === CONT TestSetKnownLeader --- PASS: TestSetKnownLeader (0.00s) === CONT TestSetIndex --- PASS: TestSetIndex (0.00s) === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.204Z [WARN] TestHTTPServer_UnixSocket_FileExists: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:21.206Z [DEBUG] TestHTTPServer_UnixSocket_FileExists.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:21.207Z [DEBUG] TestHTTPServer_UnixSocket_FileExists.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:21.212Z [INFO] TestHTTPServer_UnixSocket_FileExists.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c1dd3038-f294-c031-e6c3-8e429dc05420 Address:127.0.0.1:29909}]" writer.go:29: 2021-01-29T19:33:21.215Z [INFO] TestHTTPServer_UnixSocket_FileExists.server.raft: entering follower state: follower="Node at 127.0.0.1:29909 [Follower]" leader= writer.go:29: 2021-01-29T19:33:21.215Z [INFO] TestHTTPServer_UnixSocket_FileExists.server.serf.wan: serf: EventMemberJoin: Node-c1dd3038-f294-c031-e6c3-8e429dc05420.dc1 127.0.0.1 === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.227Z [ERROR] TestHTTPAPI_BlockEndpoints.http: Request error: method=GET url=/v1/agent/self from= error="Endpoint is blocked by agent configuration" writer.go:29: 2021-01-29T19:33:21.229Z [DEBUG] TestHTTPAPI_BlockEndpoints.http: Request finished: method=GET url=/v1/agent/checks from= latency=1.566µs writer.go:29: 2021-01-29T19:33:21.230Z [INFO] TestHTTPAPI_BlockEndpoints: Requesting shutdown writer.go:29: 2021-01-29T19:33:21.232Z [INFO] TestHTTPAPI_BlockEndpoints.server: shutting down server writer.go:29: 2021-01-29T19:33:21.233Z [DEBUG] TestHTTPAPI_BlockEndpoints.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.234Z [WARN] TestHTTPAPI_BlockEndpoints.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:21.234Z [DEBUG] TestHTTPAPI_BlockEndpoints.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.233Z [ERROR] TestHTTPAPI_BlockEndpoints.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.236Z [INFO] TestHTTPServer_UnixSocket_FileExists.server.serf.lan: serf: EventMemberJoin: Node-c1dd3038-f294-c031-e6c3-8e429dc05420 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.240Z [INFO] TestHTTPServer_UnixSocket_FileExists.server: Handled event for server in area: event=member-join server=Node-c1dd3038-f294-c031-e6c3-8e429dc05420.dc1 area=wan writer.go:29: 2021-01-29T19:33:21.242Z [INFO] TestHTTPServer_UnixSocket_FileExists.server: Adding LAN server: server="Node-c1dd3038-f294-c031-e6c3-8e429dc05420 (Addr: tcp/127.0.0.1:29909) (DC: dc1)" === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.242Z [WARN] TestHTTPAPI_BlockEndpoints.server.serf.wan: serf: Shutdown without a Leave === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.243Z [INFO] TestHTTPServer_UnixSocket_FileExists: Started DNS server: address=127.0.0.1:29904 network=udp === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.244Z [INFO] TestHTTPAPI_BlockEndpoints.server.router.manager: shutting down === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.245Z [INFO] TestHTTPServer_UnixSocket_FileExists: Started DNS server: address=127.0.0.1:29904 network=tcp === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.245Z [INFO] TestHTTPAPI_BlockEndpoints: consul server down writer.go:29: 2021-01-29T19:33:21.247Z [INFO] TestHTTPAPI_BlockEndpoints: shutdown complete === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.247Z [WARN] TestHTTPServer_UnixSocket_FileExists: Replacing socket: path=/tmp/consul-test/TestHTTPServer_UnixSocket_FileExists-consul463851154/test.sock === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.248Z [INFO] TestHTTPAPI_BlockEndpoints: Stopping server: protocol=DNS address=127.0.0.1:29898 network=tcp === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.249Z [INFO] TestHTTPServer_UnixSocket_FileExists: Started HTTP server: address=/tmp/consul-test/TestHTTPServer_UnixSocket_FileExists-consul463851154/test.sock network=unix === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.249Z [INFO] TestHTTPAPI_BlockEndpoints: Stopping server: protocol=DNS address=127.0.0.1:29898 network=udp === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.251Z [INFO] TestHTTPServer_UnixSocket_FileExists: started state syncer === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.251Z [INFO] TestHTTPAPI_BlockEndpoints: Stopping server: protocol=HTTP address=127.0.0.1:29899 network=tcp === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.275Z [WARN] TestHTTPServer_UnixSocket_FileExists.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:21.276Z [INFO] TestHTTPServer_UnixSocket_FileExists.server.raft: entering candidate state: node="Node at 127.0.0.1:29909 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:21.278Z [DEBUG] TestHTTPServer_UnixSocket_FileExists.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:21.278Z [DEBUG] TestHTTPServer_UnixSocket_FileExists.server.raft: vote granted: from=c1dd3038-f294-c031-e6c3-8e429dc05420 term=2 tally=1 writer.go:29: 2021-01-29T19:33:21.279Z [INFO] TestHTTPServer_UnixSocket_FileExists.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:21.280Z [INFO] TestHTTPServer_UnixSocket_FileExists.server.raft: entering leader state: leader="Node at 127.0.0.1:29909 [Leader]" writer.go:29: 2021-01-29T19:33:21.281Z [INFO] TestHTTPServer_UnixSocket_FileExists.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:21.282Z [DEBUG] TestHTTPServer_UnixSocket_FileExists.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29909 writer.go:29: 2021-01-29T19:33:21.282Z [INFO] TestHTTPServer_UnixSocket_FileExists.server: New leader elected: payload=Node-c1dd3038-f294-c031-e6c3-8e429dc05420 writer.go:29: 2021-01-29T19:33:21.297Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.304Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.http: Request finished: method=GET url=/v1/agent/self from= latency=1.052µs === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.316Z [INFO] TestHTTPServer_UnixSocket_FileExists.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:21.318Z [INFO] TestHTTPServer_UnixSocket_FileExists.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.323Z [DEBUG] TestHTTPServer_UnixSocket_FileExists.server: Skipping self join check for node since the cluster is too small: node=Node-c1dd3038-f294-c031-e6c3-8e429dc05420 writer.go:29: 2021-01-29T19:33:21.326Z [INFO] TestHTTPServer_UnixSocket_FileExists.server: member joined, marking health alive: member=Node-c1dd3038-f294-c031-e6c3-8e429dc05420 === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.350Z [WARN] TestHTTPAPI_TranslateAddrHeader: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:21.352Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:21.355Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:21.364Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ce5e6038-fb92-cf68-c2e7-3561d950da5a Address:127.0.0.1:29921}]" writer.go:29: 2021-01-29T19:33:21.366Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.raft: entering follower state: follower="Node at 127.0.0.1:29921 [Follower]" leader= writer.go:29: 2021-01-29T19:33:21.367Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.serf.wan: serf: EventMemberJoin: Node-ce5e6038-fb92-cf68-c2e7-3561d950da5a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.373Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.serf.lan: serf: EventMemberJoin: Node-ce5e6038-fb92-cf68-c2e7-3561d950da5a 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.376Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: Adding LAN server: server="Node-ce5e6038-fb92-cf68-c2e7-3561d950da5a (Addr: tcp/127.0.0.1:29921) (DC: dc1)" writer.go:29: 2021-01-29T19:33:21.377Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: Handled event for server in area: event=member-join server=Node-ce5e6038-fb92-cf68-c2e7-3561d950da5a.dc1 area=wan writer.go:29: 2021-01-29T19:33:21.380Z [INFO] TestHTTPAPI_TranslateAddrHeader: Started DNS server: address=127.0.0.1:29916 network=udp writer.go:29: 2021-01-29T19:33:21.381Z [INFO] TestHTTPAPI_TranslateAddrHeader: Started DNS server: address=127.0.0.1:29916 network=tcp writer.go:29: 2021-01-29T19:33:21.383Z [INFO] TestHTTPAPI_TranslateAddrHeader: Started HTTP server: address=127.0.0.1:29917 network=tcp writer.go:29: 2021-01-29T19:33:21.384Z [INFO] TestHTTPAPI_TranslateAddrHeader: started state syncer writer.go:29: 2021-01-29T19:33:21.411Z [WARN] TestHTTPAPI_TranslateAddrHeader.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:21.412Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.raft: entering candidate state: node="Node at 127.0.0.1:29921 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:21.413Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:21.414Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.server.raft: vote granted: from=ce5e6038-fb92-cf68-c2e7-3561d950da5a term=2 tally=1 writer.go:29: 2021-01-29T19:33:21.415Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:21.416Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.raft: entering leader state: leader="Node at 127.0.0.1:29921 [Leader]" writer.go:29: 2021-01-29T19:33:21.417Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:21.418Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: New leader elected: payload=Node-ce5e6038-fb92-cf68-c2e7-3561d950da5a writer.go:29: 2021-01-29T19:33:21.420Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29921 writer.go:29: 2021-01-29T19:33:21.424Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:21.427Z [DEBUG] TestHTTPAPI_TranslateAddrHeader: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:21.429Z [INFO] TestHTTPAPI_TranslateAddrHeader: Synced node info writer.go:29: 2021-01-29T19:33:21.444Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:21.446Z [INFO] TestHTTPAPI_TranslateAddrHeader.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.447Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.server: Skipping self join check for node since the cluster is too small: node=Node-ce5e6038-fb92-cf68-c2e7-3561d950da5a writer.go:29: 2021-01-29T19:33:21.449Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: member joined, marking health alive: member=Node-ce5e6038-fb92-cf68-c2e7-3561d950da5a writer.go:29: 2021-01-29T19:33:21.575Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.http: Request finished: method=GET url=/v1/agent/self from= latency=2.74µs writer.go:29: 2021-01-29T19:33:21.576Z [INFO] TestHTTPAPI_TranslateAddrHeader: Requesting shutdown === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.577Z [INFO] TestHTTPServer_UnixSocket_FileExists: Requesting shutdown writer.go:29: 2021-01-29T19:33:21.578Z [INFO] TestHTTPServer_UnixSocket_FileExists.server: shutting down server === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.578Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: shutting down server === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.579Z [DEBUG] TestHTTPServer_UnixSocket_FileExists.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.580Z [WARN] TestHTTPServer_UnixSocket_FileExists.server.serf.lan: serf: Shutdown without a Leave === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.580Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.581Z [ERROR] TestHTTPAPI_TranslateAddrHeader.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:21.583Z [WARN] TestHTTPAPI_TranslateAddrHeader.server.serf.lan: serf: Shutdown without a Leave === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.580Z [ERROR] TestHTTPServer_UnixSocket_FileExists.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:21.581Z [DEBUG] TestHTTPServer_UnixSocket_FileExists.leader: stopped routine: routine="CA root pruning" === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.583Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.leader: stopped routine: routine="CA root pruning" === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.588Z [WARN] TestHTTPServer_UnixSocket_FileExists.server.serf.wan: serf: Shutdown without a Leave === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.588Z [WARN] TestHTTPAPI_TranslateAddrHeader.server.serf.wan: serf: Shutdown without a Leave === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.591Z [INFO] TestHTTPServer_UnixSocket_FileExists.server.router.manager: shutting down === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.591Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.router.manager: shutting down === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.592Z [INFO] TestHTTPServer_UnixSocket_FileExists: consul server down === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.592Z [INFO] TestHTTPAPI_TranslateAddrHeader: consul server down === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.594Z [INFO] TestHTTPServer_UnixSocket_FileExists: shutdown complete === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.595Z [INFO] TestHTTPAPI_TranslateAddrHeader: shutdown complete === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.595Z [INFO] TestHTTPServer_UnixSocket_FileExists: Stopping server: protocol=DNS address=127.0.0.1:29904 network=tcp writer.go:29: 2021-01-29T19:33:21.596Z [INFO] TestHTTPServer_UnixSocket_FileExists: Stopping server: protocol=DNS address=127.0.0.1:29904 network=udp === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.596Z [INFO] TestHTTPAPI_TranslateAddrHeader: Stopping server: protocol=DNS address=127.0.0.1:29916 network=tcp === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:21.597Z [INFO] TestHTTPServer_UnixSocket_FileExists: Stopping server: protocol=HTTP address=/tmp/consul-test/TestHTTPServer_UnixSocket_FileExists-consul463851154/test.sock network=unix === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.598Z [INFO] TestHTTPAPI_TranslateAddrHeader: Stopping server: protocol=DNS address=127.0.0.1:29916 network=udp === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.599Z [INFO] TestHTTPAPIResponseHeaders: Waiting for endpoints to shut down === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:21.599Z [INFO] TestHTTPAPI_TranslateAddrHeader: Stopping server: protocol=HTTP address=127.0.0.1:29917 network=tcp === CONT TestHTTPAPIResponseHeaders writer.go:29: 2021-01-29T19:33:21.601Z [INFO] TestHTTPAPIResponseHeaders: Endpoints down --- PASS: TestHTTPAPIResponseHeaders (0.72s) === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:21.623Z [WARN] TestHTTPServer_UnixSocket: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:21.624Z [DEBUG] TestHTTPServer_UnixSocket.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:21.626Z [DEBUG] TestHTTPServer_UnixSocket.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:21.667Z [INFO] TestHTTPServer_UnixSocket.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cca9b405-ec60-2cee-9a2d-935320bf2a2e Address:127.0.0.1:29915}]" writer.go:29: 2021-01-29T19:33:21.670Z [INFO] TestHTTPServer_UnixSocket.server.serf.wan: serf: EventMemberJoin: Node-cca9b405-ec60-2cee-9a2d-935320bf2a2e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.673Z [INFO] TestHTTPServer_UnixSocket.server.serf.lan: serf: EventMemberJoin: Node-cca9b405-ec60-2cee-9a2d-935320bf2a2e 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.678Z [INFO] TestHTTPServer_UnixSocket.server.raft: entering follower state: follower="Node at 127.0.0.1:29915 [Follower]" leader= writer.go:29: 2021-01-29T19:33:21.678Z [INFO] TestHTTPServer_UnixSocket: Started DNS server: address=127.0.0.1:29910 network=udp writer.go:29: 2021-01-29T19:33:21.682Z [INFO] TestHTTPServer_UnixSocket.server: Handled event for server in area: event=member-join server=Node-cca9b405-ec60-2cee-9a2d-935320bf2a2e.dc1 area=wan writer.go:29: 2021-01-29T19:33:21.685Z [INFO] TestHTTPServer_UnixSocket.server: Adding LAN server: server="Node-cca9b405-ec60-2cee-9a2d-935320bf2a2e (Addr: tcp/127.0.0.1:29915) (DC: dc1)" writer.go:29: 2021-01-29T19:33:21.686Z [INFO] TestHTTPServer_UnixSocket: Started DNS server: address=127.0.0.1:29910 network=tcp writer.go:29: 2021-01-29T19:33:21.692Z [INFO] TestHTTPServer_UnixSocket: Started HTTP server: address=/tmp/consul-test/TestHTTPServer_UnixSocket-consul502569955/test.sock network=unix writer.go:29: 2021-01-29T19:33:21.694Z [INFO] TestHTTPServer_UnixSocket: started state syncer writer.go:29: 2021-01-29T19:33:21.732Z [WARN] TestHTTPServer_UnixSocket.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:21.732Z [INFO] TestHTTPServer_UnixSocket.server.raft: entering candidate state: node="Node at 127.0.0.1:29915 [Candidate]" term=2 === CONT TestHTTPAPI_BlockEndpoints writer.go:29: 2021-01-29T19:33:21.754Z [INFO] TestHTTPAPI_BlockEndpoints: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:21.756Z [INFO] TestHTTPAPI_BlockEndpoints: Endpoints down --- PASS: TestHTTPAPI_BlockEndpoints (0.70s) === CONT TestFilterNonPassing === CONT TestHealthConnectServiceNodes_PassingFilter --- PASS: TestFilterNonPassing (0.00s) === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:21.778Z [DEBUG] TestHTTPServer_UnixSocket.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:21.781Z [DEBUG] TestHTTPServer_UnixSocket.server.raft: vote granted: from=cca9b405-ec60-2cee-9a2d-935320bf2a2e term=2 tally=1 writer.go:29: 2021-01-29T19:33:21.784Z [INFO] TestHTTPServer_UnixSocket.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:21.786Z [INFO] TestHTTPServer_UnixSocket.server.raft: entering leader state: leader="Node at 127.0.0.1:29915 [Leader]" writer.go:29: 2021-01-29T19:33:21.788Z [INFO] TestHTTPServer_UnixSocket.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:21.794Z [INFO] TestHTTPServer_UnixSocket.server: New leader elected: payload=Node-cca9b405-ec60-2cee-9a2d-935320bf2a2e writer.go:29: 2021-01-29T19:33:21.811Z [DEBUG] TestHTTPServer_UnixSocket.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29915 writer.go:29: 2021-01-29T19:33:21.819Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:21.819Z [WARN] TestHealthConnectServiceNodes_PassingFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:21.824Z [DEBUG] TestHealthConnectServiceNodes_PassingFilter.tlsutil: Update: version=1 === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:21.834Z [INFO] TestHTTPServer_UnixSocket.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:21.834Z [INFO] TestHTTPServer_UnixSocket.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.835Z [DEBUG] TestHTTPServer_UnixSocket.server: Skipping self join check for node since the cluster is too small: node=Node-cca9b405-ec60-2cee-9a2d-935320bf2a2e writer.go:29: 2021-01-29T19:33:21.836Z [INFO] TestHTTPServer_UnixSocket.server: member joined, marking health alive: member=Node-cca9b405-ec60-2cee-9a2d-935320bf2a2e === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:21.843Z [DEBUG] TestHealthConnectServiceNodes_PassingFilter.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:21.866Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e9c1db18-8abf-c46a-1823-264001682a6e Address:127.0.0.1:29927}]" writer.go:29: 2021-01-29T19:33:21.878Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server.serf.wan: serf: EventMemberJoin: Node-e9c1db18-8abf-c46a-1823-264001682a6e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.885Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:29927 [Follower]" leader= writer.go:29: 2021-01-29T19:33:21.885Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server.serf.lan: serf: EventMemberJoin: Node-e9c1db18-8abf-c46a-1823-264001682a6e 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.897Z [INFO] TestHealthConnectServiceNodes_PassingFilter: Started DNS server: address=127.0.0.1:29922 network=udp writer.go:29: 2021-01-29T19:33:21.898Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server: Adding LAN server: server="Node-e9c1db18-8abf-c46a-1823-264001682a6e (Addr: tcp/127.0.0.1:29927) (DC: dc1)" writer.go:29: 2021-01-29T19:33:21.900Z [INFO] TestHealthConnectServiceNodes_PassingFilter: Started DNS server: address=127.0.0.1:29922 network=tcp writer.go:29: 2021-01-29T19:33:21.898Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server: Handled event for server in area: event=member-join server=Node-e9c1db18-8abf-c46a-1823-264001682a6e.dc1 area=wan writer.go:29: 2021-01-29T19:33:21.911Z [INFO] TestHealthConnectServiceNodes_PassingFilter: Started HTTP server: address=127.0.0.1:29923 network=tcp writer.go:29: 2021-01-29T19:33:21.913Z [INFO] TestHealthConnectServiceNodes_PassingFilter: started state syncer === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:21.947Z [DEBUG] TestHTTPServer_UnixSocket.http: Request finished: method=GET url=/v1/agent/self from=@ latency=11.10625ms === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:21.951Z [WARN] TestHealthConnectServiceNodes_PassingFilter.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:21.952Z [INFO] TestHTTPServer_UnixSocket: Requesting shutdown === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:21.952Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:29927 [Candidate]" term=2 === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:21.954Z [INFO] TestHTTPServer_UnixSocket.server: shutting down server writer.go:29: 2021-01-29T19:33:21.955Z [DEBUG] TestHTTPServer_UnixSocket.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.957Z [WARN] TestHTTPServer_UnixSocket.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:21.960Z [ERROR] TestHTTPServer_UnixSocket.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:21.962Z [DEBUG] TestHTTPServer_UnixSocket.leader: stopped routine: routine="CA root pruning" === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:21.979Z [DEBUG] TestHealthConnectServiceNodes_PassingFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:21.987Z [DEBUG] TestHealthConnectServiceNodes_PassingFilter.server.raft: vote granted: from=e9c1db18-8abf-c46a-1823-264001682a6e term=2 tally=1 writer.go:29: 2021-01-29T19:33:21.988Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:21.989Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:29927 [Leader]" writer.go:29: 2021-01-29T19:33:21.991Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server: cluster leadership acquired === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:21.980Z [WARN] TestHTTPServer_UnixSocket.server.serf.wan: serf: Shutdown without a Leave === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:21.992Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server: New leader elected: payload=Node-e9c1db18-8abf-c46a-1823-264001682a6e === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:21.997Z [INFO] TestHTTPServer_UnixSocket.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:21.999Z [INFO] TestHTTPServer_UnixSocket: consul server down writer.go:29: 2021-01-29T19:33:22.001Z [INFO] TestHTTPServer_UnixSocket: shutdown complete writer.go:29: 2021-01-29T19:33:22.002Z [INFO] TestHTTPServer_UnixSocket: Stopping server: protocol=DNS address=127.0.0.1:29910 network=tcp === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.003Z [DEBUG] TestHealthConnectServiceNodes_PassingFilter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29927 === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:22.005Z [INFO] TestHTTPServer_UnixSocket: Stopping server: protocol=DNS address=127.0.0.1:29910 network=udp === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.009Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:22.011Z [INFO] TestHTTPServer_UnixSocket: Stopping server: protocol=HTTP address=/tmp/consul-test/TestHTTPServer_UnixSocket-consul502569955/test.sock network=unix === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.034Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:22.035Z [INFO] TestHealthConnectServiceNodes_PassingFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.036Z [DEBUG] TestHealthConnectServiceNodes_PassingFilter.server: Skipping self join check for node since the cluster is too small: node=Node-e9c1db18-8abf-c46a-1823-264001682a6e writer.go:29: 2021-01-29T19:33:22.037Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server: member joined, marking health alive: member=Node-e9c1db18-8abf-c46a-1823-264001682a6e === CONT TestHTTPServer_UnixSocket_FileExists writer.go:29: 2021-01-29T19:33:22.099Z [INFO] TestHTTPServer_UnixSocket_FileExists: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:22.100Z [INFO] TestHTTPServer_UnixSocket_FileExists: Endpoints down --- PASS: TestHTTPServer_UnixSocket_FileExists (0.91s) === CONT TestHealthConnectServiceNodes_Filter === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:22.101Z [INFO] TestHTTPAPI_TranslateAddrHeader: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:22.104Z [INFO] TestHTTPAPI_TranslateAddrHeader: Endpoints down writer.go:29: 2021-01-29T19:33:22.105Z [INFO] TestHTTPAPI_TranslateAddrHeader: Requesting shutdown writer.go:29: 2021-01-29T19:33:22.108Z [INFO] TestHTTPAPI_TranslateAddrHeader.server: shutting down server writer.go:29: 2021-01-29T19:33:22.109Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.110Z [WARN] TestHTTPAPI_TranslateAddrHeader.server.serf.lan: serf: Shutdown without a Leave === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.111Z [WARN] TestHealthConnectServiceNodes_Filter: bootstrap = true: do not enable unless necessary === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:22.110Z [DEBUG] TestHTTPAPI_TranslateAddrHeader.leader: stopped routine: routine="CA root pruning" === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.113Z [DEBUG] TestHealthConnectServiceNodes_Filter.tlsutil: Update: version=1 === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:22.112Z [WARN] TestHTTPAPI_TranslateAddrHeader.server.serf.wan: serf: Shutdown without a Leave === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.116Z [DEBUG] TestHealthConnectServiceNodes_Filter.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:22.116Z [DEBUG] TestHTTPAPI_TranslateAddrHeader: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:22.117Z [DEBUG] TestHTTPAPI_TranslateAddrHeader: Node info in sync === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.118Z [INFO] TestHealthConnectServiceNodes_Filter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:71d25e6d-9846-5fe1-2879-fd53f8d0bb38 Address:127.0.0.1:29933}]" === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:22.118Z [DEBUG] TestHTTPAPI_TranslateAddrHeader: Node info in sync === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.120Z [INFO] TestHealthConnectServiceNodes_Filter.server.raft: entering follower state: follower="Node at 127.0.0.1:29933 [Follower]" leader= === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:22.117Z [INFO] TestHTTPAPI_TranslateAddrHeader.server.router.manager: shutting down === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.121Z [INFO] TestHealthConnectServiceNodes_Filter.server.serf.wan: serf: EventMemberJoin: Node-71d25e6d-9846-5fe1-2879-fd53f8d0bb38.dc1 127.0.0.1 === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:22.117Z [INFO] TestHTTPAPI_TranslateAddrHeader: consul server down writer.go:29: 2021-01-29T19:33:22.127Z [INFO] TestHTTPAPI_TranslateAddrHeader: shutdown complete writer.go:29: 2021-01-29T19:33:22.130Z [INFO] TestHTTPAPI_TranslateAddrHeader: Stopping server: protocol=DNS address=127.0.0.1:29892 network=tcp === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.128Z [INFO] TestHealthConnectServiceNodes_Filter.server.serf.lan: serf: EventMemberJoin: Node-71d25e6d-9846-5fe1-2879-fd53f8d0bb38 127.0.0.1 === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:22.132Z [INFO] TestHTTPAPI_TranslateAddrHeader: Stopping server: protocol=DNS address=127.0.0.1:29892 network=udp writer.go:29: 2021-01-29T19:33:22.136Z [INFO] TestHTTPAPI_TranslateAddrHeader: Stopping server: protocol=HTTP address=127.0.0.1:29893 network=tcp === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.137Z [INFO] TestHealthConnectServiceNodes_Filter.server: Adding LAN server: server="Node-71d25e6d-9846-5fe1-2879-fd53f8d0bb38 (Addr: tcp/127.0.0.1:29933) (DC: dc1)" writer.go:29: 2021-01-29T19:33:22.138Z [INFO] TestHealthConnectServiceNodes_Filter.server: Handled event for server in area: event=member-join server=Node-71d25e6d-9846-5fe1-2879-fd53f8d0bb38.dc1 area=wan writer.go:29: 2021-01-29T19:33:22.145Z [INFO] TestHealthConnectServiceNodes_Filter: Started DNS server: address=127.0.0.1:29928 network=tcp writer.go:29: 2021-01-29T19:33:22.146Z [INFO] TestHealthConnectServiceNodes_Filter: Started DNS server: address=127.0.0.1:29928 network=udp writer.go:29: 2021-01-29T19:33:22.148Z [INFO] TestHealthConnectServiceNodes_Filter: Started HTTP server: address=127.0.0.1:29929 network=tcp === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.151Z [DEBUG] TestHealthConnectServiceNodes_PassingFilter: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.150Z [INFO] TestHealthConnectServiceNodes_Filter: started state syncer === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.155Z [INFO] TestHealthConnectServiceNodes_PassingFilter: Synced node info writer.go:29: 2021-01-29T19:33:22.157Z [DEBUG] TestHealthConnectServiceNodes_PassingFilter: Node info in sync === RUN TestHealthConnectServiceNodes_PassingFilter/bc_no_query_value === RUN TestHealthConnectServiceNodes_PassingFilter/passing_true === RUN TestHealthConnectServiceNodes_PassingFilter/passing_false === RUN TestHealthConnectServiceNodes_PassingFilter/passing_bad === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.169Z [INFO] TestHealthConnectServiceNodes_PassingFilter: Requesting shutdown writer.go:29: 2021-01-29T19:33:22.171Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server: shutting down server writer.go:29: 2021-01-29T19:33:22.173Z [DEBUG] TestHealthConnectServiceNodes_PassingFilter.leader: stopping routine: routine="CA root pruning" === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.176Z [WARN] TestHealthConnectServiceNodes_Filter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:22.178Z [INFO] TestHealthConnectServiceNodes_Filter.server.raft: entering candidate state: node="Node at 127.0.0.1:29933 [Candidate]" term=2 === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.175Z [WARN] TestHealthConnectServiceNodes_PassingFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:22.175Z [DEBUG] TestHealthConnectServiceNodes_PassingFilter.leader: stopped routine: routine="CA root pruning" === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.181Z [DEBUG] TestHealthConnectServiceNodes_Filter.server.raft: votes: needed=1 === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.182Z [WARN] TestHealthConnectServiceNodes_PassingFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:22.184Z [INFO] TestHealthConnectServiceNodes_PassingFilter.server.router.manager: shutting down === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.183Z [DEBUG] TestHealthConnectServiceNodes_Filter.server.raft: vote granted: from=71d25e6d-9846-5fe1-2879-fd53f8d0bb38 term=2 tally=1 === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.185Z [INFO] TestHealthConnectServiceNodes_PassingFilter: consul server down writer.go:29: 2021-01-29T19:33:22.186Z [INFO] TestHealthConnectServiceNodes_PassingFilter: shutdown complete === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.186Z [INFO] TestHealthConnectServiceNodes_Filter.server.raft: election won: tally=1 === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.187Z [INFO] TestHealthConnectServiceNodes_PassingFilter: Stopping server: protocol=DNS address=127.0.0.1:29922 network=tcp writer.go:29: 2021-01-29T19:33:22.188Z [INFO] TestHealthConnectServiceNodes_PassingFilter: Stopping server: protocol=DNS address=127.0.0.1:29922 network=udp writer.go:29: 2021-01-29T19:33:22.188Z [INFO] TestHealthConnectServiceNodes_PassingFilter: Stopping server: protocol=HTTP address=127.0.0.1:29923 network=tcp === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.187Z [INFO] TestHealthConnectServiceNodes_Filter.server.raft: entering leader state: leader="Node at 127.0.0.1:29933 [Leader]" writer.go:29: 2021-01-29T19:33:22.190Z [INFO] TestHealthConnectServiceNodes_Filter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:22.190Z [INFO] TestHealthConnectServiceNodes_Filter.server: New leader elected: payload=Node-71d25e6d-9846-5fe1-2879-fd53f8d0bb38 writer.go:29: 2021-01-29T19:33:22.191Z [DEBUG] TestHealthConnectServiceNodes_Filter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29933 writer.go:29: 2021-01-29T19:33:22.198Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:22.203Z [INFO] TestHealthConnectServiceNodes_Filter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:22.204Z [INFO] TestHealthConnectServiceNodes_Filter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.205Z [DEBUG] TestHealthConnectServiceNodes_Filter.server: Skipping self join check for node since the cluster is too small: node=Node-71d25e6d-9846-5fe1-2879-fd53f8d0bb38 writer.go:29: 2021-01-29T19:33:22.206Z [INFO] TestHealthConnectServiceNodes_Filter.server: member joined, marking health alive: member=Node-71d25e6d-9846-5fe1-2879-fd53f8d0bb38 writer.go:29: 2021-01-29T19:33:22.438Z [INFO] TestHealthConnectServiceNodes_Filter: Requesting shutdown writer.go:29: 2021-01-29T19:33:22.445Z [INFO] TestHealthConnectServiceNodes_Filter.server: shutting down server writer.go:29: 2021-01-29T19:33:22.446Z [DEBUG] TestHealthConnectServiceNodes_Filter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.447Z [WARN] TestHealthConnectServiceNodes_Filter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:22.449Z [ERROR] TestHealthConnectServiceNodes_Filter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:22.449Z [DEBUG] TestHealthConnectServiceNodes_Filter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.458Z [WARN] TestHealthConnectServiceNodes_Filter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:22.462Z [INFO] TestHealthConnectServiceNodes_Filter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:22.463Z [INFO] TestHealthConnectServiceNodes_Filter: consul server down writer.go:29: 2021-01-29T19:33:22.466Z [INFO] TestHealthConnectServiceNodes_Filter: shutdown complete writer.go:29: 2021-01-29T19:33:22.467Z [INFO] TestHealthConnectServiceNodes_Filter: Stopping server: protocol=DNS address=127.0.0.1:29928 network=tcp writer.go:29: 2021-01-29T19:33:22.469Z [INFO] TestHealthConnectServiceNodes_Filter: Stopping server: protocol=DNS address=127.0.0.1:29928 network=udp writer.go:29: 2021-01-29T19:33:22.470Z [INFO] TestHealthConnectServiceNodes_Filter: Stopping server: protocol=HTTP address=127.0.0.1:29929 network=tcp === CONT TestHTTPServer_UnixSocket writer.go:29: 2021-01-29T19:33:22.519Z [INFO] TestHTTPServer_UnixSocket: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:22.520Z [INFO] TestHTTPServer_UnixSocket: Endpoints down --- PASS: TestHTTPServer_UnixSocket (0.92s) === CONT TestHealthConnectServiceNodes writer.go:29: 2021-01-29T19:33:22.532Z [WARN] TestHealthConnectServiceNodes: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:22.533Z [DEBUG] TestHealthConnectServiceNodes.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:22.536Z [DEBUG] TestHealthConnectServiceNodes.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:22.573Z [INFO] TestHealthConnectServiceNodes.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2690fe4e-cf55-c339-be3b-62831d922845 Address:127.0.0.1:29945}]" writer.go:29: 2021-01-29T19:33:22.575Z [INFO] TestHealthConnectServiceNodes.server.raft: entering follower state: follower="Node at 127.0.0.1:29945 [Follower]" leader= writer.go:29: 2021-01-29T19:33:22.576Z [INFO] TestHealthConnectServiceNodes.server.serf.wan: serf: EventMemberJoin: Node-2690fe4e-cf55-c339-be3b-62831d922845.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:22.581Z [INFO] TestHealthConnectServiceNodes.server.serf.lan: serf: EventMemberJoin: Node-2690fe4e-cf55-c339-be3b-62831d922845 127.0.0.1 writer.go:29: 2021-01-29T19:33:22.584Z [INFO] TestHealthConnectServiceNodes.server: Handled event for server in area: event=member-join server=Node-2690fe4e-cf55-c339-be3b-62831d922845.dc1 area=wan writer.go:29: 2021-01-29T19:33:22.585Z [INFO] TestHealthConnectServiceNodes: Started DNS server: address=127.0.0.1:29940 network=udp writer.go:29: 2021-01-29T19:33:22.586Z [INFO] TestHealthConnectServiceNodes: Started DNS server: address=127.0.0.1:29940 network=tcp writer.go:29: 2021-01-29T19:33:22.588Z [INFO] TestHealthConnectServiceNodes: Started HTTP server: address=127.0.0.1:29941 network=tcp writer.go:29: 2021-01-29T19:33:22.589Z [INFO] TestHealthConnectServiceNodes: started state syncer writer.go:29: 2021-01-29T19:33:22.585Z [INFO] TestHealthConnectServiceNodes.server: Adding LAN server: server="Node-2690fe4e-cf55-c339-be3b-62831d922845 (Addr: tcp/127.0.0.1:29945) (DC: dc1)" writer.go:29: 2021-01-29T19:33:22.621Z [WARN] TestHealthConnectServiceNodes.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:22.622Z [INFO] TestHealthConnectServiceNodes.server.raft: entering candidate state: node="Node at 127.0.0.1:29945 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:22.626Z [DEBUG] TestHealthConnectServiceNodes.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:22.627Z [DEBUG] TestHealthConnectServiceNodes.server.raft: vote granted: from=2690fe4e-cf55-c339-be3b-62831d922845 term=2 tally=1 writer.go:29: 2021-01-29T19:33:22.629Z [INFO] TestHealthConnectServiceNodes.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:22.630Z [INFO] TestHealthConnectServiceNodes.server.raft: entering leader state: leader="Node at 127.0.0.1:29945 [Leader]" writer.go:29: 2021-01-29T19:33:22.632Z [INFO] TestHealthConnectServiceNodes.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:22.633Z [INFO] TestHealthConnectServiceNodes.server: New leader elected: payload=Node-2690fe4e-cf55-c339-be3b-62831d922845 writer.go:29: 2021-01-29T19:33:22.635Z [DEBUG] TestHealthConnectServiceNodes.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29945 === CONT TestHTTPAPI_TranslateAddrHeader writer.go:29: 2021-01-29T19:33:22.637Z [INFO] TestHTTPAPI_TranslateAddrHeader: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:22.639Z [INFO] TestHTTPAPI_TranslateAddrHeader: Endpoints down --- PASS: TestHTTPAPI_TranslateAddrHeader (1.64s) === CONT TestHealthServiceNodes_WanTranslation === CONT TestHealthConnectServiceNodes writer.go:29: 2021-01-29T19:33:22.654Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:22.658Z [INFO] TestHealthConnectServiceNodes.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:22.659Z [INFO] TestHealthConnectServiceNodes.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.660Z [DEBUG] TestHealthConnectServiceNodes.server: Skipping self join check for node since the cluster is too small: node=Node-2690fe4e-cf55-c339-be3b-62831d922845 writer.go:29: 2021-01-29T19:33:22.661Z [INFO] TestHealthConnectServiceNodes.server: member joined, marking health alive: member=Node-2690fe4e-cf55-c339-be3b-62831d922845 === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:22.662Z [WARN] TestHealthServiceNodes_WanTranslation: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:22.663Z [WARN] TestHealthServiceNodes_WanTranslation: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:22.664Z [DEBUG] TestHealthServiceNodes_WanTranslation.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:22.666Z [DEBUG] TestHealthServiceNodes_WanTranslation.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:22.681Z [INFO] TestHealthServiceNodes_WanTranslation.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4c428e31-5b7d-9574-ca90-dd8899a1e0f6 Address:127.0.0.1:29939}]" writer.go:29: 2021-01-29T19:33:22.683Z [INFO] TestHealthServiceNodes_WanTranslation.server.raft: entering follower state: follower="Node at 127.0.0.1:29939 [Follower]" leader= writer.go:29: 2021-01-29T19:33:22.684Z [INFO] TestHealthServiceNodes_WanTranslation.server.serf.wan: serf: EventMemberJoin: Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:22.688Z [INFO] TestHealthServiceNodes_WanTranslation.server.serf.lan: serf: EventMemberJoin: Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6 127.0.0.1 writer.go:29: 2021-01-29T19:33:22.690Z [INFO] TestHealthServiceNodes_WanTranslation.server: Handled event for server in area: event=member-join server=Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6.dc1 area=wan writer.go:29: 2021-01-29T19:33:22.690Z [INFO] TestHealthServiceNodes_WanTranslation.server: Adding LAN server: server="Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6 (Addr: tcp/127.0.0.1:29939) (DC: dc1)" === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.690Z [INFO] TestHealthConnectServiceNodes_PassingFilter: Waiting for endpoints to shut down === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:22.690Z [INFO] TestHealthServiceNodes_WanTranslation: Started DNS server: address=127.0.0.1:29934 network=udp writer.go:29: 2021-01-29T19:33:22.692Z [INFO] TestHealthServiceNodes_WanTranslation: Started DNS server: address=127.0.0.1:29934 network=tcp === CONT TestHealthConnectServiceNodes_PassingFilter writer.go:29: 2021-01-29T19:33:22.692Z [INFO] TestHealthConnectServiceNodes_PassingFilter: Endpoints down --- PASS: TestHealthConnectServiceNodes_PassingFilter (0.93s) --- PASS: TestHealthConnectServiceNodes_PassingFilter/bc_no_query_value (0.00s) --- PASS: TestHealthConnectServiceNodes_PassingFilter/passing_true (0.00s) --- PASS: TestHealthConnectServiceNodes_PassingFilter/passing_false (0.00s) --- PASS: TestHealthConnectServiceNodes_PassingFilter/passing_bad (0.00s) === CONT TestHealthServiceNodes_CheckType === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:22.694Z [INFO] TestHealthServiceNodes_WanTranslation: Started HTTP server: address=127.0.0.1:29935 network=tcp writer.go:29: 2021-01-29T19:33:22.695Z [INFO] TestHealthServiceNodes_WanTranslation: started state syncer === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:22.705Z [WARN] TestHealthServiceNodes_CheckType: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:22.707Z [DEBUG] TestHealthServiceNodes_CheckType.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:22.708Z [DEBUG] TestHealthServiceNodes_CheckType.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:22.713Z [INFO] TestHealthServiceNodes_CheckType.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:377f4a84-8fd3-2339-71f7-9becbe501668 Address:127.0.0.1:29951}]" writer.go:29: 2021-01-29T19:33:22.716Z [INFO] TestHealthServiceNodes_CheckType.server.raft: entering follower state: follower="Node at 127.0.0.1:29951 [Follower]" leader= writer.go:29: 2021-01-29T19:33:22.718Z [INFO] TestHealthServiceNodes_CheckType.server.serf.wan: serf: EventMemberJoin: Node-377f4a84-8fd3-2339-71f7-9becbe501668.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:22.722Z [INFO] TestHealthServiceNodes_CheckType.server.serf.lan: serf: EventMemberJoin: Node-377f4a84-8fd3-2339-71f7-9becbe501668 127.0.0.1 writer.go:29: 2021-01-29T19:33:22.726Z [INFO] TestHealthServiceNodes_CheckType.server: Adding LAN server: server="Node-377f4a84-8fd3-2339-71f7-9becbe501668 (Addr: tcp/127.0.0.1:29951) (DC: dc1)" writer.go:29: 2021-01-29T19:33:22.728Z [INFO] TestHealthServiceNodes_CheckType.server: Handled event for server in area: event=member-join server=Node-377f4a84-8fd3-2339-71f7-9becbe501668.dc1 area=wan writer.go:29: 2021-01-29T19:33:22.730Z [INFO] TestHealthServiceNodes_CheckType: Started DNS server: address=127.0.0.1:29946 network=udp writer.go:29: 2021-01-29T19:33:22.732Z [INFO] TestHealthServiceNodes_CheckType: Started DNS server: address=127.0.0.1:29946 network=tcp writer.go:29: 2021-01-29T19:33:22.734Z [INFO] TestHealthServiceNodes_CheckType: Started HTTP server: address=127.0.0.1:29947 network=tcp writer.go:29: 2021-01-29T19:33:22.736Z [INFO] TestHealthServiceNodes_CheckType: started state syncer === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:22.739Z [WARN] TestHealthServiceNodes_WanTranslation.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:22.740Z [INFO] TestHealthServiceNodes_WanTranslation.server.raft: entering candidate state: node="Node at 127.0.0.1:29939 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:22.741Z [DEBUG] TestHealthServiceNodes_WanTranslation.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:33:22.745Z [DEBUG] TestHealthServiceNodes_WanTranslation.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:22.746Z [DEBUG] TestHealthServiceNodes_WanTranslation.server.raft: vote granted: from=4c428e31-5b7d-9574-ca90-dd8899a1e0f6 term=2 tally=1 writer.go:29: 2021-01-29T19:33:22.748Z [INFO] TestHealthServiceNodes_WanTranslation.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:22.749Z [INFO] TestHealthServiceNodes_WanTranslation.server.raft: entering leader state: leader="Node at 127.0.0.1:29939 [Leader]" writer.go:29: 2021-01-29T19:33:22.751Z [INFO] TestHealthServiceNodes_WanTranslation.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:22.753Z [INFO] TestHealthServiceNodes_WanTranslation.server: New leader elected: payload=Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6 writer.go:29: 2021-01-29T19:33:22.754Z [INFO] TestHealthServiceNodes_WanTranslation.server: initializing acls writer.go:29: 2021-01-29T19:33:22.759Z [INFO] TestHealthServiceNodes_WanTranslation.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:22.761Z [INFO] TestHealthServiceNodes_WanTranslation.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:22.763Z [INFO] TestHealthServiceNodes_WanTranslation.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:22.765Z [INFO] TestHealthServiceNodes_WanTranslation.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:22.766Z [INFO] TestHealthServiceNodes_WanTranslation.server.serf.lan: serf: EventMemberUpdate: Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6 writer.go:29: 2021-01-29T19:33:22.769Z [INFO] TestHealthServiceNodes_WanTranslation.server: Updating LAN server: server="Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6 (Addr: tcp/127.0.0.1:29939) (DC: dc1)" writer.go:29: 2021-01-29T19:33:22.769Z [INFO] TestHealthServiceNodes_WanTranslation.server.serf.wan: serf: EventMemberUpdate: Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6.dc1 writer.go:29: 2021-01-29T19:33:22.773Z [INFO] TestHealthServiceNodes_WanTranslation.server: Handled event for server in area: event=member-update server=Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6.dc1 area=wan writer.go:29: 2021-01-29T19:33:22.775Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:22.776Z [WARN] TestHealthServiceNodes_CheckType.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:22.777Z [INFO] TestHealthServiceNodes_CheckType.server.raft: entering candidate state: node="Node at 127.0.0.1:29951 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:22.779Z [DEBUG] TestHealthServiceNodes_CheckType.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:22.780Z [DEBUG] TestHealthServiceNodes_CheckType.server.raft: vote granted: from=377f4a84-8fd3-2339-71f7-9becbe501668 term=2 tally=1 writer.go:29: 2021-01-29T19:33:22.781Z [INFO] TestHealthServiceNodes_CheckType.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:22.782Z [INFO] TestHealthServiceNodes_CheckType.server.raft: entering leader state: leader="Node at 127.0.0.1:29951 [Leader]" writer.go:29: 2021-01-29T19:33:22.783Z [INFO] TestHealthServiceNodes_CheckType.server: cluster leadership acquired === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:22.781Z [INFO] TestHealthServiceNodes_WanTranslation.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:22.783Z [INFO] TestHealthServiceNodes_CheckType.server: New leader elected: payload=Node-377f4a84-8fd3-2339-71f7-9becbe501668 === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:22.786Z [INFO] TestHealthServiceNodes_WanTranslation.leader: started routine: routine="CA root pruning" === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:22.784Z [DEBUG] TestHealthServiceNodes_CheckType.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29951 === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:22.787Z [DEBUG] TestHealthServiceNodes_WanTranslation.server: Skipping self join check for node since the cluster is too small: node=Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6 === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:22.795Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHealthConnectServiceNodes writer.go:29: 2021-01-29T19:33:22.799Z [DEBUG] TestHealthConnectServiceNodes: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:22.799Z [INFO] TestHealthServiceNodes_WanTranslation.server: member joined, marking health alive: member=Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6 === CONT TestHealthConnectServiceNodes writer.go:29: 2021-01-29T19:33:22.802Z [INFO] TestHealthConnectServiceNodes: Synced node info writer.go:29: 2021-01-29T19:33:22.803Z [DEBUG] TestHealthConnectServiceNodes: Node info in sync === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:22.802Z [DEBUG] TestHealthServiceNodes_WanTranslation.server: Skipping self join check for node since the cluster is too small: node=Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6 === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:22.809Z [INFO] TestHealthServiceNodes_CheckType.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:22.810Z [INFO] TestHealthServiceNodes_CheckType.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.812Z [DEBUG] TestHealthServiceNodes_CheckType.server: Skipping self join check for node since the cluster is too small: node=Node-377f4a84-8fd3-2339-71f7-9becbe501668 writer.go:29: 2021-01-29T19:33:22.813Z [INFO] TestHealthServiceNodes_CheckType.server: member joined, marking health alive: member=Node-377f4a84-8fd3-2339-71f7-9becbe501668 === CONT TestHealthConnectServiceNodes writer.go:29: 2021-01-29T19:33:22.861Z [INFO] TestHealthConnectServiceNodes: Requesting shutdown writer.go:29: 2021-01-29T19:33:22.862Z [INFO] TestHealthConnectServiceNodes.server: shutting down server writer.go:29: 2021-01-29T19:33:22.862Z [DEBUG] TestHealthConnectServiceNodes.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.863Z [WARN] TestHealthConnectServiceNodes.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:22.863Z [DEBUG] TestHealthConnectServiceNodes.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.866Z [WARN] TestHealthConnectServiceNodes.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:22.870Z [INFO] TestHealthConnectServiceNodes.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:22.870Z [INFO] TestHealthConnectServiceNodes: consul server down writer.go:29: 2021-01-29T19:33:22.873Z [INFO] TestHealthConnectServiceNodes: shutdown complete writer.go:29: 2021-01-29T19:33:22.875Z [INFO] TestHealthConnectServiceNodes: Stopping server: protocol=DNS address=127.0.0.1:29940 network=tcp writer.go:29: 2021-01-29T19:33:22.876Z [INFO] TestHealthConnectServiceNodes: Stopping server: protocol=DNS address=127.0.0.1:29940 network=udp writer.go:29: 2021-01-29T19:33:22.878Z [INFO] TestHealthConnectServiceNodes: Stopping server: protocol=HTTP address=127.0.0.1:29941 network=tcp === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:22.928Z [DEBUG] TestHealthServiceNodes_CheckType: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:22.941Z [INFO] TestHealthServiceNodes_CheckType: Synced node info === CONT TestHealthConnectServiceNodes_Filter writer.go:29: 2021-01-29T19:33:22.971Z [INFO] TestHealthConnectServiceNodes_Filter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:22.973Z [INFO] TestHealthConnectServiceNodes_Filter: Endpoints down --- PASS: TestHealthConnectServiceNodes_Filter (0.88s) === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:22.997Z [WARN] TestHealthServiceNodes_DistanceSort: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:22.999Z [DEBUG] TestHealthServiceNodes_DistanceSort.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:23.008Z [DEBUG] TestHealthServiceNodes_DistanceSort.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:23.012Z [INFO] TestHealthServiceNodes_DistanceSort.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:24a1e896-6d46-b449-753d-785a9093217a Address:127.0.0.1:29957}]" writer.go:29: 2021-01-29T19:33:23.023Z [INFO] TestHealthServiceNodes_DistanceSort.server.raft: entering follower state: follower="Node at 127.0.0.1:29957 [Follower]" leader= writer.go:29: 2021-01-29T19:33:23.030Z [INFO] TestHealthServiceNodes_DistanceSort.server.serf.wan: serf: EventMemberJoin: Node-24a1e896-6d46-b449-753d-785a9093217a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:23.036Z [INFO] TestHealthServiceNodes_DistanceSort.server.serf.lan: serf: EventMemberJoin: Node-24a1e896-6d46-b449-753d-785a9093217a 127.0.0.1 writer.go:29: 2021-01-29T19:33:23.041Z [INFO] TestHealthServiceNodes_DistanceSort.server: Adding LAN server: server="Node-24a1e896-6d46-b449-753d-785a9093217a (Addr: tcp/127.0.0.1:29957) (DC: dc1)" writer.go:29: 2021-01-29T19:33:23.046Z [INFO] TestHealthServiceNodes_DistanceSort.server: Handled event for server in area: event=member-join server=Node-24a1e896-6d46-b449-753d-785a9093217a.dc1 area=wan === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.047Z [DEBUG] TestHealthServiceNodes_WanTranslation: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:23.052Z [INFO] TestHealthServiceNodes_WanTranslation: Synced node info === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.057Z [INFO] TestHealthServiceNodes_DistanceSort: Started DNS server: address=127.0.0.1:29952 network=tcp writer.go:29: 2021-01-29T19:33:23.060Z [INFO] TestHealthServiceNodes_DistanceSort: Started DNS server: address=127.0.0.1:29952 network=udp writer.go:29: 2021-01-29T19:33:23.062Z [INFO] TestHealthServiceNodes_DistanceSort: Started HTTP server: address=127.0.0.1:29953 network=tcp writer.go:29: 2021-01-29T19:33:23.064Z [INFO] TestHealthServiceNodes_DistanceSort: started state syncer writer.go:29: 2021-01-29T19:33:23.078Z [WARN] TestHealthServiceNodes_DistanceSort.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:23.079Z [INFO] TestHealthServiceNodes_DistanceSort.server.raft: entering candidate state: node="Node at 127.0.0.1:29957 [Candidate]" term=2 === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.094Z [WARN] TestHealthServiceNodes_WanTranslation: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:23.096Z [WARN] TestHealthServiceNodes_WanTranslation: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:23.097Z [DEBUG] TestHealthServiceNodes_WanTranslation.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:23.101Z [DEBUG] TestHealthServiceNodes_WanTranslation.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.109Z [DEBUG] TestHealthServiceNodes_DistanceSort.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:23.110Z [DEBUG] TestHealthServiceNodes_DistanceSort.server.raft: vote granted: from=24a1e896-6d46-b449-753d-785a9093217a term=2 tally=1 writer.go:29: 2021-01-29T19:33:23.111Z [INFO] TestHealthServiceNodes_DistanceSort.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:23.112Z [INFO] TestHealthServiceNodes_DistanceSort.server.raft: entering leader state: leader="Node at 127.0.0.1:29957 [Leader]" writer.go:29: 2021-01-29T19:33:23.113Z [INFO] TestHealthServiceNodes_DistanceSort.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:23.115Z [INFO] TestHealthServiceNodes_DistanceSort.server: New leader elected: payload=Node-24a1e896-6d46-b449-753d-785a9093217a === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.112Z [INFO] TestHealthServiceNodes_WanTranslation.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b80e3681-6b1f-5b95-2b83-9eac52fd4cb9 Address:127.0.0.1:29969}]" writer.go:29: 2021-01-29T19:33:23.118Z [INFO] TestHealthServiceNodes_WanTranslation.server.raft: entering follower state: follower="Node at 127.0.0.1:29969 [Follower]" leader= writer.go:29: 2021-01-29T19:33:23.125Z [INFO] TestHealthServiceNodes_WanTranslation.server.serf.wan: serf: EventMemberJoin: Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9.dc2 127.0.0.1 writer.go:29: 2021-01-29T19:33:23.132Z [INFO] TestHealthServiceNodes_WanTranslation.server.serf.lan: serf: EventMemberJoin: Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9 127.0.0.1 writer.go:29: 2021-01-29T19:33:23.136Z [INFO] TestHealthServiceNodes_WanTranslation.server: Handled event for server in area: event=member-join server=Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9.dc2 area=wan writer.go:29: 2021-01-29T19:33:23.136Z [INFO] TestHealthServiceNodes_WanTranslation.server: Adding LAN server: server="Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9 (Addr: tcp/127.0.0.1:29969) (DC: dc2)" === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.140Z [DEBUG] TestHealthServiceNodes_DistanceSort.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29957 === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.142Z [INFO] TestHealthServiceNodes_WanTranslation: Started DNS server: address=127.0.0.1:29964 network=tcp === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.147Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.148Z [INFO] TestHealthServiceNodes_WanTranslation: Started DNS server: address=127.0.0.1:29964 network=udp writer.go:29: 2021-01-29T19:33:23.152Z [INFO] TestHealthServiceNodes_WanTranslation: Started HTTP server: address=127.0.0.1:29965 network=tcp writer.go:29: 2021-01-29T19:33:23.154Z [INFO] TestHealthServiceNodes_WanTranslation: started state syncer === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.173Z [INFO] TestHealthServiceNodes_DistanceSort.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:23.174Z [INFO] TestHealthServiceNodes_DistanceSort.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:23.175Z [DEBUG] TestHealthServiceNodes_DistanceSort.server: Skipping self join check for node since the cluster is too small: node=Node-24a1e896-6d46-b449-753d-785a9093217a writer.go:29: 2021-01-29T19:33:23.176Z [INFO] TestHealthServiceNodes_DistanceSort.server: member joined, marking health alive: member=Node-24a1e896-6d46-b449-753d-785a9093217a === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.189Z [DEBUG] TestHealthServiceNodes_WanTranslation.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:23.191Z [INFO] TestHealthServiceNodes_CheckType: Requesting shutdown === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.191Z [WARN] TestHealthServiceNodes_WanTranslation.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:23.192Z [INFO] TestHealthServiceNodes_CheckType.server: shutting down server writer.go:29: 2021-01-29T19:33:23.193Z [DEBUG] TestHealthServiceNodes_CheckType.leader: stopping routine: routine="CA root pruning" === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.192Z [INFO] TestHealthServiceNodes_WanTranslation.server.raft: entering candidate state: node="Node at 127.0.0.1:29969 [Candidate]" term=2 === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:23.194Z [WARN] TestHealthServiceNodes_CheckType.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:23.195Z [DEBUG] TestHealthServiceNodes_CheckType.leader: stopped routine: routine="CA root pruning" === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.195Z [DEBUG] TestHealthServiceNodes_WanTranslation.server.raft: votes: needed=1 === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:23.197Z [WARN] TestHealthServiceNodes_CheckType.server.serf.wan: serf: Shutdown without a Leave === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.196Z [DEBUG] TestHealthServiceNodes_WanTranslation.server.raft: vote granted: from=b80e3681-6b1f-5b95-2b83-9eac52fd4cb9 term=2 tally=1 === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:23.210Z [INFO] TestHealthServiceNodes_CheckType.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:23.222Z [INFO] TestHealthServiceNodes_CheckType: consul server down writer.go:29: 2021-01-29T19:33:23.224Z [INFO] TestHealthServiceNodes_CheckType: shutdown complete writer.go:29: 2021-01-29T19:33:23.226Z [INFO] TestHealthServiceNodes_CheckType: Stopping server: protocol=DNS address=127.0.0.1:29946 network=tcp === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.207Z [INFO] TestHealthServiceNodes_WanTranslation.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:23.230Z [INFO] TestHealthServiceNodes_WanTranslation.server.raft: entering leader state: leader="Node at 127.0.0.1:29969 [Leader]" writer.go:29: 2021-01-29T19:33:23.231Z [INFO] TestHealthServiceNodes_WanTranslation.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:23.232Z [INFO] TestHealthServiceNodes_WanTranslation.server: New leader elected: payload=Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9 === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:23.228Z [INFO] TestHealthServiceNodes_CheckType: Stopping server: protocol=DNS address=127.0.0.1:29946 network=udp === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.234Z [INFO] TestHealthServiceNodes_WanTranslation.server: initializing acls === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:23.234Z [INFO] TestHealthServiceNodes_CheckType: Stopping server: protocol=HTTP address=127.0.0.1:29947 network=tcp === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.238Z [INFO] TestHealthServiceNodes_WanTranslation.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:23.240Z [INFO] TestHealthServiceNodes_WanTranslation.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:23.242Z [INFO] TestHealthServiceNodes_WanTranslation.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:23.243Z [INFO] TestHealthServiceNodes_WanTranslation.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:23.245Z [INFO] TestHealthServiceNodes_WanTranslation.server.serf.lan: serf: EventMemberUpdate: Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9 writer.go:29: 2021-01-29T19:33:23.247Z [INFO] TestHealthServiceNodes_WanTranslation.server.serf.wan: serf: EventMemberUpdate: Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9.dc2 writer.go:29: 2021-01-29T19:33:23.247Z [INFO] TestHealthServiceNodes_WanTranslation.server: Updating LAN server: server="Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9 (Addr: tcp/127.0.0.1:29969) (DC: dc2)" writer.go:29: 2021-01-29T19:33:23.249Z [INFO] TestHealthServiceNodes_WanTranslation.server: Handled event for server in area: event=member-update server=Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9.dc2 area=wan writer.go:29: 2021-01-29T19:33:23.251Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:23.257Z [INFO] TestHealthServiceNodes_WanTranslation.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:23.259Z [INFO] TestHealthServiceNodes_WanTranslation.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:23.261Z [DEBUG] TestHealthServiceNodes_WanTranslation.server: Skipping self join check for node since the cluster is too small: node=Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9 writer.go:29: 2021-01-29T19:33:23.262Z [INFO] TestHealthServiceNodes_WanTranslation.server: member joined, marking health alive: member=Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9 writer.go:29: 2021-01-29T19:33:23.265Z [DEBUG] TestHealthServiceNodes_WanTranslation.server: Skipping self join check for node since the cluster is too small: node=Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9 writer.go:29: 2021-01-29T19:33:23.277Z [DEBUG] TestHealthServiceNodes_WanTranslation: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:23.289Z [INFO] TestHealthServiceNodes_WanTranslation: Synced node info writer.go:29: 2021-01-29T19:33:23.370Z [INFO] TestHealthServiceNodes_WanTranslation: (WAN) joining: wan_addresses=[127.0.0.1:29938] writer.go:29: 2021-01-29T19:33:23.374Z [DEBUG] TestHealthServiceNodes_WanTranslation.server.memberlist.wan: memberlist: Stream connection from=127.0.0.1:37992 writer.go:29: 2021-01-29T19:33:23.375Z [DEBUG] TestHealthServiceNodes_WanTranslation.server.memberlist.wan: memberlist: Initiating push/pull sync with: 127.0.0.1:29938 === CONT TestHealthConnectServiceNodes writer.go:29: 2021-01-29T19:33:23.380Z [INFO] TestHealthConnectServiceNodes: Waiting for endpoints to shut down === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.381Z [INFO] TestHealthServiceNodes_WanTranslation.server.serf.wan: serf: EventMemberJoin: Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9.dc2 127.0.0.1 === CONT TestHealthConnectServiceNodes writer.go:29: 2021-01-29T19:33:23.382Z [INFO] TestHealthConnectServiceNodes: Endpoints down --- PASS: TestHealthConnectServiceNodes (0.86s) === CONT TestHealthServiceNodes_NodeMetaFilter === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.381Z [INFO] TestHealthServiceNodes_WanTranslation.server.serf.wan: serf: EventMemberJoin: Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:23.383Z [INFO] TestHealthServiceNodes_WanTranslation.server: Handled event for server in area: event=member-join server=Node-b80e3681-6b1f-5b95-2b83-9eac52fd4cb9.dc2 area=wan writer.go:29: 2021-01-29T19:33:23.385Z [INFO] TestHealthServiceNodes_WanTranslation: (WAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:33:23.385Z [INFO] TestHealthServiceNodes_WanTranslation.server: Handled event for server in area: event=member-join server=Node-4c428e31-5b7d-9574-ca90-dd8899a1e0f6.dc1 area=wan writer.go:29: 2021-01-29T19:33:23.391Z [DEBUG] TestHealthServiceNodes_WanTranslation.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.400Z [WARN] TestHealthServiceNodes_NodeMetaFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:23.402Z [DEBUG] TestHealthServiceNodes_NodeMetaFilter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:23.407Z [DEBUG] TestHealthServiceNodes_NodeMetaFilter.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:23.413Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:499f92c8-c1f2-3490-9687-6da90b236780 Address:127.0.0.1:29963}]" writer.go:29: 2021-01-29T19:33:23.416Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:29963 [Follower]" leader= writer.go:29: 2021-01-29T19:33:23.417Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server.serf.wan: serf: EventMemberJoin: Node-499f92c8-c1f2-3490-9687-6da90b236780.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:23.423Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server.serf.lan: serf: EventMemberJoin: Node-499f92c8-c1f2-3490-9687-6da90b236780 127.0.0.1 writer.go:29: 2021-01-29T19:33:23.427Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server: Handled event for server in area: event=member-join server=Node-499f92c8-c1f2-3490-9687-6da90b236780.dc1 area=wan writer.go:29: 2021-01-29T19:33:23.428Z [INFO] TestHealthServiceNodes_NodeMetaFilter: Started DNS server: address=127.0.0.1:29958 network=udp writer.go:29: 2021-01-29T19:33:23.431Z [INFO] TestHealthServiceNodes_NodeMetaFilter: Started DNS server: address=127.0.0.1:29958 network=tcp === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.433Z [INFO] TestHealthServiceNodes_WanTranslation: Requesting shutdown === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.435Z [INFO] TestHealthServiceNodes_NodeMetaFilter: Started HTTP server: address=127.0.0.1:29959 network=tcp === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.435Z [INFO] TestHealthServiceNodes_WanTranslation.server: shutting down server writer.go:29: 2021-01-29T19:33:23.437Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:23.438Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopping routine: routine="CA root pruning" === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.429Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server: Adding LAN server: server="Node-499f92c8-c1f2-3490-9687-6da90b236780 (Addr: tcp/127.0.0.1:29963) (DC: dc1)" === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.440Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.437Z [INFO] TestHealthServiceNodes_NodeMetaFilter: started state syncer === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.439Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:23.441Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:23.442Z [WARN] TestHealthServiceNodes_WanTranslation.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:23.442Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:23.448Z [WARN] TestHealthServiceNodes_WanTranslation.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:23.451Z [INFO] TestHealthServiceNodes_WanTranslation.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:23.453Z [INFO] TestHealthServiceNodes_WanTranslation.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:23.455Z [INFO] TestHealthServiceNodes_WanTranslation: consul server down writer.go:29: 2021-01-29T19:33:23.457Z [INFO] TestHealthServiceNodes_WanTranslation: shutdown complete writer.go:29: 2021-01-29T19:33:23.458Z [INFO] TestHealthServiceNodes_WanTranslation: Stopping server: protocol=DNS address=127.0.0.1:29964 network=tcp writer.go:29: 2021-01-29T19:33:23.460Z [INFO] TestHealthServiceNodes_WanTranslation: Stopping server: protocol=DNS address=127.0.0.1:29964 network=udp writer.go:29: 2021-01-29T19:33:23.462Z [INFO] TestHealthServiceNodes_WanTranslation: Stopping server: protocol=HTTP address=127.0.0.1:29965 network=tcp === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.464Z [DEBUG] TestHealthServiceNodes_DistanceSort: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:23.467Z [INFO] TestHealthServiceNodes_DistanceSort: Synced node info writer.go:29: 2021-01-29T19:33:23.469Z [DEBUG] TestHealthServiceNodes_DistanceSort: Node info in sync === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.482Z [WARN] TestHealthServiceNodes_NodeMetaFilter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:23.484Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:29963 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:23.489Z [DEBUG] TestHealthServiceNodes_NodeMetaFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:23.490Z [DEBUG] TestHealthServiceNodes_NodeMetaFilter.server.raft: vote granted: from=499f92c8-c1f2-3490-9687-6da90b236780 term=2 tally=1 writer.go:29: 2021-01-29T19:33:23.498Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:23.500Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:29963 [Leader]" writer.go:29: 2021-01-29T19:33:23.501Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:23.502Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server: New leader elected: payload=Node-499f92c8-c1f2-3490-9687-6da90b236780 writer.go:29: 2021-01-29T19:33:23.506Z [DEBUG] TestHealthServiceNodes_NodeMetaFilter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29963 writer.go:29: 2021-01-29T19:33:23.516Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:23.522Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:23.523Z [INFO] TestHealthServiceNodes_NodeMetaFilter.leader: started routine: routine="CA root pruning" === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.525Z [INFO] TestHealthServiceNodes_DistanceSort: Requesting shutdown writer.go:29: 2021-01-29T19:33:23.526Z [INFO] TestHealthServiceNodes_DistanceSort.server: shutting down server === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.525Z [DEBUG] TestHealthServiceNodes_NodeMetaFilter.server: Skipping self join check for node since the cluster is too small: node=Node-499f92c8-c1f2-3490-9687-6da90b236780 === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.526Z [DEBUG] TestHealthServiceNodes_DistanceSort.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:23.527Z [WARN] TestHealthServiceNodes_DistanceSort.server.serf.lan: serf: Shutdown without a Leave === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.527Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server: member joined, marking health alive: member=Node-499f92c8-c1f2-3490-9687-6da90b236780 === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.528Z [DEBUG] TestHealthServiceNodes_DistanceSort.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:23.530Z [WARN] TestHealthServiceNodes_DistanceSort.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:23.549Z [INFO] TestHealthServiceNodes_DistanceSort.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:23.550Z [INFO] TestHealthServiceNodes_DistanceSort: consul server down === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.551Z [INFO] TestHealthServiceNodes_NodeMetaFilter: Requesting shutdown === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.552Z [INFO] TestHealthServiceNodes_DistanceSort: shutdown complete === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.553Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server: shutting down server === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.554Z [INFO] TestHealthServiceNodes_DistanceSort: Stopping server: protocol=DNS address=127.0.0.1:29952 network=tcp === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.555Z [DEBUG] TestHealthServiceNodes_NodeMetaFilter.leader: stopping routine: routine="CA root pruning" === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:23.556Z [INFO] TestHealthServiceNodes_DistanceSort: Stopping server: protocol=DNS address=127.0.0.1:29952 network=udp writer.go:29: 2021-01-29T19:33:23.557Z [INFO] TestHealthServiceNodes_DistanceSort: Stopping server: protocol=HTTP address=127.0.0.1:29953 network=tcp === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.557Z [WARN] TestHealthServiceNodes_NodeMetaFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:23.560Z [DEBUG] TestHealthServiceNodes_NodeMetaFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:23.560Z [ERROR] TestHealthServiceNodes_NodeMetaFilter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:23.561Z [WARN] TestHealthServiceNodes_NodeMetaFilter.server.serf.wan: serf: Shutdown without a Leave === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.568Z [DEBUG] TestHealthServiceNodes_WanTranslation: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.570Z [INFO] TestHealthServiceNodes_NodeMetaFilter.server.router.manager: shutting down === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.570Z [DEBUG] TestHealthServiceNodes_WanTranslation: Node info in sync === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.571Z [INFO] TestHealthServiceNodes_NodeMetaFilter: consul server down === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.572Z [DEBUG] TestHealthServiceNodes_WanTranslation: Node info in sync === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:23.573Z [INFO] TestHealthServiceNodes_NodeMetaFilter: shutdown complete writer.go:29: 2021-01-29T19:33:23.574Z [INFO] TestHealthServiceNodes_NodeMetaFilter: Stopping server: protocol=DNS address=127.0.0.1:29958 network=tcp writer.go:29: 2021-01-29T19:33:23.575Z [INFO] TestHealthServiceNodes_NodeMetaFilter: Stopping server: protocol=DNS address=127.0.0.1:29958 network=udp writer.go:29: 2021-01-29T19:33:23.577Z [INFO] TestHealthServiceNodes_NodeMetaFilter: Stopping server: protocol=HTTP address=127.0.0.1:29959 network=tcp === CONT TestHealthServiceNodes_CheckType writer.go:29: 2021-01-29T19:33:23.736Z [INFO] TestHealthServiceNodes_CheckType: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:23.737Z [INFO] TestHealthServiceNodes_CheckType: Endpoints down --- PASS: TestHealthServiceNodes_CheckType (1.05s) === CONT TestHealthServiceNodes writer.go:29: 2021-01-29T19:33:23.817Z [WARN] TestHealthServiceNodes: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:23.819Z [DEBUG] TestHealthServiceNodes.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:23.825Z [DEBUG] TestHealthServiceNodes.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:23.856Z [INFO] TestHealthServiceNodes.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:461f800e-2769-450b-6724-e54e2622e5c8 Address:127.0.0.1:29975}]" writer.go:29: 2021-01-29T19:33:23.866Z [INFO] TestHealthServiceNodes.server.raft: entering follower state: follower="Node at 127.0.0.1:29975 [Follower]" leader= writer.go:29: 2021-01-29T19:33:23.871Z [INFO] TestHealthServiceNodes.server.serf.wan: serf: EventMemberJoin: Node-461f800e-2769-450b-6724-e54e2622e5c8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:23.906Z [INFO] TestHealthServiceNodes.server.serf.lan: serf: EventMemberJoin: Node-461f800e-2769-450b-6724-e54e2622e5c8 127.0.0.1 writer.go:29: 2021-01-29T19:33:23.910Z [INFO] TestHealthServiceNodes: Started DNS server: address=127.0.0.1:29970 network=udp writer.go:29: 2021-01-29T19:33:23.913Z [INFO] TestHealthServiceNodes.server: Adding LAN server: server="Node-461f800e-2769-450b-6724-e54e2622e5c8 (Addr: tcp/127.0.0.1:29975) (DC: dc1)" writer.go:29: 2021-01-29T19:33:23.915Z [INFO] TestHealthServiceNodes.server: Handled event for server in area: event=member-join server=Node-461f800e-2769-450b-6724-e54e2622e5c8.dc1 area=wan writer.go:29: 2021-01-29T19:33:23.920Z [INFO] TestHealthServiceNodes: Started DNS server: address=127.0.0.1:29970 network=tcp writer.go:29: 2021-01-29T19:33:23.925Z [INFO] TestHealthServiceNodes: Started HTTP server: address=127.0.0.1:29971 network=tcp writer.go:29: 2021-01-29T19:33:23.927Z [INFO] TestHealthServiceNodes: started state syncer writer.go:29: 2021-01-29T19:33:23.930Z [WARN] TestHealthServiceNodes.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:23.932Z [INFO] TestHealthServiceNodes.server.raft: entering candidate state: node="Node at 127.0.0.1:29975 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:23.963Z [DEBUG] TestHealthServiceNodes.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:23.964Z [DEBUG] TestHealthServiceNodes.server.raft: vote granted: from=461f800e-2769-450b-6724-e54e2622e5c8 term=2 tally=1 writer.go:29: 2021-01-29T19:33:23.966Z [INFO] TestHealthServiceNodes.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:23.967Z [INFO] TestHealthServiceNodes.server.raft: entering leader state: leader="Node at 127.0.0.1:29975 [Leader]" writer.go:29: 2021-01-29T19:33:23.976Z [INFO] TestHealthServiceNodes.server: cluster leadership acquired === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.971Z [INFO] TestHealthServiceNodes_WanTranslation: Waiting for endpoints to shut down === CONT TestHealthServiceNodes writer.go:29: 2021-01-29T19:33:23.979Z [INFO] TestHealthServiceNodes.server: New leader elected: payload=Node-461f800e-2769-450b-6724-e54e2622e5c8 === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:23.994Z [INFO] TestHealthServiceNodes_WanTranslation: Endpoints down writer.go:29: 2021-01-29T19:33:24.002Z [INFO] TestHealthServiceNodes_WanTranslation: Requesting shutdown === CONT TestHealthServiceNodes writer.go:29: 2021-01-29T19:33:23.980Z [DEBUG] TestHealthServiceNodes.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29975 === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:24.004Z [INFO] TestHealthServiceNodes_WanTranslation.server: shutting down server writer.go:29: 2021-01-29T19:33:24.005Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:24.007Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:24.009Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.014Z [WARN] TestHealthServiceNodes_WanTranslation.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.010Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:24.010Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:24.020Z [DEBUG] TestHealthServiceNodes_WanTranslation.leader: stopped routine: routine="CA root pruning" === CONT TestHealthServiceNodes writer.go:29: 2021-01-29T19:33:24.027Z [INFO] TestHealthServiceNodes: Synced node info === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:24.029Z [WARN] TestHealthServiceNodes_WanTranslation.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.035Z [INFO] TestHealthServiceNodes_WanTranslation.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:24.036Z [INFO] TestHealthServiceNodes_WanTranslation.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:24.037Z [INFO] TestHealthServiceNodes_WanTranslation: consul server down writer.go:29: 2021-01-29T19:33:24.038Z [INFO] TestHealthServiceNodes_WanTranslation: shutdown complete writer.go:29: 2021-01-29T19:33:24.039Z [INFO] TestHealthServiceNodes_WanTranslation: Stopping server: protocol=DNS address=127.0.0.1:29934 network=tcp writer.go:29: 2021-01-29T19:33:24.040Z [INFO] TestHealthServiceNodes_WanTranslation: Stopping server: protocol=DNS address=127.0.0.1:29934 network=udp writer.go:29: 2021-01-29T19:33:24.041Z [INFO] TestHealthServiceNodes_WanTranslation: Stopping server: protocol=HTTP address=127.0.0.1:29935 network=tcp === CONT TestHealthServiceNodes writer.go:29: 2021-01-29T19:33:24.044Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:24.076Z [INFO] TestHealthServiceNodes.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestHealthServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:24.079Z [INFO] TestHealthServiceNodes_NodeMetaFilter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:24.081Z [INFO] TestHealthServiceNodes_NodeMetaFilter: Endpoints down --- PASS: TestHealthServiceNodes_NodeMetaFilter (0.70s) === CONT TestHealthServiceChecks_DistanceSort === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:24.066Z [INFO] TestHealthServiceNodes_DistanceSort: Waiting for endpoints to shut down === CONT TestHealthServiceNodes writer.go:29: 2021-01-29T19:33:24.079Z [INFO] TestHealthServiceNodes.leader: started routine: routine="CA root pruning" === CONT TestHealthServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:24.084Z [INFO] TestHealthServiceNodes_DistanceSort: Endpoints down --- PASS: TestHealthServiceNodes_DistanceSort (1.12s) === CONT TestHealthServiceChecks_Filtering === CONT TestHealthServiceNodes writer.go:29: 2021-01-29T19:33:24.089Z [DEBUG] TestHealthServiceNodes.server: Skipping self join check for node since the cluster is too small: node=Node-461f800e-2769-450b-6724-e54e2622e5c8 writer.go:29: 2021-01-29T19:33:24.095Z [INFO] TestHealthServiceNodes.server: member joined, marking health alive: member=Node-461f800e-2769-450b-6724-e54e2622e5c8 === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.142Z [WARN] TestHealthServiceChecks_Filtering: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:24.151Z [DEBUG] TestHealthServiceChecks_Filtering.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:24.154Z [DEBUG] TestHealthServiceChecks_Filtering.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.159Z [WARN] TestHealthServiceChecks_DistanceSort: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:24.161Z [DEBUG] TestHealthServiceChecks_DistanceSort.tlsutil: Update: version=1 === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.167Z [INFO] TestHealthServiceChecks_Filtering.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9ccc3a15-1bce-93c6-6121-e4aee8c5e67c Address:127.0.0.1:29987}]" writer.go:29: 2021-01-29T19:33:24.170Z [INFO] TestHealthServiceChecks_Filtering.server.raft: entering follower state: follower="Node at 127.0.0.1:29987 [Follower]" leader= === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.164Z [DEBUG] TestHealthServiceChecks_DistanceSort.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.171Z [INFO] TestHealthServiceChecks_Filtering.server.serf.wan: serf: EventMemberJoin: Node-9ccc3a15-1bce-93c6-6121-e4aee8c5e67c.dc1 127.0.0.1 === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.179Z [INFO] TestHealthServiceChecks_DistanceSort.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:22fb8ad6-79ab-cf58-447b-da80916dd2f0 Address:127.0.0.1:29981}]" === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.182Z [INFO] TestHealthServiceChecks_Filtering.server.serf.lan: serf: EventMemberJoin: Node-9ccc3a15-1bce-93c6-6121-e4aee8c5e67c 127.0.0.1 === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.182Z [INFO] TestHealthServiceChecks_DistanceSort.server.raft: entering follower state: follower="Node at 127.0.0.1:29981 [Follower]" leader= === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.186Z [INFO] TestHealthServiceChecks_Filtering.server: Adding LAN server: server="Node-9ccc3a15-1bce-93c6-6121-e4aee8c5e67c (Addr: tcp/127.0.0.1:29987) (DC: dc1)" writer.go:29: 2021-01-29T19:33:24.187Z [INFO] TestHealthServiceChecks_Filtering.server: Handled event for server in area: event=member-join server=Node-9ccc3a15-1bce-93c6-6121-e4aee8c5e67c.dc1 area=wan === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.183Z [INFO] TestHealthServiceChecks_DistanceSort.server.serf.wan: serf: EventMemberJoin: Node-22fb8ad6-79ab-cf58-447b-da80916dd2f0.dc1 127.0.0.1 === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.189Z [INFO] TestHealthServiceChecks_Filtering: Started DNS server: address=127.0.0.1:29982 network=udp writer.go:29: 2021-01-29T19:33:24.191Z [INFO] TestHealthServiceChecks_Filtering: Started DNS server: address=127.0.0.1:29982 network=tcp writer.go:29: 2021-01-29T19:33:24.193Z [INFO] TestHealthServiceChecks_Filtering: Started HTTP server: address=127.0.0.1:29983 network=tcp writer.go:29: 2021-01-29T19:33:24.194Z [INFO] TestHealthServiceChecks_Filtering: started state syncer === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.190Z [INFO] TestHealthServiceChecks_DistanceSort.server.serf.lan: serf: EventMemberJoin: Node-22fb8ad6-79ab-cf58-447b-da80916dd2f0 127.0.0.1 writer.go:29: 2021-01-29T19:33:24.198Z [INFO] TestHealthServiceChecks_DistanceSort.server: Handled event for server in area: event=member-join server=Node-22fb8ad6-79ab-cf58-447b-da80916dd2f0.dc1 area=wan writer.go:29: 2021-01-29T19:33:24.199Z [INFO] TestHealthServiceChecks_DistanceSort.server: Adding LAN server: server="Node-22fb8ad6-79ab-cf58-447b-da80916dd2f0 (Addr: tcp/127.0.0.1:29981) (DC: dc1)" writer.go:29: 2021-01-29T19:33:24.201Z [INFO] TestHealthServiceChecks_DistanceSort: Started DNS server: address=127.0.0.1:29976 network=tcp writer.go:29: 2021-01-29T19:33:24.202Z [INFO] TestHealthServiceChecks_DistanceSort: Started DNS server: address=127.0.0.1:29976 network=udp writer.go:29: 2021-01-29T19:33:24.205Z [INFO] TestHealthServiceChecks_DistanceSort: Started HTTP server: address=127.0.0.1:29977 network=tcp writer.go:29: 2021-01-29T19:33:24.207Z [INFO] TestHealthServiceChecks_DistanceSort: started state syncer === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.217Z [WARN] TestHealthServiceChecks_Filtering.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:24.219Z [INFO] TestHealthServiceChecks_Filtering.server.raft: entering candidate state: node="Node at 127.0.0.1:29987 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:24.222Z [DEBUG] TestHealthServiceChecks_Filtering.server.raft: votes: needed=1 === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.227Z [WARN] TestHealthServiceChecks_DistanceSort.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:24.229Z [INFO] TestHealthServiceChecks_DistanceSort.server.raft: entering candidate state: node="Node at 127.0.0.1:29981 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:24.232Z [DEBUG] TestHealthServiceChecks_DistanceSort.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:24.233Z [DEBUG] TestHealthServiceChecks_DistanceSort.server.raft: vote granted: from=22fb8ad6-79ab-cf58-447b-da80916dd2f0 term=2 tally=1 writer.go:29: 2021-01-29T19:33:24.234Z [INFO] TestHealthServiceChecks_DistanceSort.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:24.236Z [INFO] TestHealthServiceChecks_DistanceSort.server.raft: entering leader state: leader="Node at 127.0.0.1:29981 [Leader]" writer.go:29: 2021-01-29T19:33:24.239Z [INFO] TestHealthServiceChecks_DistanceSort.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:24.240Z [INFO] TestHealthServiceChecks_DistanceSort.server: New leader elected: payload=Node-22fb8ad6-79ab-cf58-447b-da80916dd2f0 === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.224Z [DEBUG] TestHealthServiceChecks_Filtering.server.raft: vote granted: from=9ccc3a15-1bce-93c6-6121-e4aee8c5e67c term=2 tally=1 writer.go:29: 2021-01-29T19:33:24.242Z [INFO] TestHealthServiceChecks_Filtering.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:24.248Z [INFO] TestHealthServiceChecks_Filtering.server.raft: entering leader state: leader="Node at 127.0.0.1:29987 [Leader]" writer.go:29: 2021-01-29T19:33:24.253Z [INFO] TestHealthServiceChecks_Filtering.server: cluster leadership acquired === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.253Z [DEBUG] TestHealthServiceChecks_DistanceSort.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29981 === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.254Z [INFO] TestHealthServiceChecks_Filtering.server: New leader elected: payload=Node-9ccc3a15-1bce-93c6-6121-e4aee8c5e67c === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.264Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.260Z [DEBUG] TestHealthServiceChecks_Filtering.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29987 writer.go:29: 2021-01-29T19:33:24.272Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.282Z [INFO] TestHealthServiceChecks_DistanceSort.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.279Z [INFO] TestHealthServiceChecks_Filtering.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.283Z [INFO] TestHealthServiceChecks_DistanceSort.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.285Z [DEBUG] TestHealthServiceChecks_DistanceSort.server: Skipping self join check for node since the cluster is too small: node=Node-22fb8ad6-79ab-cf58-447b-da80916dd2f0 === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.284Z [INFO] TestHealthServiceChecks_Filtering.leader: started routine: routine="CA root pruning" === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.290Z [INFO] TestHealthServiceChecks_DistanceSort.server: member joined, marking health alive: member=Node-22fb8ad6-79ab-cf58-447b-da80916dd2f0 === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.290Z [DEBUG] TestHealthServiceChecks_Filtering.server: Skipping self join check for node since the cluster is too small: node=Node-9ccc3a15-1bce-93c6-6121-e4aee8c5e67c writer.go:29: 2021-01-29T19:33:24.294Z [INFO] TestHealthServiceChecks_Filtering.server: member joined, marking health alive: member=Node-9ccc3a15-1bce-93c6-6121-e4aee8c5e67c === CONT TestHealthServiceNodes writer.go:29: 2021-01-29T19:33:24.379Z [INFO] TestHealthServiceNodes: Requesting shutdown writer.go:29: 2021-01-29T19:33:24.381Z [INFO] TestHealthServiceNodes.server: shutting down server writer.go:29: 2021-01-29T19:33:24.382Z [DEBUG] TestHealthServiceNodes.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.384Z [WARN] TestHealthServiceNodes.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.386Z [DEBUG] TestHealthServiceNodes.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.400Z [WARN] TestHealthServiceNodes.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.402Z [INFO] TestHealthServiceNodes.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:24.403Z [INFO] TestHealthServiceNodes: consul server down writer.go:29: 2021-01-29T19:33:24.406Z [INFO] TestHealthServiceNodes: shutdown complete writer.go:29: 2021-01-29T19:33:24.408Z [INFO] TestHealthServiceNodes: Stopping server: protocol=DNS address=127.0.0.1:29970 network=tcp writer.go:29: 2021-01-29T19:33:24.409Z [INFO] TestHealthServiceNodes: Stopping server: protocol=DNS address=127.0.0.1:29970 network=udp writer.go:29: 2021-01-29T19:33:24.411Z [INFO] TestHealthServiceNodes: Stopping server: protocol=HTTP address=127.0.0.1:29971 network=tcp === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.459Z [INFO] TestHealthServiceChecks_Filtering: Requesting shutdown writer.go:29: 2021-01-29T19:33:24.460Z [INFO] TestHealthServiceChecks_Filtering.server: shutting down server writer.go:29: 2021-01-29T19:33:24.461Z [DEBUG] TestHealthServiceChecks_Filtering.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.464Z [WARN] TestHealthServiceChecks_Filtering.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.461Z [ERROR] TestHealthServiceChecks_Filtering.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:24.464Z [DEBUG] TestHealthServiceChecks_Filtering.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.479Z [WARN] TestHealthServiceChecks_Filtering.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.487Z [INFO] TestHealthServiceChecks_Filtering.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:24.487Z [INFO] TestHealthServiceChecks_Filtering: consul server down writer.go:29: 2021-01-29T19:33:24.489Z [INFO] TestHealthServiceChecks_Filtering: shutdown complete writer.go:29: 2021-01-29T19:33:24.490Z [INFO] TestHealthServiceChecks_Filtering: Stopping server: protocol=DNS address=127.0.0.1:29982 network=tcp writer.go:29: 2021-01-29T19:33:24.491Z [INFO] TestHealthServiceChecks_Filtering: Stopping server: protocol=DNS address=127.0.0.1:29982 network=udp writer.go:29: 2021-01-29T19:33:24.494Z [INFO] TestHealthServiceChecks_Filtering: Stopping server: protocol=HTTP address=127.0.0.1:29983 network=tcp === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.541Z [DEBUG] TestHealthServiceChecks_DistanceSort: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestHealthServiceNodes_WanTranslation writer.go:29: 2021-01-29T19:33:24.545Z [INFO] TestHealthServiceNodes_WanTranslation: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:24.546Z [INFO] TestHealthServiceNodes_WanTranslation: Endpoints down --- PASS: TestHealthServiceNodes_WanTranslation (1.91s) === CONT TestHealthServiceChecks_NodeMetaFilter writer.go:29: 2021-01-29T19:33:24.554Z [WARN] TestHealthServiceChecks_NodeMetaFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:24.555Z [DEBUG] TestHealthServiceChecks_NodeMetaFilter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:24.557Z [DEBUG] TestHealthServiceChecks_NodeMetaFilter.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.560Z [INFO] TestHealthServiceChecks_DistanceSort: Synced node info writer.go:29: 2021-01-29T19:33:24.561Z [DEBUG] TestHealthServiceChecks_DistanceSort: Node info in sync === CONT TestHealthServiceChecks_NodeMetaFilter writer.go:29: 2021-01-29T19:33:24.562Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e53badeb-7086-c8d3-c4ad-e3476fc4533f Address:127.0.0.1:29993}]" writer.go:29: 2021-01-29T19:33:24.564Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server.serf.wan: serf: EventMemberJoin: Node-e53badeb-7086-c8d3-c4ad-e3476fc4533f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:24.564Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:29993 [Follower]" leader= writer.go:29: 2021-01-29T19:33:24.567Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server.serf.lan: serf: EventMemberJoin: Node-e53badeb-7086-c8d3-c4ad-e3476fc4533f 127.0.0.1 writer.go:29: 2021-01-29T19:33:24.573Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server: Handled event for server in area: event=member-join server=Node-e53badeb-7086-c8d3-c4ad-e3476fc4533f.dc1 area=wan writer.go:29: 2021-01-29T19:33:24.574Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server: Adding LAN server: server="Node-e53badeb-7086-c8d3-c4ad-e3476fc4533f (Addr: tcp/127.0.0.1:29993) (DC: dc1)" writer.go:29: 2021-01-29T19:33:24.575Z [INFO] TestHealthServiceChecks_NodeMetaFilter: Started DNS server: address=127.0.0.1:29988 network=udp writer.go:29: 2021-01-29T19:33:24.577Z [INFO] TestHealthServiceChecks_NodeMetaFilter: Started DNS server: address=127.0.0.1:29988 network=tcp writer.go:29: 2021-01-29T19:33:24.579Z [INFO] TestHealthServiceChecks_NodeMetaFilter: Started HTTP server: address=127.0.0.1:29989 network=tcp writer.go:29: 2021-01-29T19:33:24.580Z [INFO] TestHealthServiceChecks_NodeMetaFilter: started state syncer === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:24.591Z [INFO] TestHealthServiceChecks_DistanceSort: Requesting shutdown writer.go:29: 2021-01-29T19:33:24.593Z [INFO] TestHealthServiceChecks_DistanceSort.server: shutting down server writer.go:29: 2021-01-29T19:33:24.594Z [DEBUG] TestHealthServiceChecks_DistanceSort.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.595Z [WARN] TestHealthServiceChecks_DistanceSort.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.596Z [DEBUG] TestHealthServiceChecks_DistanceSort.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.598Z [WARN] TestHealthServiceChecks_DistanceSort.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.609Z [INFO] TestHealthServiceChecks_DistanceSort.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:24.610Z [INFO] TestHealthServiceChecks_DistanceSort: consul server down writer.go:29: 2021-01-29T19:33:24.611Z [INFO] TestHealthServiceChecks_DistanceSort: shutdown complete writer.go:29: 2021-01-29T19:33:24.612Z [INFO] TestHealthServiceChecks_DistanceSort: Stopping server: protocol=DNS address=127.0.0.1:29976 network=tcp writer.go:29: 2021-01-29T19:33:24.613Z [INFO] TestHealthServiceChecks_DistanceSort: Stopping server: protocol=DNS address=127.0.0.1:29976 network=udp writer.go:29: 2021-01-29T19:33:24.615Z [INFO] TestHealthServiceChecks_DistanceSort: Stopping server: protocol=HTTP address=127.0.0.1:29977 network=tcp === CONT TestHealthServiceChecks_NodeMetaFilter writer.go:29: 2021-01-29T19:33:24.630Z [WARN] TestHealthServiceChecks_NodeMetaFilter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:24.631Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:29993 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:24.634Z [DEBUG] TestHealthServiceChecks_NodeMetaFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:24.635Z [DEBUG] TestHealthServiceChecks_NodeMetaFilter.server.raft: vote granted: from=e53badeb-7086-c8d3-c4ad-e3476fc4533f term=2 tally=1 writer.go:29: 2021-01-29T19:33:24.637Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:24.638Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:29993 [Leader]" writer.go:29: 2021-01-29T19:33:24.654Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:24.655Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server: New leader elected: payload=Node-e53badeb-7086-c8d3-c4ad-e3476fc4533f writer.go:29: 2021-01-29T19:33:24.656Z [DEBUG] TestHealthServiceChecks_NodeMetaFilter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29993 writer.go:29: 2021-01-29T19:33:24.668Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:24.676Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:24.678Z [INFO] TestHealthServiceChecks_NodeMetaFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.680Z [DEBUG] TestHealthServiceChecks_NodeMetaFilter.server: Skipping self join check for node since the cluster is too small: node=Node-e53badeb-7086-c8d3-c4ad-e3476fc4533f writer.go:29: 2021-01-29T19:33:24.682Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server: member joined, marking health alive: member=Node-e53badeb-7086-c8d3-c4ad-e3476fc4533f writer.go:29: 2021-01-29T19:33:24.850Z [INFO] TestHealthServiceChecks_NodeMetaFilter: Requesting shutdown writer.go:29: 2021-01-29T19:33:24.852Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server: shutting down server writer.go:29: 2021-01-29T19:33:24.854Z [DEBUG] TestHealthServiceChecks_NodeMetaFilter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.855Z [WARN] TestHealthServiceChecks_NodeMetaFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.854Z [ERROR] TestHealthServiceChecks_NodeMetaFilter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:24.855Z [DEBUG] TestHealthServiceChecks_NodeMetaFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.858Z [WARN] TestHealthServiceChecks_NodeMetaFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.864Z [INFO] TestHealthServiceChecks_NodeMetaFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:24.865Z [INFO] TestHealthServiceChecks_NodeMetaFilter: consul server down writer.go:29: 2021-01-29T19:33:24.869Z [INFO] TestHealthServiceChecks_NodeMetaFilter: shutdown complete writer.go:29: 2021-01-29T19:33:24.870Z [INFO] TestHealthServiceChecks_NodeMetaFilter: Stopping server: protocol=DNS address=127.0.0.1:29988 network=tcp writer.go:29: 2021-01-29T19:33:24.872Z [INFO] TestHealthServiceChecks_NodeMetaFilter: Stopping server: protocol=DNS address=127.0.0.1:29988 network=udp writer.go:29: 2021-01-29T19:33:24.874Z [INFO] TestHealthServiceChecks_NodeMetaFilter: Stopping server: protocol=HTTP address=127.0.0.1:29989 network=tcp === CONT TestHealthServiceNodes writer.go:29: 2021-01-29T19:33:24.913Z [INFO] TestHealthServiceNodes: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:24.916Z [INFO] TestHealthServiceNodes: Endpoints down --- PASS: TestHealthServiceNodes (1.18s) === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:24.932Z [WARN] TestHealthNodeChecks: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:24.935Z [DEBUG] TestHealthNodeChecks.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:24.938Z [DEBUG] TestHealthNodeChecks.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:24.943Z [INFO] TestHealthNodeChecks.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:12b7ed36-c14f-5e93-f46d-78fd8313ffe8 Address:127.0.0.1:30005}]" writer.go:29: 2021-01-29T19:33:24.945Z [INFO] TestHealthNodeChecks.server.raft: entering follower state: follower="Node at 127.0.0.1:30005 [Follower]" leader= writer.go:29: 2021-01-29T19:33:24.949Z [INFO] TestHealthNodeChecks.server.serf.wan: serf: EventMemberJoin: Node-12b7ed36-c14f-5e93-f46d-78fd8313ffe8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:24.956Z [INFO] TestHealthNodeChecks.server.serf.lan: serf: EventMemberJoin: Node-12b7ed36-c14f-5e93-f46d-78fd8313ffe8 127.0.0.1 writer.go:29: 2021-01-29T19:33:24.960Z [INFO] TestHealthNodeChecks.server: Handled event for server in area: event=member-join server=Node-12b7ed36-c14f-5e93-f46d-78fd8313ffe8.dc1 area=wan writer.go:29: 2021-01-29T19:33:24.960Z [INFO] TestHealthNodeChecks.server: Adding LAN server: server="Node-12b7ed36-c14f-5e93-f46d-78fd8313ffe8 (Addr: tcp/127.0.0.1:30005) (DC: dc1)" writer.go:29: 2021-01-29T19:33:24.966Z [INFO] TestHealthNodeChecks: Started DNS server: address=127.0.0.1:30000 network=tcp writer.go:29: 2021-01-29T19:33:24.969Z [INFO] TestHealthNodeChecks: Started DNS server: address=127.0.0.1:30000 network=udp writer.go:29: 2021-01-29T19:33:24.971Z [INFO] TestHealthNodeChecks: Started HTTP server: address=127.0.0.1:30001 network=tcp writer.go:29: 2021-01-29T19:33:24.983Z [INFO] TestHealthNodeChecks: started state syncer === CONT TestHealthServiceChecks_Filtering writer.go:29: 2021-01-29T19:33:24.995Z [INFO] TestHealthServiceChecks_Filtering: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:24.996Z [INFO] TestHealthServiceChecks_Filtering: Endpoints down --- PASS: TestHealthServiceChecks_Filtering (0.91s) === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.006Z [WARN] TestHealthServiceChecks: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:25.007Z [DEBUG] TestHealthServiceChecks.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:25.008Z [DEBUG] TestHealthServiceChecks.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.010Z [WARN] TestHealthNodeChecks.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:25.011Z [INFO] TestHealthNodeChecks.server.raft: entering candidate state: node="Node at 127.0.0.1:30005 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:25.030Z [DEBUG] TestHealthNodeChecks.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:25.031Z [DEBUG] TestHealthNodeChecks.server.raft: vote granted: from=12b7ed36-c14f-5e93-f46d-78fd8313ffe8 term=2 tally=1 === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.031Z [INFO] TestHealthServiceChecks.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3a94ae07-0bd1-886e-c94f-d3fcc490c946 Address:127.0.0.1:29999}]" === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.032Z [INFO] TestHealthNodeChecks.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:25.033Z [INFO] TestHealthNodeChecks.server.raft: entering leader state: leader="Node at 127.0.0.1:30005 [Leader]" === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.033Z [INFO] TestHealthServiceChecks.server.serf.wan: serf: EventMemberJoin: Node-3a94ae07-0bd1-886e-c94f-d3fcc490c946.dc1 127.0.0.1 === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.034Z [INFO] TestHealthNodeChecks.server: cluster leadership acquired === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.033Z [INFO] TestHealthServiceChecks.server.raft: entering follower state: follower="Node at 127.0.0.1:29999 [Follower]" leader= === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.034Z [INFO] TestHealthNodeChecks.server: New leader elected: payload=Node-12b7ed36-c14f-5e93-f46d-78fd8313ffe8 === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.037Z [INFO] TestHealthServiceChecks.server.serf.lan: serf: EventMemberJoin: Node-3a94ae07-0bd1-886e-c94f-d3fcc490c946 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.040Z [INFO] TestHealthServiceChecks.server: Adding LAN server: server="Node-3a94ae07-0bd1-886e-c94f-d3fcc490c946 (Addr: tcp/127.0.0.1:29999) (DC: dc1)" writer.go:29: 2021-01-29T19:33:25.040Z [INFO] TestHealthServiceChecks: Started DNS server: address=127.0.0.1:29994 network=udp writer.go:29: 2021-01-29T19:33:25.042Z [INFO] TestHealthServiceChecks.server: Handled event for server in area: event=member-join server=Node-3a94ae07-0bd1-886e-c94f-d3fcc490c946.dc1 area=wan writer.go:29: 2021-01-29T19:33:25.042Z [INFO] TestHealthServiceChecks: Started DNS server: address=127.0.0.1:29994 network=tcp writer.go:29: 2021-01-29T19:33:25.044Z [INFO] TestHealthServiceChecks: Started HTTP server: address=127.0.0.1:29995 network=tcp writer.go:29: 2021-01-29T19:33:25.045Z [INFO] TestHealthServiceChecks: started state syncer === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.037Z [DEBUG] TestHealthNodeChecks.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30005 writer.go:29: 2021-01-29T19:33:25.086Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:25.093Z [INFO] TestHealthNodeChecks.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:25.095Z [INFO] TestHealthNodeChecks.leader: started routine: routine="CA root pruning" === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.098Z [WARN] TestHealthServiceChecks.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:25.100Z [INFO] TestHealthServiceChecks.server.raft: entering candidate state: node="Node at 127.0.0.1:29999 [Candidate]" term=2 === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.098Z [DEBUG] TestHealthNodeChecks.server: Skipping self join check for node since the cluster is too small: node=Node-12b7ed36-c14f-5e93-f46d-78fd8313ffe8 writer.go:29: 2021-01-29T19:33:25.101Z [INFO] TestHealthNodeChecks.server: member joined, marking health alive: member=Node-12b7ed36-c14f-5e93-f46d-78fd8313ffe8 === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.108Z [DEBUG] TestHealthServiceChecks.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:25.110Z [DEBUG] TestHealthServiceChecks.server.raft: vote granted: from=3a94ae07-0bd1-886e-c94f-d3fcc490c946 term=2 tally=1 writer.go:29: 2021-01-29T19:33:25.111Z [INFO] TestHealthServiceChecks.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:25.112Z [INFO] TestHealthServiceChecks.server.raft: entering leader state: leader="Node at 127.0.0.1:29999 [Leader]" writer.go:29: 2021-01-29T19:33:25.113Z [INFO] TestHealthServiceChecks.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:25.114Z [INFO] TestHealthServiceChecks.server: New leader elected: payload=Node-3a94ae07-0bd1-886e-c94f-d3fcc490c946 === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.117Z [INFO] TestHealthNodeChecks: Requesting shutdown === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:25.116Z [INFO] TestHealthServiceChecks_DistanceSort: Waiting for endpoints to shut down === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.118Z [INFO] TestHealthNodeChecks.server: shutting down server writer.go:29: 2021-01-29T19:33:25.119Z [DEBUG] TestHealthNodeChecks.leader: stopping routine: routine="CA root pruning" === CONT TestHealthServiceChecks_DistanceSort writer.go:29: 2021-01-29T19:33:25.118Z [INFO] TestHealthServiceChecks_DistanceSort: Endpoints down --- PASS: TestHealthServiceChecks_DistanceSort (1.04s) === CONT TestHealthNodeChecks_Filtering === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.120Z [WARN] TestHealthNodeChecks.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:25.122Z [ERROR] TestHealthNodeChecks.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:25.123Z [DEBUG] TestHealthNodeChecks.leader: stopped routine: routine="CA root pruning" === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.130Z [DEBUG] TestHealthServiceChecks.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29999 === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.131Z [WARN] TestHealthNodeChecks.server.serf.wan: serf: Shutdown without a Leave === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.140Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.140Z [INFO] TestHealthNodeChecks.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:25.146Z [INFO] TestHealthNodeChecks: consul server down writer.go:29: 2021-01-29T19:33:25.149Z [INFO] TestHealthNodeChecks: shutdown complete writer.go:29: 2021-01-29T19:33:25.151Z [INFO] TestHealthNodeChecks: Stopping server: protocol=DNS address=127.0.0.1:30000 network=tcp writer.go:29: 2021-01-29T19:33:25.154Z [INFO] TestHealthNodeChecks: Stopping server: protocol=DNS address=127.0.0.1:30000 network=udp writer.go:29: 2021-01-29T19:33:25.155Z [INFO] TestHealthNodeChecks: Stopping server: protocol=HTTP address=127.0.0.1:30001 network=tcp === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.154Z [INFO] TestHealthServiceChecks.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:25.157Z [INFO] TestHealthServiceChecks.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.159Z [DEBUG] TestHealthServiceChecks.server: Skipping self join check for node since the cluster is too small: node=Node-3a94ae07-0bd1-886e-c94f-d3fcc490c946 writer.go:29: 2021-01-29T19:33:25.160Z [INFO] TestHealthServiceChecks.server: member joined, marking health alive: member=Node-3a94ae07-0bd1-886e-c94f-d3fcc490c946 === CONT TestHealthNodeChecks_Filtering writer.go:29: 2021-01-29T19:33:25.166Z [WARN] TestHealthNodeChecks_Filtering: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:25.168Z [DEBUG] TestHealthNodeChecks_Filtering.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:25.170Z [DEBUG] TestHealthNodeChecks_Filtering.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:25.180Z [INFO] TestHealthNodeChecks_Filtering.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1760350e-ba7c-7a2d-0c46-29c8c4cf9169 Address:127.0.0.1:30011}]" writer.go:29: 2021-01-29T19:33:25.183Z [INFO] TestHealthNodeChecks_Filtering.server.raft: entering follower state: follower="Node at 127.0.0.1:30011 [Follower]" leader= writer.go:29: 2021-01-29T19:33:25.185Z [INFO] TestHealthNodeChecks_Filtering.server.serf.wan: serf: EventMemberJoin: Node-1760350e-ba7c-7a2d-0c46-29c8c4cf9169.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.190Z [INFO] TestHealthNodeChecks_Filtering.server.serf.lan: serf: EventMemberJoin: Node-1760350e-ba7c-7a2d-0c46-29c8c4cf9169 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.194Z [INFO] TestHealthNodeChecks_Filtering.server: Handled event for server in area: event=member-join server=Node-1760350e-ba7c-7a2d-0c46-29c8c4cf9169.dc1 area=wan writer.go:29: 2021-01-29T19:33:25.195Z [INFO] TestHealthNodeChecks_Filtering: Started DNS server: address=127.0.0.1:30006 network=udp writer.go:29: 2021-01-29T19:33:25.197Z [INFO] TestHealthNodeChecks_Filtering: Started DNS server: address=127.0.0.1:30006 network=tcp writer.go:29: 2021-01-29T19:33:25.196Z [INFO] TestHealthNodeChecks_Filtering.server: Adding LAN server: server="Node-1760350e-ba7c-7a2d-0c46-29c8c4cf9169 (Addr: tcp/127.0.0.1:30011) (DC: dc1)" writer.go:29: 2021-01-29T19:33:25.200Z [INFO] TestHealthNodeChecks_Filtering: Started HTTP server: address=127.0.0.1:30007 network=tcp writer.go:29: 2021-01-29T19:33:25.201Z [INFO] TestHealthNodeChecks_Filtering: started state syncer writer.go:29: 2021-01-29T19:33:25.231Z [WARN] TestHealthNodeChecks_Filtering.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:25.232Z [INFO] TestHealthNodeChecks_Filtering.server.raft: entering candidate state: node="Node at 127.0.0.1:30011 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:25.233Z [DEBUG] TestHealthNodeChecks_Filtering.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:25.234Z [DEBUG] TestHealthNodeChecks_Filtering.server.raft: vote granted: from=1760350e-ba7c-7a2d-0c46-29c8c4cf9169 term=2 tally=1 writer.go:29: 2021-01-29T19:33:25.235Z [INFO] TestHealthNodeChecks_Filtering.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:25.236Z [INFO] TestHealthNodeChecks_Filtering.server.raft: entering leader state: leader="Node at 127.0.0.1:30011 [Leader]" writer.go:29: 2021-01-29T19:33:25.239Z [INFO] TestHealthNodeChecks_Filtering.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:25.240Z [INFO] TestHealthNodeChecks_Filtering.server: New leader elected: payload=Node-1760350e-ba7c-7a2d-0c46-29c8c4cf9169 writer.go:29: 2021-01-29T19:33:25.242Z [DEBUG] TestHealthNodeChecks_Filtering.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30011 writer.go:29: 2021-01-29T19:33:25.245Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:25.249Z [INFO] TestHealthNodeChecks_Filtering.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:25.250Z [INFO] TestHealthNodeChecks_Filtering.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.252Z [DEBUG] TestHealthNodeChecks_Filtering.server: Skipping self join check for node since the cluster is too small: node=Node-1760350e-ba7c-7a2d-0c46-29c8c4cf9169 writer.go:29: 2021-01-29T19:33:25.254Z [INFO] TestHealthNodeChecks_Filtering.server: member joined, marking health alive: member=Node-1760350e-ba7c-7a2d-0c46-29c8c4cf9169 === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.300Z [INFO] TestHealthServiceChecks: Requesting shutdown writer.go:29: 2021-01-29T19:33:25.301Z [INFO] TestHealthServiceChecks.server: shutting down server writer.go:29: 2021-01-29T19:33:25.301Z [DEBUG] TestHealthServiceChecks.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.302Z [WARN] TestHealthServiceChecks.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:25.302Z [ERROR] TestHealthServiceChecks.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:25.302Z [DEBUG] TestHealthServiceChecks.leader: stopped routine: routine="CA root pruning" === CONT TestHealthNodeChecks_Filtering writer.go:29: 2021-01-29T19:33:25.309Z [DEBUG] TestHealthNodeChecks_Filtering: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.339Z [WARN] TestHealthServiceChecks.server.serf.wan: serf: Shutdown without a Leave === CONT TestHealthNodeChecks_Filtering writer.go:29: 2021-01-29T19:33:25.340Z [INFO] TestHealthNodeChecks_Filtering: Synced node info writer.go:29: 2021-01-29T19:33:25.341Z [DEBUG] TestHealthNodeChecks_Filtering: Node info in sync === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.341Z [INFO] TestHealthServiceChecks.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:25.342Z [INFO] TestHealthServiceChecks: consul server down writer.go:29: 2021-01-29T19:33:25.345Z [INFO] TestHealthServiceChecks: shutdown complete writer.go:29: 2021-01-29T19:33:25.346Z [INFO] TestHealthServiceChecks: Stopping server: protocol=DNS address=127.0.0.1:29994 network=tcp writer.go:29: 2021-01-29T19:33:25.347Z [INFO] TestHealthServiceChecks: Stopping server: protocol=DNS address=127.0.0.1:29994 network=udp writer.go:29: 2021-01-29T19:33:25.349Z [INFO] TestHealthServiceChecks: Stopping server: protocol=HTTP address=127.0.0.1:29995 network=tcp === CONT TestHealthServiceChecks_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.376Z [INFO] TestHealthServiceChecks_NodeMetaFilter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:25.379Z [INFO] TestHealthServiceChecks_NodeMetaFilter: Endpoints down --- PASS: TestHealthServiceChecks_NodeMetaFilter (0.84s) === CONT TestHealthChecksInState_DistanceSort writer.go:29: 2021-01-29T19:33:25.406Z [WARN] TestHealthChecksInState_DistanceSort: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:25.418Z [DEBUG] TestHealthChecksInState_DistanceSort.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:25.421Z [DEBUG] TestHealthChecksInState_DistanceSort.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:25.426Z [INFO] TestHealthChecksInState_DistanceSort.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8b9936f7-ae63-bc9f-1a59-6e49514beb19 Address:127.0.0.1:30017}]" writer.go:29: 2021-01-29T19:33:25.429Z [INFO] TestHealthChecksInState_DistanceSort.server.raft: entering follower state: follower="Node at 127.0.0.1:30017 [Follower]" leader= writer.go:29: 2021-01-29T19:33:25.433Z [INFO] TestHealthChecksInState_DistanceSort.server.serf.wan: serf: EventMemberJoin: Node-8b9936f7-ae63-bc9f-1a59-6e49514beb19.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.443Z [INFO] TestHealthChecksInState_DistanceSort.server.serf.lan: serf: EventMemberJoin: Node-8b9936f7-ae63-bc9f-1a59-6e49514beb19 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.447Z [INFO] TestHealthChecksInState_DistanceSort.server: Adding LAN server: server="Node-8b9936f7-ae63-bc9f-1a59-6e49514beb19 (Addr: tcp/127.0.0.1:30017) (DC: dc1)" writer.go:29: 2021-01-29T19:33:25.449Z [INFO] TestHealthChecksInState_DistanceSort.server: Handled event for server in area: event=member-join server=Node-8b9936f7-ae63-bc9f-1a59-6e49514beb19.dc1 area=wan writer.go:29: 2021-01-29T19:33:25.453Z [INFO] TestHealthChecksInState_DistanceSort: Started DNS server: address=127.0.0.1:30012 network=tcp writer.go:29: 2021-01-29T19:33:25.455Z [INFO] TestHealthChecksInState_DistanceSort: Started DNS server: address=127.0.0.1:30012 network=udp writer.go:29: 2021-01-29T19:33:25.458Z [INFO] TestHealthChecksInState_DistanceSort: Started HTTP server: address=127.0.0.1:30013 network=tcp writer.go:29: 2021-01-29T19:33:25.459Z [INFO] TestHealthChecksInState_DistanceSort: started state syncer writer.go:29: 2021-01-29T19:33:25.471Z [WARN] TestHealthChecksInState_DistanceSort.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:25.473Z [INFO] TestHealthChecksInState_DistanceSort.server.raft: entering candidate state: node="Node at 127.0.0.1:30017 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:25.475Z [DEBUG] TestHealthChecksInState_DistanceSort.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:25.477Z [DEBUG] TestHealthChecksInState_DistanceSort.server.raft: vote granted: from=8b9936f7-ae63-bc9f-1a59-6e49514beb19 term=2 tally=1 writer.go:29: 2021-01-29T19:33:25.479Z [INFO] TestHealthChecksInState_DistanceSort.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:25.480Z [INFO] TestHealthChecksInState_DistanceSort.server.raft: entering leader state: leader="Node at 127.0.0.1:30017 [Leader]" writer.go:29: 2021-01-29T19:33:25.492Z [INFO] TestHealthChecksInState_DistanceSort.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:25.493Z [INFO] TestHealthChecksInState_DistanceSort.server: New leader elected: payload=Node-8b9936f7-ae63-bc9f-1a59-6e49514beb19 writer.go:29: 2021-01-29T19:33:25.494Z [DEBUG] TestHealthChecksInState_DistanceSort.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30017 writer.go:29: 2021-01-29T19:33:25.503Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:25.507Z [INFO] TestHealthChecksInState_DistanceSort.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:25.508Z [INFO] TestHealthChecksInState_DistanceSort.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.509Z [DEBUG] TestHealthChecksInState_DistanceSort.server: Skipping self join check for node since the cluster is too small: node=Node-8b9936f7-ae63-bc9f-1a59-6e49514beb19 writer.go:29: 2021-01-29T19:33:25.509Z [INFO] TestHealthChecksInState_DistanceSort.server: member joined, marking health alive: member=Node-8b9936f7-ae63-bc9f-1a59-6e49514beb19 === CONT TestHealthNodeChecks_Filtering writer.go:29: 2021-01-29T19:33:25.530Z [DEBUG] TestHealthNodeChecks_Filtering: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:25.532Z [DEBUG] TestHealthNodeChecks_Filtering: Node info in sync === CONT TestHealthChecksInState_DistanceSort writer.go:29: 2021-01-29T19:33:25.587Z [DEBUG] TestHealthChecksInState_DistanceSort: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:25.589Z [INFO] TestHealthChecksInState_DistanceSort: Synced node info writer.go:29: 2021-01-29T19:33:25.591Z [DEBUG] TestHealthChecksInState_DistanceSort: Node info in sync === CONT TestHealthNodeChecks_Filtering writer.go:29: 2021-01-29T19:33:25.620Z [INFO] TestHealthNodeChecks_Filtering: Requesting shutdown writer.go:29: 2021-01-29T19:33:25.621Z [INFO] TestHealthNodeChecks_Filtering.server: shutting down server writer.go:29: 2021-01-29T19:33:25.622Z [DEBUG] TestHealthNodeChecks_Filtering.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.623Z [WARN] TestHealthNodeChecks_Filtering.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:25.623Z [DEBUG] TestHealthNodeChecks_Filtering.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.625Z [WARN] TestHealthNodeChecks_Filtering.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:25.630Z [INFO] TestHealthNodeChecks_Filtering.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:25.630Z [INFO] TestHealthNodeChecks_Filtering: consul server down writer.go:29: 2021-01-29T19:33:25.633Z [INFO] TestHealthNodeChecks_Filtering: shutdown complete writer.go:29: 2021-01-29T19:33:25.635Z [INFO] TestHealthNodeChecks_Filtering: Stopping server: protocol=DNS address=127.0.0.1:30006 network=tcp writer.go:29: 2021-01-29T19:33:25.636Z [INFO] TestHealthNodeChecks_Filtering: Stopping server: protocol=DNS address=127.0.0.1:30006 network=udp writer.go:29: 2021-01-29T19:33:25.637Z [INFO] TestHealthNodeChecks_Filtering: Stopping server: protocol=HTTP address=127.0.0.1:30007 network=tcp === CONT TestHealthNodeChecks writer.go:29: 2021-01-29T19:33:25.657Z [INFO] TestHealthNodeChecks: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:25.659Z [INFO] TestHealthNodeChecks: Endpoints down --- PASS: TestHealthNodeChecks (0.74s) === CONT TestHealthChecksInState_Filter writer.go:29: 2021-01-29T19:33:25.681Z [WARN] TestHealthChecksInState_Filter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:25.683Z [DEBUG] TestHealthChecksInState_Filter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:25.685Z [DEBUG] TestHealthChecksInState_Filter.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:25.690Z [INFO] TestHealthChecksInState_Filter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ba7196e6-502b-61f1-6ab4-15edc3d85176 Address:127.0.0.1:30023}]" writer.go:29: 2021-01-29T19:33:25.692Z [INFO] TestHealthChecksInState_Filter.server.raft: entering follower state: follower="Node at 127.0.0.1:30023 [Follower]" leader= writer.go:29: 2021-01-29T19:33:25.693Z [INFO] TestHealthChecksInState_Filter.server.serf.wan: serf: EventMemberJoin: Node-ba7196e6-502b-61f1-6ab4-15edc3d85176.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.696Z [INFO] TestHealthChecksInState_Filter.server.serf.lan: serf: EventMemberJoin: Node-ba7196e6-502b-61f1-6ab4-15edc3d85176 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.698Z [INFO] TestHealthChecksInState_Filter.server: Handled event for server in area: event=member-join server=Node-ba7196e6-502b-61f1-6ab4-15edc3d85176.dc1 area=wan writer.go:29: 2021-01-29T19:33:25.698Z [INFO] TestHealthChecksInState_Filter.server: Adding LAN server: server="Node-ba7196e6-502b-61f1-6ab4-15edc3d85176 (Addr: tcp/127.0.0.1:30023) (DC: dc1)" writer.go:29: 2021-01-29T19:33:25.699Z [INFO] TestHealthChecksInState_Filter: Started DNS server: address=127.0.0.1:30018 network=udp writer.go:29: 2021-01-29T19:33:25.702Z [INFO] TestHealthChecksInState_Filter: Started DNS server: address=127.0.0.1:30018 network=tcp writer.go:29: 2021-01-29T19:33:25.706Z [INFO] TestHealthChecksInState_Filter: Started HTTP server: address=127.0.0.1:30019 network=tcp writer.go:29: 2021-01-29T19:33:25.707Z [INFO] TestHealthChecksInState_Filter: started state syncer writer.go:29: 2021-01-29T19:33:25.740Z [WARN] TestHealthChecksInState_Filter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:25.741Z [INFO] TestHealthChecksInState_Filter.server.raft: entering candidate state: node="Node at 127.0.0.1:30023 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:25.743Z [DEBUG] TestHealthChecksInState_Filter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:25.745Z [DEBUG] TestHealthChecksInState_Filter.server.raft: vote granted: from=ba7196e6-502b-61f1-6ab4-15edc3d85176 term=2 tally=1 writer.go:29: 2021-01-29T19:33:25.746Z [INFO] TestHealthChecksInState_Filter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:25.748Z [INFO] TestHealthChecksInState_Filter.server.raft: entering leader state: leader="Node at 127.0.0.1:30023 [Leader]" writer.go:29: 2021-01-29T19:33:25.749Z [INFO] TestHealthChecksInState_Filter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:25.750Z [INFO] TestHealthChecksInState_Filter.server: New leader elected: payload=Node-ba7196e6-502b-61f1-6ab4-15edc3d85176 writer.go:29: 2021-01-29T19:33:25.751Z [DEBUG] TestHealthChecksInState_Filter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30023 writer.go:29: 2021-01-29T19:33:25.760Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:25.767Z [INFO] TestHealthChecksInState_Filter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:25.768Z [INFO] TestHealthChecksInState_Filter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.770Z [DEBUG] TestHealthChecksInState_Filter.server: Skipping self join check for node since the cluster is too small: node=Node-ba7196e6-502b-61f1-6ab4-15edc3d85176 writer.go:29: 2021-01-29T19:33:25.772Z [INFO] TestHealthChecksInState_Filter.server: member joined, marking health alive: member=Node-ba7196e6-502b-61f1-6ab4-15edc3d85176 writer.go:29: 2021-01-29T19:33:25.846Z [INFO] TestHealthChecksInState_Filter: Requesting shutdown writer.go:29: 2021-01-29T19:33:25.847Z [INFO] TestHealthChecksInState_Filter.server: shutting down server writer.go:29: 2021-01-29T19:33:25.849Z [DEBUG] TestHealthChecksInState_Filter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.850Z [WARN] TestHealthChecksInState_Filter.server.serf.lan: serf: Shutdown without a Leave === CONT TestHealthServiceChecks writer.go:29: 2021-01-29T19:33:25.852Z [INFO] TestHealthServiceChecks: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:25.852Z [INFO] TestHealthServiceChecks: Endpoints down --- PASS: TestHealthServiceChecks (0.86s) === CONT TestHealthChecksInState_NodeMetaFilter === CONT TestHealthChecksInState_Filter writer.go:29: 2021-01-29T19:33:25.850Z [ERROR] TestHealthChecksInState_Filter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:25.851Z [DEBUG] TestHealthChecksInState_Filter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.859Z [WARN] TestHealthChecksInState_Filter.server.serf.wan: serf: Shutdown without a Leave === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.861Z [WARN] TestHealthChecksInState_NodeMetaFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:25.862Z [DEBUG] TestHealthChecksInState_NodeMetaFilter.tlsutil: Update: version=1 === CONT TestHealthChecksInState_Filter writer.go:29: 2021-01-29T19:33:25.862Z [INFO] TestHealthChecksInState_Filter.server.router.manager: shutting down === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.864Z [DEBUG] TestHealthChecksInState_NodeMetaFilter.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestHealthChecksInState_Filter writer.go:29: 2021-01-29T19:33:25.862Z [INFO] TestHealthChecksInState_Filter: consul server down writer.go:29: 2021-01-29T19:33:25.866Z [INFO] TestHealthChecksInState_Filter: shutdown complete writer.go:29: 2021-01-29T19:33:25.868Z [INFO] TestHealthChecksInState_Filter: Stopping server: protocol=DNS address=127.0.0.1:30018 network=tcp === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.870Z [INFO] TestHealthChecksInState_NodeMetaFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:18a37481-00fd-cda8-4096-50130f1b9d7a Address:127.0.0.1:30029}]" === CONT TestHealthChecksInState_Filter writer.go:29: 2021-01-29T19:33:25.870Z [INFO] TestHealthChecksInState_Filter: Stopping server: protocol=DNS address=127.0.0.1:30018 network=udp writer.go:29: 2021-01-29T19:33:25.872Z [INFO] TestHealthChecksInState_Filter: Stopping server: protocol=HTTP address=127.0.0.1:30019 network=tcp === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.872Z [INFO] TestHealthChecksInState_NodeMetaFilter.server.serf.wan: serf: EventMemberJoin: Node-18a37481-00fd-cda8-4096-50130f1b9d7a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.882Z [INFO] TestHealthChecksInState_NodeMetaFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:30029 [Follower]" leader= writer.go:29: 2021-01-29T19:33:25.885Z [INFO] TestHealthChecksInState_NodeMetaFilter.server.serf.lan: serf: EventMemberJoin: Node-18a37481-00fd-cda8-4096-50130f1b9d7a 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.900Z [INFO] TestHealthChecksInState_NodeMetaFilter.server: Adding LAN server: server="Node-18a37481-00fd-cda8-4096-50130f1b9d7a (Addr: tcp/127.0.0.1:30029) (DC: dc1)" writer.go:29: 2021-01-29T19:33:25.900Z [INFO] TestHealthChecksInState_NodeMetaFilter.server: Handled event for server in area: event=member-join server=Node-18a37481-00fd-cda8-4096-50130f1b9d7a.dc1 area=wan writer.go:29: 2021-01-29T19:33:25.905Z [INFO] TestHealthChecksInState_NodeMetaFilter: Started DNS server: address=127.0.0.1:30024 network=tcp writer.go:29: 2021-01-29T19:33:25.911Z [INFO] TestHealthChecksInState_NodeMetaFilter: Started DNS server: address=127.0.0.1:30024 network=udp writer.go:29: 2021-01-29T19:33:25.917Z [INFO] TestHealthChecksInState_NodeMetaFilter: Started HTTP server: address=127.0.0.1:30025 network=tcp writer.go:29: 2021-01-29T19:33:25.923Z [INFO] TestHealthChecksInState_NodeMetaFilter: started state syncer writer.go:29: 2021-01-29T19:33:25.945Z [WARN] TestHealthChecksInState_NodeMetaFilter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:25.947Z [INFO] TestHealthChecksInState_NodeMetaFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:30029 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:25.949Z [DEBUG] TestHealthChecksInState_NodeMetaFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:25.951Z [DEBUG] TestHealthChecksInState_NodeMetaFilter.server.raft: vote granted: from=18a37481-00fd-cda8-4096-50130f1b9d7a term=2 tally=1 === CONT TestHealthChecksInState_DistanceSort writer.go:29: 2021-01-29T19:33:25.951Z [INFO] TestHealthChecksInState_DistanceSort: Requesting shutdown === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.952Z [INFO] TestHealthChecksInState_NodeMetaFilter.server.raft: election won: tally=1 === CONT TestHealthChecksInState_DistanceSort writer.go:29: 2021-01-29T19:33:25.953Z [INFO] TestHealthChecksInState_DistanceSort.server: shutting down server === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.953Z [INFO] TestHealthChecksInState_NodeMetaFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:30029 [Leader]" writer.go:29: 2021-01-29T19:33:25.954Z [INFO] TestHealthChecksInState_NodeMetaFilter.server: cluster leadership acquired === CONT TestHealthChecksInState_DistanceSort writer.go:29: 2021-01-29T19:33:25.954Z [DEBUG] TestHealthChecksInState_DistanceSort.leader: stopping routine: routine="CA root pruning" === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.955Z [INFO] TestHealthChecksInState_NodeMetaFilter.server: New leader elected: payload=Node-18a37481-00fd-cda8-4096-50130f1b9d7a === CONT TestHealthChecksInState_DistanceSort writer.go:29: 2021-01-29T19:33:25.955Z [WARN] TestHealthChecksInState_DistanceSort.server.serf.lan: serf: Shutdown without a Leave === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.956Z [DEBUG] TestHealthChecksInState_NodeMetaFilter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30029 === CONT TestHealthChecksInState_DistanceSort writer.go:29: 2021-01-29T19:33:25.956Z [DEBUG] TestHealthChecksInState_DistanceSort.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.959Z [WARN] TestHealthChecksInState_DistanceSort.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:25.970Z [INFO] TestHealthChecksInState_DistanceSort.server.router.manager: shutting down === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.970Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestHealthChecksInState_DistanceSort writer.go:29: 2021-01-29T19:33:25.970Z [INFO] TestHealthChecksInState_DistanceSort: consul server down writer.go:29: 2021-01-29T19:33:25.972Z [INFO] TestHealthChecksInState_DistanceSort: shutdown complete writer.go:29: 2021-01-29T19:33:25.974Z [INFO] TestHealthChecksInState_DistanceSort: Stopping server: protocol=DNS address=127.0.0.1:30012 network=tcp writer.go:29: 2021-01-29T19:33:25.975Z [INFO] TestHealthChecksInState_DistanceSort: Stopping server: protocol=DNS address=127.0.0.1:30012 network=udp === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.976Z [INFO] TestHealthChecksInState_NodeMetaFilter.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestHealthChecksInState_DistanceSort writer.go:29: 2021-01-29T19:33:25.976Z [INFO] TestHealthChecksInState_DistanceSort: Stopping server: protocol=HTTP address=127.0.0.1:30013 network=tcp === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:25.977Z [INFO] TestHealthChecksInState_NodeMetaFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.979Z [DEBUG] TestHealthChecksInState_NodeMetaFilter.server: Skipping self join check for node since the cluster is too small: node=Node-18a37481-00fd-cda8-4096-50130f1b9d7a writer.go:29: 2021-01-29T19:33:25.980Z [INFO] TestHealthChecksInState_NodeMetaFilter.server: member joined, marking health alive: member=Node-18a37481-00fd-cda8-4096-50130f1b9d7a writer.go:29: 2021-01-29T19:33:26.018Z [INFO] TestHealthChecksInState_NodeMetaFilter: Requesting shutdown writer.go:29: 2021-01-29T19:33:26.020Z [INFO] TestHealthChecksInState_NodeMetaFilter.server: shutting down server writer.go:29: 2021-01-29T19:33:26.021Z [DEBUG] TestHealthChecksInState_NodeMetaFilter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.023Z [WARN] TestHealthChecksInState_NodeMetaFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.021Z [ERROR] TestHealthChecksInState_NodeMetaFilter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:26.023Z [DEBUG] TestHealthChecksInState_NodeMetaFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.036Z [WARN] TestHealthChecksInState_NodeMetaFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.047Z [INFO] TestHealthChecksInState_NodeMetaFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:26.047Z [INFO] TestHealthChecksInState_NodeMetaFilter: consul server down writer.go:29: 2021-01-29T19:33:26.051Z [INFO] TestHealthChecksInState_NodeMetaFilter: shutdown complete writer.go:29: 2021-01-29T19:33:26.052Z [INFO] TestHealthChecksInState_NodeMetaFilter: Stopping server: protocol=DNS address=127.0.0.1:30024 network=tcp writer.go:29: 2021-01-29T19:33:26.055Z [INFO] TestHealthChecksInState_NodeMetaFilter: Stopping server: protocol=DNS address=127.0.0.1:30024 network=udp writer.go:29: 2021-01-29T19:33:26.057Z [INFO] TestHealthChecksInState_NodeMetaFilter: Stopping server: protocol=HTTP address=127.0.0.1:30025 network=tcp === CONT TestHealthNodeChecks_Filtering writer.go:29: 2021-01-29T19:33:26.139Z [INFO] TestHealthNodeChecks_Filtering: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:26.140Z [INFO] TestHealthNodeChecks_Filtering: Endpoints down --- PASS: TestHealthNodeChecks_Filtering (1.02s) === CONT TestUUIDToUint64 === CONT TestEventList_EventBufOrder --- PASS: TestUUIDToUint64 (0.00s) === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.192Z [WARN] TestEventList_EventBufOrder: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:26.194Z [DEBUG] TestEventList_EventBufOrder.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:26.197Z [DEBUG] TestEventList_EventBufOrder.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:26.220Z [INFO] TestEventList_EventBufOrder.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4bb9d8be-9ec2-236f-6fa5-f96e132c69d8 Address:127.0.0.1:30035}]" writer.go:29: 2021-01-29T19:33:26.223Z [INFO] TestEventList_EventBufOrder.server.serf.wan: serf: EventMemberJoin: Node-4bb9d8be-9ec2-236f-6fa5-f96e132c69d8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:26.227Z [INFO] TestEventList_EventBufOrder.server.serf.lan: serf: EventMemberJoin: Node-4bb9d8be-9ec2-236f-6fa5-f96e132c69d8 127.0.0.1 writer.go:29: 2021-01-29T19:33:26.224Z [INFO] TestEventList_EventBufOrder.server.raft: entering follower state: follower="Node at 127.0.0.1:30035 [Follower]" leader= writer.go:29: 2021-01-29T19:33:26.234Z [INFO] TestEventList_EventBufOrder.server: Adding LAN server: server="Node-4bb9d8be-9ec2-236f-6fa5-f96e132c69d8 (Addr: tcp/127.0.0.1:30035) (DC: dc1)" writer.go:29: 2021-01-29T19:33:26.236Z [INFO] TestEventList_EventBufOrder.server: Handled event for server in area: event=member-join server=Node-4bb9d8be-9ec2-236f-6fa5-f96e132c69d8.dc1 area=wan writer.go:29: 2021-01-29T19:33:26.243Z [INFO] TestEventList_EventBufOrder: Started DNS server: address=127.0.0.1:30030 network=udp writer.go:29: 2021-01-29T19:33:26.247Z [INFO] TestEventList_EventBufOrder: Started DNS server: address=127.0.0.1:30030 network=tcp writer.go:29: 2021-01-29T19:33:26.253Z [INFO] TestEventList_EventBufOrder: Started HTTP server: address=127.0.0.1:30031 network=tcp writer.go:29: 2021-01-29T19:33:26.254Z [INFO] TestEventList_EventBufOrder: started state syncer writer.go:29: 2021-01-29T19:33:26.302Z [WARN] TestEventList_EventBufOrder.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:26.303Z [INFO] TestEventList_EventBufOrder.server.raft: entering candidate state: node="Node at 127.0.0.1:30035 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:26.342Z [DEBUG] TestEventList_EventBufOrder.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:26.344Z [DEBUG] TestEventList_EventBufOrder.server.raft: vote granted: from=4bb9d8be-9ec2-236f-6fa5-f96e132c69d8 term=2 tally=1 writer.go:29: 2021-01-29T19:33:26.352Z [INFO] TestEventList_EventBufOrder.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:26.365Z [INFO] TestEventList_EventBufOrder.server.raft: entering leader state: leader="Node at 127.0.0.1:30035 [Leader]" writer.go:29: 2021-01-29T19:33:26.369Z [INFO] TestEventList_EventBufOrder.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:26.372Z [INFO] TestEventList_EventBufOrder.server: New leader elected: payload=Node-4bb9d8be-9ec2-236f-6fa5-f96e132c69d8 === CONT TestHealthChecksInState_Filter writer.go:29: 2021-01-29T19:33:26.374Z [INFO] TestHealthChecksInState_Filter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:26.375Z [INFO] TestHealthChecksInState_Filter: Endpoints down --- PASS: TestHealthChecksInState_Filter (0.72s) === CONT TestEventList_Blocking === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.384Z [DEBUG] TestEventList_EventBufOrder.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30035 writer.go:29: 2021-01-29T19:33:26.397Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.409Z [WARN] TestEventList_Blocking: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:26.410Z [DEBUG] TestEventList_Blocking.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:26.430Z [DEBUG] TestEventList_Blocking.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.427Z [INFO] TestEventList_EventBufOrder.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:26.434Z [INFO] TestEventList_EventBufOrder.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.441Z [DEBUG] TestEventList_EventBufOrder.server: Skipping self join check for node since the cluster is too small: node=Node-4bb9d8be-9ec2-236f-6fa5-f96e132c69d8 writer.go:29: 2021-01-29T19:33:26.448Z [INFO] TestEventList_EventBufOrder.server: member joined, marking health alive: member=Node-4bb9d8be-9ec2-236f-6fa5-f96e132c69d8 writer.go:29: 2021-01-29T19:33:26.450Z [INFO] TestEventList_EventBufOrder: Synced node info === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.466Z [INFO] TestEventList_Blocking.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3bd8f76f-23e4-84c9-6949-7b7159b5b54b Address:127.0.0.1:30041}]" writer.go:29: 2021-01-29T19:33:26.471Z [INFO] TestEventList_Blocking.server.raft: entering follower state: follower="Node at 127.0.0.1:30041 [Follower]" leader= writer.go:29: 2021-01-29T19:33:26.473Z [INFO] TestEventList_Blocking.server.serf.wan: serf: EventMemberJoin: Node-3bd8f76f-23e4-84c9-6949-7b7159b5b54b.dc1 127.0.0.1 === CONT TestHealthChecksInState_DistanceSort writer.go:29: 2021-01-29T19:33:26.478Z [INFO] TestHealthChecksInState_DistanceSort: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:26.479Z [INFO] TestHealthChecksInState_DistanceSort: Endpoints down --- PASS: TestHealthChecksInState_DistanceSort (1.10s) === CONT TestEventList_ACLFilter === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.480Z [INFO] TestEventList_Blocking.server.serf.lan: serf: EventMemberJoin: Node-3bd8f76f-23e4-84c9-6949-7b7159b5b54b 127.0.0.1 writer.go:29: 2021-01-29T19:33:26.484Z [INFO] TestEventList_Blocking: Started DNS server: address=127.0.0.1:30036 network=udp writer.go:29: 2021-01-29T19:33:26.487Z [INFO] TestEventList_Blocking.server: Adding LAN server: server="Node-3bd8f76f-23e4-84c9-6949-7b7159b5b54b (Addr: tcp/127.0.0.1:30041) (DC: dc1)" writer.go:29: 2021-01-29T19:33:26.488Z [INFO] TestEventList_Blocking.server: Handled event for server in area: event=member-join server=Node-3bd8f76f-23e4-84c9-6949-7b7159b5b54b.dc1 area=wan === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.489Z [WARN] TestEventList_ACLFilter: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:26.490Z [WARN] TestEventList_ACLFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:26.491Z [DEBUG] TestEventList_ACLFilter.tlsutil: Update: version=1 === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.492Z [INFO] TestEventList_Blocking: Started DNS server: address=127.0.0.1:30036 network=tcp writer.go:29: 2021-01-29T19:33:26.494Z [INFO] TestEventList_Blocking: Started HTTP server: address=127.0.0.1:30037 network=tcp writer.go:29: 2021-01-29T19:33:26.496Z [INFO] TestEventList_Blocking: started state syncer === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.493Z [DEBUG] TestEventList_ACLFilter.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:26.502Z [INFO] TestEventList_ACLFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d182a765-b17e-703e-4185-d96ceb31d798 Address:127.0.0.1:30047}]" writer.go:29: 2021-01-29T19:33:26.504Z [INFO] TestEventList_ACLFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:30047 [Follower]" leader= writer.go:29: 2021-01-29T19:33:26.505Z [INFO] TestEventList_ACLFilter.server.serf.wan: serf: EventMemberJoin: Node-d182a765-b17e-703e-4185-d96ceb31d798.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:26.510Z [INFO] TestEventList_ACLFilter.server.serf.lan: serf: EventMemberJoin: Node-d182a765-b17e-703e-4185-d96ceb31d798 127.0.0.1 writer.go:29: 2021-01-29T19:33:26.513Z [INFO] TestEventList_ACLFilter.server: Handled event for server in area: event=member-join server=Node-d182a765-b17e-703e-4185-d96ceb31d798.dc1 area=wan writer.go:29: 2021-01-29T19:33:26.513Z [INFO] TestEventList_ACLFilter.server: Adding LAN server: server="Node-d182a765-b17e-703e-4185-d96ceb31d798 (Addr: tcp/127.0.0.1:30047) (DC: dc1)" writer.go:29: 2021-01-29T19:33:26.515Z [INFO] TestEventList_ACLFilter: Started DNS server: address=127.0.0.1:30042 network=udp writer.go:29: 2021-01-29T19:33:26.517Z [INFO] TestEventList_ACLFilter: Started DNS server: address=127.0.0.1:30042 network=tcp writer.go:29: 2021-01-29T19:33:26.519Z [INFO] TestEventList_ACLFilter: Started HTTP server: address=127.0.0.1:30043 network=tcp writer.go:29: 2021-01-29T19:33:26.520Z [INFO] TestEventList_ACLFilter: started state syncer === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.537Z [WARN] TestEventList_Blocking.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:26.539Z [INFO] TestEventList_Blocking.server.raft: entering candidate state: node="Node at 127.0.0.1:30041 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:26.542Z [DEBUG] TestEventList_Blocking.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:26.544Z [DEBUG] TestEventList_Blocking.server.raft: vote granted: from=3bd8f76f-23e4-84c9-6949-7b7159b5b54b term=2 tally=1 writer.go:29: 2021-01-29T19:33:26.545Z [INFO] TestEventList_Blocking.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:26.550Z [INFO] TestEventList_Blocking.server.raft: entering leader state: leader="Node at 127.0.0.1:30041 [Leader]" writer.go:29: 2021-01-29T19:33:26.552Z [INFO] TestEventList_Blocking.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:26.553Z [INFO] TestEventList_Blocking.server: New leader elected: payload=Node-3bd8f76f-23e4-84c9-6949-7b7159b5b54b === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:26.559Z [INFO] TestHealthChecksInState_NodeMetaFilter: Waiting for endpoints to shut down === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.557Z [DEBUG] TestEventList_Blocking.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30041 === CONT TestHealthChecksInState_NodeMetaFilter writer.go:29: 2021-01-29T19:33:26.560Z [INFO] TestHealthChecksInState_NodeMetaFilter: Endpoints down --- PASS: TestHealthChecksInState_NodeMetaFilter (0.71s) === CONT TestEventList_Filter === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.565Z [DEBUG] TestEventList_ACLFilter.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.568Z [WARN] TestEventList_Filter: bootstrap = true: do not enable unless necessary === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.567Z [WARN] TestEventList_ACLFilter.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.569Z [DEBUG] TestEventList_Filter.tlsutil: Update: version=1 === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.569Z [INFO] TestEventList_ACLFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:30047 [Candidate]" term=2 === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.570Z [DEBUG] TestEventList_Filter.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.572Z [DEBUG] TestEventList_ACLFilter.server.raft: votes: needed=1 === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.571Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.575Z [DEBUG] TestEventList_ACLFilter.server.raft: vote granted: from=d182a765-b17e-703e-4185-d96ceb31d798 term=2 tally=1 writer.go:29: 2021-01-29T19:33:26.577Z [INFO] TestEventList_ACLFilter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:26.579Z [INFO] TestEventList_ACLFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:30047 [Leader]" writer.go:29: 2021-01-29T19:33:26.581Z [INFO] TestEventList_ACLFilter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:26.582Z [INFO] TestEventList_ACLFilter.server: New leader elected: payload=Node-d182a765-b17e-703e-4185-d96ceb31d798 === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.590Z [INFO] TestEventList_Filter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1e761716-5609-058a-6d48-e7292e1613c8 Address:127.0.0.1:30053}]" === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.583Z [INFO] TestEventList_ACLFilter.server: initializing acls === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.594Z [INFO] TestEventList_Filter.server.raft: entering follower state: follower="Node at 127.0.0.1:30053 [Follower]" leader= === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.599Z [INFO] TestEventList_Blocking.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:26.599Z [INFO] TestEventList_Blocking.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.600Z [DEBUG] TestEventList_Blocking.server: Skipping self join check for node since the cluster is too small: node=Node-3bd8f76f-23e4-84c9-6949-7b7159b5b54b writer.go:29: 2021-01-29T19:33:26.601Z [INFO] TestEventList_Blocking.server: member joined, marking health alive: member=Node-3bd8f76f-23e4-84c9-6949-7b7159b5b54b === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.599Z [INFO] TestEventList_ACLFilter.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:26.603Z [WARN] TestEventList_ACLFilter.server: Configuring a non-UUID master token is deprecated === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.599Z [INFO] TestEventList_Filter.server.serf.wan: serf: EventMemberJoin: Node-1e761716-5609-058a-6d48-e7292e1613c8.dc1 127.0.0.1 === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.608Z [INFO] TestEventList_ACLFilter.server: Bootstrapped ACL master token from configuration === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.607Z [INFO] TestEventList_Filter.server.serf.lan: serf: EventMemberJoin: Node-1e761716-5609-058a-6d48-e7292e1613c8 127.0.0.1 writer.go:29: 2021-01-29T19:33:26.611Z [INFO] TestEventList_Filter.server: Handled event for server in area: event=member-join server=Node-1e761716-5609-058a-6d48-e7292e1613c8.dc1 area=wan === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.611Z [INFO] TestEventList_ACLFilter.server: Created ACL anonymous token from configuration === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.612Z [INFO] TestEventList_Filter.server: Adding LAN server: server="Node-1e761716-5609-058a-6d48-e7292e1613c8 (Addr: tcp/127.0.0.1:30053) (DC: dc1)" === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.613Z [INFO] TestEventList_ACLFilter.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:26.617Z [INFO] TestEventList_ACLFilter.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:26.619Z [INFO] TestEventList_ACLFilter.server.serf.lan: serf: EventMemberUpdate: Node-d182a765-b17e-703e-4185-d96ceb31d798 writer.go:29: 2021-01-29T19:33:26.621Z [INFO] TestEventList_ACLFilter.server: Updating LAN server: server="Node-d182a765-b17e-703e-4185-d96ceb31d798 (Addr: tcp/127.0.0.1:30047) (DC: dc1)" writer.go:29: 2021-01-29T19:33:26.621Z [INFO] TestEventList_ACLFilter.server.serf.wan: serf: EventMemberUpdate: Node-d182a765-b17e-703e-4185-d96ceb31d798.dc1 === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.619Z [INFO] TestEventList_Filter: Started DNS server: address=127.0.0.1:30048 network=tcp === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.631Z [INFO] TestEventList_ACLFilter.server: Handled event for server in area: event=member-update server=Node-d182a765-b17e-703e-4185-d96ceb31d798.dc1 area=wan === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.632Z [INFO] TestEventList_Filter: Started DNS server: address=127.0.0.1:30048 network=udp === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.635Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.638Z [INFO] TestEventList_Filter: Started HTTP server: address=127.0.0.1:30049 network=tcp writer.go:29: 2021-01-29T19:33:26.640Z [INFO] TestEventList_Filter: started state syncer writer.go:29: 2021-01-29T19:33:26.641Z [WARN] TestEventList_Filter.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.645Z [INFO] TestEventList_ACLFilter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:26.646Z [INFO] TestEventList_ACLFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.648Z [DEBUG] TestEventList_ACLFilter.server: Skipping self join check for node since the cluster is too small: node=Node-d182a765-b17e-703e-4185-d96ceb31d798 writer.go:29: 2021-01-29T19:33:26.649Z [INFO] TestEventList_ACLFilter.server: member joined, marking health alive: member=Node-d182a765-b17e-703e-4185-d96ceb31d798 writer.go:29: 2021-01-29T19:33:26.652Z [DEBUG] TestEventList_ACLFilter.server: Skipping self join check for node since the cluster is too small: node=Node-d182a765-b17e-703e-4185-d96ceb31d798 === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.644Z [INFO] TestEventList_Filter.server.raft: entering candidate state: node="Node at 127.0.0.1:30053 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:26.656Z [DEBUG] TestEventList_Filter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:26.658Z [DEBUG] TestEventList_Filter.server.raft: vote granted: from=1e761716-5609-058a-6d48-e7292e1613c8 term=2 tally=1 writer.go:29: 2021-01-29T19:33:26.659Z [INFO] TestEventList_Filter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:26.661Z [INFO] TestEventList_Filter.server.raft: entering leader state: leader="Node at 127.0.0.1:30053 [Leader]" writer.go:29: 2021-01-29T19:33:26.663Z [INFO] TestEventList_Filter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:26.664Z [INFO] TestEventList_Filter.server: New leader elected: payload=Node-1e761716-5609-058a-6d48-e7292e1613c8 writer.go:29: 2021-01-29T19:33:26.665Z [DEBUG] TestEventList_Filter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30053 writer.go:29: 2021-01-29T19:33:26.674Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:26.679Z [INFO] TestEventList_Filter.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.681Z [DEBUG] TestEventList_ACLFilter: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.681Z [INFO] TestEventList_Filter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.683Z [DEBUG] TestEventList_Filter.server: Skipping self join check for node since the cluster is too small: node=Node-1e761716-5609-058a-6d48-e7292e1613c8 === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.684Z [INFO] TestEventList_ACLFilter: Synced node info === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.685Z [INFO] TestEventList_Filter.server: member joined, marking health alive: member=Node-1e761716-5609-058a-6d48-e7292e1613c8 === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.686Z [DEBUG] TestEventList_ACLFilter: Node info in sync === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.692Z [DEBUG] TestEventList_EventBufOrder.server: User event: event=foo writer.go:29: 2021-01-29T19:33:26.695Z [DEBUG] TestEventList_EventBufOrder.server: User event: event=bar writer.go:29: 2021-01-29T19:33:26.698Z [DEBUG] TestEventList_EventBufOrder.server: User event: event=foo writer.go:29: 2021-01-29T19:33:26.696Z [DEBUG] TestEventList_EventBufOrder: new event: event_name=foo event_id=b69e5ffe-562e-0162-1ccb-d3fcff26f55e === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.703Z [DEBUG] TestEventList_Filter.server: User event: event=test === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.704Z [DEBUG] TestEventList_EventBufOrder: new event: event_name=bar event_id=e28bf062-bf67-6265-dbd0-c558359e6574 writer.go:29: 2021-01-29T19:33:26.704Z [DEBUG] TestEventList_EventBufOrder: new event: event_name=foo event_id=c3cf1143-4fa5-4a06-9105-2d57e7daf4b3 === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.704Z [DEBUG] TestEventList_Filter.server: User event: event=foo === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.701Z [DEBUG] TestEventList_EventBufOrder.server: User event: event=foo writer.go:29: 2021-01-29T19:33:26.706Z [DEBUG] TestEventList_EventBufOrder.server: User event: event=bar writer.go:29: 2021-01-29T19:33:26.707Z [DEBUG] TestEventList_EventBufOrder: new event: event_name=foo event_id=da7a8e67-f05d-2934-efe3-ac9facb42a4c === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.706Z [DEBUG] TestEventList_Filter: new event: event_name=test event_id=15c70e44-f7da-aa07-1a34-4566576718f1 === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.707Z [DEBUG] TestEventList_EventBufOrder: new event: event_name=bar event_id=0de7b790-35fc-b7ce-d649-5a049ebcdbc8 === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.708Z [DEBUG] TestEventList_Filter: new event: event_name=foo event_id=144b6898-6af6-db54-54de-28198c145da7 === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.719Z [INFO] TestEventList_EventBufOrder: Requesting shutdown writer.go:29: 2021-01-29T19:33:26.720Z [INFO] TestEventList_EventBufOrder.server: shutting down server writer.go:29: 2021-01-29T19:33:26.721Z [DEBUG] TestEventList_EventBufOrder.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.722Z [WARN] TestEventList_EventBufOrder.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.722Z [DEBUG] TestEventList_EventBufOrder.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.725Z [WARN] TestEventList_EventBufOrder.server.serf.wan: serf: Shutdown without a Leave === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.728Z [INFO] TestEventList_Filter: Requesting shutdown === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.729Z [INFO] TestEventList_EventBufOrder.server.router.manager: shutting down === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.729Z [INFO] TestEventList_Filter.server: shutting down server === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.730Z [INFO] TestEventList_EventBufOrder: consul server down === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.731Z [DEBUG] TestEventList_Filter.leader: stopping routine: routine="CA root pruning" === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.732Z [INFO] TestEventList_EventBufOrder: shutdown complete === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.732Z [WARN] TestEventList_Filter.server.serf.lan: serf: Shutdown without a Leave === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.733Z [INFO] TestEventList_EventBufOrder: Stopping server: protocol=DNS address=127.0.0.1:30030 network=tcp === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.731Z [ERROR] TestEventList_Filter.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:26.735Z [INFO] TestEventList_EventBufOrder: Stopping server: protocol=DNS address=127.0.0.1:30030 network=udp writer.go:29: 2021-01-29T19:33:26.736Z [INFO] TestEventList_EventBufOrder: Stopping server: protocol=HTTP address=127.0.0.1:30031 network=tcp === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:26.733Z [DEBUG] TestEventList_Filter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.739Z [WARN] TestEventList_Filter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.741Z [INFO] TestEventList_Filter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:26.742Z [INFO] TestEventList_Filter: consul server down writer.go:29: 2021-01-29T19:33:26.744Z [INFO] TestEventList_Filter: shutdown complete writer.go:29: 2021-01-29T19:33:26.745Z [INFO] TestEventList_Filter: Stopping server: protocol=DNS address=127.0.0.1:30048 network=tcp writer.go:29: 2021-01-29T19:33:26.746Z [INFO] TestEventList_Filter: Stopping server: protocol=DNS address=127.0.0.1:30048 network=udp writer.go:29: 2021-01-29T19:33:26.747Z [INFO] TestEventList_Filter: Stopping server: protocol=HTTP address=127.0.0.1:30049 network=tcp === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.829Z [DEBUG] TestEventList_Blocking.server: User event: event=test writer.go:29: 2021-01-29T19:33:26.831Z [DEBUG] TestEventList_Blocking: new event: event_name=test event_id=b7fbd385-532b-55ad-3ae2-d7516c69efe9 writer.go:29: 2021-01-29T19:33:26.830Z [DEBUG] TestEventList_Blocking: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.833Z [DEBUG] TestEventList_ACLFilter.acl: dropping node from result due to ACLs: node=Node-d182a765-b17e-703e-4185-d96ceb31d798 writer.go:29: 2021-01-29T19:33:26.838Z [DEBUG] TestEventList_ACLFilter.acl: dropping node from result due to ACLs: node=Node-d182a765-b17e-703e-4185-d96ceb31d798 === RUN TestEventList_ACLFilter/no_token === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.841Z [DEBUG] TestEventList_ACLFilter.server: User event: event=foo writer.go:29: 2021-01-29T19:33:26.842Z [DEBUG] TestEventList_ACLFilter: new event: event_name=foo event_id=b4d744bc-87ac-8e9b-2ea0-db6f115f6f48 === RUN TestEventList_ACLFilter/root_token === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.847Z [INFO] TestEventList_ACLFilter: Requesting shutdown writer.go:29: 2021-01-29T19:33:26.848Z [INFO] TestEventList_ACLFilter.server: shutting down server === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.850Z [INFO] TestEventList_Blocking: Synced node info === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.850Z [DEBUG] TestEventList_ACLFilter.leader: stopping routine: routine="CA root pruning" === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.852Z [DEBUG] TestEventList_Blocking: Node info in sync === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:26.853Z [DEBUG] TestEventList_ACLFilter.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:26.855Z [DEBUG] TestEventList_ACLFilter.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:26.853Z [DEBUG] TestEventList_ACLFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.855Z [DEBUG] TestEventList_ACLFilter.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:26.857Z [WARN] TestEventList_ACLFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.857Z [DEBUG] TestEventList_ACLFilter.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:26.864Z [WARN] TestEventList_ACLFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.868Z [INFO] TestEventList_ACLFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:26.868Z [INFO] TestEventList_ACLFilter: consul server down writer.go:29: 2021-01-29T19:33:26.871Z [INFO] TestEventList_ACLFilter: shutdown complete writer.go:29: 2021-01-29T19:33:26.872Z [INFO] TestEventList_ACLFilter: Stopping server: protocol=DNS address=127.0.0.1:30042 network=tcp writer.go:29: 2021-01-29T19:33:26.874Z [INFO] TestEventList_ACLFilter: Stopping server: protocol=DNS address=127.0.0.1:30042 network=udp writer.go:29: 2021-01-29T19:33:26.876Z [INFO] TestEventList_ACLFilter: Stopping server: protocol=HTTP address=127.0.0.1:30043 network=tcp === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:26.883Z [DEBUG] TestEventList_Blocking.server: User event: event=second writer.go:29: 2021-01-29T19:33:26.885Z [DEBUG] TestEventList_Blocking: new event: event_name=second event_id=fe1cd283-6e03-f82d-61b5-e5b4f632e54f writer.go:29: 2021-01-29T19:33:26.913Z [INFO] TestEventList_Blocking: Requesting shutdown writer.go:29: 2021-01-29T19:33:26.915Z [INFO] TestEventList_Blocking.server: shutting down server writer.go:29: 2021-01-29T19:33:26.917Z [DEBUG] TestEventList_Blocking.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.919Z [WARN] TestEventList_Blocking.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.919Z [DEBUG] TestEventList_Blocking.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.922Z [WARN] TestEventList_Blocking.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.926Z [INFO] TestEventList_Blocking.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:26.927Z [INFO] TestEventList_Blocking: consul server down writer.go:29: 2021-01-29T19:33:26.929Z [INFO] TestEventList_Blocking: shutdown complete writer.go:29: 2021-01-29T19:33:26.931Z [INFO] TestEventList_Blocking: Stopping server: protocol=DNS address=127.0.0.1:30036 network=tcp writer.go:29: 2021-01-29T19:33:26.933Z [INFO] TestEventList_Blocking: Stopping server: protocol=DNS address=127.0.0.1:30036 network=udp writer.go:29: 2021-01-29T19:33:26.935Z [INFO] TestEventList_Blocking: Stopping server: protocol=HTTP address=127.0.0.1:30037 network=tcp === CONT TestEventList_EventBufOrder writer.go:29: 2021-01-29T19:33:27.238Z [INFO] TestEventList_EventBufOrder: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:27.240Z [INFO] TestEventList_EventBufOrder: Endpoints down --- PASS: TestEventList_EventBufOrder (1.10s) === CONT TestEventList === CONT TestEventList_Filter writer.go:29: 2021-01-29T19:33:27.249Z [INFO] TestEventList_Filter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:27.251Z [INFO] TestEventList_Filter: Endpoints down --- PASS: TestEventList_Filter (0.69s) === CONT TestEventFire_token === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.264Z [WARN] TestEventList: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:27.266Z [DEBUG] TestEventList.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:27.271Z [DEBUG] TestEventList.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.275Z [WARN] TestEventFire_token: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:27.276Z [WARN] TestEventFire_token: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:27.278Z [DEBUG] TestEventFire_token.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:27.280Z [DEBUG] TestEventFire_token.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:27.310Z [INFO] TestEventFire_token.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:adaffc0f-33ef-113c-91cd-44a2f61c2efc Address:127.0.0.1:30065}]" === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.314Z [INFO] TestEventList.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ea01e3ba-9f22-f994-1ac2-0acf0c74edaa Address:127.0.0.1:30059}]" writer.go:29: 2021-01-29T19:33:27.316Z [INFO] TestEventList.server.raft: entering follower state: follower="Node at 127.0.0.1:30059 [Follower]" leader= === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.313Z [INFO] TestEventFire_token.server.raft: entering follower state: follower="Node at 127.0.0.1:30065 [Follower]" leader= === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.317Z [INFO] TestEventList.server.serf.wan: serf: EventMemberJoin: Node-ea01e3ba-9f22-f994-1ac2-0acf0c74edaa.dc1 127.0.0.1 === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.314Z [INFO] TestEventFire_token.server.serf.wan: serf: EventMemberJoin: Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc.dc1 127.0.0.1 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.340Z [INFO] TestEventList.server.serf.lan: serf: EventMemberJoin: Node-ea01e3ba-9f22-f994-1ac2-0acf0c74edaa 127.0.0.1 writer.go:29: 2021-01-29T19:33:27.343Z [INFO] TestEventList.server: Adding LAN server: server="Node-ea01e3ba-9f22-f994-1ac2-0acf0c74edaa (Addr: tcp/127.0.0.1:30059) (DC: dc1)" === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.344Z [INFO] TestEventFire_token.server.serf.lan: serf: EventMemberJoin: Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc 127.0.0.1 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.343Z [INFO] TestEventList: Started DNS server: address=127.0.0.1:30054 network=udp writer.go:29: 2021-01-29T19:33:27.345Z [INFO] TestEventList.server: Handled event for server in area: event=member-join server=Node-ea01e3ba-9f22-f994-1ac2-0acf0c74edaa.dc1 area=wan === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.347Z [INFO] TestEventFire_token: Started DNS server: address=127.0.0.1:30060 network=udp writer.go:29: 2021-01-29T19:33:27.348Z [INFO] TestEventFire_token.server: Adding LAN server: server="Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc (Addr: tcp/127.0.0.1:30065) (DC: dc1)" writer.go:29: 2021-01-29T19:33:27.350Z [INFO] TestEventFire_token.server: Handled event for server in area: event=member-join server=Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc.dc1 area=wan writer.go:29: 2021-01-29T19:33:27.349Z [INFO] TestEventFire_token: Started DNS server: address=127.0.0.1:30060 network=tcp === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.347Z [INFO] TestEventList: Started DNS server: address=127.0.0.1:30054 network=tcp === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.352Z [INFO] TestEventFire_token: Started HTTP server: address=127.0.0.1:30061 network=tcp writer.go:29: 2021-01-29T19:33:27.354Z [INFO] TestEventFire_token: started state syncer === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.356Z [INFO] TestEventList: Started HTTP server: address=127.0.0.1:30055 network=tcp writer.go:29: 2021-01-29T19:33:27.367Z [INFO] TestEventList: started state syncer === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:27.378Z [INFO] TestEventList_ACLFilter: Waiting for endpoints to shut down === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.381Z [WARN] TestEventFire_token.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:27.382Z [INFO] TestEventFire_token.server.raft: entering candidate state: node="Node at 127.0.0.1:30065 [Candidate]" term=2 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.387Z [WARN] TestEventList.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:27.389Z [INFO] TestEventList.server.raft: entering candidate state: node="Node at 127.0.0.1:30059 [Candidate]" term=2 === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.387Z [DEBUG] TestEventFire_token.server.raft: votes: needed=1 === CONT TestEventList_ACLFilter writer.go:29: 2021-01-29T19:33:27.380Z [INFO] TestEventList_ACLFilter: Endpoints down --- PASS: TestEventList_ACLFilter (0.92s) --- PASS: TestEventList_ACLFilter/no_token (0.00s) --- PASS: TestEventList_ACLFilter/root_token (0.00s) === CONT TestEventFire === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.392Z [DEBUG] TestEventFire_token.server.raft: vote granted: from=adaffc0f-33ef-113c-91cd-44a2f61c2efc term=2 tally=1 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.400Z [DEBUG] TestEventList.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:27.401Z [DEBUG] TestEventList.server.raft: vote granted: from=ea01e3ba-9f22-f994-1ac2-0acf0c74edaa term=2 tally=1 writer.go:29: 2021-01-29T19:33:27.403Z [INFO] TestEventList.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:27.404Z [INFO] TestEventList.server.raft: entering leader state: leader="Node at 127.0.0.1:30059 [Leader]" writer.go:29: 2021-01-29T19:33:27.406Z [INFO] TestEventList.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:27.407Z [INFO] TestEventList.server: New leader elected: payload=Node-ea01e3ba-9f22-f994-1ac2-0acf0c74edaa === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.401Z [INFO] TestEventFire_token.server.raft: election won: tally=1 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.408Z [DEBUG] TestEventList.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30059 === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.411Z [INFO] TestEventFire_token.server.raft: entering leader state: leader="Node at 127.0.0.1:30065 [Leader]" === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.425Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.401Z [DEBUG] TestEventFire_token.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.432Z [INFO] TestEventList.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:27.434Z [INFO] TestEventList.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:27.435Z [DEBUG] TestEventList.server: Skipping self join check for node since the cluster is too small: node=Node-ea01e3ba-9f22-f994-1ac2-0acf0c74edaa === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:27.437Z [INFO] TestEventList_Blocking: Waiting for endpoints to shut down === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.436Z [INFO] TestEventList.server: member joined, marking health alive: member=Node-ea01e3ba-9f22-f994-1ac2-0acf0c74edaa === CONT TestEventList_Blocking writer.go:29: 2021-01-29T19:33:27.438Z [INFO] TestEventList_Blocking: Endpoints down --- PASS: TestEventList_Blocking (1.06s) === CONT TestDNS_ConfigReload === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.432Z [INFO] TestEventFire_token.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:27.433Z [INFO] TestEventFire_token.server: New leader elected: payload=Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.449Z [WARN] TestDNS_ConfigReload: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:27.450Z [DEBUG] TestDNS_ConfigReload.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:27.452Z [DEBUG] TestDNS_ConfigReload.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:27.457Z [INFO] TestDNS_ConfigReload.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e4d42275-17af-2add-0141-7d62f8b2d0bf Address:127.0.0.1:30077}]" === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.433Z [INFO] TestEventFire_token.server: initializing acls === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.459Z [DEBUG] TestEventList: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.459Z [INFO] TestDNS_ConfigReload.server.raft: entering follower state: follower="Node at 127.0.0.1:30077 [Follower]" leader= === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.461Z [INFO] TestEventList: Synced node info === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.459Z [INFO] TestDNS_ConfigReload.server.serf.wan: serf: EventMemberJoin: Node-e4d42275-17af-2add-0141-7d62f8b2d0bf.dc1 127.0.0.1 === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.466Z [WARN] TestEventFire: bootstrap = true: do not enable unless necessary === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.466Z [INFO] TestDNS_ConfigReload.server.serf.lan: serf: EventMemberJoin: Node-e4d42275-17af-2add-0141-7d62f8b2d0bf 127.0.0.1 === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.467Z [DEBUG] TestEventFire.tlsutil: Update: version=1 === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.464Z [INFO] TestEventFire_token.server: Created ACL 'global-management' policy === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.469Z [INFO] TestDNS_ConfigReload.server: Adding LAN server: server="Node-e4d42275-17af-2add-0141-7d62f8b2d0bf (Addr: tcp/127.0.0.1:30077) (DC: dc1)" === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.469Z [DEBUG] TestEventFire.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.470Z [DEBUG] TestDNS_ConfigReload.dns: recursor enabled writer.go:29: 2021-01-29T19:33:27.471Z [INFO] TestDNS_ConfigReload: Started DNS server: address=127.0.0.1:30072 network=udp writer.go:29: 2021-01-29T19:33:27.472Z [DEBUG] TestDNS_ConfigReload.dns: recursor enabled writer.go:29: 2021-01-29T19:33:27.473Z [INFO] TestDNS_ConfigReload: Started DNS server: address=127.0.0.1:30072 network=tcp writer.go:29: 2021-01-29T19:33:27.475Z [INFO] TestDNS_ConfigReload: Started HTTP server: address=127.0.0.1:30073 network=tcp writer.go:29: 2021-01-29T19:33:27.477Z [INFO] TestDNS_ConfigReload: started state syncer writer.go:29: 2021-01-29T19:33:27.482Z [INFO] TestDNS_ConfigReload.server: Handled event for server in area: event=member-join server=Node-e4d42275-17af-2add-0141-7d62f8b2d0bf.dc1 area=wan === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.469Z [WARN] TestEventFire_token.server: Configuring a non-UUID master token is deprecated === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.493Z [INFO] TestEventFire.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bb237dec-db7e-eb94-04bb-2416344c4684 Address:127.0.0.1:30071}]" writer.go:29: 2021-01-29T19:33:27.495Z [INFO] TestEventFire.server.raft: entering follower state: follower="Node at 127.0.0.1:30071 [Follower]" leader= === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.494Z [INFO] TestEventFire_token.server: Bootstrapped ACL master token from configuration === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.496Z [INFO] TestEventFire.server.serf.wan: serf: EventMemberJoin: Node-bb237dec-db7e-eb94-04bb-2416344c4684.dc1 127.0.0.1 === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.500Z [INFO] TestEventFire_token.server: initializing acls === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.501Z [DEBUG] TestEventList.server: User event: event=test === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.501Z [WARN] TestEventFire_token.server: Configuring a non-UUID master token is deprecated === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.503Z [DEBUG] TestEventList: new event: event_name=test event_id=932185d9-016f-c6b1-d207-92fd36116a03 === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.502Z [INFO] TestEventFire.server.serf.lan: serf: EventMemberJoin: Node-bb237dec-db7e-eb94-04bb-2416344c4684 127.0.0.1 writer.go:29: 2021-01-29T19:33:27.507Z [INFO] TestEventFire: Started DNS server: address=127.0.0.1:30066 network=udp writer.go:29: 2021-01-29T19:33:27.508Z [INFO] TestEventFire.server: Adding LAN server: server="Node-bb237dec-db7e-eb94-04bb-2416344c4684 (Addr: tcp/127.0.0.1:30071) (DC: dc1)" writer.go:29: 2021-01-29T19:33:27.510Z [INFO] TestEventFire.server: Handled event for server in area: event=member-join server=Node-bb237dec-db7e-eb94-04bb-2416344c4684.dc1 area=wan writer.go:29: 2021-01-29T19:33:27.513Z [INFO] TestEventFire: Started DNS server: address=127.0.0.1:30066 network=tcp === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.511Z [WARN] TestDNS_ConfigReload.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.518Z [INFO] TestEventFire: Started HTTP server: address=127.0.0.1:30067 network=tcp === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.516Z [INFO] TestDNS_ConfigReload.server.raft: entering candidate state: node="Node at 127.0.0.1:30077 [Candidate]" term=2 === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.520Z [INFO] TestEventFire: started state syncer === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.523Z [INFO] TestEventFire_token.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:27.524Z [INFO] TestEventFire_token.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:27.526Z [INFO] TestEventFire_token.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:27.528Z [INFO] TestEventFire_token.server.serf.lan: serf: EventMemberUpdate: Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc writer.go:29: 2021-01-29T19:33:27.530Z [INFO] TestEventFire_token.server.serf.wan: serf: EventMemberUpdate: Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc.dc1 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.533Z [INFO] TestEventList: Requesting shutdown writer.go:29: 2021-01-29T19:33:27.535Z [INFO] TestEventList.server: shutting down server === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.531Z [INFO] TestEventFire_token.server: Created ACL anonymous token from configuration === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.536Z [DEBUG] TestEventList.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.532Z [DEBUG] TestDNS_ConfigReload.server.raft: votes: needed=1 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.537Z [WARN] TestEventList.server.serf.lan: serf: Shutdown without a Leave === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.536Z [DEBUG] TestEventFire_token.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:33:27.533Z [INFO] TestEventFire_token.server: Updating LAN server: server="Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc (Addr: tcp/127.0.0.1:30065) (DC: dc1)" === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.538Z [DEBUG] TestDNS_ConfigReload.server.raft: vote granted: from=e4d42275-17af-2add-0141-7d62f8b2d0bf term=2 tally=1 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.537Z [DEBUG] TestEventList.leader: stopped routine: routine="CA root pruning" === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.535Z [INFO] TestEventFire_token.server: Handled event for server in area: event=member-update server=Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc.dc1 area=wan === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.546Z [INFO] TestDNS_ConfigReload.server.raft: election won: tally=1 === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.538Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.552Z [WARN] TestEventList.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.550Z [INFO] TestDNS_ConfigReload.server.raft: entering leader state: leader="Node at 127.0.0.1:30077 [Leader]" === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.557Z [INFO] TestEventList.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:27.559Z [INFO] TestEventList: consul server down writer.go:29: 2021-01-29T19:33:27.560Z [INFO] TestEventList: shutdown complete === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.562Z [WARN] TestEventFire.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.562Z [INFO] TestDNS_ConfigReload.server: cluster leadership acquired === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.562Z [INFO] TestEventFire.server.raft: entering candidate state: node="Node at 127.0.0.1:30071 [Candidate]" term=2 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.561Z [INFO] TestEventList: Stopping server: protocol=DNS address=127.0.0.1:30054 network=tcp === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.563Z [INFO] TestDNS_ConfigReload.server: New leader elected: payload=Node-e4d42275-17af-2add-0141-7d62f8b2d0bf === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.564Z [DEBUG] TestEventFire.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:27.565Z [DEBUG] TestEventFire.server.raft: vote granted: from=bb237dec-db7e-eb94-04bb-2416344c4684 term=2 tally=1 writer.go:29: 2021-01-29T19:33:27.566Z [INFO] TestEventFire.server.raft: election won: tally=1 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.564Z [INFO] TestEventList: Stopping server: protocol=DNS address=127.0.0.1:30054 network=udp === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.566Z [INFO] TestEventFire.server.raft: entering leader state: leader="Node at 127.0.0.1:30071 [Leader]" writer.go:29: 2021-01-29T19:33:27.568Z [INFO] TestEventFire.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:27.569Z [INFO] TestEventFire.server: New leader elected: payload=Node-bb237dec-db7e-eb94-04bb-2416344c4684 === CONT TestEventList writer.go:29: 2021-01-29T19:33:27.567Z [INFO] TestEventList: Stopping server: protocol=HTTP address=127.0.0.1:30055 network=tcp === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.540Z [INFO] TestEventFire_token.server.serf.lan: serf: EventMemberUpdate: Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc writer.go:29: 2021-01-29T19:33:27.571Z [INFO] TestEventFire_token.server.serf.wan: serf: EventMemberUpdate: Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc.dc1 writer.go:29: 2021-01-29T19:33:27.565Z [INFO] TestEventFire_token.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:27.571Z [INFO] TestEventFire_token.server: Updating LAN server: server="Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc (Addr: tcp/127.0.0.1:30065) (DC: dc1)" writer.go:29: 2021-01-29T19:33:27.572Z [INFO] TestEventFire_token.server: Handled event for server in area: event=member-update server=Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc.dc1 area=wan writer.go:29: 2021-01-29T19:33:27.573Z [INFO] TestEventFire_token.leader: started routine: routine="CA root pruning" === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.567Z [DEBUG] TestDNS_ConfigReload.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30077 === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.575Z [DEBUG] TestEventFire_token.server: Skipping self join check for node since the cluster is too small: node=Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc writer.go:29: 2021-01-29T19:33:27.576Z [INFO] TestEventFire_token.server: member joined, marking health alive: member=Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.577Z [DEBUG] TestEventFire.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30071 === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.581Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.597Z [DEBUG] TestEventFire_token.server: Skipping self join check for node since the cluster is too small: node=Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.598Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.597Z [INFO] TestDNS_ConfigReload.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.598Z [DEBUG] TestEventFire_token.server: Skipping self join check for node since the cluster is too small: node=Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc writer.go:29: 2021-01-29T19:33:27.600Z [DEBUG] TestEventFire_token.acl: dropping node from result due to ACLs: node=Node-adaffc0f-33ef-113c-91cd-44a2f61c2efc === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.599Z [INFO] TestDNS_ConfigReload.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:27.603Z [DEBUG] TestDNS_ConfigReload.server: Skipping self join check for node since the cluster is too small: node=Node-e4d42275-17af-2add-0141-7d62f8b2d0bf === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.607Z [WARN] TestEventFire_token.server.internal: user event blocked by ACLs: event=foo accessorID= === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.605Z [INFO] TestDNS_ConfigReload.server: member joined, marking health alive: member=Node-e4d42275-17af-2add-0141-7d62f8b2d0bf === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.608Z [WARN] TestEventFire_token.server.internal: user event blocked by ACLs: event=bar accessorID= writer.go:29: 2021-01-29T19:33:27.610Z [INFO] TestEventFire_token: Requesting shutdown === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.612Z [INFO] TestEventFire.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.610Z [DEBUG] TestEventFire_token.server: User event: event=baz === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.613Z [INFO] TestEventFire.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:27.613Z [DEBUG] TestEventFire.server: Skipping self join check for node since the cluster is too small: node=Node-bb237dec-db7e-eb94-04bb-2416344c4684 === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.610Z [INFO] TestEventFire_token.server: shutting down server === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.614Z [INFO] TestEventFire.server: member joined, marking health alive: member=Node-bb237dec-db7e-eb94-04bb-2416344c4684 === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.615Z [DEBUG] TestEventFire_token.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:27.617Z [DEBUG] TestEventFire_token.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:27.614Z [DEBUG] TestEventFire_token: new event: event_name=baz event_id=90064686-50a2-e79f-d777-c29acb47d730 writer.go:29: 2021-01-29T19:33:27.615Z [ERROR] TestEventFire_token.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:27.617Z [DEBUG] TestEventFire_token.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:27.619Z [DEBUG] TestEventFire_token.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.626Z [DEBUG] TestDNS_ConfigReload: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.619Z [DEBUG] TestEventFire_token.leader: stopped routine: routine="acl token reaping" === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.627Z [INFO] TestDNS_ConfigReload: Synced node info === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:27.626Z [WARN] TestEventFire_token.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:27.626Z [DEBUG] TestEventFire_token.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:27.631Z [WARN] TestEventFire_token.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:27.633Z [INFO] TestEventFire_token.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:27.633Z [INFO] TestEventFire_token: consul server down writer.go:29: 2021-01-29T19:33:27.636Z [INFO] TestEventFire_token: shutdown complete writer.go:29: 2021-01-29T19:33:27.637Z [INFO] TestEventFire_token: Stopping server: protocol=DNS address=127.0.0.1:30060 network=tcp writer.go:29: 2021-01-29T19:33:27.639Z [INFO] TestEventFire_token: Stopping server: protocol=DNS address=127.0.0.1:30060 network=udp writer.go:29: 2021-01-29T19:33:27.640Z [INFO] TestEventFire_token: Stopping server: protocol=HTTP address=127.0.0.1:30061 network=tcp === CONT TestDNS_ConfigReload writer.go:29: 2021-01-29T19:33:27.778Z [DEBUG] TestDNS_ConfigReload.tlsutil: Update: version=2 writer.go:29: 2021-01-29T19:33:27.835Z [INFO] TestDNS_ConfigReload: Requesting shutdown writer.go:29: 2021-01-29T19:33:27.848Z [INFO] TestDNS_ConfigReload.server: shutting down server writer.go:29: 2021-01-29T19:33:27.849Z [DEBUG] TestDNS_ConfigReload.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:27.851Z [WARN] TestDNS_ConfigReload.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:27.851Z [DEBUG] TestDNS_ConfigReload.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:27.855Z [WARN] TestDNS_ConfigReload.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:27.859Z [INFO] TestDNS_ConfigReload.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:27.859Z [INFO] TestDNS_ConfigReload: consul server down writer.go:29: 2021-01-29T19:33:27.862Z [INFO] TestDNS_ConfigReload: shutdown complete writer.go:29: 2021-01-29T19:33:27.864Z [INFO] TestDNS_ConfigReload: Stopping server: protocol=DNS address=127.0.0.1:30072 network=tcp writer.go:29: 2021-01-29T19:33:27.866Z [INFO] TestDNS_ConfigReload: Stopping server: protocol=DNS address=127.0.0.1:30072 network=udp writer.go:29: 2021-01-29T19:33:27.867Z [INFO] TestDNS_ConfigReload: Stopping server: protocol=HTTP address=127.0.0.1:30073 network=tcp writer.go:29: 2021-01-29T19:33:27.869Z [INFO] TestDNS_ConfigReload: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:27.871Z [INFO] TestDNS_ConfigReload: Endpoints down --- PASS: TestDNS_ConfigReload (0.44s) === CONT TestDNS_trimUDPResponse_TrimSizeEDNS === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.882Z [DEBUG] TestEventFire: Skipping remote check since it is managed automatically: check=serfHealth --- PASS: TestDNS_trimUDPResponse_TrimSizeEDNS (0.01s) === CONT TestDNS_Compression_Recurse === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.886Z [INFO] TestEventFire: Synced node info writer.go:29: 2021-01-29T19:33:27.888Z [DEBUG] TestEventFire: Node info in sync === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:27.900Z [WARN] TestDNS_Compression_Recurse: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:27.902Z [DEBUG] TestDNS_Compression_Recurse.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:27.904Z [DEBUG] TestDNS_Compression_Recurse.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:27.930Z [INFO] TestDNS_Compression_Recurse.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:987ed4ce-f600-38ae-0ae2-797ab5d32764 Address:127.0.0.1:30083}]" writer.go:29: 2021-01-29T19:33:27.933Z [INFO] TestDNS_Compression_Recurse.server.raft: entering follower state: follower="Node at 127.0.0.1:30083 [Follower]" leader= === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.941Z [INFO] TestEventFire: Requesting shutdown writer.go:29: 2021-01-29T19:33:27.942Z [DEBUG] TestEventFire.server: User event: event=test writer.go:29: 2021-01-29T19:33:27.944Z [INFO] TestEventFire.server: shutting down server writer.go:29: 2021-01-29T19:33:27.947Z [DEBUG] TestEventFire.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:27.935Z [INFO] TestDNS_Compression_Recurse.server.serf.wan: serf: EventMemberJoin: Node-987ed4ce-f600-38ae-0ae2-797ab5d32764.dc1 127.0.0.1 === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.948Z [WARN] TestEventFire.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:27.948Z [DEBUG] TestEventFire.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:27.952Z [WARN] TestEventFire.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:27.956Z [INFO] TestEventFire.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:27.956Z [INFO] TestEventFire: consul server down writer.go:29: 2021-01-29T19:33:27.958Z [INFO] TestEventFire: shutdown complete writer.go:29: 2021-01-29T19:33:27.959Z [INFO] TestEventFire: Stopping server: protocol=DNS address=127.0.0.1:30066 network=tcp === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:27.957Z [INFO] TestDNS_Compression_Recurse.server.serf.lan: serf: EventMemberJoin: Node-987ed4ce-f600-38ae-0ae2-797ab5d32764 127.0.0.1 === CONT TestEventFire writer.go:29: 2021-01-29T19:33:27.960Z [INFO] TestEventFire: Stopping server: protocol=DNS address=127.0.0.1:30066 network=udp writer.go:29: 2021-01-29T19:33:27.961Z [INFO] TestEventFire: Stopping server: protocol=HTTP address=127.0.0.1:30067 network=tcp === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:27.962Z [INFO] TestDNS_Compression_Recurse.server: Handled event for server in area: event=member-join server=Node-987ed4ce-f600-38ae-0ae2-797ab5d32764.dc1 area=wan writer.go:29: 2021-01-29T19:33:27.962Z [INFO] TestDNS_Compression_Recurse.server: Adding LAN server: server="Node-987ed4ce-f600-38ae-0ae2-797ab5d32764 (Addr: tcp/127.0.0.1:30083) (DC: dc1)" writer.go:29: 2021-01-29T19:33:27.963Z [DEBUG] TestDNS_Compression_Recurse.dns: recursor enabled writer.go:29: 2021-01-29T19:33:27.963Z [DEBUG] TestDNS_Compression_Recurse.dns: recursor enabled writer.go:29: 2021-01-29T19:33:27.967Z [INFO] TestDNS_Compression_Recurse: Started DNS server: address=127.0.0.1:30078 network=udp writer.go:29: 2021-01-29T19:33:27.971Z [INFO] TestDNS_Compression_Recurse: Started DNS server: address=127.0.0.1:30078 network=tcp writer.go:29: 2021-01-29T19:33:27.973Z [INFO] TestDNS_Compression_Recurse: Started HTTP server: address=127.0.0.1:30079 network=tcp writer.go:29: 2021-01-29T19:33:27.974Z [INFO] TestDNS_Compression_Recurse: started state syncer writer.go:29: 2021-01-29T19:33:27.985Z [WARN] TestDNS_Compression_Recurse.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:27.986Z [INFO] TestDNS_Compression_Recurse.server.raft: entering candidate state: node="Node at 127.0.0.1:30083 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:27.989Z [DEBUG] TestDNS_Compression_Recurse.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:27.990Z [DEBUG] TestDNS_Compression_Recurse.server.raft: vote granted: from=987ed4ce-f600-38ae-0ae2-797ab5d32764 term=2 tally=1 writer.go:29: 2021-01-29T19:33:27.992Z [INFO] TestDNS_Compression_Recurse.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:27.993Z [INFO] TestDNS_Compression_Recurse.server.raft: entering leader state: leader="Node at 127.0.0.1:30083 [Leader]" writer.go:29: 2021-01-29T19:33:27.995Z [INFO] TestDNS_Compression_Recurse.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:27.996Z [INFO] TestDNS_Compression_Recurse.server: New leader elected: payload=Node-987ed4ce-f600-38ae-0ae2-797ab5d32764 writer.go:29: 2021-01-29T19:33:27.997Z [DEBUG] TestDNS_Compression_Recurse.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30083 writer.go:29: 2021-01-29T19:33:28.014Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:28.044Z [INFO] TestDNS_Compression_Recurse.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:28.047Z [INFO] TestDNS_Compression_Recurse.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:28.052Z [DEBUG] TestDNS_Compression_Recurse.server: Skipping self join check for node since the cluster is too small: node=Node-987ed4ce-f600-38ae-0ae2-797ab5d32764 writer.go:29: 2021-01-29T19:33:28.056Z [INFO] TestDNS_Compression_Recurse.server: member joined, marking health alive: member=Node-987ed4ce-f600-38ae-0ae2-797ab5d32764 === CONT TestEventList writer.go:29: 2021-01-29T19:33:28.070Z [INFO] TestEventList: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:28.072Z [INFO] TestEventList: Endpoints down --- PASS: TestEventList (0.83s) === CONT TestDNS_Compression_ReverseLookup === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:28.094Z [DEBUG] TestDNS_Compression_Recurse.dns: recurse succeeded for question: question="{apple.com. 255 1}" rtt=728.831µs recursor=127.0.0.1:42447 writer.go:29: 2021-01-29T19:33:28.095Z [DEBUG] TestDNS_Compression_Recurse.dns: request served from client: question="{apple.com. 255 1}" network=udp latency=1.709563ms client=127.0.0.1:59471 client_network=udp writer.go:29: 2021-01-29T19:33:28.098Z [DEBUG] TestDNS_Compression_Recurse.dns: recurse succeeded for question: question="{apple.com. 255 1}" rtt=971.076µs recursor=127.0.0.1:42447 writer.go:29: 2021-01-29T19:33:28.100Z [DEBUG] TestDNS_Compression_Recurse.dns: request served from client: question="{apple.com. 255 1}" network=udp latency=2.904152ms client=127.0.0.1:59471 client_network=udp === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.114Z [WARN] TestDNS_Compression_ReverseLookup: bootstrap = true: do not enable unless necessary === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:28.102Z [INFO] TestDNS_Compression_Recurse: Requesting shutdown === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.116Z [DEBUG] TestDNS_Compression_ReverseLookup.tlsutil: Update: version=1 === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:28.116Z [INFO] TestDNS_Compression_Recurse.server: shutting down server writer.go:29: 2021-01-29T19:33:28.120Z [DEBUG] TestDNS_Compression_Recurse.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.127Z [DEBUG] TestDNS_Compression_ReverseLookup.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:28.123Z [WARN] TestDNS_Compression_Recurse.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:28.134Z [DEBUG] TestDNS_Compression_Recurse.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:28.140Z [ERROR] TestDNS_Compression_Recurse.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:28.141Z [WARN] TestDNS_Compression_Recurse.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:28.162Z [INFO] TestDNS_Compression_Recurse.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:28.164Z [INFO] TestDNS_Compression_Recurse: consul server down === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:28.151Z [INFO] TestEventFire_token: Waiting for endpoints to shut down === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:28.166Z [INFO] TestDNS_Compression_Recurse: shutdown complete writer.go:29: 2021-01-29T19:33:28.167Z [INFO] TestDNS_Compression_Recurse: Stopping server: protocol=DNS address=127.0.0.1:30078 network=tcp === CONT TestEventFire_token writer.go:29: 2021-01-29T19:33:28.166Z [INFO] TestEventFire_token: Endpoints down --- PASS: TestEventFire_token (0.92s) === CONT TestDNS_Compression_Query === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:28.169Z [INFO] TestDNS_Compression_Recurse: Stopping server: protocol=DNS address=127.0.0.1:30078 network=udp writer.go:29: 2021-01-29T19:33:28.170Z [INFO] TestDNS_Compression_Recurse: Stopping server: protocol=HTTP address=127.0.0.1:30079 network=tcp === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.234Z [INFO] TestDNS_Compression_ReverseLookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:49e6c615-94c9-b277-5f10-b75d3a32ae6f Address:127.0.0.1:30089}]" writer.go:29: 2021-01-29T19:33:28.242Z [INFO] TestDNS_Compression_ReverseLookup.server.serf.wan: serf: EventMemberJoin: Node-49e6c615-94c9-b277-5f10-b75d3a32ae6f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:28.247Z [INFO] TestDNS_Compression_ReverseLookup.server.raft: entering follower state: follower="Node at 127.0.0.1:30089 [Follower]" leader= writer.go:29: 2021-01-29T19:33:28.265Z [INFO] TestDNS_Compression_ReverseLookup.server.serf.lan: serf: EventMemberJoin: Node-49e6c615-94c9-b277-5f10-b75d3a32ae6f 127.0.0.1 === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.270Z [WARN] TestDNS_Compression_Query: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:28.277Z [DEBUG] TestDNS_Compression_Query.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:28.280Z [DEBUG] TestDNS_Compression_Query.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.285Z [INFO] TestDNS_Compression_ReverseLookup.server: Adding LAN server: server="Node-49e6c615-94c9-b277-5f10-b75d3a32ae6f (Addr: tcp/127.0.0.1:30089) (DC: dc1)" writer.go:29: 2021-01-29T19:33:28.287Z [INFO] TestDNS_Compression_ReverseLookup.server: Handled event for server in area: event=member-join server=Node-49e6c615-94c9-b277-5f10-b75d3a32ae6f.dc1 area=wan writer.go:29: 2021-01-29T19:33:28.292Z [INFO] TestDNS_Compression_ReverseLookup: Started DNS server: address=127.0.0.1:30084 network=tcp writer.go:29: 2021-01-29T19:33:28.303Z [INFO] TestDNS_Compression_ReverseLookup: Started DNS server: address=127.0.0.1:30084 network=udp writer.go:29: 2021-01-29T19:33:28.307Z [INFO] TestDNS_Compression_ReverseLookup: Started HTTP server: address=127.0.0.1:30085 network=tcp writer.go:29: 2021-01-29T19:33:28.309Z [INFO] TestDNS_Compression_ReverseLookup: started state syncer writer.go:29: 2021-01-29T19:33:28.343Z [WARN] TestDNS_Compression_ReverseLookup.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:28.344Z [INFO] TestDNS_Compression_ReverseLookup.server.raft: entering candidate state: node="Node at 127.0.0.1:30089 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:28.352Z [DEBUG] TestDNS_Compression_ReverseLookup.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:28.354Z [DEBUG] TestDNS_Compression_ReverseLookup.server.raft: vote granted: from=49e6c615-94c9-b277-5f10-b75d3a32ae6f term=2 tally=1 writer.go:29: 2021-01-29T19:33:28.355Z [INFO] TestDNS_Compression_ReverseLookup.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:28.357Z [INFO] TestDNS_Compression_ReverseLookup.server.raft: entering leader state: leader="Node at 127.0.0.1:30089 [Leader]" === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.358Z [INFO] TestDNS_Compression_Query.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ffd69fff-d81a-ec39-fcef-eef7e575be23 Address:127.0.0.1:30095}]" === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.358Z [INFO] TestDNS_Compression_ReverseLookup.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:28.361Z [INFO] TestDNS_Compression_ReverseLookup.server: New leader elected: payload=Node-49e6c615-94c9-b277-5f10-b75d3a32ae6f === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.361Z [INFO] TestDNS_Compression_Query.server.raft: entering follower state: follower="Node at 127.0.0.1:30095 [Follower]" leader= writer.go:29: 2021-01-29T19:33:28.363Z [INFO] TestDNS_Compression_Query.server.serf.wan: serf: EventMemberJoin: Node-ffd69fff-d81a-ec39-fcef-eef7e575be23.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:28.370Z [INFO] TestDNS_Compression_Query.server.serf.lan: serf: EventMemberJoin: Node-ffd69fff-d81a-ec39-fcef-eef7e575be23 127.0.0.1 writer.go:29: 2021-01-29T19:33:28.390Z [INFO] TestDNS_Compression_Query: Started DNS server: address=127.0.0.1:30090 network=udp writer.go:29: 2021-01-29T19:33:28.391Z [INFO] TestDNS_Compression_Query.server: Handled event for server in area: event=member-join server=Node-ffd69fff-d81a-ec39-fcef-eef7e575be23.dc1 area=wan writer.go:29: 2021-01-29T19:33:28.432Z [WARN] TestDNS_Compression_Query.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:28.433Z [INFO] TestDNS_Compression_Query.server.raft: entering candidate state: node="Node at 127.0.0.1:30095 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:28.432Z [INFO] TestDNS_Compression_Query.server: Adding LAN server: server="Node-ffd69fff-d81a-ec39-fcef-eef7e575be23 (Addr: tcp/127.0.0.1:30095) (DC: dc1)" writer.go:29: 2021-01-29T19:33:28.455Z [INFO] TestDNS_Compression_Query: Started DNS server: address=127.0.0.1:30090 network=tcp === CONT TestEventFire writer.go:29: 2021-01-29T19:33:28.462Z [INFO] TestEventFire: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:28.470Z [INFO] TestEventFire: Endpoints down --- PASS: TestEventFire (1.08s) === CONT TestDNS_Compression_trimUDPResponse === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.472Z [DEBUG] TestDNS_Compression_ReverseLookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30089 === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.474Z [INFO] TestDNS_Compression_Query: Started HTTP server: address=127.0.0.1:30091 network=tcp writer.go:29: 2021-01-29T19:33:28.476Z [INFO] TestDNS_Compression_Query: started state syncer --- PASS: TestDNS_Compression_trimUDPResponse (0.01s) === CONT TestAgent_HTTPCheck_TLSSkipVerify === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.492Z [DEBUG] TestDNS_Compression_Query.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:28.493Z [DEBUG] TestDNS_Compression_Query.server.raft: vote granted: from=ffd69fff-d81a-ec39-fcef-eef7e575be23 term=2 tally=1 writer.go:29: 2021-01-29T19:33:28.494Z [INFO] TestDNS_Compression_Query.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:28.496Z [INFO] TestDNS_Compression_Query.server.raft: entering leader state: leader="Node at 127.0.0.1:30095 [Leader]" === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.497Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.497Z [INFO] TestDNS_Compression_Query.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:28.501Z [INFO] TestDNS_Compression_Query.server: New leader elected: payload=Node-ffd69fff-d81a-ec39-fcef-eef7e575be23 === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.502Z [WARN] TestAgent_HTTPCheck_TLSSkipVerify: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:28.503Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:28.505Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.508Z [DEBUG] TestDNS_Compression_Query.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30095 writer.go:29: 2021-01-29T19:33:28.539Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.540Z [INFO] TestDNS_Compression_ReverseLookup.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:28.541Z [INFO] TestDNS_Compression_ReverseLookup.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:28.542Z [DEBUG] TestDNS_Compression_ReverseLookup.server: Skipping self join check for node since the cluster is too small: node=Node-49e6c615-94c9-b277-5f10-b75d3a32ae6f writer.go:29: 2021-01-29T19:33:28.543Z [INFO] TestDNS_Compression_ReverseLookup.server: member joined, marking health alive: member=Node-49e6c615-94c9-b277-5f10-b75d3a32ae6f === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.547Z [INFO] TestDNS_Compression_Query.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:28.548Z [INFO] TestDNS_Compression_Query.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:28.550Z [DEBUG] TestDNS_Compression_Query.server: Skipping self join check for node since the cluster is too small: node=Node-ffd69fff-d81a-ec39-fcef-eef7e575be23 writer.go:29: 2021-01-29T19:33:28.551Z [INFO] TestDNS_Compression_Query.server: member joined, marking health alive: member=Node-ffd69fff-d81a-ec39-fcef-eef7e575be23 === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.552Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:eef8319f-976e-c551-59c7-687c2d0ebfd7 Address:127.0.0.1:30101}]" writer.go:29: 2021-01-29T19:33:28.554Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server.raft: entering follower state: follower="Node at 127.0.0.1:30101 [Follower]" leader= writer.go:29: 2021-01-29T19:33:28.555Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server.serf.wan: serf: EventMemberJoin: Node-eef8319f-976e-c551-59c7-687c2d0ebfd7.dc1 127.0.0.1 === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.562Z [DEBUG] TestDNS_Compression_ReverseLookup.dns: request served from client: question="{2.0.0.127.in-addr.arpa. 255 1}" latency=65.111µs client=127.0.0.1:56359 client_network=udp writer.go:29: 2021-01-29T19:33:28.563Z [DEBUG] TestDNS_Compression_ReverseLookup.dns: request served from client: question="{2.0.0.127.in-addr.arpa. 255 1}" latency=103.448µs client=127.0.0.1:56359 client_network=udp writer.go:29: 2021-01-29T19:33:28.563Z [INFO] TestDNS_Compression_ReverseLookup: Requesting shutdown === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.565Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server.serf.lan: serf: EventMemberJoin: Node-eef8319f-976e-c551-59c7-687c2d0ebfd7 127.0.0.1 === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.567Z [INFO] TestDNS_Compression_ReverseLookup.server: shutting down server writer.go:29: 2021-01-29T19:33:28.569Z [DEBUG] TestDNS_Compression_ReverseLookup.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.570Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server: Adding LAN server: server="Node-eef8319f-976e-c551-59c7-687c2d0ebfd7 (Addr: tcp/127.0.0.1:30101) (DC: dc1)" === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.570Z [WARN] TestDNS_Compression_ReverseLookup.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.572Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server: Handled event for server in area: event=member-join server=Node-eef8319f-976e-c551-59c7-687c2d0ebfd7.dc1 area=wan === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.572Z [ERROR] TestDNS_Compression_ReverseLookup.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:28.572Z [DEBUG] TestDNS_Compression_ReverseLookup.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:28.575Z [WARN] TestDNS_Compression_ReverseLookup.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:28.584Z [INFO] TestDNS_Compression_ReverseLookup.server.router.manager: shutting down === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.585Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Started DNS server: address=127.0.0.1:30096 network=udp === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.587Z [INFO] TestDNS_Compression_ReverseLookup: consul server down writer.go:29: 2021-01-29T19:33:28.588Z [INFO] TestDNS_Compression_ReverseLookup: shutdown complete === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.587Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Started DNS server: address=127.0.0.1:30096 network=tcp === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.589Z [INFO] TestDNS_Compression_ReverseLookup: Stopping server: protocol=DNS address=127.0.0.1:30084 network=tcp === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.591Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Started HTTP server: address=127.0.0.1:30097 network=tcp writer.go:29: 2021-01-29T19:33:28.592Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: started state syncer === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.590Z [INFO] TestDNS_Compression_ReverseLookup: Stopping server: protocol=DNS address=127.0.0.1:30084 network=udp === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.595Z [WARN] TestAgent_HTTPCheck_TLSSkipVerify.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:28.597Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server.raft: entering candidate state: node="Node at 127.0.0.1:30101 [Candidate]" term=2 === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:28.594Z [INFO] TestDNS_Compression_ReverseLookup: Stopping server: protocol=HTTP address=127.0.0.1:30085 network=tcp === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.599Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:28.600Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify.server.raft: vote granted: from=eef8319f-976e-c551-59c7-687c2d0ebfd7 term=2 tally=1 writer.go:29: 2021-01-29T19:33:28.601Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:28.603Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server.raft: entering leader state: leader="Node at 127.0.0.1:30101 [Leader]" writer.go:29: 2021-01-29T19:33:28.604Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:28.606Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server: New leader elected: payload=Node-eef8319f-976e-c551-59c7-687c2d0ebfd7 writer.go:29: 2021-01-29T19:33:28.613Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30101 writer.go:29: 2021-01-29T19:33:28.619Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:28.625Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:28.626Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:28.627Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify.server: Skipping self join check for node since the cluster is too small: node=Node-eef8319f-976e-c551-59c7-687c2d0ebfd7 writer.go:29: 2021-01-29T19:33:28.636Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server: member joined, marking health alive: member=Node-eef8319f-976e-c551-59c7-687c2d0ebfd7 writer.go:29: 2021-01-29T19:33:28.682Z [WARN] TestAgent_HTTPCheck_TLSSkipVerify: check has interval below minimum: check=tls minimum_interval=1s === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:28.683Z [INFO] TestDNS_Compression_Recurse: Waiting for endpoints to shut down === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.684Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify.tlsutil: OutgoingTLSConfigForCheck: version=1 === CONT TestDNS_Compression_Recurse writer.go:29: 2021-01-29T19:33:28.685Z [INFO] TestDNS_Compression_Recurse: Endpoints down --- PASS: TestDNS_Compression_Recurse (0.80s) === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:28.694Z [WARN] TestAgent_Service_MaintenanceMode: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:28.695Z [DEBUG] TestAgent_Service_MaintenanceMode.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:28.697Z [DEBUG] TestAgent_Service_MaintenanceMode.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:28.720Z [INFO] TestAgent_Service_MaintenanceMode.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c745bc89-2d48-b6df-66d4-ffab333dc0fa Address:127.0.0.1:30107}]" writer.go:29: 2021-01-29T19:33:28.722Z [INFO] TestAgent_Service_MaintenanceMode.server.serf.wan: serf: EventMemberJoin: Node-c745bc89-2d48-b6df-66d4-ffab333dc0fa.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:28.725Z [INFO] TestAgent_Service_MaintenanceMode.server.serf.lan: serf: EventMemberJoin: Node-c745bc89-2d48-b6df-66d4-ffab333dc0fa 127.0.0.1 writer.go:29: 2021-01-29T19:33:28.727Z [INFO] TestAgent_Service_MaintenanceMode: Started DNS server: address=127.0.0.1:30102 network=udp writer.go:29: 2021-01-29T19:33:28.728Z [INFO] TestAgent_Service_MaintenanceMode.server.raft: entering follower state: follower="Node at 127.0.0.1:30107 [Follower]" leader= writer.go:29: 2021-01-29T19:33:28.731Z [INFO] TestAgent_Service_MaintenanceMode.server: Adding LAN server: server="Node-c745bc89-2d48-b6df-66d4-ffab333dc0fa (Addr: tcp/127.0.0.1:30107) (DC: dc1)" writer.go:29: 2021-01-29T19:33:28.730Z [INFO] TestAgent_Service_MaintenanceMode: Started DNS server: address=127.0.0.1:30102 network=tcp writer.go:29: 2021-01-29T19:33:28.730Z [INFO] TestAgent_Service_MaintenanceMode.server: Handled event for server in area: event=member-join server=Node-c745bc89-2d48-b6df-66d4-ffab333dc0fa.dc1 area=wan writer.go:29: 2021-01-29T19:33:28.739Z [INFO] TestAgent_Service_MaintenanceMode: Started HTTP server: address=127.0.0.1:30103 network=tcp writer.go:29: 2021-01-29T19:33:28.743Z [INFO] TestAgent_Service_MaintenanceMode: started state syncer === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.751Z [DEBUG] TestDNS_Compression_Query.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=118.057µs client=127.0.0.1:59659 client_network=udp writer.go:29: 2021-01-29T19:33:28.752Z [DEBUG] TestDNS_Compression_Query.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=58.6µs client=127.0.0.1:59659 client_network=udp writer.go:29: 2021-01-29T19:33:28.752Z [DEBUG] TestDNS_Compression_Query.dns: request served from client: name=59a57f8f-30ea-0063-7743-7902fca2da36.query.consul. type=SRV class=IN latency=98.379µs client=127.0.0.1:36248 client_network=udp writer.go:29: 2021-01-29T19:33:28.762Z [DEBUG] TestDNS_Compression_Query.dns: request served from client: name=59a57f8f-30ea-0063-7743-7902fca2da36.query.consul. type=SRV class=IN latency=141.318µs client=127.0.0.1:36248 client_network=udp writer.go:29: 2021-01-29T19:33:28.762Z [INFO] TestDNS_Compression_Query: Requesting shutdown writer.go:29: 2021-01-29T19:33:28.765Z [INFO] TestDNS_Compression_Query.server: shutting down server writer.go:29: 2021-01-29T19:33:28.767Z [DEBUG] TestDNS_Compression_Query.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:28.767Z [ERROR] TestDNS_Compression_Query.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:28.768Z [WARN] TestDNS_Compression_Query.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:28.770Z [WARN] TestAgent_Service_MaintenanceMode.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.769Z [DEBUG] TestDNS_Compression_Query.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:28.771Z [INFO] TestAgent_Service_MaintenanceMode.server.raft: entering candidate state: node="Node at 127.0.0.1:30107 [Candidate]" term=2 === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.772Z [WARN] TestDNS_Compression_Query.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.773Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:28.775Z [DEBUG] TestAgent_Service_MaintenanceMode.server.raft: votes: needed=1 === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.774Z [INFO] TestDNS_Compression_Query.server.router.manager: shutting down === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:28.776Z [DEBUG] TestAgent_Service_MaintenanceMode.server.raft: vote granted: from=c745bc89-2d48-b6df-66d4-ffab333dc0fa term=2 tally=1 === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.774Z [INFO] TestDNS_Compression_Query: consul server down === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:28.778Z [INFO] TestAgent_Service_MaintenanceMode.server.raft: election won: tally=1 === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.778Z [INFO] TestDNS_Compression_Query: shutdown complete writer.go:29: 2021-01-29T19:33:28.779Z [INFO] TestDNS_Compression_Query: Stopping server: protocol=DNS address=127.0.0.1:30090 network=tcp === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.779Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Synced node info === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:28.779Z [INFO] TestAgent_Service_MaintenanceMode.server.raft: entering leader state: leader="Node at 127.0.0.1:30107 [Leader]" writer.go:29: 2021-01-29T19:33:28.781Z [INFO] TestAgent_Service_MaintenanceMode.server: cluster leadership acquired === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.781Z [INFO] TestDNS_Compression_Query: Stopping server: protocol=DNS address=127.0.0.1:30090 network=udp === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:28.782Z [INFO] TestAgent_Service_MaintenanceMode.server: New leader elected: payload=Node-c745bc89-2d48-b6df-66d4-ffab333dc0fa === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.782Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Synced check: check=tls === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:28.783Z [INFO] TestDNS_Compression_Query: Stopping server: protocol=HTTP address=127.0.0.1:30091 network=tcp === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:28.784Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify: Node info in sync writer.go:29: 2021-01-29T19:33:28.785Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify: Check in sync: check=tls === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:28.786Z [DEBUG] TestAgent_Service_MaintenanceMode.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30107 writer.go:29: 2021-01-29T19:33:28.791Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:28.798Z [INFO] TestAgent_Service_MaintenanceMode.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:28.800Z [INFO] TestAgent_Service_MaintenanceMode.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:28.802Z [DEBUG] TestAgent_Service_MaintenanceMode.server: Skipping self join check for node since the cluster is too small: node=Node-c745bc89-2d48-b6df-66d4-ffab333dc0fa writer.go:29: 2021-01-29T19:33:28.803Z [INFO] TestAgent_Service_MaintenanceMode.server: member joined, marking health alive: member=Node-c745bc89-2d48-b6df-66d4-ffab333dc0fa writer.go:29: 2021-01-29T19:33:28.938Z [DEBUG] TestAgent_Service_MaintenanceMode: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:28.940Z [INFO] TestAgent_Service_MaintenanceMode: Synced node info writer.go:29: 2021-01-29T19:33:28.941Z [DEBUG] TestAgent_Service_MaintenanceMode: Node info in sync writer.go:29: 2021-01-29T19:33:29.077Z [INFO] TestAgent_Service_MaintenanceMode: Service entered maintenance mode: service=redis writer.go:29: 2021-01-29T19:33:29.078Z [DEBUG] TestAgent_Service_MaintenanceMode: removed check: check=_service_maintenance:redis writer.go:29: 2021-01-29T19:33:29.080Z [INFO] TestAgent_Service_MaintenanceMode: Service left maintenance mode: service=redis writer.go:29: 2021-01-29T19:33:29.082Z [INFO] TestAgent_Service_MaintenanceMode: Service entered maintenance mode: service=redis writer.go:29: 2021-01-29T19:33:29.084Z [INFO] TestAgent_Service_MaintenanceMode: Requesting shutdown writer.go:29: 2021-01-29T19:33:29.086Z [INFO] TestAgent_Service_MaintenanceMode.server: shutting down server writer.go:29: 2021-01-29T19:33:29.088Z [DEBUG] TestAgent_Service_MaintenanceMode.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:29.089Z [WARN] TestAgent_Service_MaintenanceMode.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:29.089Z [DEBUG] TestAgent_Service_MaintenanceMode.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:29.093Z [WARN] TestAgent_Service_MaintenanceMode.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:29.096Z [INFO] TestAgent_Service_MaintenanceMode.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:29.096Z [INFO] TestAgent_Service_MaintenanceMode: consul server down === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:29.099Z [INFO] TestDNS_Compression_ReverseLookup: Waiting for endpoints to shut down === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:29.099Z [INFO] TestAgent_Service_MaintenanceMode: shutdown complete === CONT TestDNS_Compression_ReverseLookup writer.go:29: 2021-01-29T19:33:29.101Z [INFO] TestDNS_Compression_ReverseLookup: Endpoints down --- PASS: TestDNS_Compression_ReverseLookup (1.03s) === CONT TestDNS_syncExtra === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:29.101Z [INFO] TestAgent_Service_MaintenanceMode: Stopping server: protocol=DNS address=127.0.0.1:30102 network=tcp --- PASS: TestDNS_syncExtra (0.00s) === CONT TestDNS_trimUDPResponse_TrimSize === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:29.102Z [INFO] TestAgent_Service_MaintenanceMode: Stopping server: protocol=DNS address=127.0.0.1:30102 network=udp writer.go:29: 2021-01-29T19:33:29.104Z [INFO] TestAgent_Service_MaintenanceMode: Stopping server: protocol=HTTP address=127.0.0.1:30103 network=tcp --- PASS: TestDNS_trimUDPResponse_TrimSize (0.01s) === CONT TestDNS_trimUDPResponse_TrimLimit --- PASS: TestDNS_trimUDPResponse_TrimLimit (0.01s) === CONT TestDNS_trimUDPResponse_NoTrim --- PASS: TestDNS_trimUDPResponse_NoTrim (0.01s) === CONT TestDNS_PreparedQuery_AgentSource writer.go:29: 2021-01-29T19:33:29.137Z [WARN] TestDNS_PreparedQuery_AgentSource: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:29.139Z [DEBUG] TestDNS_PreparedQuery_AgentSource.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:29.141Z [DEBUG] TestDNS_PreparedQuery_AgentSource.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:29.159Z [INFO] TestDNS_PreparedQuery_AgentSource.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ca44e6e3-7422-88c8-83fb-61c86a5ca5fb Address:127.0.0.1:30113}]" writer.go:29: 2021-01-29T19:33:29.161Z [INFO] TestDNS_PreparedQuery_AgentSource.server.raft: entering follower state: follower="Node at 127.0.0.1:30113 [Follower]" leader= writer.go:29: 2021-01-29T19:33:29.165Z [INFO] TestDNS_PreparedQuery_AgentSource.server.serf.wan: serf: EventMemberJoin: Node-ca44e6e3-7422-88c8-83fb-61c86a5ca5fb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:29.171Z [INFO] TestDNS_PreparedQuery_AgentSource.server.serf.lan: serf: EventMemberJoin: Node-ca44e6e3-7422-88c8-83fb-61c86a5ca5fb 127.0.0.1 writer.go:29: 2021-01-29T19:33:29.174Z [INFO] TestDNS_PreparedQuery_AgentSource.server: Adding LAN server: server="Node-ca44e6e3-7422-88c8-83fb-61c86a5ca5fb (Addr: tcp/127.0.0.1:30113) (DC: dc1)" writer.go:29: 2021-01-29T19:33:29.175Z [INFO] TestDNS_PreparedQuery_AgentSource.server: Handled event for server in area: event=member-join server=Node-ca44e6e3-7422-88c8-83fb-61c86a5ca5fb.dc1 area=wan writer.go:29: 2021-01-29T19:33:29.177Z [INFO] TestDNS_PreparedQuery_AgentSource: Started DNS server: address=127.0.0.1:30108 network=udp writer.go:29: 2021-01-29T19:33:29.187Z [INFO] TestDNS_PreparedQuery_AgentSource: Started DNS server: address=127.0.0.1:30108 network=tcp writer.go:29: 2021-01-29T19:33:29.200Z [INFO] TestDNS_PreparedQuery_AgentSource: Started HTTP server: address=127.0.0.1:30109 network=tcp writer.go:29: 2021-01-29T19:33:29.202Z [INFO] TestDNS_PreparedQuery_AgentSource: started state syncer writer.go:29: 2021-01-29T19:33:29.227Z [WARN] TestDNS_PreparedQuery_AgentSource.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:29.229Z [INFO] TestDNS_PreparedQuery_AgentSource.server.raft: entering candidate state: node="Node at 127.0.0.1:30113 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:29.232Z [DEBUG] TestDNS_PreparedQuery_AgentSource.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:29.234Z [DEBUG] TestDNS_PreparedQuery_AgentSource.server.raft: vote granted: from=ca44e6e3-7422-88c8-83fb-61c86a5ca5fb term=2 tally=1 writer.go:29: 2021-01-29T19:33:29.236Z [INFO] TestDNS_PreparedQuery_AgentSource.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:29.238Z [INFO] TestDNS_PreparedQuery_AgentSource.server.raft: entering leader state: leader="Node at 127.0.0.1:30113 [Leader]" writer.go:29: 2021-01-29T19:33:29.241Z [INFO] TestDNS_PreparedQuery_AgentSource.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:29.242Z [INFO] TestDNS_PreparedQuery_AgentSource.server: New leader elected: payload=Node-ca44e6e3-7422-88c8-83fb-61c86a5ca5fb writer.go:29: 2021-01-29T19:33:29.255Z [DEBUG] TestDNS_PreparedQuery_AgentSource.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30113 writer.go:29: 2021-01-29T19:33:29.262Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:29.280Z [INFO] TestDNS_PreparedQuery_AgentSource.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:29.282Z [INFO] TestDNS_PreparedQuery_AgentSource.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:29.284Z [DEBUG] TestDNS_PreparedQuery_AgentSource.server: Skipping self join check for node since the cluster is too small: node=Node-ca44e6e3-7422-88c8-83fb-61c86a5ca5fb === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:29.284Z [INFO] TestDNS_Compression_Query: Waiting for endpoints to shut down === CONT TestDNS_PreparedQuery_AgentSource writer.go:29: 2021-01-29T19:33:29.286Z [INFO] TestDNS_PreparedQuery_AgentSource.server: member joined, marking health alive: member=Node-ca44e6e3-7422-88c8-83fb-61c86a5ca5fb === CONT TestDNS_Compression_Query writer.go:29: 2021-01-29T19:33:29.286Z [INFO] TestDNS_Compression_Query: Endpoints down --- PASS: TestDNS_Compression_Query (1.12s) === CONT TestDNS_InvalidQueries writer.go:29: 2021-01-29T19:33:29.332Z [WARN] TestDNS_InvalidQueries: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:29.334Z [DEBUG] TestDNS_InvalidQueries.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:29.336Z [DEBUG] TestDNS_InvalidQueries.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:29.354Z [INFO] TestDNS_InvalidQueries.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d25d82eb-62e8-b021-3603-4ee4d7180bc4 Address:127.0.0.1:30119}]" writer.go:29: 2021-01-29T19:33:29.357Z [INFO] TestDNS_InvalidQueries.server.raft: entering follower state: follower="Node at 127.0.0.1:30119 [Follower]" leader= writer.go:29: 2021-01-29T19:33:29.357Z [INFO] TestDNS_InvalidQueries.server.serf.wan: serf: EventMemberJoin: Node-d25d82eb-62e8-b021-3603-4ee4d7180bc4.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:29.365Z [INFO] TestDNS_InvalidQueries.server.serf.lan: serf: EventMemberJoin: Node-d25d82eb-62e8-b021-3603-4ee4d7180bc4 127.0.0.1 writer.go:29: 2021-01-29T19:33:29.372Z [INFO] TestDNS_InvalidQueries.server: Adding LAN server: server="Node-d25d82eb-62e8-b021-3603-4ee4d7180bc4 (Addr: tcp/127.0.0.1:30119) (DC: dc1)" writer.go:29: 2021-01-29T19:33:29.373Z [INFO] TestDNS_InvalidQueries.server: Handled event for server in area: event=member-join server=Node-d25d82eb-62e8-b021-3603-4ee4d7180bc4.dc1 area=wan writer.go:29: 2021-01-29T19:33:29.379Z [INFO] TestDNS_InvalidQueries: Started DNS server: address=127.0.0.1:30114 network=udp writer.go:29: 2021-01-29T19:33:29.385Z [INFO] TestDNS_InvalidQueries: Started DNS server: address=127.0.0.1:30114 network=tcp writer.go:29: 2021-01-29T19:33:29.392Z [INFO] TestDNS_InvalidQueries: Started HTTP server: address=127.0.0.1:30115 network=tcp writer.go:29: 2021-01-29T19:33:29.396Z [INFO] TestDNS_InvalidQueries: started state syncer writer.go:29: 2021-01-29T19:33:29.420Z [WARN] TestDNS_InvalidQueries.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:29.422Z [INFO] TestDNS_InvalidQueries.server.raft: entering candidate state: node="Node at 127.0.0.1:30119 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:29.424Z [DEBUG] TestDNS_InvalidQueries.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:29.426Z [DEBUG] TestDNS_InvalidQueries.server.raft: vote granted: from=d25d82eb-62e8-b021-3603-4ee4d7180bc4 term=2 tally=1 writer.go:29: 2021-01-29T19:33:29.428Z [INFO] TestDNS_InvalidQueries.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:29.437Z [INFO] TestDNS_InvalidQueries.server.raft: entering leader state: leader="Node at 127.0.0.1:30119 [Leader]" writer.go:29: 2021-01-29T19:33:29.439Z [INFO] TestDNS_InvalidQueries.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:29.440Z [INFO] TestDNS_InvalidQueries.server: New leader elected: payload=Node-d25d82eb-62e8-b021-3603-4ee4d7180bc4 writer.go:29: 2021-01-29T19:33:29.441Z [DEBUG] TestDNS_InvalidQueries.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30119 writer.go:29: 2021-01-29T19:33:29.450Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:29.457Z [INFO] TestDNS_InvalidQueries.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:29.460Z [INFO] TestDNS_InvalidQueries.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:29.462Z [DEBUG] TestDNS_InvalidQueries.server: Skipping self join check for node since the cluster is too small: node=Node-d25d82eb-62e8-b021-3603-4ee4d7180bc4 writer.go:29: 2021-01-29T19:33:29.465Z [INFO] TestDNS_InvalidQueries.server: member joined, marking health alive: member=Node-d25d82eb-62e8-b021-3603-4ee4d7180bc4 === CONT TestDNS_PreparedQuery_AgentSource writer.go:29: 2021-01-29T19:33:29.525Z [DEBUG] TestDNS_PreparedQuery_AgentSource: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:29.527Z [INFO] TestDNS_PreparedQuery_AgentSource: Synced node info === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:29.544Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify: Check status updated: check=tls status=passing writer.go:29: 2021-01-29T19:33:29.561Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Requesting shutdown writer.go:29: 2021-01-29T19:33:29.564Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server: shutting down server writer.go:29: 2021-01-29T19:33:29.565Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:29.567Z [WARN] TestAgent_HTTPCheck_TLSSkipVerify.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:29.569Z [DEBUG] TestAgent_HTTPCheck_TLSSkipVerify.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:29.573Z [WARN] TestAgent_HTTPCheck_TLSSkipVerify.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:29.578Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:29.582Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: consul server down writer.go:29: 2021-01-29T19:33:29.586Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: shutdown complete === CONT TestDNS_PreparedQuery_AgentSource writer.go:29: 2021-01-29T19:33:29.587Z [WARN] TestDNS_PreparedQuery_AgentSource.server: endpoint injected; this should only be used for testing === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:29.590Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Stopping server: protocol=DNS address=127.0.0.1:30096 network=tcp === CONT TestDNS_PreparedQuery_AgentSource writer.go:29: 2021-01-29T19:33:29.595Z [DEBUG] TestDNS_PreparedQuery_AgentSource.dns: request served from client: name=foo.query.consul. type=SRV class=IN latency=73.171µs client=127.0.0.1:38426 client_network=udp === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:29.595Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Stopping server: protocol=DNS address=127.0.0.1:30096 network=udp === CONT TestDNS_PreparedQuery_AgentSource writer.go:29: 2021-01-29T19:33:29.598Z [INFO] TestDNS_PreparedQuery_AgentSource: Requesting shutdown === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:29.606Z [INFO] TestAgent_Service_MaintenanceMode: Waiting for endpoints to shut down === CONT TestDNS_PreparedQuery_AgentSource writer.go:29: 2021-01-29T19:33:29.609Z [INFO] TestDNS_PreparedQuery_AgentSource.server: shutting down server writer.go:29: 2021-01-29T19:33:29.610Z [DEBUG] TestDNS_PreparedQuery_AgentSource.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_Service_MaintenanceMode writer.go:29: 2021-01-29T19:33:29.609Z [INFO] TestAgent_Service_MaintenanceMode: Endpoints down --- PASS: TestAgent_Service_MaintenanceMode (0.93s) === CONT TestDNS_PreparedQuery_AllowStale === CONT TestDNS_PreparedQuery_AgentSource writer.go:29: 2021-01-29T19:33:29.611Z [WARN] TestDNS_PreparedQuery_AgentSource.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:29.614Z [DEBUG] TestDNS_PreparedQuery_AgentSource.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:29.599Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Stopping server: protocol=HTTP address=127.0.0.1:30097 network=tcp === CONT TestDNS_PreparedQuery_AgentSource writer.go:29: 2021-01-29T19:33:29.636Z [WARN] TestDNS_PreparedQuery_AgentSource.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:29.649Z [INFO] TestDNS_PreparedQuery_AgentSource.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:29.652Z [INFO] TestDNS_PreparedQuery_AgentSource: consul server down writer.go:29: 2021-01-29T19:33:29.653Z [INFO] TestDNS_PreparedQuery_AgentSource: shutdown complete writer.go:29: 2021-01-29T19:33:29.655Z [INFO] TestDNS_PreparedQuery_AgentSource: Stopping server: protocol=DNS address=127.0.0.1:30108 network=tcp writer.go:29: 2021-01-29T19:33:29.656Z [INFO] TestDNS_PreparedQuery_AgentSource: Stopping server: protocol=DNS address=127.0.0.1:30108 network=udp writer.go:29: 2021-01-29T19:33:29.658Z [INFO] TestDNS_PreparedQuery_AgentSource: Stopping server: protocol=HTTP address=127.0.0.1:30109 network=tcp === CONT TestDNS_InvalidQueries writer.go:29: 2021-01-29T19:33:29.678Z [WARN] TestDNS_InvalidQueries.dns: QName invalid: qname= writer.go:29: 2021-01-29T19:33:29.680Z [DEBUG] TestDNS_InvalidQueries.dns: request served from client: name=consul. type=SRV class=IN latency=1.775168ms client=127.0.0.1:43829 client_network=udp writer.go:29: 2021-01-29T19:33:29.682Z [WARN] TestDNS_InvalidQueries.dns: QName invalid: qname=node. writer.go:29: 2021-01-29T19:33:29.685Z [DEBUG] TestDNS_InvalidQueries.dns: request served from client: name=node.consul. type=SRV class=IN latency=2.753135ms client=127.0.0.1:56084 client_network=udp writer.go:29: 2021-01-29T19:33:29.685Z [WARN] TestDNS_InvalidQueries.dns: QName invalid: qname=service. writer.go:29: 2021-01-29T19:33:29.691Z [WARN] TestDNS_InvalidQueries.dns: QName invalid: qname=query. writer.go:29: 2021-01-29T19:33:29.693Z [DEBUG] TestDNS_InvalidQueries.dns: request served from client: name=service.consul. type=SRV class=IN latency=8.01633ms client=127.0.0.1:55821 client_network=udp writer.go:29: 2021-01-29T19:33:29.694Z [DEBUG] TestDNS_InvalidQueries.dns: request served from client: name=query.consul. type=SRV class=IN latency=2.417554ms client=127.0.0.1:40501 client_network=udp writer.go:29: 2021-01-29T19:33:29.694Z [WARN] TestDNS_InvalidQueries.dns: QName invalid: qname=foo.node.dc1.extra.more. writer.go:29: 2021-01-29T19:33:29.716Z [DEBUG] TestDNS_InvalidQueries: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:29.723Z [DEBUG] TestDNS_InvalidQueries.dns: request served from client: name=foo.node.dc1.extra.more.consul. type=SRV class=IN latency=28.92087ms client=127.0.0.1:34284 client_network=udp writer.go:29: 2021-01-29T19:33:29.732Z [INFO] TestDNS_InvalidQueries: Synced node info writer.go:29: 2021-01-29T19:33:29.736Z [DEBUG] TestDNS_InvalidQueries: Node info in sync === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:29.721Z [WARN] TestDNS_PreparedQuery_AllowStale: bootstrap = true: do not enable unless necessary === CONT TestDNS_InvalidQueries writer.go:29: 2021-01-29T19:33:29.737Z [WARN] TestDNS_InvalidQueries.dns: QName invalid: qname=foo.service.dc1.extra.more. writer.go:29: 2021-01-29T19:33:29.746Z [WARN] TestDNS_InvalidQueries.dns: QName invalid: qname=foo.query.dc1.extra.more. === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:29.742Z [DEBUG] TestDNS_PreparedQuery_AllowStale.tlsutil: Update: version=1 === CONT TestDNS_InvalidQueries writer.go:29: 2021-01-29T19:33:29.747Z [DEBUG] TestDNS_InvalidQueries.dns: request served from client: name=foo.service.dc1.extra.more.consul. type=SRV class=IN latency=16.241685ms client=127.0.0.1:45908 client_network=udp writer.go:29: 2021-01-29T19:33:29.750Z [DEBUG] TestDNS_InvalidQueries.dns: request served from client: name=foo.query.dc1.extra.more.consul. type=SRV class=IN latency=4.242642ms client=127.0.0.1:57557 client_network=udp writer.go:29: 2021-01-29T19:33:29.751Z [INFO] TestDNS_InvalidQueries: Requesting shutdown === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:29.753Z [DEBUG] TestDNS_PreparedQuery_AllowStale.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_InvalidQueries writer.go:29: 2021-01-29T19:33:29.761Z [INFO] TestDNS_InvalidQueries.server: shutting down server writer.go:29: 2021-01-29T19:33:29.769Z [DEBUG] TestDNS_InvalidQueries.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:29.773Z [DEBUG] TestDNS_InvalidQueries.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:29.781Z [INFO] TestDNS_PreparedQuery_AllowStale.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:06fdf0f3-521e-a70e-798a-859a3e859b45 Address:127.0.0.1:30125}]" === CONT TestDNS_InvalidQueries writer.go:29: 2021-01-29T19:33:29.776Z [WARN] TestDNS_InvalidQueries.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:29.784Z [INFO] TestDNS_PreparedQuery_AllowStale.server.raft: entering follower state: follower="Node at 127.0.0.1:30125 [Follower]" leader= writer.go:29: 2021-01-29T19:33:29.793Z [INFO] TestDNS_PreparedQuery_AllowStale.server.serf.wan: serf: EventMemberJoin: Node-06fdf0f3-521e-a70e-798a-859a3e859b45.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:29.797Z [INFO] TestDNS_PreparedQuery_AllowStale.server.serf.lan: serf: EventMemberJoin: Node-06fdf0f3-521e-a70e-798a-859a3e859b45 127.0.0.1 writer.go:29: 2021-01-29T19:33:29.800Z [INFO] TestDNS_PreparedQuery_AllowStale.server: Adding LAN server: server="Node-06fdf0f3-521e-a70e-798a-859a3e859b45 (Addr: tcp/127.0.0.1:30125) (DC: dc1)" === CONT TestDNS_InvalidQueries writer.go:29: 2021-01-29T19:33:29.799Z [WARN] TestDNS_InvalidQueries.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:29.803Z [INFO] TestDNS_PreparedQuery_AllowStale.server: Handled event for server in area: event=member-join server=Node-06fdf0f3-521e-a70e-798a-859a3e859b45.dc1 area=wan === CONT TestDNS_InvalidQueries writer.go:29: 2021-01-29T19:33:29.805Z [INFO] TestDNS_InvalidQueries.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:29.806Z [INFO] TestDNS_InvalidQueries: consul server down writer.go:29: 2021-01-29T19:33:29.809Z [INFO] TestDNS_InvalidQueries: shutdown complete writer.go:29: 2021-01-29T19:33:29.810Z [INFO] TestDNS_InvalidQueries: Stopping server: protocol=DNS address=127.0.0.1:30114 network=tcp writer.go:29: 2021-01-29T19:33:29.812Z [INFO] TestDNS_InvalidQueries: Stopping server: protocol=DNS address=127.0.0.1:30114 network=udp === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:29.809Z [INFO] TestDNS_PreparedQuery_AllowStale: Started DNS server: address=127.0.0.1:30120 network=tcp === CONT TestDNS_InvalidQueries writer.go:29: 2021-01-29T19:33:29.822Z [INFO] TestDNS_InvalidQueries: Stopping server: protocol=HTTP address=127.0.0.1:30115 network=tcp === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:29.823Z [INFO] TestDNS_PreparedQuery_AllowStale: Started DNS server: address=127.0.0.1:30120 network=udp writer.go:29: 2021-01-29T19:33:29.829Z [INFO] TestDNS_PreparedQuery_AllowStale: Started HTTP server: address=127.0.0.1:30121 network=tcp writer.go:29: 2021-01-29T19:33:29.836Z [INFO] TestDNS_PreparedQuery_AllowStale: started state syncer writer.go:29: 2021-01-29T19:33:29.858Z [WARN] TestDNS_PreparedQuery_AllowStale.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:29.859Z [INFO] TestDNS_PreparedQuery_AllowStale.server.raft: entering candidate state: node="Node at 127.0.0.1:30125 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:29.861Z [DEBUG] TestDNS_PreparedQuery_AllowStale.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:29.863Z [DEBUG] TestDNS_PreparedQuery_AllowStale.server.raft: vote granted: from=06fdf0f3-521e-a70e-798a-859a3e859b45 term=2 tally=1 writer.go:29: 2021-01-29T19:33:29.864Z [INFO] TestDNS_PreparedQuery_AllowStale.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:29.865Z [INFO] TestDNS_PreparedQuery_AllowStale.server.raft: entering leader state: leader="Node at 127.0.0.1:30125 [Leader]" writer.go:29: 2021-01-29T19:33:29.866Z [INFO] TestDNS_PreparedQuery_AllowStale.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:29.867Z [INFO] TestDNS_PreparedQuery_AllowStale.server: New leader elected: payload=Node-06fdf0f3-521e-a70e-798a-859a3e859b45 writer.go:29: 2021-01-29T19:33:29.869Z [DEBUG] TestDNS_PreparedQuery_AllowStale.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30125 writer.go:29: 2021-01-29T19:33:29.875Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:29.885Z [INFO] TestDNS_PreparedQuery_AllowStale.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:29.886Z [INFO] TestDNS_PreparedQuery_AllowStale.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:29.887Z [DEBUG] TestDNS_PreparedQuery_AllowStale.server: Skipping self join check for node since the cluster is too small: node=Node-06fdf0f3-521e-a70e-798a-859a3e859b45 writer.go:29: 2021-01-29T19:33:29.889Z [INFO] TestDNS_PreparedQuery_AllowStale.server: member joined, marking health alive: member=Node-06fdf0f3-521e-a70e-798a-859a3e859b45 writer.go:29: 2021-01-29T19:33:30.098Z [WARN] TestDNS_PreparedQuery_AllowStale.server: endpoint injected; this should only be used for testing writer.go:29: 2021-01-29T19:33:30.113Z [WARN] TestDNS_PreparedQuery_AllowStale.dns: Query results too stale, re-requesting writer.go:29: 2021-01-29T19:33:30.115Z [DEBUG] TestDNS_PreparedQuery_AllowStale.dns: request served from client: name=nope.query.consul. type=SRV class=IN latency=1.761321ms client=127.0.0.1:34924 client_network=udp writer.go:29: 2021-01-29T19:33:30.115Z [INFO] TestDNS_PreparedQuery_AllowStale: Requesting shutdown writer.go:29: 2021-01-29T19:33:30.118Z [INFO] TestDNS_PreparedQuery_AllowStale.server: shutting down server writer.go:29: 2021-01-29T19:33:30.120Z [DEBUG] TestDNS_PreparedQuery_AllowStale.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.122Z [DEBUG] TestDNS_PreparedQuery_AllowStale.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.122Z [ERROR] TestDNS_PreparedQuery_AllowStale.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:30.126Z [WARN] TestDNS_PreparedQuery_AllowStale.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_HTTPCheck_TLSSkipVerify writer.go:29: 2021-01-29T19:33:30.129Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:30.131Z [INFO] TestAgent_HTTPCheck_TLSSkipVerify: Endpoints down --- PASS: TestAgent_HTTPCheck_TLSSkipVerify (1.65s) === CONT TestDNS_AltDomains_Overlap === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:30.131Z [WARN] TestDNS_PreparedQuery_AllowStale.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:30.134Z [INFO] TestDNS_PreparedQuery_AllowStale.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:30.137Z [INFO] TestDNS_PreparedQuery_AllowStale: consul server down writer.go:29: 2021-01-29T19:33:30.138Z [INFO] TestDNS_PreparedQuery_AllowStale: shutdown complete writer.go:29: 2021-01-29T19:33:30.139Z [INFO] TestDNS_PreparedQuery_AllowStale: Stopping server: protocol=DNS address=127.0.0.1:30120 network=tcp writer.go:29: 2021-01-29T19:33:30.140Z [INFO] TestDNS_PreparedQuery_AllowStale: Stopping server: protocol=DNS address=127.0.0.1:30120 network=udp writer.go:29: 2021-01-29T19:33:30.143Z [INFO] TestDNS_PreparedQuery_AllowStale: Stopping server: protocol=HTTP address=127.0.0.1:30121 network=tcp === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.159Z [WARN] TestDNS_AltDomains_Overlap: bootstrap = true: do not enable unless necessary === CONT TestDNS_PreparedQuery_AgentSource writer.go:29: 2021-01-29T19:33:30.160Z [INFO] TestDNS_PreparedQuery_AgentSource: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:30.164Z [INFO] TestDNS_PreparedQuery_AgentSource: Endpoints down --- PASS: TestDNS_PreparedQuery_AgentSource (1.04s) === CONT TestDNS_AltDomains_SOA === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.164Z [DEBUG] TestDNS_AltDomains_Overlap.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:30.170Z [DEBUG] TestDNS_AltDomains_Overlap.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:30.198Z [INFO] TestDNS_AltDomains_Overlap.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ba88f2c2-ecfa-f715-fba9-cc3c780dcb11 Address:127.0.0.1:30131}]" === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.204Z [WARN] TestDNS_AltDomains_SOA: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:30.206Z [DEBUG] TestDNS_AltDomains_SOA.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:30.209Z [DEBUG] TestDNS_AltDomains_SOA.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.201Z [INFO] TestDNS_AltDomains_Overlap.server.raft: entering follower state: follower="Node at 127.0.0.1:30131 [Follower]" leader= writer.go:29: 2021-01-29T19:33:30.201Z [INFO] TestDNS_AltDomains_Overlap.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.220Z [INFO] TestDNS_AltDomains_Overlap.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.228Z [INFO] TestDNS_AltDomains_Overlap.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:30131) (DC: dc1)" writer.go:29: 2021-01-29T19:33:30.230Z [INFO] TestDNS_AltDomains_Overlap.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:33:30.234Z [INFO] TestDNS_AltDomains_Overlap: Started DNS server: address=127.0.0.1:30126 network=udp writer.go:29: 2021-01-29T19:33:30.235Z [INFO] TestDNS_AltDomains_Overlap: Started DNS server: address=127.0.0.1:30126 network=tcp writer.go:29: 2021-01-29T19:33:30.241Z [INFO] TestDNS_AltDomains_Overlap: Started HTTP server: address=127.0.0.1:30127 network=tcp === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.242Z [INFO] TestDNS_AltDomains_SOA.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:37b9c0e4-6a7e-a1d5-7e79-894661781729 Address:127.0.0.1:30137}]" === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.249Z [INFO] TestDNS_AltDomains_Overlap: started state syncer === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.256Z [INFO] TestDNS_AltDomains_SOA.server.raft: entering follower state: follower="Node at 127.0.0.1:30137 [Follower]" leader= === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.267Z [WARN] TestDNS_AltDomains_Overlap.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.260Z [INFO] TestDNS_AltDomains_SOA.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.279Z [INFO] TestDNS_AltDomains_Overlap.server.raft: entering candidate state: node="Node at 127.0.0.1:30131 [Candidate]" term=2 === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.297Z [INFO] TestDNS_AltDomains_SOA.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.303Z [DEBUG] TestDNS_AltDomains_Overlap.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:30.306Z [DEBUG] TestDNS_AltDomains_Overlap.server.raft: vote granted: from=ba88f2c2-ecfa-f715-fba9-cc3c780dcb11 term=2 tally=1 writer.go:29: 2021-01-29T19:33:30.309Z [INFO] TestDNS_AltDomains_Overlap.server.raft: election won: tally=1 === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.312Z [INFO] TestDNS_AltDomains_SOA.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:30137) (DC: dc1)" writer.go:29: 2021-01-29T19:33:30.318Z [INFO] TestDNS_AltDomains_SOA.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.313Z [INFO] TestDNS_AltDomains_Overlap.server.raft: entering leader state: leader="Node at 127.0.0.1:30131 [Leader]" writer.go:29: 2021-01-29T19:33:30.324Z [INFO] TestDNS_AltDomains_Overlap.server: cluster leadership acquired === CONT TestDNS_InvalidQueries writer.go:29: 2021-01-29T19:33:30.325Z [INFO] TestDNS_InvalidQueries: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:30.325Z [INFO] TestDNS_InvalidQueries: Endpoints down --- PASS: TestDNS_InvalidQueries (1.04s) === CONT TestDNS_AltDomains_Service === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.324Z [INFO] TestDNS_AltDomains_SOA: Started DNS server: address=127.0.0.1:30132 network=udp writer.go:29: 2021-01-29T19:33:30.334Z [INFO] TestDNS_AltDomains_SOA: Started DNS server: address=127.0.0.1:30132 network=tcp === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.333Z [INFO] TestDNS_AltDomains_Overlap.server: New leader elected: payload=test-node === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.337Z [INFO] TestDNS_AltDomains_SOA: Started HTTP server: address=127.0.0.1:30133 network=tcp writer.go:29: 2021-01-29T19:33:30.342Z [INFO] TestDNS_AltDomains_SOA: started state syncer writer.go:29: 2021-01-29T19:33:30.345Z [WARN] TestDNS_AltDomains_SOA.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:30.351Z [INFO] TestDNS_AltDomains_SOA.server.raft: entering candidate state: node="Node at 127.0.0.1:30137 [Candidate]" term=2 === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.348Z [DEBUG] TestDNS_AltDomains_Overlap.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30131 === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.361Z [DEBUG] TestDNS_AltDomains_SOA.server.raft: votes: needed=1 === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.367Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.367Z [DEBUG] TestDNS_AltDomains_SOA.server.raft: vote granted: from=37b9c0e4-6a7e-a1d5-7e79-894661781729 term=2 tally=1 writer.go:29: 2021-01-29T19:33:30.371Z [INFO] TestDNS_AltDomains_SOA.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:30.373Z [INFO] TestDNS_AltDomains_SOA.server.raft: entering leader state: leader="Node at 127.0.0.1:30137 [Leader]" writer.go:29: 2021-01-29T19:33:30.380Z [INFO] TestDNS_AltDomains_SOA.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:30.381Z [INFO] TestDNS_AltDomains_SOA.server: New leader elected: payload=test-node === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.388Z [INFO] TestDNS_AltDomains_Overlap.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:30.389Z [INFO] TestDNS_AltDomains_Overlap.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.390Z [DEBUG] TestDNS_AltDomains_Overlap.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:33:30.391Z [INFO] TestDNS_AltDomains_Overlap.server: member joined, marking health alive: member=test-node writer.go:29: 2021-01-29T19:33:30.407Z [DEBUG] TestDNS_AltDomains_Overlap.dns: request served from client: name=test-node.node.consul. type=A class=IN latency=160.506µs client=127.0.0.1:59271 client_network=udp === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.413Z [DEBUG] TestDNS_AltDomains_SOA.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30137 === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.418Z [DEBUG] TestDNS_AltDomains_Overlap.dns: request served from client: name=test-node.node.test.consul. type=A class=IN latency=175.52µs client=127.0.0.1:36639 client_network=udp writer.go:29: 2021-01-29T19:33:30.426Z [DEBUG] TestDNS_AltDomains_Overlap.dns: request served from client: name=test-node.node.dc1.consul. type=A class=IN latency=186.225µs client=127.0.0.1:50379 client_network=udp === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.418Z [WARN] TestDNS_AltDomains_Service: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:30.431Z [DEBUG] TestDNS_AltDomains_Service.tlsutil: Update: version=1 === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.429Z [DEBUG] TestDNS_AltDomains_Overlap.dns: request served from client: name=test-node.node.dc1.test.consul. type=A class=IN latency=151.646µs client=127.0.0.1:59863 client_network=udp === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.434Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.430Z [INFO] TestDNS_AltDomains_Overlap: Requesting shutdown writer.go:29: 2021-01-29T19:33:30.435Z [INFO] TestDNS_AltDomains_Overlap.server: shutting down server writer.go:29: 2021-01-29T19:33:30.437Z [DEBUG] TestDNS_AltDomains_Overlap.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.437Z [DEBUG] TestDNS_AltDomains_Service.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.439Z [WARN] TestDNS_AltDomains_Overlap.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.441Z [INFO] TestDNS_AltDomains_SOA.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.440Z [ERROR] TestDNS_AltDomains_Overlap.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:30.466Z [WARN] TestDNS_AltDomains_Overlap.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.464Z [INFO] TestDNS_AltDomains_SOA.leader: started routine: routine="CA root pruning" === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.440Z [DEBUG] TestDNS_AltDomains_Overlap.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.469Z [INFO] TestDNS_AltDomains_Overlap.server.router.manager: shutting down === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.468Z [DEBUG] TestDNS_AltDomains_SOA.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.469Z [INFO] TestDNS_AltDomains_Overlap: consul server down writer.go:29: 2021-01-29T19:33:30.473Z [INFO] TestDNS_AltDomains_Overlap: shutdown complete === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.474Z [INFO] TestDNS_AltDomains_Service.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4d94a764-d901-b8d1-2c63-45d00dd7557e Address:127.0.0.1:30143}]" === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.472Z [INFO] TestDNS_AltDomains_SOA.server: member joined, marking health alive: member=test-node === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.474Z [INFO] TestDNS_AltDomains_Overlap: Stopping server: protocol=DNS address=127.0.0.1:30126 network=tcp writer.go:29: 2021-01-29T19:33:30.476Z [INFO] TestDNS_AltDomains_Overlap: Stopping server: protocol=DNS address=127.0.0.1:30126 network=udp === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.476Z [INFO] TestDNS_AltDomains_Service.server.serf.wan: serf: EventMemberJoin: Node-4d94a764-d901-b8d1-2c63-45d00dd7557e.dc1 127.0.0.1 === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.478Z [INFO] TestDNS_AltDomains_Overlap: Stopping server: protocol=HTTP address=127.0.0.1:30127 network=tcp === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.481Z [INFO] TestDNS_AltDomains_Service.server.serf.lan: serf: EventMemberJoin: Node-4d94a764-d901-b8d1-2c63-45d00dd7557e 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.487Z [INFO] TestDNS_AltDomains_Service: Started DNS server: address=127.0.0.1:30138 network=udp writer.go:29: 2021-01-29T19:33:30.477Z [INFO] TestDNS_AltDomains_Service.server.raft: entering follower state: follower="Node at 127.0.0.1:30143 [Follower]" leader= writer.go:29: 2021-01-29T19:33:30.493Z [INFO] TestDNS_AltDomains_Service.server: Handled event for server in area: event=member-join server=Node-4d94a764-d901-b8d1-2c63-45d00dd7557e.dc1 area=wan writer.go:29: 2021-01-29T19:33:30.495Z [INFO] TestDNS_AltDomains_Service: Started DNS server: address=127.0.0.1:30138 network=tcp writer.go:29: 2021-01-29T19:33:30.503Z [INFO] TestDNS_AltDomains_Service.server: Adding LAN server: server="Node-4d94a764-d901-b8d1-2c63-45d00dd7557e (Addr: tcp/127.0.0.1:30143) (DC: dc1)" writer.go:29: 2021-01-29T19:33:30.548Z [WARN] TestDNS_AltDomains_Service.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:30.550Z [INFO] TestDNS_AltDomains_Service.server.raft: entering candidate state: node="Node at 127.0.0.1:30143 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:30.552Z [INFO] TestDNS_AltDomains_Service: Started HTTP server: address=127.0.0.1:30139 network=tcp writer.go:29: 2021-01-29T19:33:30.553Z [INFO] TestDNS_AltDomains_Service: started state syncer writer.go:29: 2021-01-29T19:33:30.581Z [DEBUG] TestDNS_AltDomains_Service.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:30.582Z [DEBUG] TestDNS_AltDomains_Service.server.raft: vote granted: from=4d94a764-d901-b8d1-2c63-45d00dd7557e term=2 tally=1 writer.go:29: 2021-01-29T19:33:30.583Z [INFO] TestDNS_AltDomains_Service.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:30.584Z [INFO] TestDNS_AltDomains_Service.server.raft: entering leader state: leader="Node at 127.0.0.1:30143 [Leader]" writer.go:29: 2021-01-29T19:33:30.585Z [INFO] TestDNS_AltDomains_Service.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:30.586Z [INFO] TestDNS_AltDomains_Service.server: New leader elected: payload=Node-4d94a764-d901-b8d1-2c63-45d00dd7557e writer.go:29: 2021-01-29T19:33:30.587Z [DEBUG] TestDNS_AltDomains_Service.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30143 writer.go:29: 2021-01-29T19:33:30.592Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:30.598Z [INFO] TestDNS_AltDomains_Service.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:30.600Z [INFO] TestDNS_AltDomains_Service.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.601Z [DEBUG] TestDNS_AltDomains_Service.server: Skipping self join check for node since the cluster is too small: node=Node-4d94a764-d901-b8d1-2c63-45d00dd7557e writer.go:29: 2021-01-29T19:33:30.603Z [INFO] TestDNS_AltDomains_Service.server: member joined, marking health alive: member=Node-4d94a764-d901-b8d1-2c63-45d00dd7557e writer.go:29: 2021-01-29T19:33:30.624Z [DEBUG] TestDNS_AltDomains_Service.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=171.102µs client=127.0.0.1:45539 client_network=udp writer.go:29: 2021-01-29T19:33:30.625Z [DEBUG] TestDNS_AltDomains_Service.dns: request served from client: name=db.service.test-domain. type=SRV class=IN latency=95.059µs client=127.0.0.1:44494 client_network=udp writer.go:29: 2021-01-29T19:33:30.628Z [DEBUG] TestDNS_AltDomains_Service.dns: request served from client: name=db.service.dc1.consul. type=SRV class=IN latency=109.61µs client=127.0.0.1:50879 client_network=udp === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.636Z [DEBUG] TestDNS_AltDomains_SOA.dns: request served from client: name=test-node.node.consul. type=SOA class=IN latency=188.018µs client=127.0.0.1:40851 client_network=udp === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.637Z [DEBUG] TestDNS_AltDomains_Service.dns: request served from client: name=db.service.dc1.test-domain. type=SRV class=IN latency=159.277µs client=127.0.0.1:40183 client_network=udp === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.636Z [DEBUG] TestDNS_AltDomains_SOA.dns: request served from client: name=test-node.node.test-domain. type=SOA class=IN latency=81.445µs client=127.0.0.1:38391 client_network=udp writer.go:29: 2021-01-29T19:33:30.636Z [INFO] TestDNS_AltDomains_SOA: Requesting shutdown writer.go:29: 2021-01-29T19:33:30.642Z [INFO] TestDNS_AltDomains_SOA.server: shutting down server writer.go:29: 2021-01-29T19:33:30.643Z [DEBUG] TestDNS_AltDomains_SOA.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.644Z [WARN] TestDNS_AltDomains_SOA.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:30.643Z [ERROR] TestDNS_AltDomains_SOA.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:30.645Z [INFO] TestDNS_PreparedQuery_AllowStale: Waiting for endpoints to shut down === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.637Z [INFO] TestDNS_AltDomains_Service: Requesting shutdown === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.644Z [DEBUG] TestDNS_AltDomains_SOA.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_PreparedQuery_AllowStale writer.go:29: 2021-01-29T19:33:30.647Z [INFO] TestDNS_PreparedQuery_AllowStale: Endpoints down --- PASS: TestDNS_PreparedQuery_AllowStale (1.04s) === CONT TestDNS_NonExistingLookupEmptyAorAAAA === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.647Z [DEBUG] TestDNS_AltDomains_Service: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.646Z [WARN] TestDNS_AltDomains_SOA.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.649Z [INFO] TestDNS_AltDomains_Service.server: shutting down server === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.652Z [INFO] TestDNS_AltDomains_SOA.server.router.manager: shutting down === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.653Z [DEBUG] TestDNS_AltDomains_Service.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.657Z [INFO] TestDNS_AltDomains_SOA: consul server down === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.650Z [INFO] TestDNS_AltDomains_Service: Synced node info === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:30.659Z [WARN] TestDNS_NonExistingLookupEmptyAorAAAA: bootstrap = true: do not enable unless necessary === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.658Z [INFO] TestDNS_AltDomains_SOA: shutdown complete === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.658Z [WARN] TestDNS_AltDomains_Service.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:30.660Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.tlsutil: Update: version=1 === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.660Z [INFO] TestDNS_AltDomains_SOA: Stopping server: protocol=DNS address=127.0.0.1:30132 network=tcp === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.661Z [DEBUG] TestDNS_AltDomains_Service: Node info in sync === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:30.661Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:30.662Z [INFO] TestDNS_AltDomains_SOA: Stopping server: protocol=DNS address=127.0.0.1:30132 network=udp writer.go:29: 2021-01-29T19:33:30.666Z [INFO] TestDNS_AltDomains_SOA: Stopping server: protocol=HTTP address=127.0.0.1:30133 network=tcp === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:30.667Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:91217704-4771-0698-e2cb-447be3f9fcd7 Address:127.0.0.1:30149}]" === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.662Z [DEBUG] TestDNS_AltDomains_Service.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.663Z [WARN] TestDNS_AltDomains_Service.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:30.670Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server.raft: entering follower state: follower="Node at 127.0.0.1:30149 [Follower]" leader= === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:30.675Z [INFO] TestDNS_AltDomains_Service.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:30.676Z [INFO] TestDNS_AltDomains_Service: consul server down writer.go:29: 2021-01-29T19:33:30.680Z [INFO] TestDNS_AltDomains_Service: shutdown complete writer.go:29: 2021-01-29T19:33:30.682Z [INFO] TestDNS_AltDomains_Service: Stopping server: protocol=DNS address=127.0.0.1:30138 network=tcp writer.go:29: 2021-01-29T19:33:30.684Z [INFO] TestDNS_AltDomains_Service: Stopping server: protocol=DNS address=127.0.0.1:30138 network=udp writer.go:29: 2021-01-29T19:33:30.686Z [INFO] TestDNS_AltDomains_Service: Stopping server: protocol=HTTP address=127.0.0.1:30139 network=tcp === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:30.682Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server.serf.wan: serf: EventMemberJoin: Node-91217704-4771-0698-e2cb-447be3f9fcd7.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.691Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server.serf.lan: serf: EventMemberJoin: Node-91217704-4771-0698-e2cb-447be3f9fcd7 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.694Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server: Adding LAN server: server="Node-91217704-4771-0698-e2cb-447be3f9fcd7 (Addr: tcp/127.0.0.1:30149) (DC: dc1)" writer.go:29: 2021-01-29T19:33:30.695Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server: Handled event for server in area: event=member-join server=Node-91217704-4771-0698-e2cb-447be3f9fcd7.dc1 area=wan writer.go:29: 2021-01-29T19:33:30.696Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: Started DNS server: address=127.0.0.1:30144 network=tcp writer.go:29: 2021-01-29T19:33:30.699Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: Started DNS server: address=127.0.0.1:30144 network=udp writer.go:29: 2021-01-29T19:33:30.702Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: Started HTTP server: address=127.0.0.1:30145 network=tcp writer.go:29: 2021-01-29T19:33:30.703Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: started state syncer writer.go:29: 2021-01-29T19:33:30.713Z [WARN] TestDNS_NonExistingLookupEmptyAorAAAA.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:30.715Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server.raft: entering candidate state: node="Node at 127.0.0.1:30149 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:30.718Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:30.719Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.server.raft: vote granted: from=91217704-4771-0698-e2cb-447be3f9fcd7 term=2 tally=1 writer.go:29: 2021-01-29T19:33:30.721Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:30.726Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server.raft: entering leader state: leader="Node at 127.0.0.1:30149 [Leader]" writer.go:29: 2021-01-29T19:33:30.727Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:30.728Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server: New leader elected: payload=Node-91217704-4771-0698-e2cb-447be3f9fcd7 writer.go:29: 2021-01-29T19:33:30.730Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30149 writer.go:29: 2021-01-29T19:33:30.737Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:30.771Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:30.773Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.774Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.server: Skipping self join check for node since the cluster is too small: node=Node-91217704-4771-0698-e2cb-447be3f9fcd7 writer.go:29: 2021-01-29T19:33:30.776Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server: member joined, marking health alive: member=Node-91217704-4771-0698-e2cb-447be3f9fcd7 === CONT TestDNS_AltDomains_Overlap writer.go:29: 2021-01-29T19:33:30.980Z [INFO] TestDNS_AltDomains_Overlap: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:30.981Z [INFO] TestDNS_AltDomains_Overlap: Endpoints down --- PASS: TestDNS_AltDomains_Overlap (0.85s) === CONT TestDNS_NonExistingLookup writer.go:29: 2021-01-29T19:33:30.991Z [WARN] TestDNS_NonExistingLookup: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:30.993Z [DEBUG] TestDNS_NonExistingLookup.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:30.995Z [DEBUG] TestDNS_NonExistingLookup.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:31.003Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.dns: request served from client: name=webv4.service.consul. type=AAAA class=IN latency=90.552µs client=127.0.0.1:41199 client_network=udp writer.go:29: 2021-01-29T19:33:31.005Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.dns: request served from client: name=webv4.query.consul. type=AAAA class=IN latency=83.238µs client=127.0.0.1:38121 client_network=udp writer.go:29: 2021-01-29T19:33:31.008Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.dns: request served from client: name=webv6.service.consul. type=A class=IN latency=111.269µs client=127.0.0.1:54559 client_network=udp writer.go:29: 2021-01-29T19:33:31.010Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.dns: request served from client: name=webv6.query.consul. type=A class=IN latency=98.704µs client=127.0.0.1:56748 client_network=udp writer.go:29: 2021-01-29T19:33:31.011Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: Requesting shutdown writer.go:29: 2021-01-29T19:33:31.021Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server: shutting down server writer.go:29: 2021-01-29T19:33:31.023Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.024Z [WARN] TestDNS_NonExistingLookupEmptyAorAAAA.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:31.023Z [ERROR] TestDNS_NonExistingLookupEmptyAorAAAA.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:31.024Z [DEBUG] TestDNS_NonExistingLookupEmptyAorAAAA.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_NonExistingLookup writer.go:29: 2021-01-29T19:33:31.029Z [INFO] TestDNS_NonExistingLookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cd3eaf88-7043-ea7b-de0e-2a1450e960a3 Address:127.0.0.1:30155}]" === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:31.027Z [WARN] TestDNS_NonExistingLookupEmptyAorAAAA.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_NonExistingLookup writer.go:29: 2021-01-29T19:33:31.031Z [INFO] TestDNS_NonExistingLookup.server.serf.wan: serf: EventMemberJoin: Node-cd3eaf88-7043-ea7b-de0e-2a1450e960a3.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:31.031Z [INFO] TestDNS_NonExistingLookup.server.raft: entering follower state: follower="Node at 127.0.0.1:30155 [Follower]" leader= === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:31.035Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:31.036Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: consul server down === CONT TestDNS_NonExistingLookup writer.go:29: 2021-01-29T19:33:31.034Z [INFO] TestDNS_NonExistingLookup.server.serf.lan: serf: EventMemberJoin: Node-cd3eaf88-7043-ea7b-de0e-2a1450e960a3 127.0.0.1 === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:31.037Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: shutdown complete writer.go:29: 2021-01-29T19:33:31.037Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: Stopping server: protocol=DNS address=127.0.0.1:30144 network=tcp writer.go:29: 2021-01-29T19:33:31.038Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: Stopping server: protocol=DNS address=127.0.0.1:30144 network=udp writer.go:29: 2021-01-29T19:33:31.039Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: Stopping server: protocol=HTTP address=127.0.0.1:30145 network=tcp === CONT TestDNS_NonExistingLookup writer.go:29: 2021-01-29T19:33:31.040Z [INFO] TestDNS_NonExistingLookup.server: Adding LAN server: server="Node-cd3eaf88-7043-ea7b-de0e-2a1450e960a3 (Addr: tcp/127.0.0.1:30155) (DC: dc1)" writer.go:29: 2021-01-29T19:33:31.041Z [INFO] TestDNS_NonExistingLookup.server: Handled event for server in area: event=member-join server=Node-cd3eaf88-7043-ea7b-de0e-2a1450e960a3.dc1 area=wan writer.go:29: 2021-01-29T19:33:31.042Z [INFO] TestDNS_NonExistingLookup: Started DNS server: address=127.0.0.1:30150 network=tcp writer.go:29: 2021-01-29T19:33:31.044Z [INFO] TestDNS_NonExistingLookup: Started DNS server: address=127.0.0.1:30150 network=udp writer.go:29: 2021-01-29T19:33:31.046Z [INFO] TestDNS_NonExistingLookup: Started HTTP server: address=127.0.0.1:30151 network=tcp writer.go:29: 2021-01-29T19:33:31.048Z [INFO] TestDNS_NonExistingLookup: started state syncer writer.go:29: 2021-01-29T19:33:31.097Z [WARN] TestDNS_NonExistingLookup.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:31.099Z [INFO] TestDNS_NonExistingLookup.server.raft: entering candidate state: node="Node at 127.0.0.1:30155 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:31.102Z [DEBUG] TestDNS_NonExistingLookup.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:31.103Z [DEBUG] TestDNS_NonExistingLookup.server.raft: vote granted: from=cd3eaf88-7043-ea7b-de0e-2a1450e960a3 term=2 tally=1 writer.go:29: 2021-01-29T19:33:31.105Z [INFO] TestDNS_NonExistingLookup.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:31.106Z [INFO] TestDNS_NonExistingLookup.server.raft: entering leader state: leader="Node at 127.0.0.1:30155 [Leader]" writer.go:29: 2021-01-29T19:33:31.108Z [INFO] TestDNS_NonExistingLookup.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:31.109Z [INFO] TestDNS_NonExistingLookup.server: New leader elected: payload=Node-cd3eaf88-7043-ea7b-de0e-2a1450e960a3 writer.go:29: 2021-01-29T19:33:31.110Z [DEBUG] TestDNS_NonExistingLookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30155 writer.go:29: 2021-01-29T19:33:31.116Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:31.124Z [INFO] TestDNS_NonExistingLookup.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:31.126Z [INFO] TestDNS_NonExistingLookup.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.127Z [DEBUG] TestDNS_NonExistingLookup.server: Skipping self join check for node since the cluster is too small: node=Node-cd3eaf88-7043-ea7b-de0e-2a1450e960a3 writer.go:29: 2021-01-29T19:33:31.129Z [INFO] TestDNS_NonExistingLookup.server: member joined, marking health alive: member=Node-cd3eaf88-7043-ea7b-de0e-2a1450e960a3 === CONT TestDNS_AltDomains_SOA writer.go:29: 2021-01-29T19:33:31.167Z [INFO] TestDNS_AltDomains_SOA: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:31.169Z [INFO] TestDNS_AltDomains_SOA: Endpoints down --- PASS: TestDNS_AltDomains_SOA (1.00s) === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.179Z [WARN] TestDNS_AddressLookup: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:31.181Z [DEBUG] TestDNS_AddressLookup.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:31.183Z [DEBUG] TestDNS_AddressLookup.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:31.188Z [INFO] TestDNS_AltDomains_Service: Waiting for endpoints to shut down === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.188Z [INFO] TestDNS_AddressLookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:afb16dad-fe04-a0fa-0638-d14b3fb87d17 Address:127.0.0.1:30161}]" === CONT TestDNS_AltDomains_Service writer.go:29: 2021-01-29T19:33:31.189Z [INFO] TestDNS_AltDomains_Service: Endpoints down --- PASS: TestDNS_AltDomains_Service (0.86s) === CONT TestDNS_ServiceLookup_FilterACL === RUN TestDNS_ServiceLookup_FilterACL/ACLToken_==_root === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.191Z [INFO] TestDNS_AddressLookup.server.serf.wan: serf: EventMemberJoin: Node-afb16dad-fe04-a0fa-0638-d14b3fb87d17.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:31.193Z [INFO] TestDNS_AddressLookup.server.raft: entering follower state: follower="Node at 127.0.0.1:30161 [Follower]" leader= writer.go:29: 2021-01-29T19:33:31.193Z [INFO] TestDNS_AddressLookup.server.serf.lan: serf: EventMemberJoin: Node-afb16dad-fe04-a0fa-0638-d14b3fb87d17 127.0.0.1 writer.go:29: 2021-01-29T19:33:31.200Z [INFO] TestDNS_AddressLookup.server: Handled event for server in area: event=member-join server=Node-afb16dad-fe04-a0fa-0638-d14b3fb87d17.dc1 area=wan writer.go:29: 2021-01-29T19:33:31.200Z [INFO] TestDNS_AddressLookup.server: Adding LAN server: server="Node-afb16dad-fe04-a0fa-0638-d14b3fb87d17 (Addr: tcp/127.0.0.1:30161) (DC: dc1)" writer.go:29: 2021-01-29T19:33:31.203Z [INFO] TestDNS_AddressLookup: Started DNS server: address=127.0.0.1:30156 network=tcp writer.go:29: 2021-01-29T19:33:31.204Z [INFO] TestDNS_AddressLookup: Started DNS server: address=127.0.0.1:30156 network=udp writer.go:29: 2021-01-29T19:33:31.206Z [INFO] TestDNS_AddressLookup: Started HTTP server: address=127.0.0.1:30157 network=tcp === CONT TestDNS_NonExistingLookup writer.go:29: 2021-01-29T19:33:31.207Z [DEBUG] TestDNS_NonExistingLookup: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.207Z [INFO] TestDNS_AddressLookup: started state syncer === CONT TestDNS_NonExistingLookup writer.go:29: 2021-01-29T19:33:31.209Z [INFO] TestDNS_NonExistingLookup: Synced node info === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_root writer.go:29: 2021-01-29T19:33:31.213Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:31.214Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:31.215Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:31.216Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:31.222Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:080464dd-2fee-c1a3-69c9-5a596b8716a9 Address:127.0.0.1:30167}]" writer.go:29: 2021-01-29T19:33:31.223Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.raft: entering follower state: follower="Node at 127.0.0.1:30167 [Follower]" leader= writer.go:29: 2021-01-29T19:33:31.224Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.serf.wan: serf: EventMemberJoin: Node-080464dd-2fee-c1a3-69c9-5a596b8716a9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:31.231Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.serf.lan: serf: EventMemberJoin: Node-080464dd-2fee-c1a3-69c9-5a596b8716a9 127.0.0.1 writer.go:29: 2021-01-29T19:33:31.234Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Adding LAN server: server="Node-080464dd-2fee-c1a3-69c9-5a596b8716a9 (Addr: tcp/127.0.0.1:30167) (DC: dc1)" writer.go:29: 2021-01-29T19:33:31.235Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Handled event for server in area: event=member-join server=Node-080464dd-2fee-c1a3-69c9-5a596b8716a9.dc1 area=wan === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.236Z [WARN] TestDNS_AddressLookup.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_root writer.go:29: 2021-01-29T19:33:31.236Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: Started DNS server: address=127.0.0.1:30162 network=udp === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.238Z [INFO] TestDNS_AddressLookup.server.raft: entering candidate state: node="Node at 127.0.0.1:30161 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_root writer.go:29: 2021-01-29T19:33:31.238Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: Started DNS server: address=127.0.0.1:30162 network=tcp writer.go:29: 2021-01-29T19:33:31.241Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: Started HTTP server: address=127.0.0.1:30163 network=tcp writer.go:29: 2021-01-29T19:33:31.243Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: started state syncer === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.245Z [DEBUG] TestDNS_AddressLookup.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:31.246Z [DEBUG] TestDNS_AddressLookup.server.raft: vote granted: from=afb16dad-fe04-a0fa-0638-d14b3fb87d17 term=2 tally=1 writer.go:29: 2021-01-29T19:33:31.246Z [INFO] TestDNS_AddressLookup.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:31.247Z [INFO] TestDNS_AddressLookup.server.raft: entering leader state: leader="Node at 127.0.0.1:30161 [Leader]" writer.go:29: 2021-01-29T19:33:31.248Z [INFO] TestDNS_AddressLookup.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:31.249Z [INFO] TestDNS_AddressLookup.server: New leader elected: payload=Node-afb16dad-fe04-a0fa-0638-d14b3fb87d17 writer.go:29: 2021-01-29T19:33:31.249Z [DEBUG] TestDNS_AddressLookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30161 writer.go:29: 2021-01-29T19:33:31.258Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:31.264Z [INFO] TestDNS_AddressLookup.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:31.265Z [INFO] TestDNS_AddressLookup.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.268Z [DEBUG] TestDNS_AddressLookup.server: Skipping self join check for node since the cluster is too small: node=Node-afb16dad-fe04-a0fa-0638-d14b3fb87d17 writer.go:29: 2021-01-29T19:33:31.270Z [INFO] TestDNS_AddressLookup.server: member joined, marking health alive: member=Node-afb16dad-fe04-a0fa-0638-d14b3fb87d17 === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_root writer.go:29: 2021-01-29T19:33:31.274Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:31.276Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.raft: entering candidate state: node="Node at 127.0.0.1:30167 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:31.286Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:33:31.290Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:31.292Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.raft: vote granted: from=080464dd-2fee-c1a3-69c9-5a596b8716a9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:31.293Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:31.295Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.raft: entering leader state: leader="Node at 127.0.0.1:30167 [Leader]" writer.go:29: 2021-01-29T19:33:31.297Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:31.298Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: New leader elected: payload=Node-080464dd-2fee-c1a3-69c9-5a596b8716a9 writer.go:29: 2021-01-29T19:33:31.299Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: initializing acls writer.go:29: 2021-01-29T19:33:31.305Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:31.306Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:33:31.309Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:33:31.311Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:31.313Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:31.314Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:31.317Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.serf.lan: serf: EventMemberUpdate: Node-080464dd-2fee-c1a3-69c9-5a596b8716a9 writer.go:29: 2021-01-29T19:33:31.319Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.serf.wan: serf: EventMemberUpdate: Node-080464dd-2fee-c1a3-69c9-5a596b8716a9.dc1 writer.go:29: 2021-01-29T19:33:31.319Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Updating LAN server: server="Node-080464dd-2fee-c1a3-69c9-5a596b8716a9 (Addr: tcp/127.0.0.1:30167) (DC: dc1)" writer.go:29: 2021-01-29T19:33:31.321Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Handled event for server in area: event=member-update server=Node-080464dd-2fee-c1a3-69c9-5a596b8716a9.dc1 area=wan writer.go:29: 2021-01-29T19:33:31.325Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:31.339Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:31.340Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.341Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Skipping self join check for node since the cluster is too small: node=Node-080464dd-2fee-c1a3-69c9-5a596b8716a9 writer.go:29: 2021-01-29T19:33:31.342Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: member joined, marking health alive: member=Node-080464dd-2fee-c1a3-69c9-5a596b8716a9 writer.go:29: 2021-01-29T19:33:31.344Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: Skipping self join check for node since the cluster is too small: node=Node-080464dd-2fee-c1a3-69c9-5a596b8716a9 writer.go:29: 2021-01-29T19:33:31.351Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.acl: dropping node from result due to ACLs: node=Node-080464dd-2fee-c1a3-69c9-5a596b8716a9 writer.go:29: 2021-01-29T19:33:31.358Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.dns: request served from client: name=foo.service.consul. type=A class=IN latency=197.947µs client=127.0.0.1:46572 client_network=udp writer.go:29: 2021-01-29T19:33:31.358Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: Requesting shutdown writer.go:29: 2021-01-29T19:33:31.361Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server: shutting down server writer.go:29: 2021-01-29T19:33:31.362Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:31.364Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:31.362Z [ERROR] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:31.365Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:31.365Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.365Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:31.370Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:31.370Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.374Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:31.379Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:31.380Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: consul server down writer.go:29: 2021-01-29T19:33:31.381Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: shutdown complete writer.go:29: 2021-01-29T19:33:31.382Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: Stopping server: protocol=DNS address=127.0.0.1:30162 network=tcp writer.go:29: 2021-01-29T19:33:31.383Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: Stopping server: protocol=DNS address=127.0.0.1:30162 network=udp writer.go:29: 2021-01-29T19:33:31.384Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: Stopping server: protocol=HTTP address=127.0.0.1:30163 network=tcp === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.406Z [DEBUG] TestDNS_AddressLookup: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:31.410Z [INFO] TestDNS_AddressLookup: Synced node info === CONT TestDNS_NonExistingLookup writer.go:29: 2021-01-29T19:33:31.493Z [WARN] TestDNS_NonExistingLookup.dns: QName invalid: qname=nonexisting. writer.go:29: 2021-01-29T19:33:31.495Z [DEBUG] TestDNS_NonExistingLookup.dns: request served from client: name=nonexisting.consul. type=ANY class=IN latency=1.606134ms client=127.0.0.1:43604 client_network=udp writer.go:29: 2021-01-29T19:33:31.497Z [INFO] TestDNS_NonExistingLookup: Requesting shutdown writer.go:29: 2021-01-29T19:33:31.498Z [INFO] TestDNS_NonExistingLookup.server: shutting down server writer.go:29: 2021-01-29T19:33:31.500Z [DEBUG] TestDNS_NonExistingLookup.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.502Z [DEBUG] TestDNS_NonExistingLookup.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.503Z [WARN] TestDNS_NonExistingLookup.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:31.507Z [WARN] TestDNS_NonExistingLookup.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:31.511Z [INFO] TestDNS_NonExistingLookup.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:31.511Z [INFO] TestDNS_NonExistingLookup: consul server down writer.go:29: 2021-01-29T19:33:31.514Z [INFO] TestDNS_NonExistingLookup: shutdown complete writer.go:29: 2021-01-29T19:33:31.520Z [INFO] TestDNS_NonExistingLookup: Stopping server: protocol=DNS address=127.0.0.1:30150 network=tcp writer.go:29: 2021-01-29T19:33:31.522Z [INFO] TestDNS_NonExistingLookup: Stopping server: protocol=DNS address=127.0.0.1:30150 network=udp writer.go:29: 2021-01-29T19:33:31.523Z [INFO] TestDNS_NonExistingLookup: Stopping server: protocol=HTTP address=127.0.0.1:30151 network=tcp === CONT TestDNS_NonExistingLookupEmptyAorAAAA writer.go:29: 2021-01-29T19:33:31.540Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:31.543Z [INFO] TestDNS_NonExistingLookupEmptyAorAAAA: Endpoints down --- PASS: TestDNS_NonExistingLookupEmptyAorAAAA (0.90s) === CONT TestDNS_ServiceLookup_SRV_RFC_TCP_Default writer.go:29: 2021-01-29T19:33:31.560Z [WARN] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:31.562Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:31.564Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:31.569Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0f793519-f2f5-d7e8-fb11-a2b9b1080442 Address:127.0.0.1:30173}]" writer.go:29: 2021-01-29T19:33:31.571Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.raft: entering follower state: follower="Node at 127.0.0.1:30173 [Follower]" leader= writer.go:29: 2021-01-29T19:33:31.572Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.serf.wan: serf: EventMemberJoin: Node-0f793519-f2f5-d7e8-fb11-a2b9b1080442.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:31.578Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.serf.lan: serf: EventMemberJoin: Node-0f793519-f2f5-d7e8-fb11-a2b9b1080442 127.0.0.1 writer.go:29: 2021-01-29T19:33:31.582Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server: Adding LAN server: server="Node-0f793519-f2f5-d7e8-fb11-a2b9b1080442 (Addr: tcp/127.0.0.1:30173) (DC: dc1)" writer.go:29: 2021-01-29T19:33:31.585Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server: Handled event for server in area: event=member-join server=Node-0f793519-f2f5-d7e8-fb11-a2b9b1080442.dc1 area=wan writer.go:29: 2021-01-29T19:33:31.587Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: Started DNS server: address=127.0.0.1:30168 network=tcp writer.go:29: 2021-01-29T19:33:31.590Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: Started DNS server: address=127.0.0.1:30168 network=udp writer.go:29: 2021-01-29T19:33:31.593Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: Started HTTP server: address=127.0.0.1:30169 network=tcp writer.go:29: 2021-01-29T19:33:31.594Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: started state syncer === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.625Z [DEBUG] TestDNS_AddressLookup.dns: request served from client: name=7f000001.addr.dc1.consul. type=SRV class=IN latency=34.883µs client=127.0.0.1:59838 client_network=udp writer.go:29: 2021-01-29T19:33:31.627Z [INFO] TestDNS_AddressLookup: Requesting shutdown writer.go:29: 2021-01-29T19:33:31.628Z [INFO] TestDNS_AddressLookup.server: shutting down server writer.go:29: 2021-01-29T19:33:31.630Z [DEBUG] TestDNS_AddressLookup.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.631Z [WARN] TestDNS_AddressLookup.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:31.631Z [DEBUG] TestDNS_AddressLookup.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.636Z [WARN] TestDNS_AddressLookup.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:31.640Z [INFO] TestDNS_AddressLookup.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:31.640Z [INFO] TestDNS_AddressLookup: consul server down === CONT TestDNS_ServiceLookup_SRV_RFC_TCP_Default writer.go:29: 2021-01-29T19:33:31.642Z [WARN] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.643Z [INFO] TestDNS_AddressLookup: shutdown complete === CONT TestDNS_ServiceLookup_SRV_RFC_TCP_Default writer.go:29: 2021-01-29T19:33:31.644Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.raft: entering candidate state: node="Node at 127.0.0.1:30173 [Candidate]" term=2 === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:31.644Z [INFO] TestDNS_AddressLookup: Stopping server: protocol=DNS address=127.0.0.1:30156 network=tcp writer.go:29: 2021-01-29T19:33:31.647Z [INFO] TestDNS_AddressLookup: Stopping server: protocol=DNS address=127.0.0.1:30156 network=udp writer.go:29: 2021-01-29T19:33:31.648Z [INFO] TestDNS_AddressLookup: Stopping server: protocol=HTTP address=127.0.0.1:30157 network=tcp === CONT TestDNS_ServiceLookup_SRV_RFC_TCP_Default writer.go:29: 2021-01-29T19:33:31.650Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:31.651Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.raft: vote granted: from=0f793519-f2f5-d7e8-fb11-a2b9b1080442 term=2 tally=1 writer.go:29: 2021-01-29T19:33:31.653Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:31.654Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.raft: entering leader state: leader="Node at 127.0.0.1:30173 [Leader]" writer.go:29: 2021-01-29T19:33:31.656Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:31.658Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server: New leader elected: payload=Node-0f793519-f2f5-d7e8-fb11-a2b9b1080442 writer.go:29: 2021-01-29T19:33:31.661Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30173 writer.go:29: 2021-01-29T19:33:31.684Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:31.691Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:31.693Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.694Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server: Skipping self join check for node since the cluster is too small: node=Node-0f793519-f2f5-d7e8-fb11-a2b9b1080442 writer.go:29: 2021-01-29T19:33:31.698Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server: member joined, marking health alive: member=Node-0f793519-f2f5-d7e8-fb11-a2b9b1080442 writer.go:29: 2021-01-29T19:33:31.728Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.dns: request served from client: name=_db._tcp.service.dc1.consul. type=SRV class=IN latency=134.872µs client=127.0.0.1:58577 client_network=udp writer.go:29: 2021-01-29T19:33:31.729Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.dns: request served from client: name=_db._tcp.service.consul. type=SRV class=IN latency=121.435µs client=127.0.0.1:36046 client_network=udp writer.go:29: 2021-01-29T19:33:31.730Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.dns: request served from client: name=_db._tcp.dc1.consul. type=SRV class=IN latency=48.284µs client=127.0.0.1:37443 client_network=udp writer.go:29: 2021-01-29T19:33:31.730Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.dns: request served from client: name=_db._tcp.consul. type=SRV class=IN latency=99.655µs client=127.0.0.1:41587 client_network=udp writer.go:29: 2021-01-29T19:33:31.731Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: Requesting shutdown writer.go:29: 2021-01-29T19:33:31.736Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server: shutting down server writer.go:29: 2021-01-29T19:33:31.737Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.738Z [WARN] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:31.737Z [ERROR] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:31.738Z [DEBUG] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:31.741Z [WARN] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:31.746Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:31.746Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: consul server down writer.go:29: 2021-01-29T19:33:31.749Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: shutdown complete writer.go:29: 2021-01-29T19:33:31.750Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: Stopping server: protocol=DNS address=127.0.0.1:30168 network=tcp writer.go:29: 2021-01-29T19:33:31.752Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: Stopping server: protocol=DNS address=127.0.0.1:30168 network=udp writer.go:29: 2021-01-29T19:33:31.754Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: Stopping server: protocol=HTTP address=127.0.0.1:30169 network=tcp === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_root writer.go:29: 2021-01-29T19:33:31.886Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:31.888Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_root: Endpoints down === RUN TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:31.918Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:31.919Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:31.920Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:31.922Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:31.926Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a9196612-a7c5-37ab-535f-ffb77bdbe26f Address:127.0.0.1:30179}]" writer.go:29: 2021-01-29T19:33:31.928Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.raft: entering follower state: follower="Node at 127.0.0.1:30179 [Follower]" leader= writer.go:29: 2021-01-29T19:33:31.929Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.serf.wan: serf: EventMemberJoin: Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:31.934Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.serf.lan: serf: EventMemberJoin: Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f 127.0.0.1 writer.go:29: 2021-01-29T19:33:31.937Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Adding LAN server: server="Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f (Addr: tcp/127.0.0.1:30179) (DC: dc1)" writer.go:29: 2021-01-29T19:33:31.938Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Handled event for server in area: event=member-join server=Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f.dc1 area=wan writer.go:29: 2021-01-29T19:33:31.939Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: Started DNS server: address=127.0.0.1:30174 network=udp writer.go:29: 2021-01-29T19:33:31.944Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: Started DNS server: address=127.0.0.1:30174 network=tcp writer.go:29: 2021-01-29T19:33:31.947Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: Started HTTP server: address=127.0.0.1:30175 network=tcp writer.go:29: 2021-01-29T19:33:31.949Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: started state syncer writer.go:29: 2021-01-29T19:33:31.994Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:31.995Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.raft: entering candidate state: node="Node at 127.0.0.1:30179 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:31.995Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:33:32.000Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:32.001Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.raft: vote granted: from=a9196612-a7c5-37ab-535f-ffb77bdbe26f term=2 tally=1 writer.go:29: 2021-01-29T19:33:32.002Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:32.003Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.raft: entering leader state: leader="Node at 127.0.0.1:30179 [Leader]" writer.go:29: 2021-01-29T19:33:32.004Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:32.005Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: New leader elected: payload=Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f writer.go:29: 2021-01-29T19:33:32.021Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: initializing acls === CONT TestDNS_NonExistingLookup writer.go:29: 2021-01-29T19:33:32.025Z [INFO] TestDNS_NonExistingLookup: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.026Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Created ACL 'global-management' policy === CONT TestDNS_NonExistingLookup writer.go:29: 2021-01-29T19:33:32.027Z [INFO] TestDNS_NonExistingLookup: Endpoints down --- PASS: TestDNS_NonExistingLookup (1.05s) === CONT TestDNS_PreparedQuery_TTL === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.028Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:33:32.031Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:33:32.034Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:32.036Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:32.038Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:32.040Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.serf.lan: serf: EventMemberUpdate: Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f writer.go:29: 2021-01-29T19:33:32.042Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.serf.wan: serf: EventMemberUpdate: Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f.dc1 writer.go:29: 2021-01-29T19:33:32.043Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Updating LAN server: server="Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f (Addr: tcp/127.0.0.1:30179) (DC: dc1)" writer.go:29: 2021-01-29T19:33:32.045Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Handled event for server in area: event=member-update server=Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f.dc1 area=wan writer.go:29: 2021-01-29T19:33:32.047Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.048Z [WARN] TestDNS_PreparedQuery_TTL: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:32.051Z [DEBUG] TestDNS_PreparedQuery_TTL.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:32.058Z [DEBUG] TestDNS_PreparedQuery_TTL.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:32.067Z [INFO] TestDNS_PreparedQuery_TTL.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:87353a5a-ba32-dddb-4ad9-9641ef5a4bc0 Address:127.0.0.1:30185}]" writer.go:29: 2021-01-29T19:33:32.070Z [INFO] TestDNS_PreparedQuery_TTL.server.raft: entering follower state: follower="Node at 127.0.0.1:30185 [Follower]" leader= writer.go:29: 2021-01-29T19:33:32.071Z [INFO] TestDNS_PreparedQuery_TTL.server.serf.wan: serf: EventMemberJoin: Node-87353a5a-ba32-dddb-4ad9-9641ef5a4bc0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:32.073Z [INFO] TestDNS_PreparedQuery_TTL.server.serf.lan: serf: EventMemberJoin: Node-87353a5a-ba32-dddb-4ad9-9641ef5a4bc0 127.0.0.1 writer.go:29: 2021-01-29T19:33:32.076Z [INFO] TestDNS_PreparedQuery_TTL: Started DNS server: address=127.0.0.1:30180 network=udp writer.go:29: 2021-01-29T19:33:32.077Z [INFO] TestDNS_PreparedQuery_TTL.server: Adding LAN server: server="Node-87353a5a-ba32-dddb-4ad9-9641ef5a4bc0 (Addr: tcp/127.0.0.1:30185) (DC: dc1)" writer.go:29: 2021-01-29T19:33:32.078Z [INFO] TestDNS_PreparedQuery_TTL.server: Handled event for server in area: event=member-join server=Node-87353a5a-ba32-dddb-4ad9-9641ef5a4bc0.dc1 area=wan writer.go:29: 2021-01-29T19:33:32.080Z [INFO] TestDNS_PreparedQuery_TTL: Started DNS server: address=127.0.0.1:30180 network=tcp writer.go:29: 2021-01-29T19:33:32.084Z [INFO] TestDNS_PreparedQuery_TTL: Started HTTP server: address=127.0.0.1:30181 network=tcp writer.go:29: 2021-01-29T19:33:32.085Z [INFO] TestDNS_PreparedQuery_TTL: started state syncer writer.go:29: 2021-01-29T19:33:32.120Z [WARN] TestDNS_PreparedQuery_TTL.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:32.121Z [INFO] TestDNS_PreparedQuery_TTL.server.raft: entering candidate state: node="Node at 127.0.0.1:30185 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.123Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:32.125Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.126Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Skipping self join check for node since the cluster is too small: node=Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f writer.go:29: 2021-01-29T19:33:32.128Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: member joined, marking health alive: member=Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f writer.go:29: 2021-01-29T19:33:32.134Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: Skipping self join check for node since the cluster is too small: node=Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.136Z [DEBUG] TestDNS_PreparedQuery_TTL.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:32.139Z [DEBUG] TestDNS_PreparedQuery_TTL.server.raft: vote granted: from=87353a5a-ba32-dddb-4ad9-9641ef5a4bc0 term=2 tally=1 writer.go:29: 2021-01-29T19:33:32.146Z [INFO] TestDNS_PreparedQuery_TTL.server.raft: election won: tally=1 === CONT TestDNS_AddressLookup writer.go:29: 2021-01-29T19:33:32.149Z [INFO] TestDNS_AddressLookup: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:32.151Z [INFO] TestDNS_AddressLookup: Endpoints down --- PASS: TestDNS_AddressLookup (0.98s) === CONT TestDNS_ServiceLookup_TTL === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.152Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.acl: dropping node from result due to ACLs: node=Node-a9196612-a7c5-37ab-535f-ffb77bdbe26f === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.148Z [INFO] TestDNS_PreparedQuery_TTL.server.raft: entering leader state: leader="Node at 127.0.0.1:30185 [Leader]" === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.157Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.acl: dropping node from result due to ACLs: node=foo === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.156Z [INFO] TestDNS_PreparedQuery_TTL.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.158Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.dns: request served from client: name=foo.service.consul. type=A class=IN latency=1.500084ms client=127.0.0.1:40659 client_network=udp writer.go:29: 2021-01-29T19:33:32.159Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: Requesting shutdown writer.go:29: 2021-01-29T19:33:32.160Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server: shutting down server === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.160Z [INFO] TestDNS_PreparedQuery_TTL.server: New leader elected: payload=Node-87353a5a-ba32-dddb-4ad9-9641ef5a4bc0 === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.161Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:32.162Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:32.163Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.163Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:32.161Z [ERROR] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:32.166Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:32.167Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:32.167Z [DEBUG] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.180Z [WARN] TestDNS_ServiceLookup_TTL: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:32.182Z [DEBUG] TestDNS_ServiceLookup_TTL.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:32.184Z [DEBUG] TestDNS_ServiceLookup_TTL.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.190Z [WARN] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.191Z [DEBUG] TestDNS_PreparedQuery_TTL.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30185 === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.199Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:32.200Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: consul server down writer.go:29: 2021-01-29T19:33:32.203Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: shutdown complete writer.go:29: 2021-01-29T19:33:32.204Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: Stopping server: protocol=DNS address=127.0.0.1:30174 network=tcp writer.go:29: 2021-01-29T19:33:32.208Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: Stopping server: protocol=DNS address=127.0.0.1:30174 network=udp === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.210Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.211Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: Stopping server: protocol=HTTP address=127.0.0.1:30175 network=tcp === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.251Z [INFO] TestDNS_PreparedQuery_TTL.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:32.252Z [INFO] TestDNS_PreparedQuery_TTL.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.253Z [DEBUG] TestDNS_PreparedQuery_TTL.server: Skipping self join check for node since the cluster is too small: node=Node-87353a5a-ba32-dddb-4ad9-9641ef5a4bc0 writer.go:29: 2021-01-29T19:33:32.255Z [INFO] TestDNS_PreparedQuery_TTL.server: member joined, marking health alive: member=Node-87353a5a-ba32-dddb-4ad9-9641ef5a4bc0 === CONT TestDNS_ServiceLookup_SRV_RFC_TCP_Default writer.go:29: 2021-01-29T19:33:32.257Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:32.258Z [INFO] TestDNS_ServiceLookup_SRV_RFC_TCP_Default: Endpoints down --- PASS: TestDNS_ServiceLookup_SRV_RFC_TCP_Default (0.71s) === CONT TestDNS_NodeLookup_TTL === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.261Z [INFO] TestDNS_ServiceLookup_TTL.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:485e7183-3814-07f5-7e7a-a22c5babf127 Address:127.0.0.1:30191}]" writer.go:29: 2021-01-29T19:33:32.266Z [INFO] TestDNS_ServiceLookup_TTL.server.serf.wan: serf: EventMemberJoin: Node-485e7183-3814-07f5-7e7a-a22c5babf127.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:32.270Z [INFO] TestDNS_ServiceLookup_TTL.server.raft: entering follower state: follower="Node at 127.0.0.1:30191 [Follower]" leader= writer.go:29: 2021-01-29T19:33:32.277Z [INFO] TestDNS_ServiceLookup_TTL.server.serf.lan: serf: EventMemberJoin: Node-485e7183-3814-07f5-7e7a-a22c5babf127 127.0.0.1 writer.go:29: 2021-01-29T19:33:32.290Z [INFO] TestDNS_ServiceLookup_TTL: Started DNS server: address=127.0.0.1:30186 network=udp writer.go:29: 2021-01-29T19:33:32.293Z [INFO] TestDNS_ServiceLookup_TTL.server: Handled event for server in area: event=member-join server=Node-485e7183-3814-07f5-7e7a-a22c5babf127.dc1 area=wan writer.go:29: 2021-01-29T19:33:32.303Z [INFO] TestDNS_ServiceLookup_TTL: Started DNS server: address=127.0.0.1:30186 network=tcp writer.go:29: 2021-01-29T19:33:32.307Z [INFO] TestDNS_ServiceLookup_TTL: Started HTTP server: address=127.0.0.1:30187 network=tcp writer.go:29: 2021-01-29T19:33:32.308Z [INFO] TestDNS_ServiceLookup_TTL: started state syncer === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.314Z [DEBUG] TestDNS_PreparedQuery_TTL: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.319Z [INFO] TestDNS_ServiceLookup_TTL.server: Adding LAN server: server="Node-485e7183-3814-07f5-7e7a-a22c5babf127 (Addr: tcp/127.0.0.1:30191) (DC: dc1)" === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.327Z [WARN] TestDNS_NodeLookup_TTL: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:32.349Z [DEBUG] TestDNS_NodeLookup_TTL.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.343Z [WARN] TestDNS_ServiceLookup_TTL.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.353Z [DEBUG] TestDNS_NodeLookup_TTL.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.354Z [INFO] TestDNS_ServiceLookup_TTL.server.raft: entering candidate state: node="Node at 127.0.0.1:30191 [Candidate]" term=2 === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.357Z [INFO] TestDNS_PreparedQuery_TTL: Synced node info === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.374Z [DEBUG] TestDNS_ServiceLookup_TTL.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:32.377Z [DEBUG] TestDNS_ServiceLookup_TTL.server.raft: vote granted: from=485e7183-3814-07f5-7e7a-a22c5babf127 term=2 tally=1 writer.go:29: 2021-01-29T19:33:32.378Z [INFO] TestDNS_ServiceLookup_TTL.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:32.379Z [INFO] TestDNS_ServiceLookup_TTL.server.raft: entering leader state: leader="Node at 127.0.0.1:30191 [Leader]" writer.go:29: 2021-01-29T19:33:32.383Z [INFO] TestDNS_ServiceLookup_TTL.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:32.386Z [INFO] TestDNS_ServiceLookup_TTL.server: New leader elected: payload=Node-485e7183-3814-07f5-7e7a-a22c5babf127 === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.391Z [INFO] TestDNS_NodeLookup_TTL.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8f935cef-7ef1-8be9-28e1-1b39f8b46e44 Address:127.0.0.1:30197}]" === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.388Z [DEBUG] TestDNS_ServiceLookup_TTL.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30191 === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.394Z [INFO] TestDNS_NodeLookup_TTL.server.raft: entering follower state: follower="Node at 127.0.0.1:30197 [Follower]" leader= writer.go:29: 2021-01-29T19:33:32.394Z [INFO] TestDNS_NodeLookup_TTL.server.serf.wan: serf: EventMemberJoin: Node-8f935cef-7ef1-8be9-28e1-1b39f8b46e44.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:32.403Z [INFO] TestDNS_NodeLookup_TTL.server.serf.lan: serf: EventMemberJoin: Node-8f935cef-7ef1-8be9-28e1-1b39f8b46e44 127.0.0.1 writer.go:29: 2021-01-29T19:33:32.408Z [DEBUG] TestDNS_NodeLookup_TTL.dns: recursor enabled writer.go:29: 2021-01-29T19:33:32.411Z [INFO] TestDNS_NodeLookup_TTL: Started DNS server: address=127.0.0.1:30192 network=udp writer.go:29: 2021-01-29T19:33:32.408Z [INFO] TestDNS_NodeLookup_TTL.server: Adding LAN server: server="Node-8f935cef-7ef1-8be9-28e1-1b39f8b46e44 (Addr: tcp/127.0.0.1:30197) (DC: dc1)" writer.go:29: 2021-01-29T19:33:32.409Z [INFO] TestDNS_NodeLookup_TTL.server: Handled event for server in area: event=member-join server=Node-8f935cef-7ef1-8be9-28e1-1b39f8b46e44.dc1 area=wan writer.go:29: 2021-01-29T19:33:32.411Z [DEBUG] TestDNS_NodeLookup_TTL.dns: recursor enabled writer.go:29: 2021-01-29T19:33:32.418Z [INFO] TestDNS_NodeLookup_TTL: Started DNS server: address=127.0.0.1:30192 network=tcp === RUN TestDNS_PreparedQuery_TTL/db.query.consul. === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.446Z [INFO] TestDNS_NodeLookup_TTL: Started HTTP server: address=127.0.0.1:30193 network=tcp writer.go:29: 2021-01-29T19:33:32.449Z [INFO] TestDNS_NodeLookup_TTL: started state syncer writer.go:29: 2021-01-29T19:33:32.469Z [WARN] TestDNS_NodeLookup_TTL.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:32.471Z [INFO] TestDNS_NodeLookup_TTL.server.raft: entering candidate state: node="Node at 127.0.0.1:30197 [Candidate]" term=2 === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.461Z [DEBUG] TestDNS_PreparedQuery_TTL.dns: request served from client: name=db.query.consul. type=SRV class=IN latency=420.519µs client=127.0.0.1:47600 client_network=udp === RUN TestDNS_PreparedQuery_TTL/db-ttl.query.consul. === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.483Z [DEBUG] TestDNS_PreparedQuery_TTL.dns: request served from client: name=db-ttl.query.consul. type=SRV class=IN latency=123.663µs client=127.0.0.1:33410 client_network=udp === RUN TestDNS_PreparedQuery_TTL/dblb.query.consul. === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.530Z [DEBUG] TestDNS_NodeLookup_TTL.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.530Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.532Z [DEBUG] TestDNS_NodeLookup_TTL.server.raft: vote granted: from=8f935cef-7ef1-8be9-28e1-1b39f8b46e44 term=2 tally=1 writer.go:29: 2021-01-29T19:33:32.535Z [INFO] TestDNS_NodeLookup_TTL.server.raft: election won: tally=1 === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.538Z [DEBUG] TestDNS_PreparedQuery_TTL.dns: request served from client: name=dblb.query.consul. type=SRV class=IN latency=24.041569ms client=127.0.0.1:46794 client_network=udp === RUN TestDNS_PreparedQuery_TTL/dblb-ttl.query.consul. === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.542Z [INFO] TestDNS_NodeLookup_TTL.server.raft: entering leader state: leader="Node at 127.0.0.1:30197 [Leader]" writer.go:29: 2021-01-29T19:33:32.547Z [INFO] TestDNS_NodeLookup_TTL.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:32.551Z [INFO] TestDNS_NodeLookup_TTL.server: New leader elected: payload=Node-8f935cef-7ef1-8be9-28e1-1b39f8b46e44 === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.552Z [DEBUG] TestDNS_PreparedQuery_TTL.dns: request served from client: name=dblb-ttl.query.consul. type=SRV class=IN latency=95.347µs client=127.0.0.1:34492 client_network=udp === RUN TestDNS_PreparedQuery_TTL/dk.query.consul. === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.555Z [DEBUG] TestDNS_NodeLookup_TTL.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30197 === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.557Z [DEBUG] TestDNS_PreparedQuery_TTL.dns: request served from client: name=dk.query.consul. type=SRV class=IN latency=53.112µs client=127.0.0.1:39773 client_network=udp === RUN TestDNS_PreparedQuery_TTL/dk-ttl.query.consul. === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.560Z [INFO] TestDNS_ServiceLookup_TTL.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:32.562Z [INFO] TestDNS_ServiceLookup_TTL.leader: started routine: routine="CA root pruning" === RUN TestDNS_ServiceLookup_TTL/db.service.consul. === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.563Z [DEBUG] TestDNS_ServiceLookup_TTL.server: Skipping self join check for node since the cluster is too small: node=Node-485e7183-3814-07f5-7e7a-a22c5babf127 writer.go:29: 2021-01-29T19:33:32.569Z [INFO] TestDNS_ServiceLookup_TTL.server: member joined, marking health alive: member=Node-485e7183-3814-07f5-7e7a-a22c5babf127 === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.587Z [DEBUG] TestDNS_PreparedQuery_TTL.dns: request served from client: name=dk-ttl.query.consul. type=SRV class=IN latency=100.848µs client=127.0.0.1:39767 client_network=udp === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.588Z [DEBUG] TestDNS_ServiceLookup_TTL.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=100.579µs client=127.0.0.1:43905 client_network=udp === RUN TestDNS_ServiceLookup_TTL/dblb.service.consul. === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.591Z [DEBUG] TestDNS_ServiceLookup_TTL.dns: request served from client: name=dblb.service.consul. type=SRV class=IN latency=44.897µs client=127.0.0.1:53050 client_network=udp === RUN TestDNS_ServiceLookup_TTL/dk.service.consul. === RUN TestDNS_PreparedQuery_TTL/api.query.consul. === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.594Z [DEBUG] TestDNS_ServiceLookup_TTL.dns: request served from client: name=dk.service.consul. type=SRV class=IN latency=175.7µs client=127.0.0.1:54335 client_network=udp === RUN TestDNS_ServiceLookup_TTL/api.service.consul. === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.606Z [DEBUG] TestDNS_ServiceLookup_TTL.dns: request served from client: name=api.service.consul. type=SRV class=IN latency=198.077µs client=127.0.0.1:43819 client_network=udp writer.go:29: 2021-01-29T19:33:32.608Z [INFO] TestDNS_ServiceLookup_TTL: Requesting shutdown === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.612Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.611Z [INFO] TestDNS_ServiceLookup_TTL.server: shutting down server writer.go:29: 2021-01-29T19:33:32.613Z [DEBUG] TestDNS_ServiceLookup_TTL.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.614Z [WARN] TestDNS_ServiceLookup_TTL.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:32.615Z [DEBUG] TestDNS_ServiceLookup_TTL.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.615Z [ERROR] TestDNS_ServiceLookup_TTL.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.606Z [DEBUG] TestDNS_PreparedQuery_TTL.dns: request served from client: name=api.query.consul. type=SRV class=IN latency=89.258µs client=127.0.0.1:52021 client_network=udp === RUN TestDNS_PreparedQuery_TTL/api-ttl.query.consul. === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.616Z [WARN] TestDNS_ServiceLookup_TTL.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.618Z [INFO] TestDNS_NodeLookup_TTL: Synced node info writer.go:29: 2021-01-29T19:33:32.619Z [DEBUG] TestDNS_NodeLookup_TTL: Node info in sync === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.620Z [DEBUG] TestDNS_PreparedQuery_TTL.dns: request served from client: name=api-ttl.query.consul. type=SRV class=IN latency=130.045µs client=127.0.0.1:46332 client_network=udp === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.624Z [INFO] TestDNS_NodeLookup_TTL.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:32.625Z [INFO] TestDNS_NodeLookup_TTL.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.625Z [INFO] TestDNS_ServiceLookup_TTL.server.router.manager: shutting down === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.626Z [DEBUG] TestDNS_NodeLookup_TTL.server: Skipping self join check for node since the cluster is too small: node=Node-8f935cef-7ef1-8be9-28e1-1b39f8b46e44 === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.627Z [INFO] TestDNS_ServiceLookup_TTL: consul server down writer.go:29: 2021-01-29T19:33:32.628Z [INFO] TestDNS_ServiceLookup_TTL: shutdown complete === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.627Z [INFO] TestDNS_NodeLookup_TTL.server: member joined, marking health alive: member=Node-8f935cef-7ef1-8be9-28e1-1b39f8b46e44 === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.629Z [INFO] TestDNS_ServiceLookup_TTL: Stopping server: protocol=DNS address=127.0.0.1:30186 network=tcp === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.622Z [INFO] TestDNS_PreparedQuery_TTL: Requesting shutdown writer.go:29: 2021-01-29T19:33:32.631Z [INFO] TestDNS_PreparedQuery_TTL.server: shutting down server writer.go:29: 2021-01-29T19:33:32.632Z [DEBUG] TestDNS_PreparedQuery_TTL.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.633Z [WARN] TestDNS_PreparedQuery_TTL.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:32.633Z [DEBUG] TestDNS_PreparedQuery_TTL.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.637Z [WARN] TestDNS_PreparedQuery_TTL.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:32.630Z [INFO] TestDNS_ServiceLookup_TTL: Stopping server: protocol=DNS address=127.0.0.1:30186 network=udp writer.go:29: 2021-01-29T19:33:32.643Z [INFO] TestDNS_ServiceLookup_TTL: Stopping server: protocol=HTTP address=127.0.0.1:30187 network=tcp === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:32.643Z [INFO] TestDNS_PreparedQuery_TTL.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:32.645Z [INFO] TestDNS_PreparedQuery_TTL: consul server down writer.go:29: 2021-01-29T19:33:32.646Z [INFO] TestDNS_PreparedQuery_TTL: shutdown complete writer.go:29: 2021-01-29T19:33:32.648Z [INFO] TestDNS_PreparedQuery_TTL: Stopping server: protocol=DNS address=127.0.0.1:30180 network=tcp writer.go:29: 2021-01-29T19:33:32.649Z [INFO] TestDNS_PreparedQuery_TTL: Stopping server: protocol=DNS address=127.0.0.1:30180 network=udp writer.go:29: 2021-01-29T19:33:32.651Z [INFO] TestDNS_PreparedQuery_TTL: Stopping server: protocol=HTTP address=127.0.0.1:30181 network=tcp === CONT TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous writer.go:29: 2021-01-29T19:33:32.713Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:32.714Z [INFO] TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous: Endpoints down --- PASS: TestDNS_ServiceLookup_FilterACL (1.52s) --- PASS: TestDNS_ServiceLookup_FilterACL/ACLToken_==_root (0.70s) --- PASS: TestDNS_ServiceLookup_FilterACL/ACLToken_==_anonymous (0.82s) === CONT TestDNS_ServiceLookup_ServiceAddress_CNAME writer.go:29: 2021-01-29T19:33:32.725Z [WARN] TestDNS_ServiceLookup_ServiceAddress_CNAME: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:32.726Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:32.728Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:32.740Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0c4f3a96-af6c-41fc-8b77-75bd9188cc0d Address:127.0.0.1:30203}]" writer.go:29: 2021-01-29T19:33:32.742Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.raft: entering follower state: follower="Node at 127.0.0.1:30203 [Follower]" leader= writer.go:29: 2021-01-29T19:33:32.742Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.serf.wan: serf: EventMemberJoin: Node-0c4f3a96-af6c-41fc-8b77-75bd9188cc0d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:32.751Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.serf.lan: serf: EventMemberJoin: Node-0c4f3a96-af6c-41fc-8b77-75bd9188cc0d 127.0.0.1 writer.go:29: 2021-01-29T19:33:32.754Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server: Handled event for server in area: event=member-join server=Node-0c4f3a96-af6c-41fc-8b77-75bd9188cc0d.dc1 area=wan writer.go:29: 2021-01-29T19:33:32.754Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server: Adding LAN server: server="Node-0c4f3a96-af6c-41fc-8b77-75bd9188cc0d (Addr: tcp/127.0.0.1:30203) (DC: dc1)" writer.go:29: 2021-01-29T19:33:32.755Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.dns: recursor enabled writer.go:29: 2021-01-29T19:33:32.760Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: Started DNS server: address=127.0.0.1:30198 network=udp writer.go:29: 2021-01-29T19:33:32.755Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.dns: recursor enabled writer.go:29: 2021-01-29T19:33:32.763Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: Started DNS server: address=127.0.0.1:30198 network=tcp writer.go:29: 2021-01-29T19:33:32.765Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: Started HTTP server: address=127.0.0.1:30199 network=tcp writer.go:29: 2021-01-29T19:33:32.768Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: started state syncer writer.go:29: 2021-01-29T19:33:32.805Z [WARN] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:32.806Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.raft: entering candidate state: node="Node at 127.0.0.1:30203 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:32.819Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:32.820Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.raft: vote granted: from=0c4f3a96-af6c-41fc-8b77-75bd9188cc0d term=2 tally=1 writer.go:29: 2021-01-29T19:33:32.822Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:32.824Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.raft: entering leader state: leader="Node at 127.0.0.1:30203 [Leader]" writer.go:29: 2021-01-29T19:33:32.827Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:32.829Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server: New leader elected: payload=Node-0c4f3a96-af6c-41fc-8b77-75bd9188cc0d writer.go:29: 2021-01-29T19:33:32.834Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30203 === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.843Z [DEBUG] TestDNS_NodeLookup_TTL.dns: request served from client: name=foo.node.consul. type=ANY class=IN latency=81.279µs client=127.0.0.1:49659 client_network=udp writer.go:29: 2021-01-29T19:33:32.845Z [DEBUG] TestDNS_NodeLookup_TTL.dns: request served from client: name=bar.node.consul. type=ANY class=IN latency=117.93µs client=127.0.0.1:40314 client_network=udp === CONT TestDNS_ServiceLookup_ServiceAddress_CNAME writer.go:29: 2021-01-29T19:33:32.849Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:32.849Z [DEBUG] TestDNS_NodeLookup_TTL.dns: cname recurse RTT for name: name=www.google.com. rtt=827.779µs writer.go:29: 2021-01-29T19:33:32.853Z [INFO] TestDNS_NodeLookup_TTL: Requesting shutdown writer.go:29: 2021-01-29T19:33:32.856Z [INFO] TestDNS_NodeLookup_TTL.server: shutting down server writer.go:29: 2021-01-29T19:33:32.853Z [DEBUG] TestDNS_NodeLookup_TTL.dns: request served from client: name=google.node.consul. type=ANY class=IN latency=4.8917ms client=127.0.0.1:40264 client_network=udp writer.go:29: 2021-01-29T19:33:32.857Z [DEBUG] TestDNS_NodeLookup_TTL.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.877Z [WARN] TestDNS_NodeLookup_TTL.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:32.877Z [DEBUG] TestDNS_NodeLookup_TTL.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.880Z [WARN] TestDNS_NodeLookup_TTL.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:32.885Z [INFO] TestDNS_NodeLookup_TTL.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:32.887Z [INFO] TestDNS_NodeLookup_TTL: consul server down writer.go:29: 2021-01-29T19:33:32.889Z [INFO] TestDNS_NodeLookup_TTL: shutdown complete writer.go:29: 2021-01-29T19:33:32.890Z [INFO] TestDNS_NodeLookup_TTL: Stopping server: protocol=DNS address=127.0.0.1:30192 network=tcp writer.go:29: 2021-01-29T19:33:32.892Z [INFO] TestDNS_NodeLookup_TTL: Stopping server: protocol=DNS address=127.0.0.1:30192 network=udp writer.go:29: 2021-01-29T19:33:32.893Z [INFO] TestDNS_NodeLookup_TTL: Stopping server: protocol=HTTP address=127.0.0.1:30193 network=tcp === CONT TestDNS_ServiceLookup_ServiceAddress_CNAME writer.go:29: 2021-01-29T19:33:32.905Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:32.912Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.924Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.server: Skipping self join check for node since the cluster is too small: node=Node-0c4f3a96-af6c-41fc-8b77-75bd9188cc0d writer.go:29: 2021-01-29T19:33:32.931Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server: member joined, marking health alive: member=Node-0c4f3a96-af6c-41fc-8b77-75bd9188cc0d writer.go:29: 2021-01-29T19:33:33.102Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:33.106Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: Synced node info writer.go:29: 2021-01-29T19:33:33.108Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME: Node info in sync writer.go:29: 2021-01-29T19:33:33.108Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.dns: cname recurse RTT for name: name=www.google.com. rtt=722.95µs writer.go:29: 2021-01-29T19:33:33.112Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.dns: request served from client: name=search.service.consul. type=ANY class=IN latency=5.424631ms client=127.0.0.1:55452 client_network=udp writer.go:29: 2021-01-29T19:33:33.113Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.dns: cname recurse RTT for name: name=www.google.com. rtt=541.69µs writer.go:29: 2021-01-29T19:33:33.119Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.dns: request served from client: name=957e95c4-628f-9af2-92cb-f5ff4e256458.query.consul. type=ANY class=IN latency=6.907355ms client=127.0.0.1:38502 client_network=udp writer.go:29: 2021-01-29T19:33:33.119Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: Requesting shutdown writer.go:29: 2021-01-29T19:33:33.123Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server: shutting down server writer.go:29: 2021-01-29T19:33:33.124Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:33.126Z [WARN] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:33.126Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_CNAME.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:33.132Z [WARN] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:33.135Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:33.135Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: consul server down writer.go:29: 2021-01-29T19:33:33.138Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: shutdown complete writer.go:29: 2021-01-29T19:33:33.139Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: Stopping server: protocol=DNS address=127.0.0.1:30198 network=tcp writer.go:29: 2021-01-29T19:33:33.141Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: Stopping server: protocol=DNS address=127.0.0.1:30198 network=udp writer.go:29: 2021-01-29T19:33:33.143Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: Stopping server: protocol=HTTP address=127.0.0.1:30199 network=tcp === CONT TestDNS_ServiceLookup_TTL writer.go:29: 2021-01-29T19:33:33.144Z [INFO] TestDNS_ServiceLookup_TTL: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:33.146Z [INFO] TestDNS_ServiceLookup_TTL: Endpoints down --- PASS: TestDNS_ServiceLookup_TTL (1.00s) --- PASS: TestDNS_ServiceLookup_TTL/db.service.consul. (0.01s) --- PASS: TestDNS_ServiceLookup_TTL/dblb.service.consul. (0.00s) --- PASS: TestDNS_ServiceLookup_TTL/dk.service.consul. (0.00s) --- PASS: TestDNS_ServiceLookup_TTL/api.service.consul. (0.00s) === CONT TestDNS_ServiceLookup_AnswerLimits === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:33.153Z [INFO] TestDNS_PreparedQuery_TTL: Waiting for endpoints to shut down === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} === CONT TestDNS_PreparedQuery_TTL writer.go:29: 2021-01-29T19:33:33.155Z [INFO] TestDNS_PreparedQuery_TTL: Endpoints down --- PASS: TestDNS_PreparedQuery_TTL (1.13s) --- PASS: TestDNS_PreparedQuery_TTL/db.query.consul. (0.05s) --- PASS: TestDNS_PreparedQuery_TTL/db-ttl.query.consul. (0.00s) --- PASS: TestDNS_PreparedQuery_TTL/dblb.query.consul. (0.05s) --- PASS: TestDNS_PreparedQuery_TTL/dblb-ttl.query.consul. (0.01s) --- PASS: TestDNS_PreparedQuery_TTL/dk.query.consul. (0.00s) --- PASS: TestDNS_PreparedQuery_TTL/dk-ttl.query.consul. (0.03s) --- PASS: TestDNS_PreparedQuery_TTL/api.query.consul. (0.00s) --- PASS: TestDNS_PreparedQuery_TTL/api-ttl.query.consul. (0.00s) === CONT TestDNS_ServiceLookup_LargeResponses === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.172Z [WARN] TestDNS_ServiceLookup_LargeResponses: bootstrap = true: do not enable unless necessary === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.174Z [DEBUG] TestDNS_ServiceLookup_LargeResponses.tlsutil: Update: version=1 === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.176Z [DEBUG] TestDNS_ServiceLookup_LargeResponses.tlsutil: OutgoingRPCWrapper: version=1 === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.186Z [INFO] TestDNS_ServiceLookup_LargeResponses.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1c6ce58c-1947-02d3-e94d-a2b8ed43ebf9 Address:127.0.0.1:30209}]" === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.190Z [INFO] TestDNS_ServiceLookup_LargeResponses.server.raft: entering follower state: follower="Node at 127.0.0.1:30209 [Follower]" leader= === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.192Z [INFO] TestDNS_ServiceLookup_LargeResponses.server.serf.wan: serf: EventMemberJoin: Node-1c6ce58c-1947-02d3-e94d-a2b8ed43ebf9.dc1 127.0.0.1 === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} === RUN TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} === PAUSE TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} === CONT TestDNS_ServiceLookup_Truncate === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.197Z [INFO] TestDNS_ServiceLookup_LargeResponses.server.serf.lan: serf: EventMemberJoin: Node-1c6ce58c-1947-02d3-e94d-a2b8ed43ebf9 127.0.0.1 writer.go:29: 2021-01-29T19:33:33.201Z [INFO] TestDNS_ServiceLookup_LargeResponses.server: Adding LAN server: server="Node-1c6ce58c-1947-02d3-e94d-a2b8ed43ebf9 (Addr: tcp/127.0.0.1:30209) (DC: dc1)" writer.go:29: 2021-01-29T19:33:33.202Z [INFO] TestDNS_ServiceLookup_LargeResponses.server: Handled event for server in area: event=member-join server=Node-1c6ce58c-1947-02d3-e94d-a2b8ed43ebf9.dc1 area=wan writer.go:29: 2021-01-29T19:33:33.204Z [INFO] TestDNS_ServiceLookup_LargeResponses: Started DNS server: address=127.0.0.1:30204 network=tcp === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.210Z [WARN] TestDNS_ServiceLookup_Truncate: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.209Z [INFO] TestDNS_ServiceLookup_LargeResponses: Started DNS server: address=127.0.0.1:30204 network=udp === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.211Z [DEBUG] TestDNS_ServiceLookup_Truncate.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.221Z [INFO] TestDNS_ServiceLookup_LargeResponses: Started HTTP server: address=127.0.0.1:30205 network=tcp === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.223Z [DEBUG] TestDNS_ServiceLookup_Truncate.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.225Z [INFO] TestDNS_ServiceLookup_LargeResponses: started state syncer === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.232Z [INFO] TestDNS_ServiceLookup_Truncate.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:73fe6e0e-ceaf-c7c9-b9e0-ace326b60900 Address:127.0.0.1:30233}]" writer.go:29: 2021-01-29T19:33:33.235Z [INFO] TestDNS_ServiceLookup_Truncate.server.raft: entering follower state: follower="Node at 127.0.0.1:30233 [Follower]" leader= writer.go:29: 2021-01-29T19:33:33.235Z [INFO] TestDNS_ServiceLookup_Truncate.server.serf.wan: serf: EventMemberJoin: Node-73fe6e0e-ceaf-c7c9-b9e0-ace326b60900.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:33.241Z [INFO] TestDNS_ServiceLookup_Truncate.server.serf.lan: serf: EventMemberJoin: Node-73fe6e0e-ceaf-c7c9-b9e0-ace326b60900 127.0.0.1 writer.go:29: 2021-01-29T19:33:33.246Z [INFO] TestDNS_ServiceLookup_Truncate.server: Handled event for server in area: event=member-join server=Node-73fe6e0e-ceaf-c7c9-b9e0-ace326b60900.dc1 area=wan writer.go:29: 2021-01-29T19:33:33.246Z [INFO] TestDNS_ServiceLookup_Truncate.server: Adding LAN server: server="Node-73fe6e0e-ceaf-c7c9-b9e0-ace326b60900 (Addr: tcp/127.0.0.1:30233) (DC: dc1)" writer.go:29: 2021-01-29T19:33:33.251Z [INFO] TestDNS_ServiceLookup_Truncate: Started DNS server: address=127.0.0.1:30228 network=tcp writer.go:29: 2021-01-29T19:33:33.255Z [INFO] TestDNS_ServiceLookup_Truncate: Started DNS server: address=127.0.0.1:30228 network=udp writer.go:29: 2021-01-29T19:33:33.257Z [INFO] TestDNS_ServiceLookup_Truncate: Started HTTP server: address=127.0.0.1:30229 network=tcp writer.go:29: 2021-01-29T19:33:33.259Z [INFO] TestDNS_ServiceLookup_Truncate: started state syncer === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.261Z [WARN] TestDNS_ServiceLookup_LargeResponses.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:33.264Z [INFO] TestDNS_ServiceLookup_LargeResponses.server.raft: entering candidate state: node="Node at 127.0.0.1:30209 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:33.267Z [DEBUG] TestDNS_ServiceLookup_LargeResponses.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:33.269Z [DEBUG] TestDNS_ServiceLookup_LargeResponses.server.raft: vote granted: from=1c6ce58c-1947-02d3-e94d-a2b8ed43ebf9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:33.270Z [INFO] TestDNS_ServiceLookup_LargeResponses.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:33.272Z [INFO] TestDNS_ServiceLookup_LargeResponses.server.raft: entering leader state: leader="Node at 127.0.0.1:30209 [Leader]" writer.go:29: 2021-01-29T19:33:33.273Z [INFO] TestDNS_ServiceLookup_LargeResponses.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:33.274Z [INFO] TestDNS_ServiceLookup_LargeResponses.server: New leader elected: payload=Node-1c6ce58c-1947-02d3-e94d-a2b8ed43ebf9 writer.go:29: 2021-01-29T19:33:33.276Z [DEBUG] TestDNS_ServiceLookup_LargeResponses.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30209 writer.go:29: 2021-01-29T19:33:33.282Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.284Z [WARN] TestDNS_ServiceLookup_Truncate.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:33.285Z [INFO] TestDNS_ServiceLookup_Truncate.server.raft: entering candidate state: node="Node at 127.0.0.1:30233 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:33.289Z [DEBUG] TestDNS_ServiceLookup_Truncate.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:33.290Z [DEBUG] TestDNS_ServiceLookup_Truncate.server.raft: vote granted: from=73fe6e0e-ceaf-c7c9-b9e0-ace326b60900 term=2 tally=1 writer.go:29: 2021-01-29T19:33:33.291Z [INFO] TestDNS_ServiceLookup_Truncate.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:33.292Z [INFO] TestDNS_ServiceLookup_Truncate.server.raft: entering leader state: leader="Node at 127.0.0.1:30233 [Leader]" writer.go:29: 2021-01-29T19:33:33.293Z [INFO] TestDNS_ServiceLookup_Truncate.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:33.293Z [INFO] TestDNS_ServiceLookup_Truncate.server: New leader elected: payload=Node-73fe6e0e-ceaf-c7c9-b9e0-ace326b60900 writer.go:29: 2021-01-29T19:33:33.294Z [DEBUG] TestDNS_ServiceLookup_Truncate.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30233 writer.go:29: 2021-01-29T19:33:33.298Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.298Z [INFO] TestDNS_ServiceLookup_LargeResponses.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:33.300Z [INFO] TestDNS_ServiceLookup_LargeResponses.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:33.301Z [DEBUG] TestDNS_ServiceLookup_LargeResponses.server: Skipping self join check for node since the cluster is too small: node=Node-1c6ce58c-1947-02d3-e94d-a2b8ed43ebf9 === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.303Z [INFO] TestDNS_ServiceLookup_Truncate.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.303Z [INFO] TestDNS_ServiceLookup_LargeResponses.server: member joined, marking health alive: member=Node-1c6ce58c-1947-02d3-e94d-a2b8ed43ebf9 === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.304Z [INFO] TestDNS_ServiceLookup_Truncate.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:33.305Z [DEBUG] TestDNS_ServiceLookup_Truncate.server: Skipping self join check for node since the cluster is too small: node=Node-73fe6e0e-ceaf-c7c9-b9e0-ace326b60900 writer.go:29: 2021-01-29T19:33:33.306Z [INFO] TestDNS_ServiceLookup_Truncate.server: member joined, marking health alive: member=Node-73fe6e0e-ceaf-c7c9-b9e0-ace326b60900 === CONT TestDNS_NodeLookup_TTL writer.go:29: 2021-01-29T19:33:33.395Z [INFO] TestDNS_NodeLookup_TTL: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:33.396Z [INFO] TestDNS_NodeLookup_TTL: Endpoints down --- PASS: TestDNS_NodeLookup_TTL (1.14s) === CONT TestBinarySearch === RUN TestBinarySearch/binarySearch_12 === RUN TestBinarySearch/binarySearch_256 === RUN TestBinarySearch/binarySearch_512 === RUN TestBinarySearch/binarySearch_8192 === RUN TestBinarySearch/binarySearch_65535 === RUN TestBinarySearch/binarySearch_12#01 === RUN TestBinarySearch/binarySearch_256#01 === RUN TestBinarySearch/binarySearch_512#01 === RUN TestBinarySearch/binarySearch_8192#01 === RUN TestBinarySearch/binarySearch_65535#01 === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.559Z [DEBUG] TestDNS_ServiceLookup_LargeResponses: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:33.560Z [DEBUG] TestDNS_ServiceLookup_LargeResponses.dns: request served from client: name=_this-is-a-very-very-very-very-very-long-name-for-a-service._master.service.consul. type=SRV class=IN latency=199.844µs client=127.0.0.1:54527 client_network=udp writer.go:29: 2021-01-29T19:33:33.560Z [INFO] TestDNS_ServiceLookup_LargeResponses: Synced node info === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.562Z [DEBUG] TestDNS_ServiceLookup_Truncate: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.563Z [DEBUG] TestDNS_ServiceLookup_LargeResponses: Node info in sync writer.go:29: 2021-01-29T19:33:33.561Z [DEBUG] TestDNS_ServiceLookup_LargeResponses.dns: request served from client: name=this-is-a-very-very-very-very-very-long-name-for-a-service.query.consul. type=SRV class=IN latency=79.861µs client=127.0.0.1:50547 client_network=udp writer.go:29: 2021-01-29T19:33:33.561Z [INFO] TestDNS_ServiceLookup_LargeResponses: Requesting shutdown writer.go:29: 2021-01-29T19:33:33.565Z [INFO] TestDNS_ServiceLookup_LargeResponses.server: shutting down server writer.go:29: 2021-01-29T19:33:33.566Z [DEBUG] TestDNS_ServiceLookup_LargeResponses.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.567Z [DEBUG] TestDNS_ServiceLookup_Truncate.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=2.294061ms client=127.0.0.1:35581 client_network=udp === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.567Z [WARN] TestDNS_ServiceLookup_LargeResponses.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.568Z [INFO] TestDNS_ServiceLookup_Truncate: Synced node info === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.568Z [DEBUG] TestDNS_ServiceLookup_LargeResponses.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.570Z [DEBUG] TestDNS_ServiceLookup_Truncate.dns: request served from client: name=38bef147-4b0d-bcab-bb22-3b65ca3acf30.query.consul. type=ANY class=IN latency=733.388µs client=127.0.0.1:36833 client_network=udp writer.go:29: 2021-01-29T19:33:33.570Z [INFO] TestDNS_ServiceLookup_Truncate: Requesting shutdown writer.go:29: 2021-01-29T19:33:33.572Z [INFO] TestDNS_ServiceLookup_Truncate.server: shutting down server writer.go:29: 2021-01-29T19:33:33.573Z [DEBUG] TestDNS_ServiceLookup_Truncate.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.573Z [WARN] TestDNS_ServiceLookup_LargeResponses.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.574Z [WARN] TestDNS_ServiceLookup_Truncate.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.574Z [INFO] TestDNS_ServiceLookup_LargeResponses.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:33.575Z [INFO] TestDNS_ServiceLookup_LargeResponses: consul server down === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.574Z [DEBUG] TestDNS_ServiceLookup_Truncate.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.576Z [INFO] TestDNS_ServiceLookup_LargeResponses: shutdown complete writer.go:29: 2021-01-29T19:33:33.578Z [INFO] TestDNS_ServiceLookup_LargeResponses: Stopping server: protocol=DNS address=127.0.0.1:30204 network=tcp --- PASS: TestBinarySearch (0.18s) --- PASS: TestBinarySearch/binarySearch_12 (0.01s) --- PASS: TestBinarySearch/binarySearch_256 (0.01s) --- PASS: TestBinarySearch/binarySearch_512 (0.01s) --- PASS: TestBinarySearch/binarySearch_8192 (0.02s) --- PASS: TestBinarySearch/binarySearch_65535 (0.02s) --- PASS: TestBinarySearch/binarySearch_12#01 (0.01s) --- PASS: TestBinarySearch/binarySearch_256#01 (0.02s) --- PASS: TestBinarySearch/binarySearch_512#01 (0.01s) --- PASS: TestBinarySearch/binarySearch_8192#01 (0.01s) --- PASS: TestBinarySearch/binarySearch_65535#01 (0.04s) === CONT TestDNS_ServiceLookup_Randomize === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:33.581Z [INFO] TestDNS_ServiceLookup_LargeResponses: Stopping server: protocol=DNS address=127.0.0.1:30204 network=udp writer.go:29: 2021-01-29T19:33:33.585Z [INFO] TestDNS_ServiceLookup_LargeResponses: Stopping server: protocol=HTTP address=127.0.0.1:30205 network=tcp === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:33.591Z [WARN] TestDNS_ServiceLookup_Truncate.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:33.608Z [INFO] TestDNS_ServiceLookup_Truncate.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:33.615Z [INFO] TestDNS_ServiceLookup_Truncate: consul server down writer.go:29: 2021-01-29T19:33:33.617Z [INFO] TestDNS_ServiceLookup_Truncate: shutdown complete writer.go:29: 2021-01-29T19:33:33.619Z [INFO] TestDNS_ServiceLookup_Truncate: Stopping server: protocol=DNS address=127.0.0.1:30228 network=tcp writer.go:29: 2021-01-29T19:33:33.621Z [INFO] TestDNS_ServiceLookup_Truncate: Stopping server: protocol=DNS address=127.0.0.1:30228 network=udp writer.go:29: 2021-01-29T19:33:33.622Z [INFO] TestDNS_ServiceLookup_Truncate: Stopping server: protocol=HTTP address=127.0.0.1:30229 network=tcp === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.641Z [WARN] TestDNS_ServiceLookup_Randomize: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:33.646Z [DEBUG] TestDNS_ServiceLookup_Randomize.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_ServiceAddress_CNAME writer.go:29: 2021-01-29T19:33:33.647Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:33.650Z [INFO] TestDNS_ServiceLookup_ServiceAddress_CNAME: Endpoints down --- PASS: TestDNS_ServiceLookup_ServiceAddress_CNAME (0.94s) === CONT TestDNS_ServiceLookup_OnlyPassing === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.657Z [DEBUG] TestDNS_ServiceLookup_Randomize.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:33.692Z [INFO] TestDNS_ServiceLookup_Randomize.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7d8a705e-c362-9985-130b-d06f7cd44a25 Address:127.0.0.1:30215}]" writer.go:29: 2021-01-29T19:33:33.697Z [INFO] TestDNS_ServiceLookup_Randomize.server.raft: entering follower state: follower="Node at 127.0.0.1:30215 [Follower]" leader= === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.701Z [WARN] TestDNS_ServiceLookup_OnlyPassing: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.701Z [INFO] TestDNS_ServiceLookup_Randomize.server.serf.wan: serf: EventMemberJoin: Node-7d8a705e-c362-9985-130b-d06f7cd44a25.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.702Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:33.705Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.706Z [INFO] TestDNS_ServiceLookup_Randomize.server.serf.lan: serf: EventMemberJoin: Node-7d8a705e-c362-9985-130b-d06f7cd44a25 127.0.0.1 writer.go:29: 2021-01-29T19:33:33.717Z [INFO] TestDNS_ServiceLookup_Randomize.server: Handled event for server in area: event=member-join server=Node-7d8a705e-c362-9985-130b-d06f7cd44a25.dc1 area=wan writer.go:29: 2021-01-29T19:33:33.718Z [INFO] TestDNS_ServiceLookup_Randomize: Started DNS server: address=127.0.0.1:30210 network=udp writer.go:29: 2021-01-29T19:33:33.719Z [INFO] TestDNS_ServiceLookup_Randomize.server: Adding LAN server: server="Node-7d8a705e-c362-9985-130b-d06f7cd44a25 (Addr: tcp/127.0.0.1:30215) (DC: dc1)" writer.go:29: 2021-01-29T19:33:33.724Z [INFO] TestDNS_ServiceLookup_Randomize: Started DNS server: address=127.0.0.1:30210 network=tcp writer.go:29: 2021-01-29T19:33:33.739Z [WARN] TestDNS_ServiceLookup_Randomize.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:33.742Z [INFO] TestDNS_ServiceLookup_Randomize.server.raft: entering candidate state: node="Node at 127.0.0.1:30215 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:33.746Z [DEBUG] TestDNS_ServiceLookup_Randomize.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:33.748Z [DEBUG] TestDNS_ServiceLookup_Randomize.server.raft: vote granted: from=7d8a705e-c362-9985-130b-d06f7cd44a25 term=2 tally=1 === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.750Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ac9781a8-53a7-9f8e-3499-2b9e0268ecd1 Address:127.0.0.1:30227}]" === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.750Z [INFO] TestDNS_ServiceLookup_Randomize.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:33.751Z [INFO] TestDNS_ServiceLookup_Randomize.server.raft: entering leader state: leader="Node at 127.0.0.1:30215 [Leader]" writer.go:29: 2021-01-29T19:33:33.750Z [INFO] TestDNS_ServiceLookup_Randomize: Started HTTP server: address=127.0.0.1:30211 network=tcp writer.go:29: 2021-01-29T19:33:33.758Z [INFO] TestDNS_ServiceLookup_Randomize: started state syncer === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.761Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server.serf.wan: serf: EventMemberJoin: Node-ac9781a8-53a7-9f8e-3499-2b9e0268ecd1.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.762Z [INFO] TestDNS_ServiceLookup_Randomize.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:33.763Z [INFO] TestDNS_ServiceLookup_Randomize.server: New leader elected: payload=Node-7d8a705e-c362-9985-130b-d06f7cd44a25 === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.765Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server.raft: entering follower state: follower="Node at 127.0.0.1:30227 [Follower]" leader= === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.778Z [DEBUG] TestDNS_ServiceLookup_Randomize.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30215 === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.782Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server.serf.lan: serf: EventMemberJoin: Node-ac9781a8-53a7-9f8e-3499-2b9e0268ecd1 127.0.0.1 === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.782Z [INFO] TestDNS_ServiceLookup_Randomize: Synced node info writer.go:29: 2021-01-29T19:33:33.783Z [DEBUG] TestDNS_ServiceLookup_Randomize: Node info in sync === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.797Z [INFO] TestDNS_ServiceLookup_OnlyPassing: Started DNS server: address=127.0.0.1:30222 network=udp writer.go:29: 2021-01-29T19:33:33.822Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server: Adding LAN server: server="Node-ac9781a8-53a7-9f8e-3499-2b9e0268ecd1 (Addr: tcp/127.0.0.1:30227) (DC: dc1)" === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.820Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.823Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server: Handled event for server in area: event=member-join server=Node-ac9781a8-53a7-9f8e-3499-2b9e0268ecd1.dc1 area=wan writer.go:29: 2021-01-29T19:33:33.824Z [INFO] TestDNS_ServiceLookup_OnlyPassing: Started DNS server: address=127.0.0.1:30222 network=tcp writer.go:29: 2021-01-29T19:33:33.827Z [INFO] TestDNS_ServiceLookup_OnlyPassing: Started HTTP server: address=127.0.0.1:30223 network=tcp writer.go:29: 2021-01-29T19:33:33.828Z [INFO] TestDNS_ServiceLookup_OnlyPassing: started state syncer writer.go:29: 2021-01-29T19:33:33.833Z [WARN] TestDNS_ServiceLookup_OnlyPassing.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:33.834Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server.raft: entering candidate state: node="Node at 127.0.0.1:30227 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.836Z [INFO] TestDNS_ServiceLookup_Randomize.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.836Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.837Z [INFO] TestDNS_ServiceLookup_Randomize.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.837Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.server.raft: vote granted: from=ac9781a8-53a7-9f8e-3499-2b9e0268ecd1 term=2 tally=1 writer.go:29: 2021-01-29T19:33:33.838Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:33.839Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server.raft: entering leader state: leader="Node at 127.0.0.1:30227 [Leader]" === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.838Z [DEBUG] TestDNS_ServiceLookup_Randomize.server: Skipping self join check for node since the cluster is too small: node=Node-7d8a705e-c362-9985-130b-d06f7cd44a25 === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.840Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:33.840Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server: New leader elected: payload=Node-ac9781a8-53a7-9f8e-3499-2b9e0268ecd1 === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.840Z [INFO] TestDNS_ServiceLookup_Randomize.server: member joined, marking health alive: member=Node-7d8a705e-c362-9985-130b-d06f7cd44a25 === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:33.841Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30227 writer.go:29: 2021-01-29T19:33:33.851Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:33.856Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:33.857Z [INFO] TestDNS_ServiceLookup_OnlyPassing.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:33.859Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.server: Skipping self join check for node since the cluster is too small: node=Node-ac9781a8-53a7-9f8e-3499-2b9e0268ecd1 writer.go:29: 2021-01-29T19:33:33.861Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server: member joined, marking health alive: member=Node-ac9781a8-53a7-9f8e-3499-2b9e0268ecd1 writer.go:29: 2021-01-29T19:33:33.881Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.dns: request served from client: name=db.service.consul. type=ANY class=IN latency=117.287µs client=127.0.0.1:45030 client_network=udp writer.go:29: 2021-01-29T19:33:33.882Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.dns: request served from client: name=61d8a937-e1bc-e5be-126b-6e6d47833d9c.query.consul. type=ANY class=IN latency=116.965µs client=127.0.0.1:36560 client_network=udp writer.go:29: 2021-01-29T19:33:33.882Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.tlsutil: Update: version=2 writer.go:29: 2021-01-29T19:33:33.887Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.dns: request served from client: name=db.service.consul. type=ANY class=IN latency=111.415µs client=127.0.0.1:54973 client_network=udp writer.go:29: 2021-01-29T19:33:33.890Z [INFO] TestDNS_ServiceLookup_OnlyPassing: Requesting shutdown writer.go:29: 2021-01-29T19:33:33.890Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server: shutting down server writer.go:29: 2021-01-29T19:33:33.892Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:33.892Z [ERROR] TestDNS_ServiceLookup_OnlyPassing.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:33.894Z [WARN] TestDNS_ServiceLookup_OnlyPassing.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:33.894Z [DEBUG] TestDNS_ServiceLookup_OnlyPassing.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:33.897Z [WARN] TestDNS_ServiceLookup_OnlyPassing.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:33.899Z [INFO] TestDNS_ServiceLookup_OnlyPassing.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:33.899Z [INFO] TestDNS_ServiceLookup_OnlyPassing: consul server down writer.go:29: 2021-01-29T19:33:33.904Z [INFO] TestDNS_ServiceLookup_OnlyPassing: shutdown complete writer.go:29: 2021-01-29T19:33:33.906Z [INFO] TestDNS_ServiceLookup_OnlyPassing: Stopping server: protocol=DNS address=127.0.0.1:30222 network=tcp writer.go:29: 2021-01-29T19:33:33.908Z [INFO] TestDNS_ServiceLookup_OnlyPassing: Stopping server: protocol=DNS address=127.0.0.1:30222 network=udp writer.go:29: 2021-01-29T19:33:33.909Z [INFO] TestDNS_ServiceLookup_OnlyPassing: Stopping server: protocol=HTTP address=127.0.0.1:30223 network=tcp === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:33.926Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=1.175467ms client=127.0.0.1:37045 client_network=udp writer.go:29: 2021-01-29T19:33:33.928Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=1.48505ms client=127.0.0.1:46195 client_network=udp writer.go:29: 2021-01-29T19:33:33.929Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=677.87µs client=127.0.0.1:59108 client_network=udp writer.go:29: 2021-01-29T19:33:33.932Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=2.930709ms client=127.0.0.1:33350 client_network=udp writer.go:29: 2021-01-29T19:33:33.933Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=2.412923ms client=127.0.0.1:42779 client_network=udp writer.go:29: 2021-01-29T19:33:33.935Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=2.421252ms client=127.0.0.1:43048 client_network=udp writer.go:29: 2021-01-29T19:33:33.935Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=1.446556ms client=127.0.0.1:59072 client_network=udp writer.go:29: 2021-01-29T19:33:33.937Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=1.121373ms client=127.0.0.1:57716 client_network=udp writer.go:29: 2021-01-29T19:33:33.938Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=739.232µs client=127.0.0.1:33681 client_network=udp writer.go:29: 2021-01-29T19:33:33.940Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=web.service.consul. type=ANY class=IN latency=834.24µs client=127.0.0.1:49983 client_network=udp writer.go:29: 2021-01-29T19:33:33.941Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=c6210b8c-4892-4cc7-0545-9288575728be.query.consul. type=ANY class=IN latency=750.78µs client=127.0.0.1:36237 client_network=udp writer.go:29: 2021-01-29T19:33:33.943Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=c6210b8c-4892-4cc7-0545-9288575728be.query.consul. type=ANY class=IN latency=1.013162ms client=127.0.0.1:52614 client_network=udp writer.go:29: 2021-01-29T19:33:33.953Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=c6210b8c-4892-4cc7-0545-9288575728be.query.consul. type=ANY class=IN latency=1.162453ms client=127.0.0.1:55366 client_network=udp writer.go:29: 2021-01-29T19:33:33.955Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=c6210b8c-4892-4cc7-0545-9288575728be.query.consul. type=ANY class=IN latency=663.288µs client=127.0.0.1:37617 client_network=udp writer.go:29: 2021-01-29T19:33:33.954Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=c6210b8c-4892-4cc7-0545-9288575728be.query.consul. type=ANY class=IN latency=722.828µs client=127.0.0.1:54337 client_network=udp writer.go:29: 2021-01-29T19:33:33.959Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=c6210b8c-4892-4cc7-0545-9288575728be.query.consul. type=ANY class=IN latency=2.684738ms client=127.0.0.1:41013 client_network=udp writer.go:29: 2021-01-29T19:33:33.960Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=c6210b8c-4892-4cc7-0545-9288575728be.query.consul. type=ANY class=IN latency=2.061043ms client=127.0.0.1:36401 client_network=udp writer.go:29: 2021-01-29T19:33:33.961Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=c6210b8c-4892-4cc7-0545-9288575728be.query.consul. type=ANY class=IN latency=680.306µs client=127.0.0.1:35128 client_network=udp writer.go:29: 2021-01-29T19:33:33.963Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=c6210b8c-4892-4cc7-0545-9288575728be.query.consul. type=ANY class=IN latency=923.276µs client=127.0.0.1:54449 client_network=udp writer.go:29: 2021-01-29T19:33:33.964Z [DEBUG] TestDNS_ServiceLookup_Randomize.dns: request served from client: name=c6210b8c-4892-4cc7-0545-9288575728be.query.consul. type=ANY class=IN latency=710.266µs client=127.0.0.1:38466 client_network=udp writer.go:29: 2021-01-29T19:33:33.964Z [INFO] TestDNS_ServiceLookup_Randomize: Requesting shutdown writer.go:29: 2021-01-29T19:33:33.968Z [INFO] TestDNS_ServiceLookup_Randomize.server: shutting down server writer.go:29: 2021-01-29T19:33:33.969Z [DEBUG] TestDNS_ServiceLookup_Randomize.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:33.970Z [WARN] TestDNS_ServiceLookup_Randomize.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:33.971Z [DEBUG] TestDNS_ServiceLookup_Randomize.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:33.972Z [WARN] TestDNS_ServiceLookup_Randomize.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:33.974Z [INFO] TestDNS_ServiceLookup_Randomize.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:33.975Z [INFO] TestDNS_ServiceLookup_Randomize: consul server down writer.go:29: 2021-01-29T19:33:33.976Z [INFO] TestDNS_ServiceLookup_Randomize: shutdown complete writer.go:29: 2021-01-29T19:33:33.977Z [INFO] TestDNS_ServiceLookup_Randomize: Stopping server: protocol=DNS address=127.0.0.1:30210 network=tcp writer.go:29: 2021-01-29T19:33:33.978Z [INFO] TestDNS_ServiceLookup_Randomize: Stopping server: protocol=DNS address=127.0.0.1:30210 network=udp writer.go:29: 2021-01-29T19:33:33.980Z [INFO] TestDNS_ServiceLookup_Randomize: Stopping server: protocol=HTTP address=127.0.0.1:30211 network=tcp === CONT TestDNS_ServiceLookup_LargeResponses writer.go:29: 2021-01-29T19:33:34.087Z [INFO] TestDNS_ServiceLookup_LargeResponses: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:34.089Z [INFO] TestDNS_ServiceLookup_LargeResponses: Endpoints down --- PASS: TestDNS_ServiceLookup_LargeResponses (0.93s) === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.101Z [WARN] TestDNS_ServiceLookup_OnlyFailing: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:34.102Z [DEBUG] TestDNS_ServiceLookup_OnlyFailing.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:34.105Z [DEBUG] TestDNS_ServiceLookup_OnlyFailing.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:34.123Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9830ad8f-b299-3c12-755d-87b3c12f94da Address:127.0.0.1:30239}]" === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:34.124Z [INFO] TestDNS_ServiceLookup_Truncate: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.125Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server.raft: entering follower state: follower="Node at 127.0.0.1:30239 [Follower]" leader= === CONT TestDNS_ServiceLookup_Truncate writer.go:29: 2021-01-29T19:33:34.125Z [INFO] TestDNS_ServiceLookup_Truncate: Endpoints down --- PASS: TestDNS_ServiceLookup_Truncate (0.93s) === CONT TestDNS_ServiceLookup_FilterCritical === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.125Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server.serf.wan: serf: EventMemberJoin: Node-9830ad8f-b299-3c12-755d-87b3c12f94da.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:34.130Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server.serf.lan: serf: EventMemberJoin: Node-9830ad8f-b299-3c12-755d-87b3c12f94da 127.0.0.1 writer.go:29: 2021-01-29T19:33:34.132Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server: Handled event for server in area: event=member-join server=Node-9830ad8f-b299-3c12-755d-87b3c12f94da.dc1 area=wan writer.go:29: 2021-01-29T19:33:34.132Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server: Adding LAN server: server="Node-9830ad8f-b299-3c12-755d-87b3c12f94da (Addr: tcp/127.0.0.1:30239) (DC: dc1)" writer.go:29: 2021-01-29T19:33:34.133Z [INFO] TestDNS_ServiceLookup_OnlyFailing: Started DNS server: address=127.0.0.1:30234 network=udp writer.go:29: 2021-01-29T19:33:34.136Z [INFO] TestDNS_ServiceLookup_OnlyFailing: Started DNS server: address=127.0.0.1:30234 network=tcp writer.go:29: 2021-01-29T19:33:34.139Z [INFO] TestDNS_ServiceLookup_OnlyFailing: Started HTTP server: address=127.0.0.1:30235 network=tcp writer.go:29: 2021-01-29T19:33:34.141Z [INFO] TestDNS_ServiceLookup_OnlyFailing: started state syncer === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.148Z [WARN] TestDNS_ServiceLookup_FilterCritical: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:34.149Z [DEBUG] TestDNS_ServiceLookup_FilterCritical.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:34.150Z [DEBUG] TestDNS_ServiceLookup_FilterCritical.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:34.157Z [INFO] TestDNS_ServiceLookup_FilterCritical.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3962d960-f4e8-4d91-3ae2-ba1e9c46755e Address:127.0.0.1:30245}]" writer.go:29: 2021-01-29T19:33:34.160Z [INFO] TestDNS_ServiceLookup_FilterCritical.server.raft: entering follower state: follower="Node at 127.0.0.1:30245 [Follower]" leader= writer.go:29: 2021-01-29T19:33:34.161Z [INFO] TestDNS_ServiceLookup_FilterCritical.server.serf.wan: serf: EventMemberJoin: Node-3962d960-f4e8-4d91-3ae2-ba1e9c46755e.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.175Z [WARN] TestDNS_ServiceLookup_OnlyFailing.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:34.176Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server.raft: entering candidate state: node="Node at 127.0.0.1:30239 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:34.178Z [DEBUG] TestDNS_ServiceLookup_OnlyFailing.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:34.179Z [DEBUG] TestDNS_ServiceLookup_OnlyFailing.server.raft: vote granted: from=9830ad8f-b299-3c12-755d-87b3c12f94da term=2 tally=1 writer.go:29: 2021-01-29T19:33:34.179Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:34.181Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server.raft: entering leader state: leader="Node at 127.0.0.1:30239 [Leader]" writer.go:29: 2021-01-29T19:33:34.182Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:34.182Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server: New leader elected: payload=Node-9830ad8f-b299-3c12-755d-87b3c12f94da === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.182Z [INFO] TestDNS_ServiceLookup_FilterCritical.server.serf.lan: serf: EventMemberJoin: Node-3962d960-f4e8-4d91-3ae2-ba1e9c46755e 127.0.0.1 === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.185Z [DEBUG] TestDNS_ServiceLookup_OnlyFailing.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30239 === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.187Z [INFO] TestDNS_ServiceLookup_FilterCritical.server: Adding LAN server: server="Node-3962d960-f4e8-4d91-3ae2-ba1e9c46755e (Addr: tcp/127.0.0.1:30245) (DC: dc1)" writer.go:29: 2021-01-29T19:33:34.188Z [INFO] TestDNS_ServiceLookup_FilterCritical.server: Handled event for server in area: event=member-join server=Node-3962d960-f4e8-4d91-3ae2-ba1e9c46755e.dc1 area=wan === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.188Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:34.193Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.194Z [INFO] TestDNS_ServiceLookup_FilterCritical: Started DNS server: address=127.0.0.1:30240 network=tcp === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.194Z [INFO] TestDNS_ServiceLookup_OnlyFailing.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.196Z [INFO] TestDNS_ServiceLookup_FilterCritical: Started DNS server: address=127.0.0.1:30240 network=udp === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.196Z [DEBUG] TestDNS_ServiceLookup_OnlyFailing.server: Skipping self join check for node since the cluster is too small: node=Node-9830ad8f-b299-3c12-755d-87b3c12f94da === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.197Z [INFO] TestDNS_ServiceLookup_FilterCritical: Started HTTP server: address=127.0.0.1:30241 network=tcp writer.go:29: 2021-01-29T19:33:34.198Z [INFO] TestDNS_ServiceLookup_FilterCritical: started state syncer === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.197Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server: member joined, marking health alive: member=Node-9830ad8f-b299-3c12-755d-87b3c12f94da === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.225Z [WARN] TestDNS_ServiceLookup_FilterCritical.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:34.226Z [INFO] TestDNS_ServiceLookup_FilterCritical.server.raft: entering candidate state: node="Node at 127.0.0.1:30245 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:34.228Z [DEBUG] TestDNS_ServiceLookup_FilterCritical.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:34.229Z [DEBUG] TestDNS_ServiceLookup_FilterCritical.server.raft: vote granted: from=3962d960-f4e8-4d91-3ae2-ba1e9c46755e term=2 tally=1 writer.go:29: 2021-01-29T19:33:34.230Z [INFO] TestDNS_ServiceLookup_FilterCritical.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:34.231Z [INFO] TestDNS_ServiceLookup_FilterCritical.server.raft: entering leader state: leader="Node at 127.0.0.1:30245 [Leader]" writer.go:29: 2021-01-29T19:33:34.233Z [INFO] TestDNS_ServiceLookup_FilterCritical.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:34.234Z [INFO] TestDNS_ServiceLookup_FilterCritical.server: New leader elected: payload=Node-3962d960-f4e8-4d91-3ae2-ba1e9c46755e writer.go:29: 2021-01-29T19:33:34.234Z [DEBUG] TestDNS_ServiceLookup_FilterCritical.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30245 writer.go:29: 2021-01-29T19:33:34.240Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:34.247Z [INFO] TestDNS_ServiceLookup_FilterCritical.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:34.248Z [INFO] TestDNS_ServiceLookup_FilterCritical.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:34.250Z [DEBUG] TestDNS_ServiceLookup_FilterCritical.server: Skipping self join check for node since the cluster is too small: node=Node-3962d960-f4e8-4d91-3ae2-ba1e9c46755e writer.go:29: 2021-01-29T19:33:34.251Z [INFO] TestDNS_ServiceLookup_FilterCritical.server: member joined, marking health alive: member=Node-3962d960-f4e8-4d91-3ae2-ba1e9c46755e writer.go:29: 2021-01-29T19:33:34.267Z [DEBUG] TestDNS_ServiceLookup_FilterCritical.dns: request served from client: name=db.service.consul. type=ANY class=IN latency=114.677µs client=127.0.0.1:35192 client_network=udp writer.go:29: 2021-01-29T19:33:34.268Z [INFO] TestDNS_ServiceLookup_FilterCritical: Requesting shutdown writer.go:29: 2021-01-29T19:33:34.268Z [DEBUG] TestDNS_ServiceLookup_FilterCritical.dns: request served from client: name=511b4f7c-563b-d98f-49be-809795190532.query.consul. type=ANY class=IN latency=206.568µs client=127.0.0.1:35459 client_network=udp writer.go:29: 2021-01-29T19:33:34.269Z [INFO] TestDNS_ServiceLookup_FilterCritical.server: shutting down server writer.go:29: 2021-01-29T19:33:34.272Z [DEBUG] TestDNS_ServiceLookup_FilterCritical.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:34.273Z [WARN] TestDNS_ServiceLookup_FilterCritical.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:34.272Z [ERROR] TestDNS_ServiceLookup_FilterCritical.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.277Z [DEBUG] TestDNS_ServiceLookup_OnlyFailing.dns: request served from client: name=db.service.consul. type=ANY class=IN latency=49.948µs client=127.0.0.1:41837 client_network=udp === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.273Z [DEBUG] TestDNS_ServiceLookup_FilterCritical.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:34.276Z [WARN] TestDNS_ServiceLookup_FilterCritical.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.279Z [DEBUG] TestDNS_ServiceLookup_OnlyFailing.dns: request served from client: name=b8e0a63c-687f-cc4d-01f2-b34e1aeb4f99.query.consul. type=ANY class=IN latency=1.191412ms client=127.0.0.1:39982 client_network=udp writer.go:29: 2021-01-29T19:33:34.279Z [INFO] TestDNS_ServiceLookup_OnlyFailing: Requesting shutdown === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.281Z [INFO] TestDNS_ServiceLookup_FilterCritical.server.router.manager: shutting down === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.282Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server: shutting down server writer.go:29: 2021-01-29T19:33:34.283Z [DEBUG] TestDNS_ServiceLookup_OnlyFailing.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.281Z [INFO] TestDNS_ServiceLookup_FilterCritical: consul server down === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.284Z [WARN] TestDNS_ServiceLookup_OnlyFailing.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.284Z [INFO] TestDNS_ServiceLookup_FilterCritical: shutdown complete === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.283Z [ERROR] TestDNS_ServiceLookup_OnlyFailing.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.285Z [INFO] TestDNS_ServiceLookup_FilterCritical: Stopping server: protocol=DNS address=127.0.0.1:30240 network=tcp === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.284Z [DEBUG] TestDNS_ServiceLookup_OnlyFailing.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.287Z [INFO] TestDNS_ServiceLookup_FilterCritical: Stopping server: protocol=DNS address=127.0.0.1:30240 network=udp === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.286Z [WARN] TestDNS_ServiceLookup_OnlyFailing.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.288Z [INFO] TestDNS_ServiceLookup_FilterCritical: Stopping server: protocol=HTTP address=127.0.0.1:30241 network=tcp === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.288Z [INFO] TestDNS_ServiceLookup_OnlyFailing.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:34.289Z [INFO] TestDNS_ServiceLookup_OnlyFailing: consul server down writer.go:29: 2021-01-29T19:33:34.291Z [INFO] TestDNS_ServiceLookup_OnlyFailing: shutdown complete writer.go:29: 2021-01-29T19:33:34.293Z [INFO] TestDNS_ServiceLookup_OnlyFailing: Stopping server: protocol=DNS address=127.0.0.1:30234 network=tcp writer.go:29: 2021-01-29T19:33:34.295Z [INFO] TestDNS_ServiceLookup_OnlyFailing: Stopping server: protocol=DNS address=127.0.0.1:30234 network=udp writer.go:29: 2021-01-29T19:33:34.298Z [INFO] TestDNS_ServiceLookup_OnlyFailing: Stopping server: protocol=HTTP address=127.0.0.1:30235 network=tcp === CONT TestDNS_ServiceLookup_OnlyPassing writer.go:29: 2021-01-29T19:33:34.420Z [INFO] TestDNS_ServiceLookup_OnlyPassing: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:34.422Z [INFO] TestDNS_ServiceLookup_OnlyPassing: Endpoints down --- PASS: TestDNS_ServiceLookup_OnlyPassing (0.78s) === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:34.437Z [WARN] TestDNS_RecursorTimeout: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:34.438Z [DEBUG] TestDNS_RecursorTimeout.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:34.441Z [DEBUG] TestDNS_RecursorTimeout.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:34.451Z [INFO] TestDNS_RecursorTimeout.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:886ee5cd-ddfe-2e56-3e92-592d779e16d3 Address:127.0.0.1:30221}]" writer.go:29: 2021-01-29T19:33:34.454Z [INFO] TestDNS_RecursorTimeout.server.raft: entering follower state: follower="Node at 127.0.0.1:30221 [Follower]" leader= writer.go:29: 2021-01-29T19:33:34.455Z [INFO] TestDNS_RecursorTimeout.server.serf.wan: serf: EventMemberJoin: Node-886ee5cd-ddfe-2e56-3e92-592d779e16d3.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:34.458Z [INFO] TestDNS_RecursorTimeout.server.serf.lan: serf: EventMemberJoin: Node-886ee5cd-ddfe-2e56-3e92-592d779e16d3 127.0.0.1 writer.go:29: 2021-01-29T19:33:34.461Z [DEBUG] TestDNS_RecursorTimeout.dns: recursor enabled writer.go:29: 2021-01-29T19:33:34.462Z [INFO] TestDNS_RecursorTimeout: Started DNS server: address=127.0.0.1:30216 network=udp writer.go:29: 2021-01-29T19:33:34.463Z [DEBUG] TestDNS_RecursorTimeout.dns: recursor enabled writer.go:29: 2021-01-29T19:33:34.466Z [INFO] TestDNS_RecursorTimeout: Started DNS server: address=127.0.0.1:30216 network=tcp writer.go:29: 2021-01-29T19:33:34.461Z [INFO] TestDNS_RecursorTimeout.server: Adding LAN server: server="Node-886ee5cd-ddfe-2e56-3e92-592d779e16d3 (Addr: tcp/127.0.0.1:30221) (DC: dc1)" writer.go:29: 2021-01-29T19:33:34.463Z [INFO] TestDNS_RecursorTimeout.server: Handled event for server in area: event=member-join server=Node-886ee5cd-ddfe-2e56-3e92-592d779e16d3.dc1 area=wan writer.go:29: 2021-01-29T19:33:34.469Z [INFO] TestDNS_RecursorTimeout: Started HTTP server: address=127.0.0.1:30217 network=tcp writer.go:29: 2021-01-29T19:33:34.473Z [INFO] TestDNS_RecursorTimeout: started state syncer === CONT TestDNS_ServiceLookup_Randomize writer.go:29: 2021-01-29T19:33:34.482Z [INFO] TestDNS_ServiceLookup_Randomize: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:34.483Z [INFO] TestDNS_ServiceLookup_Randomize: Endpoints down --- PASS: TestDNS_ServiceLookup_Randomize (0.90s) === CONT TestDNS_Recurse_Truncation === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:34.497Z [WARN] TestDNS_RecursorTimeout.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:34.498Z [INFO] TestDNS_RecursorTimeout.server.raft: entering candidate state: node="Node at 127.0.0.1:30221 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:34.500Z [DEBUG] TestDNS_RecursorTimeout.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:34.501Z [DEBUG] TestDNS_RecursorTimeout.server.raft: vote granted: from=886ee5cd-ddfe-2e56-3e92-592d779e16d3 term=2 tally=1 writer.go:29: 2021-01-29T19:33:34.502Z [INFO] TestDNS_RecursorTimeout.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:34.502Z [INFO] TestDNS_RecursorTimeout.server.raft: entering leader state: leader="Node at 127.0.0.1:30221 [Leader]" writer.go:29: 2021-01-29T19:33:34.503Z [INFO] TestDNS_RecursorTimeout.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:34.504Z [INFO] TestDNS_RecursorTimeout.server: New leader elected: payload=Node-886ee5cd-ddfe-2e56-3e92-592d779e16d3 === CONT TestDNS_Recurse_Truncation writer.go:29: 2021-01-29T19:33:34.502Z [WARN] TestDNS_Recurse_Truncation: bootstrap = true: do not enable unless necessary === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:34.505Z [DEBUG] TestDNS_RecursorTimeout.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30221 === CONT TestDNS_Recurse_Truncation writer.go:29: 2021-01-29T19:33:34.506Z [DEBUG] TestDNS_Recurse_Truncation.tlsutil: Update: version=1 === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:34.509Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_Recurse_Truncation writer.go:29: 2021-01-29T19:33:34.508Z [DEBUG] TestDNS_Recurse_Truncation.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:34.522Z [INFO] TestDNS_RecursorTimeout.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:34.523Z [INFO] TestDNS_RecursorTimeout.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:34.523Z [DEBUG] TestDNS_RecursorTimeout.server: Skipping self join check for node since the cluster is too small: node=Node-886ee5cd-ddfe-2e56-3e92-592d779e16d3 writer.go:29: 2021-01-29T19:33:34.524Z [INFO] TestDNS_RecursorTimeout.server: member joined, marking health alive: member=Node-886ee5cd-ddfe-2e56-3e92-592d779e16d3 === CONT TestDNS_Recurse_Truncation writer.go:29: 2021-01-29T19:33:34.512Z [INFO] TestDNS_Recurse_Truncation.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c578e10b-ec64-28ee-af50-cdc4c4af1bf9 Address:127.0.0.1:30257}]" writer.go:29: 2021-01-29T19:33:34.528Z [INFO] TestDNS_Recurse_Truncation.server.raft: entering follower state: follower="Node at 127.0.0.1:30257 [Follower]" leader= writer.go:29: 2021-01-29T19:33:34.530Z [INFO] TestDNS_Recurse_Truncation.server.serf.wan: serf: EventMemberJoin: Node-c578e10b-ec64-28ee-af50-cdc4c4af1bf9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:34.544Z [INFO] TestDNS_Recurse_Truncation.server.serf.lan: serf: EventMemberJoin: Node-c578e10b-ec64-28ee-af50-cdc4c4af1bf9 127.0.0.1 writer.go:29: 2021-01-29T19:33:34.549Z [INFO] TestDNS_Recurse_Truncation.server: Adding LAN server: server="Node-c578e10b-ec64-28ee-af50-cdc4c4af1bf9 (Addr: tcp/127.0.0.1:30257) (DC: dc1)" writer.go:29: 2021-01-29T19:33:34.550Z [INFO] TestDNS_Recurse_Truncation.server: Handled event for server in area: event=member-join server=Node-c578e10b-ec64-28ee-af50-cdc4c4af1bf9.dc1 area=wan writer.go:29: 2021-01-29T19:33:34.553Z [DEBUG] TestDNS_Recurse_Truncation.dns: recursor enabled writer.go:29: 2021-01-29T19:33:34.554Z [INFO] TestDNS_Recurse_Truncation: Started DNS server: address=127.0.0.1:30252 network=tcp writer.go:29: 2021-01-29T19:33:34.554Z [DEBUG] TestDNS_Recurse_Truncation.dns: recursor enabled writer.go:29: 2021-01-29T19:33:34.556Z [INFO] TestDNS_Recurse_Truncation: Started DNS server: address=127.0.0.1:30252 network=udp writer.go:29: 2021-01-29T19:33:34.557Z [INFO] TestDNS_Recurse_Truncation: Started HTTP server: address=127.0.0.1:30253 network=tcp writer.go:29: 2021-01-29T19:33:34.559Z [INFO] TestDNS_Recurse_Truncation: started state syncer === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:34.566Z [DEBUG] TestDNS_RecursorTimeout: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_Recurse_Truncation writer.go:29: 2021-01-29T19:33:34.567Z [WARN] TestDNS_Recurse_Truncation.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:34.568Z [INFO] TestDNS_Recurse_Truncation.server.raft: entering candidate state: node="Node at 127.0.0.1:30257 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:34.570Z [DEBUG] TestDNS_Recurse_Truncation.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:34.571Z [DEBUG] TestDNS_Recurse_Truncation.server.raft: vote granted: from=c578e10b-ec64-28ee-af50-cdc4c4af1bf9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:34.572Z [INFO] TestDNS_Recurse_Truncation.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:34.572Z [INFO] TestDNS_Recurse_Truncation.server.raft: entering leader state: leader="Node at 127.0.0.1:30257 [Leader]" writer.go:29: 2021-01-29T19:33:34.573Z [INFO] TestDNS_Recurse_Truncation.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:34.574Z [INFO] TestDNS_Recurse_Truncation.server: New leader elected: payload=Node-c578e10b-ec64-28ee-af50-cdc4c4af1bf9 writer.go:29: 2021-01-29T19:33:34.576Z [DEBUG] TestDNS_Recurse_Truncation.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30257 === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:34.570Z [INFO] TestDNS_RecursorTimeout: Synced node info === CONT TestDNS_Recurse_Truncation writer.go:29: 2021-01-29T19:33:34.581Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:34.586Z [INFO] TestDNS_Recurse_Truncation.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:34.587Z [INFO] TestDNS_Recurse_Truncation.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:34.588Z [DEBUG] TestDNS_Recurse_Truncation.server: Skipping self join check for node since the cluster is too small: node=Node-c578e10b-ec64-28ee-af50-cdc4c4af1bf9 writer.go:29: 2021-01-29T19:33:34.589Z [INFO] TestDNS_Recurse_Truncation.server: member joined, marking health alive: member=Node-c578e10b-ec64-28ee-af50-cdc4c4af1bf9 writer.go:29: 2021-01-29T19:33:34.692Z [DEBUG] TestDNS_Recurse_Truncation: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:34.695Z [INFO] TestDNS_Recurse_Truncation: Synced node info writer.go:29: 2021-01-29T19:33:34.784Z [DEBUG] TestDNS_Recurse_Truncation.dns: recurse succeeded for question: question="{apple.com. 255 1}" rtt=379.427µs recursor=127.0.0.1:56133 writer.go:29: 2021-01-29T19:33:34.785Z [DEBUG] TestDNS_Recurse_Truncation.dns: request served from client: question="{apple.com. 255 1}" network=udp latency=1.480328ms client=127.0.0.1:52432 client_network=udp writer.go:29: 2021-01-29T19:33:34.785Z [INFO] TestDNS_Recurse_Truncation: Requesting shutdown writer.go:29: 2021-01-29T19:33:34.787Z [INFO] TestDNS_Recurse_Truncation.server: shutting down server writer.go:29: 2021-01-29T19:33:34.788Z [DEBUG] TestDNS_Recurse_Truncation.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.789Z [INFO] TestDNS_ServiceLookup_FilterCritical: Waiting for endpoints to shut down === CONT TestDNS_Recurse_Truncation writer.go:29: 2021-01-29T19:33:34.789Z [WARN] TestDNS_Recurse_Truncation.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:34.789Z [DEBUG] TestDNS_Recurse_Truncation.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:34.793Z [WARN] TestDNS_Recurse_Truncation.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:34.795Z [INFO] TestDNS_Recurse_Truncation.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:34.796Z [INFO] TestDNS_Recurse_Truncation: consul server down writer.go:29: 2021-01-29T19:33:34.797Z [INFO] TestDNS_Recurse_Truncation: shutdown complete writer.go:29: 2021-01-29T19:33:34.798Z [INFO] TestDNS_Recurse_Truncation: Stopping server: protocol=DNS address=127.0.0.1:30252 network=tcp writer.go:29: 2021-01-29T19:33:34.799Z [INFO] TestDNS_Recurse_Truncation: Stopping server: protocol=DNS address=127.0.0.1:30252 network=udp === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.799Z [INFO] TestDNS_ServiceLookup_OnlyFailing: Waiting for endpoints to shut down === CONT TestDNS_Recurse_Truncation writer.go:29: 2021-01-29T19:33:34.800Z [INFO] TestDNS_Recurse_Truncation: Stopping server: protocol=HTTP address=127.0.0.1:30253 network=tcp === CONT TestDNS_ServiceLookup_OnlyFailing writer.go:29: 2021-01-29T19:33:34.801Z [INFO] TestDNS_ServiceLookup_OnlyFailing: Endpoints down --- PASS: TestDNS_ServiceLookup_OnlyFailing (0.71s) === CONT TestDNS_Recurse === CONT TestDNS_ServiceLookup_FilterCritical writer.go:29: 2021-01-29T19:33:34.791Z [INFO] TestDNS_ServiceLookup_FilterCritical: Endpoints down --- PASS: TestDNS_ServiceLookup_FilterCritical (0.68s) === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.863Z [WARN] TestDNS_ServiceLookup_Dedup_SRV: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:34.866Z [DEBUG] TestDNS_ServiceLookup_Dedup_SRV.tlsutil: Update: version=1 === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:34.867Z [WARN] TestDNS_Recurse: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.869Z [DEBUG] TestDNS_ServiceLookup_Dedup_SRV.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:34.870Z [DEBUG] TestDNS_Recurse.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:34.874Z [DEBUG] TestDNS_Recurse.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.893Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f22da219-13c1-8187-000b-3210b6c27e0e Address:127.0.0.1:30263}]" writer.go:29: 2021-01-29T19:33:34.899Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server.serf.wan: serf: EventMemberJoin: Node-f22da219-13c1-8187-000b-3210b6c27e0e.dc1 127.0.0.1 === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:34.895Z [INFO] TestDNS_Recurse.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:39fda104-17cf-cbe4-ee9e-dcd30e242455 Address:127.0.0.1:30269}]" === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.903Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server.serf.lan: serf: EventMemberJoin: Node-f22da219-13c1-8187-000b-3210b6c27e0e 127.0.0.1 writer.go:29: 2021-01-29T19:33:34.904Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server.raft: entering follower state: follower="Node at 127.0.0.1:30263 [Follower]" leader= writer.go:29: 2021-01-29T19:33:34.907Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: Started DNS server: address=127.0.0.1:30258 network=udp === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:34.911Z [INFO] TestDNS_Recurse.server.serf.wan: serf: EventMemberJoin: Node-39fda104-17cf-cbe4-ee9e-dcd30e242455.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.918Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server: Adding LAN server: server="Node-f22da219-13c1-8187-000b-3210b6c27e0e (Addr: tcp/127.0.0.1:30263) (DC: dc1)" writer.go:29: 2021-01-29T19:33:34.921Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server: Handled event for server in area: event=member-join server=Node-f22da219-13c1-8187-000b-3210b6c27e0e.dc1 area=wan === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:34.915Z [INFO] TestDNS_Recurse.server.raft: entering follower state: follower="Node at 127.0.0.1:30269 [Follower]" leader= === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.929Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: Started DNS server: address=127.0.0.1:30258 network=tcp === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:34.929Z [INFO] TestDNS_Recurse.server.serf.lan: serf: EventMemberJoin: Node-39fda104-17cf-cbe4-ee9e-dcd30e242455 127.0.0.1 === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.935Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: Started HTTP server: address=127.0.0.1:30259 network=tcp === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:34.937Z [DEBUG] TestDNS_Recurse.dns: recursor enabled writer.go:29: 2021-01-29T19:33:34.938Z [INFO] TestDNS_Recurse.server: Adding LAN server: server="Node-39fda104-17cf-cbe4-ee9e-dcd30e242455 (Addr: tcp/127.0.0.1:30269) (DC: dc1)" === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.937Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: started state syncer === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:34.939Z [INFO] TestDNS_Recurse: Started DNS server: address=127.0.0.1:30264 network=udp writer.go:29: 2021-01-29T19:33:34.940Z [DEBUG] TestDNS_Recurse.dns: recursor enabled writer.go:29: 2021-01-29T19:33:34.941Z [INFO] TestDNS_Recurse.server: Handled event for server in area: event=member-join server=Node-39fda104-17cf-cbe4-ee9e-dcd30e242455.dc1 area=wan writer.go:29: 2021-01-29T19:33:34.946Z [INFO] TestDNS_Recurse: Started DNS server: address=127.0.0.1:30264 network=tcp writer.go:29: 2021-01-29T19:33:34.954Z [INFO] TestDNS_Recurse: Started HTTP server: address=127.0.0.1:30265 network=tcp writer.go:29: 2021-01-29T19:33:34.956Z [INFO] TestDNS_Recurse: started state syncer writer.go:29: 2021-01-29T19:33:34.968Z [WARN] TestDNS_Recurse.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:34.969Z [INFO] TestDNS_Recurse.server.raft: entering candidate state: node="Node at 127.0.0.1:30269 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.969Z [WARN] TestDNS_ServiceLookup_Dedup_SRV.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:34.984Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server.raft: entering candidate state: node="Node at 127.0.0.1:30263 [Candidate]" term=2 === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:34.985Z [DEBUG] TestDNS_Recurse.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:34.986Z [DEBUG] TestDNS_Recurse.server.raft: vote granted: from=39fda104-17cf-cbe4-ee9e-dcd30e242455 term=2 tally=1 writer.go:29: 2021-01-29T19:33:34.987Z [INFO] TestDNS_Recurse.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:34.990Z [INFO] TestDNS_Recurse.server.raft: entering leader state: leader="Node at 127.0.0.1:30269 [Leader]" writer.go:29: 2021-01-29T19:33:34.992Z [INFO] TestDNS_Recurse.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.988Z [DEBUG] TestDNS_ServiceLookup_Dedup_SRV.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:34.994Z [DEBUG] TestDNS_ServiceLookup_Dedup_SRV.server.raft: vote granted: from=f22da219-13c1-8187-000b-3210b6c27e0e term=2 tally=1 === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:34.994Z [INFO] TestDNS_Recurse.server: New leader elected: payload=Node-39fda104-17cf-cbe4-ee9e-dcd30e242455 === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:34.996Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:34.999Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server.raft: entering leader state: leader="Node at 127.0.0.1:30263 [Leader]" writer.go:29: 2021-01-29T19:33:35.002Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server: cluster leadership acquired === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:35.003Z [DEBUG] TestDNS_Recurse.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30269 === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:35.006Z [DEBUG] TestDNS_ServiceLookup_Dedup_SRV.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30263 writer.go:29: 2021-01-29T19:33:35.008Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server: New leader elected: payload=Node-f22da219-13c1-8187-000b-3210b6c27e0e === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:35.015Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:35.025Z [INFO] TestDNS_Recurse.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:35.027Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:35.026Z [INFO] TestDNS_Recurse.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.029Z [DEBUG] TestDNS_Recurse.server: Skipping self join check for node since the cluster is too small: node=Node-39fda104-17cf-cbe4-ee9e-dcd30e242455 writer.go:29: 2021-01-29T19:33:35.032Z [INFO] TestDNS_Recurse.server: member joined, marking health alive: member=Node-39fda104-17cf-cbe4-ee9e-dcd30e242455 === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:35.054Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:35.055Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.056Z [DEBUG] TestDNS_ServiceLookup_Dedup_SRV.server: Skipping self join check for node since the cluster is too small: node=Node-f22da219-13c1-8187-000b-3210b6c27e0e writer.go:29: 2021-01-29T19:33:35.056Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server: member joined, marking health alive: member=Node-f22da219-13c1-8187-000b-3210b6c27e0e writer.go:29: 2021-01-29T19:33:35.134Z [DEBUG] TestDNS_ServiceLookup_Dedup_SRV.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=378.879µs client=127.0.0.1:45278 client_network=udp writer.go:29: 2021-01-29T19:33:35.136Z [DEBUG] TestDNS_ServiceLookup_Dedup_SRV.dns: request served from client: name=176cc180-250e-b4eb-4064-a2565df71f59.query.consul. type=SRV class=IN latency=237.711µs client=127.0.0.1:38285 client_network=udp writer.go:29: 2021-01-29T19:33:35.136Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: Requesting shutdown writer.go:29: 2021-01-29T19:33:35.140Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server: shutting down server writer.go:29: 2021-01-29T19:33:35.142Z [DEBUG] TestDNS_ServiceLookup_Dedup_SRV.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.143Z [WARN] TestDNS_ServiceLookup_Dedup_SRV.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:35.142Z [ERROR] TestDNS_ServiceLookup_Dedup_SRV.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:35.143Z [DEBUG] TestDNS_ServiceLookup_Dedup_SRV.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.150Z [WARN] TestDNS_ServiceLookup_Dedup_SRV.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:35.153Z [INFO] TestDNS_ServiceLookup_Dedup_SRV.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:35.153Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: consul server down writer.go:29: 2021-01-29T19:33:35.156Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: shutdown complete writer.go:29: 2021-01-29T19:33:35.157Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: Stopping server: protocol=DNS address=127.0.0.1:30258 network=tcp writer.go:29: 2021-01-29T19:33:35.159Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: Stopping server: protocol=DNS address=127.0.0.1:30258 network=udp writer.go:29: 2021-01-29T19:33:35.161Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: Stopping server: protocol=HTTP address=127.0.0.1:30259 network=tcp === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:35.186Z [DEBUG] TestDNS_Recurse.dns: recurse succeeded for question: question="{apple.com. 255 1}" rtt=681.318µs recursor=127.0.0.1:49345 writer.go:29: 2021-01-29T19:33:35.187Z [DEBUG] TestDNS_Recurse.dns: request served from client: question="{apple.com. 255 1}" network=udp latency=1.662339ms client=127.0.0.1:58292 client_network=udp writer.go:29: 2021-01-29T19:33:35.187Z [INFO] TestDNS_Recurse: Requesting shutdown writer.go:29: 2021-01-29T19:33:35.188Z [INFO] TestDNS_Recurse.server: shutting down server writer.go:29: 2021-01-29T19:33:35.189Z [DEBUG] TestDNS_Recurse.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.189Z [ERROR] TestDNS_Recurse.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:35.190Z [WARN] TestDNS_Recurse.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:35.190Z [DEBUG] TestDNS_Recurse.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.194Z [WARN] TestDNS_Recurse.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:35.199Z [INFO] TestDNS_Recurse.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:35.199Z [INFO] TestDNS_Recurse: consul server down writer.go:29: 2021-01-29T19:33:35.202Z [INFO] TestDNS_Recurse: shutdown complete writer.go:29: 2021-01-29T19:33:35.205Z [INFO] TestDNS_Recurse: Stopping server: protocol=DNS address=127.0.0.1:30264 network=tcp writer.go:29: 2021-01-29T19:33:35.207Z [INFO] TestDNS_Recurse: Stopping server: protocol=DNS address=127.0.0.1:30264 network=udp writer.go:29: 2021-01-29T19:33:35.209Z [INFO] TestDNS_Recurse: Stopping server: protocol=HTTP address=127.0.0.1:30265 network=tcp === CONT TestDNS_Recurse_Truncation writer.go:29: 2021-01-29T19:33:35.302Z [INFO] TestDNS_Recurse_Truncation: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:35.303Z [INFO] TestDNS_Recurse_Truncation: Endpoints down --- PASS: TestDNS_Recurse_Truncation (0.82s) === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.318Z [WARN] TestDNS_ServiceLookup_PreparedQueryNamePeriod: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:35.320Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:35.323Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:35.329Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6b3ccbac-54d6-636a-5a50-0e0ec9d0559d Address:127.0.0.1:30275}]" writer.go:29: 2021-01-29T19:33:35.330Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.raft: entering follower state: follower="Node at 127.0.0.1:30275 [Follower]" leader= writer.go:29: 2021-01-29T19:33:35.331Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.serf.wan: serf: EventMemberJoin: Node-6b3ccbac-54d6-636a-5a50-0e0ec9d0559d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:35.335Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.serf.lan: serf: EventMemberJoin: Node-6b3ccbac-54d6-636a-5a50-0e0ec9d0559d 127.0.0.1 writer.go:29: 2021-01-29T19:33:35.337Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server: Handled event for server in area: event=member-join server=Node-6b3ccbac-54d6-636a-5a50-0e0ec9d0559d.dc1 area=wan writer.go:29: 2021-01-29T19:33:35.337Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server: Adding LAN server: server="Node-6b3ccbac-54d6-636a-5a50-0e0ec9d0559d (Addr: tcp/127.0.0.1:30275) (DC: dc1)" writer.go:29: 2021-01-29T19:33:35.337Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Started DNS server: address=127.0.0.1:30270 network=udp writer.go:29: 2021-01-29T19:33:35.341Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Started DNS server: address=127.0.0.1:30270 network=tcp writer.go:29: 2021-01-29T19:33:35.343Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Started HTTP server: address=127.0.0.1:30271 network=tcp writer.go:29: 2021-01-29T19:33:35.344Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: started state syncer writer.go:29: 2021-01-29T19:33:35.384Z [WARN] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:35.386Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.raft: entering candidate state: node="Node at 127.0.0.1:30275 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:35.402Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:35.405Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.raft: vote granted: from=6b3ccbac-54d6-636a-5a50-0e0ec9d0559d term=2 tally=1 writer.go:29: 2021-01-29T19:33:35.407Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:35.408Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.raft: entering leader state: leader="Node at 127.0.0.1:30275 [Leader]" writer.go:29: 2021-01-29T19:33:35.410Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:35.411Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server: New leader elected: payload=Node-6b3ccbac-54d6-636a-5a50-0e0ec9d0559d writer.go:29: 2021-01-29T19:33:35.413Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30275 writer.go:29: 2021-01-29T19:33:35.427Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:35.434Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:35.436Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.437Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server: Skipping self join check for node since the cluster is too small: node=Node-6b3ccbac-54d6-636a-5a50-0e0ec9d0559d writer.go:29: 2021-01-29T19:33:35.439Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server: member joined, marking health alive: member=Node-6b3ccbac-54d6-636a-5a50-0e0ec9d0559d === CONT TestDNS_ServiceLookup_Dedup_SRV writer.go:29: 2021-01-29T19:33:35.664Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:35.666Z [INFO] TestDNS_ServiceLookup_Dedup_SRV: Endpoints down --- PASS: TestDNS_ServiceLookup_Dedup_SRV (0.86s) === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.681Z [WARN] TestDNS_PreparedQueryNearIP: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:35.682Z [DEBUG] TestDNS_PreparedQueryNearIP.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:35.685Z [DEBUG] TestDNS_PreparedQueryNearIP.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:35.691Z [INFO] TestDNS_PreparedQueryNearIP.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5b943fac-c3ab-5f9c-68dd-dfd6e85a24ba Address:127.0.0.1:30251}]" writer.go:29: 2021-01-29T19:33:35.695Z [INFO] TestDNS_PreparedQueryNearIP.server.raft: entering follower state: follower="Node at 127.0.0.1:30251 [Follower]" leader= writer.go:29: 2021-01-29T19:33:35.696Z [INFO] TestDNS_PreparedQueryNearIP.server.serf.wan: serf: EventMemberJoin: Node-5b943fac-c3ab-5f9c-68dd-dfd6e85a24ba.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.702Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.705Z [INFO] TestDNS_PreparedQueryNearIP.server.serf.lan: serf: EventMemberJoin: Node-5b943fac-c3ab-5f9c-68dd-dfd6e85a24ba 127.0.0.1 === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.706Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Synced node info === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.709Z [INFO] TestDNS_PreparedQueryNearIP.server: Adding LAN server: server="Node-5b943fac-c3ab-5f9c-68dd-dfd6e85a24ba (Addr: tcp/127.0.0.1:30251) (DC: dc1)" === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.708Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Node info in sync === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.709Z [INFO] TestDNS_PreparedQueryNearIP.server: Handled event for server in area: event=member-join server=Node-5b943fac-c3ab-5f9c-68dd-dfd6e85a24ba.dc1 area=wan writer.go:29: 2021-01-29T19:33:35.710Z [INFO] TestDNS_PreparedQueryNearIP: Started DNS server: address=127.0.0.1:30246 network=udp writer.go:29: 2021-01-29T19:33:35.711Z [INFO] TestDNS_PreparedQueryNearIP: Started DNS server: address=127.0.0.1:30246 network=tcp === CONT TestDNS_Recurse writer.go:29: 2021-01-29T19:33:35.711Z [INFO] TestDNS_Recurse: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:35.712Z [INFO] TestDNS_Recurse: Endpoints down --- PASS: TestDNS_Recurse (0.91s) === CONT TestDNS_PreparedQueryNearIPEDNS === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.713Z [INFO] TestDNS_PreparedQueryNearIP: Started HTTP server: address=127.0.0.1:30247 network=tcp writer.go:29: 2021-01-29T19:33:35.714Z [INFO] TestDNS_PreparedQueryNearIP: started state syncer === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.721Z [WARN] TestDNS_PreparedQueryNearIPEDNS: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:35.725Z [DEBUG] TestDNS_PreparedQueryNearIPEDNS.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.723Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod.dns: request served from client: name=some.query.we.like.query.consul. type=SRV class=IN latency=153.996µs client=127.0.0.1:57176 client_network=udp === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.726Z [DEBUG] TestDNS_PreparedQueryNearIPEDNS.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.723Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Requesting shutdown writer.go:29: 2021-01-29T19:33:35.728Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server: shutting down server writer.go:29: 2021-01-29T19:33:35.730Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.734Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1f9ff3e0-4560-708f-2b24-31d23ab2ab0a Address:127.0.0.1:30287}]" === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.734Z [WARN] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.736Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server.serf.wan: serf: EventMemberJoin: Node-1f9ff3e0-4560-708f-2b24-31d23ab2ab0a.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.734Z [DEBUG] TestDNS_ServiceLookup_PreparedQueryNamePeriod.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.737Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server.raft: entering follower state: follower="Node at 127.0.0.1:30287 [Follower]" leader= writer.go:29: 2021-01-29T19:33:35.738Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server.serf.lan: serf: EventMemberJoin: Node-1f9ff3e0-4560-708f-2b24-31d23ab2ab0a 127.0.0.1 === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.742Z [WARN] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.744Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server: Adding LAN server: server="Node-1f9ff3e0-4560-708f-2b24-31d23ab2ab0a (Addr: tcp/127.0.0.1:30287) (DC: dc1)" === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.745Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod.server.router.manager: shutting down === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.746Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server: Handled event for server in area: event=member-join server=Node-1f9ff3e0-4560-708f-2b24-31d23ab2ab0a.dc1 area=wan === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.745Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: consul server down === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.747Z [INFO] TestDNS_PreparedQueryNearIPEDNS: Started DNS server: address=127.0.0.1:30282 network=tcp === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.750Z [WARN] TestDNS_PreparedQueryNearIP.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:35.751Z [INFO] TestDNS_PreparedQueryNearIP.server.raft: entering candidate state: node="Node at 127.0.0.1:30251 [Candidate]" term=2 === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.751Z [INFO] TestDNS_PreparedQueryNearIPEDNS: Started DNS server: address=127.0.0.1:30282 network=udp writer.go:29: 2021-01-29T19:33:35.754Z [INFO] TestDNS_PreparedQueryNearIPEDNS: Started HTTP server: address=127.0.0.1:30283 network=tcp writer.go:29: 2021-01-29T19:33:35.755Z [INFO] TestDNS_PreparedQueryNearIPEDNS: started state syncer === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.753Z [DEBUG] TestDNS_PreparedQueryNearIP.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.750Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: shutdown complete === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.758Z [DEBUG] TestDNS_PreparedQueryNearIP.server.raft: vote granted: from=5b943fac-c3ab-5f9c-68dd-dfd6e85a24ba term=2 tally=1 writer.go:29: 2021-01-29T19:33:35.759Z [INFO] TestDNS_PreparedQueryNearIP.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:35.760Z [INFO] TestDNS_PreparedQueryNearIP.server.raft: entering leader state: leader="Node at 127.0.0.1:30251 [Leader]" writer.go:29: 2021-01-29T19:33:35.761Z [INFO] TestDNS_PreparedQueryNearIP.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.759Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Stopping server: protocol=DNS address=127.0.0.1:30270 network=tcp === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.762Z [INFO] TestDNS_PreparedQueryNearIP.server: New leader elected: payload=Node-5b943fac-c3ab-5f9c-68dd-dfd6e85a24ba === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:35.769Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Stopping server: protocol=DNS address=127.0.0.1:30270 network=udp writer.go:29: 2021-01-29T19:33:35.772Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Stopping server: protocol=HTTP address=127.0.0.1:30271 network=tcp === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.762Z [DEBUG] TestDNS_PreparedQueryNearIP.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30251 writer.go:29: 2021-01-29T19:33:35.783Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:35.789Z [INFO] TestDNS_PreparedQueryNearIP.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:35.791Z [INFO] TestDNS_PreparedQueryNearIP.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.793Z [DEBUG] TestDNS_PreparedQueryNearIP.server: Skipping self join check for node since the cluster is too small: node=Node-5b943fac-c3ab-5f9c-68dd-dfd6e85a24ba === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.795Z [WARN] TestDNS_PreparedQueryNearIPEDNS.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:35.794Z [INFO] TestDNS_PreparedQueryNearIP.server: member joined, marking health alive: member=Node-5b943fac-c3ab-5f9c-68dd-dfd6e85a24ba === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:35.796Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server.raft: entering candidate state: node="Node at 127.0.0.1:30287 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:35.798Z [DEBUG] TestDNS_PreparedQueryNearIPEDNS.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:35.799Z [DEBUG] TestDNS_PreparedQueryNearIPEDNS.server.raft: vote granted: from=1f9ff3e0-4560-708f-2b24-31d23ab2ab0a term=2 tally=1 writer.go:29: 2021-01-29T19:33:35.800Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:35.801Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server.raft: entering leader state: leader="Node at 127.0.0.1:30287 [Leader]" writer.go:29: 2021-01-29T19:33:35.802Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:35.803Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server: New leader elected: payload=Node-1f9ff3e0-4560-708f-2b24-31d23ab2ab0a writer.go:29: 2021-01-29T19:33:35.804Z [DEBUG] TestDNS_PreparedQueryNearIPEDNS.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30287 writer.go:29: 2021-01-29T19:33:35.810Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:35.816Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:35.817Z [INFO] TestDNS_PreparedQueryNearIPEDNS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.819Z [DEBUG] TestDNS_PreparedQueryNearIPEDNS.server: Skipping self join check for node since the cluster is too small: node=Node-1f9ff3e0-4560-708f-2b24-31d23ab2ab0a writer.go:29: 2021-01-29T19:33:35.821Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server: member joined, marking health alive: member=Node-1f9ff3e0-4560-708f-2b24-31d23ab2ab0a Added 3 service nodes writer.go:29: 2021-01-29T19:33:35.963Z [DEBUG] TestDNS_PreparedQueryNearIPEDNS.dns: request served from client: name=some.query.we.like.query.consul. type=A class=IN latency=125.137µs client=127.0.0.1:33173 client_network=udp writer.go:29: 2021-01-29T19:33:35.963Z [INFO] TestDNS_PreparedQueryNearIPEDNS: Requesting shutdown writer.go:29: 2021-01-29T19:33:35.965Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server: shutting down server writer.go:29: 2021-01-29T19:33:35.966Z [DEBUG] TestDNS_PreparedQueryNearIPEDNS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.967Z [ERROR] TestDNS_PreparedQueryNearIPEDNS.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:35.968Z [WARN] TestDNS_PreparedQueryNearIPEDNS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:35.968Z [DEBUG] TestDNS_PreparedQueryNearIPEDNS.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.972Z [WARN] TestDNS_PreparedQueryNearIPEDNS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:35.974Z [INFO] TestDNS_PreparedQueryNearIPEDNS.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:35.975Z [INFO] TestDNS_PreparedQueryNearIPEDNS: consul server down writer.go:29: 2021-01-29T19:33:35.977Z [INFO] TestDNS_PreparedQueryNearIPEDNS: shutdown complete writer.go:29: 2021-01-29T19:33:35.979Z [INFO] TestDNS_PreparedQueryNearIPEDNS: Stopping server: protocol=DNS address=127.0.0.1:30282 network=tcp writer.go:29: 2021-01-29T19:33:35.980Z [INFO] TestDNS_PreparedQueryNearIPEDNS: Stopping server: protocol=DNS address=127.0.0.1:30282 network=udp writer.go:29: 2021-01-29T19:33:35.985Z [INFO] TestDNS_PreparedQueryNearIPEDNS: Stopping server: protocol=HTTP address=127.0.0.1:30283 network=tcp Added 3 service nodes === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:36.063Z [DEBUG] TestDNS_PreparedQueryNearIP.dns: request served from client: name=some.query.we.like.query.consul. type=A class=IN latency=202.269µs client=127.0.0.1:36271 client_network=udp writer.go:29: 2021-01-29T19:33:36.092Z [DEBUG] TestDNS_PreparedQueryNearIP: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:36.093Z [INFO] TestDNS_PreparedQueryNearIP: Synced node info writer.go:29: 2021-01-29T19:33:36.096Z [DEBUG] TestDNS_PreparedQueryNearIP.dns: request served from client: name=some.query.we.like.query.consul. type=A class=IN latency=133.84µs client=127.0.0.1:52629 client_network=udp writer.go:29: 2021-01-29T19:33:36.124Z [DEBUG] TestDNS_PreparedQueryNearIP.dns: request served from client: name=some.query.we.like.query.consul. type=A class=IN latency=185.24µs client=127.0.0.1:53565 client_network=udp writer.go:29: 2021-01-29T19:33:36.151Z [DEBUG] TestDNS_PreparedQueryNearIP.dns: request served from client: name=some.query.we.like.query.consul. type=A class=IN latency=142.745µs client=127.0.0.1:36871 client_network=udp writer.go:29: 2021-01-29T19:33:36.178Z [DEBUG] TestDNS_PreparedQueryNearIP.dns: request served from client: name=some.query.we.like.query.consul. type=A class=IN latency=125.776µs client=127.0.0.1:50936 client_network=udp writer.go:29: 2021-01-29T19:33:36.180Z [INFO] TestDNS_PreparedQueryNearIP: Requesting shutdown writer.go:29: 2021-01-29T19:33:36.182Z [INFO] TestDNS_PreparedQueryNearIP.server: shutting down server writer.go:29: 2021-01-29T19:33:36.183Z [DEBUG] TestDNS_PreparedQueryNearIP.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.184Z [WARN] TestDNS_PreparedQueryNearIP.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:36.184Z [DEBUG] TestDNS_PreparedQueryNearIP.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.186Z [WARN] TestDNS_PreparedQueryNearIP.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:36.192Z [INFO] TestDNS_PreparedQueryNearIP.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:36.192Z [INFO] TestDNS_PreparedQueryNearIP: consul server down writer.go:29: 2021-01-29T19:33:36.194Z [INFO] TestDNS_PreparedQueryNearIP: shutdown complete writer.go:29: 2021-01-29T19:33:36.195Z [INFO] TestDNS_PreparedQueryNearIP: Stopping server: protocol=DNS address=127.0.0.1:30246 network=tcp writer.go:29: 2021-01-29T19:33:36.196Z [INFO] TestDNS_PreparedQueryNearIP: Stopping server: protocol=DNS address=127.0.0.1:30246 network=udp writer.go:29: 2021-01-29T19:33:36.201Z [INFO] TestDNS_PreparedQueryNearIP: Stopping server: protocol=HTTP address=127.0.0.1:30247 network=tcp === CONT TestDNS_ServiceLookup_PreparedQueryNamePeriod writer.go:29: 2021-01-29T19:33:36.273Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:36.284Z [INFO] TestDNS_ServiceLookup_PreparedQueryNamePeriod: Endpoints down --- PASS: TestDNS_ServiceLookup_PreparedQueryNamePeriod (0.98s) === CONT TestDNS_ServiceLookup_TagPeriod writer.go:29: 2021-01-29T19:33:36.312Z [WARN] TestDNS_ServiceLookup_TagPeriod: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:36.314Z [DEBUG] TestDNS_ServiceLookup_TagPeriod.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:36.323Z [DEBUG] TestDNS_ServiceLookup_TagPeriod.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:36.335Z [INFO] TestDNS_ServiceLookup_TagPeriod.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b2d7a533-dcf3-3897-c61a-272ce97a71dd Address:127.0.0.1:30281}]" writer.go:29: 2021-01-29T19:33:36.339Z [INFO] TestDNS_ServiceLookup_TagPeriod.server.raft: entering follower state: follower="Node at 127.0.0.1:30281 [Follower]" leader= writer.go:29: 2021-01-29T19:33:36.341Z [INFO] TestDNS_ServiceLookup_TagPeriod.server.serf.wan: serf: EventMemberJoin: Node-b2d7a533-dcf3-3897-c61a-272ce97a71dd.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:36.346Z [INFO] TestDNS_ServiceLookup_TagPeriod.server.serf.lan: serf: EventMemberJoin: Node-b2d7a533-dcf3-3897-c61a-272ce97a71dd 127.0.0.1 writer.go:29: 2021-01-29T19:33:36.350Z [INFO] TestDNS_ServiceLookup_TagPeriod.server: Handled event for server in area: event=member-join server=Node-b2d7a533-dcf3-3897-c61a-272ce97a71dd.dc1 area=wan writer.go:29: 2021-01-29T19:33:36.350Z [INFO] TestDNS_ServiceLookup_TagPeriod.server: Adding LAN server: server="Node-b2d7a533-dcf3-3897-c61a-272ce97a71dd (Addr: tcp/127.0.0.1:30281) (DC: dc1)" writer.go:29: 2021-01-29T19:33:36.351Z [INFO] TestDNS_ServiceLookup_TagPeriod: Started DNS server: address=127.0.0.1:30276 network=udp writer.go:29: 2021-01-29T19:33:36.356Z [INFO] TestDNS_ServiceLookup_TagPeriod: Started DNS server: address=127.0.0.1:30276 network=tcp writer.go:29: 2021-01-29T19:33:36.360Z [INFO] TestDNS_ServiceLookup_TagPeriod: Started HTTP server: address=127.0.0.1:30277 network=tcp writer.go:29: 2021-01-29T19:33:36.361Z [INFO] TestDNS_ServiceLookup_TagPeriod: started state syncer writer.go:29: 2021-01-29T19:33:36.400Z [WARN] TestDNS_ServiceLookup_TagPeriod.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:36.401Z [INFO] TestDNS_ServiceLookup_TagPeriod.server.raft: entering candidate state: node="Node at 127.0.0.1:30281 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:36.403Z [DEBUG] TestDNS_ServiceLookup_TagPeriod.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:36.404Z [DEBUG] TestDNS_ServiceLookup_TagPeriod.server.raft: vote granted: from=b2d7a533-dcf3-3897-c61a-272ce97a71dd term=2 tally=1 writer.go:29: 2021-01-29T19:33:36.404Z [INFO] TestDNS_ServiceLookup_TagPeriod.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:36.406Z [INFO] TestDNS_ServiceLookup_TagPeriod.server.raft: entering leader state: leader="Node at 127.0.0.1:30281 [Leader]" writer.go:29: 2021-01-29T19:33:36.407Z [INFO] TestDNS_ServiceLookup_TagPeriod.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:36.407Z [INFO] TestDNS_ServiceLookup_TagPeriod.server: New leader elected: payload=Node-b2d7a533-dcf3-3897-c61a-272ce97a71dd writer.go:29: 2021-01-29T19:33:36.408Z [DEBUG] TestDNS_ServiceLookup_TagPeriod.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30281 writer.go:29: 2021-01-29T19:33:36.413Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:36.429Z [INFO] TestDNS_ServiceLookup_TagPeriod.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:36.431Z [INFO] TestDNS_ServiceLookup_TagPeriod.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.432Z [DEBUG] TestDNS_ServiceLookup_TagPeriod.server: Skipping self join check for node since the cluster is too small: node=Node-b2d7a533-dcf3-3897-c61a-272ce97a71dd writer.go:29: 2021-01-29T19:33:36.434Z [INFO] TestDNS_ServiceLookup_TagPeriod.server: member joined, marking health alive: member=Node-b2d7a533-dcf3-3897-c61a-272ce97a71dd === CONT TestDNS_PreparedQueryNearIPEDNS writer.go:29: 2021-01-29T19:33:36.487Z [INFO] TestDNS_PreparedQueryNearIPEDNS: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:36.488Z [INFO] TestDNS_PreparedQueryNearIPEDNS: Endpoints down --- PASS: TestDNS_PreparedQueryNearIPEDNS (0.78s) === CONT TestDNS_CaseInsensitiveServiceLookup === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:36.508Z [DEBUG] TestDNS_RecursorTimeout.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.513Z [WARN] TestDNS_CaseInsensitiveServiceLookup: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:36.515Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:36.517Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:36.527Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ca2fe8a0-143e-80f6-e895-f1d35d275d80 Address:127.0.0.1:30299}]" writer.go:29: 2021-01-29T19:33:36.529Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server.raft: entering follower state: follower="Node at 127.0.0.1:30299 [Follower]" leader= writer.go:29: 2021-01-29T19:33:36.534Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server.serf.wan: serf: EventMemberJoin: Node-ca2fe8a0-143e-80f6-e895-f1d35d275d80.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:36.539Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server.serf.lan: serf: EventMemberJoin: Node-ca2fe8a0-143e-80f6-e895-f1d35d275d80 127.0.0.1 writer.go:29: 2021-01-29T19:33:36.543Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server: Handled event for server in area: event=member-join server=Node-ca2fe8a0-143e-80f6-e895-f1d35d275d80.dc1 area=wan writer.go:29: 2021-01-29T19:33:36.543Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server: Adding LAN server: server="Node-ca2fe8a0-143e-80f6-e895-f1d35d275d80 (Addr: tcp/127.0.0.1:30299) (DC: dc1)" writer.go:29: 2021-01-29T19:33:36.546Z [INFO] TestDNS_CaseInsensitiveServiceLookup: Started DNS server: address=127.0.0.1:30294 network=tcp writer.go:29: 2021-01-29T19:33:36.548Z [INFO] TestDNS_CaseInsensitiveServiceLookup: Started DNS server: address=127.0.0.1:30294 network=udp writer.go:29: 2021-01-29T19:33:36.550Z [INFO] TestDNS_CaseInsensitiveServiceLookup: Started HTTP server: address=127.0.0.1:30295 network=tcp writer.go:29: 2021-01-29T19:33:36.552Z [INFO] TestDNS_CaseInsensitiveServiceLookup: started state syncer writer.go:29: 2021-01-29T19:33:36.591Z [WARN] TestDNS_CaseInsensitiveServiceLookup.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:36.592Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server.raft: entering candidate state: node="Node at 127.0.0.1:30299 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:36.599Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:36.600Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.server.raft: vote granted: from=ca2fe8a0-143e-80f6-e895-f1d35d275d80 term=2 tally=1 writer.go:29: 2021-01-29T19:33:36.601Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_TagPeriod writer.go:29: 2021-01-29T19:33:36.600Z [DEBUG] TestDNS_ServiceLookup_TagPeriod.dns: request served from client: name=v1.master2.db.service.consul. type=SRV class=IN latency=131.344µs client=127.0.0.1:60276 client_network=udp === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.602Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server.raft: entering leader state: leader="Node at 127.0.0.1:30299 [Leader]" writer.go:29: 2021-01-29T19:33:36.603Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_TagPeriod writer.go:29: 2021-01-29T19:33:36.601Z [DEBUG] TestDNS_ServiceLookup_TagPeriod.dns: request served from client: name=v1.master.db.service.consul. type=SRV class=IN latency=77.051µs client=127.0.0.1:42529 client_network=udp === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.603Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server: New leader elected: payload=Node-ca2fe8a0-143e-80f6-e895-f1d35d275d80 === CONT TestDNS_ServiceLookup_TagPeriod writer.go:29: 2021-01-29T19:33:36.601Z [INFO] TestDNS_ServiceLookup_TagPeriod: Requesting shutdown writer.go:29: 2021-01-29T19:33:36.605Z [INFO] TestDNS_ServiceLookup_TagPeriod.server: shutting down server writer.go:29: 2021-01-29T19:33:36.606Z [DEBUG] TestDNS_ServiceLookup_TagPeriod.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.606Z [WARN] TestDNS_ServiceLookup_TagPeriod.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.605Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30299 === CONT TestDNS_ServiceLookup_TagPeriod writer.go:29: 2021-01-29T19:33:36.606Z [ERROR] TestDNS_ServiceLookup_TagPeriod.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:36.606Z [DEBUG] TestDNS_ServiceLookup_TagPeriod.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.609Z [WARN] TestDNS_ServiceLookup_TagPeriod.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.612Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_TagPeriod writer.go:29: 2021-01-29T19:33:36.616Z [INFO] TestDNS_ServiceLookup_TagPeriod.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:36.616Z [INFO] TestDNS_ServiceLookup_TagPeriod: consul server down writer.go:29: 2021-01-29T19:33:36.618Z [INFO] TestDNS_ServiceLookup_TagPeriod: shutdown complete writer.go:29: 2021-01-29T19:33:36.619Z [INFO] TestDNS_ServiceLookup_TagPeriod: Stopping server: protocol=DNS address=127.0.0.1:30276 network=tcp writer.go:29: 2021-01-29T19:33:36.620Z [INFO] TestDNS_ServiceLookup_TagPeriod: Stopping server: protocol=DNS address=127.0.0.1:30276 network=udp writer.go:29: 2021-01-29T19:33:36.621Z [INFO] TestDNS_ServiceLookup_TagPeriod: Stopping server: protocol=HTTP address=127.0.0.1:30277 network=tcp === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.636Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:36.637Z [INFO] TestDNS_CaseInsensitiveServiceLookup.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.638Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.server: Skipping self join check for node since the cluster is too small: node=Node-ca2fe8a0-143e-80f6-e895-f1d35d275d80 writer.go:29: 2021-01-29T19:33:36.639Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server: member joined, marking health alive: member=Node-ca2fe8a0-143e-80f6-e895-f1d35d275d80 === CONT TestDNS_PreparedQueryNearIP writer.go:29: 2021-01-29T19:33:36.702Z [INFO] TestDNS_PreparedQueryNearIP: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:36.704Z [INFO] TestDNS_PreparedQueryNearIP: Endpoints down --- PASS: TestDNS_PreparedQueryNearIP (1.04s) === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:36.718Z [WARN] TestDNS_Lookup_TaggedIPAddresses: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:36.721Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:36.727Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.730Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.dns: request served from client: name=master.db.service.consul. type=SRV class=IN latency=94.299µs client=127.0.0.1:53844 client_network=udp writer.go:29: 2021-01-29T19:33:36.731Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.dns: request served from client: name=mASTER.dB.service.consul. type=SRV class=IN latency=106.359µs client=127.0.0.1:49937 client_network=udp writer.go:29: 2021-01-29T19:33:36.731Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.dns: request served from client: name=MASTER.dB.service.consul. type=SRV class=IN latency=90.451µs client=127.0.0.1:55708 client_network=udp writer.go:29: 2021-01-29T19:33:36.732Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=103.776µs client=127.0.0.1:52993 client_network=udp writer.go:29: 2021-01-29T19:33:36.733Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.dns: request served from client: name=DB.service.consul. type=SRV class=IN latency=93.495µs client=127.0.0.1:48971 client_network=udp writer.go:29: 2021-01-29T19:33:36.734Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.dns: request served from client: name=Db.service.consul. type=SRV class=IN latency=80.638µs client=127.0.0.1:54330 client_network=udp === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:36.737Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fa63907d-666a-c9ed-9aea-de78bcd9aedd Address:127.0.0.1:30305}]" === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.735Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.dns: request served from client: name=somequery.query.consul. type=SRV class=IN latency=70.085µs client=127.0.0.1:49485 client_network=udp writer.go:29: 2021-01-29T19:33:36.736Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.dns: request served from client: name=SomeQuery.query.consul. type=SRV class=IN latency=45.712µs client=127.0.0.1:36772 client_network=udp === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:36.739Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server.raft: entering follower state: follower="Node at 127.0.0.1:30305 [Follower]" leader= === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.736Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.dns: request served from client: name=SOMEQUERY.query.consul. type=SRV class=IN latency=58.417µs client=127.0.0.1:36592 client_network=udp === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:36.739Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server.serf.wan: serf: EventMemberJoin: Node-fa63907d-666a-c9ed-9aea-de78bcd9aedd.dc1 127.0.0.1 === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.736Z [INFO] TestDNS_CaseInsensitiveServiceLookup: Requesting shutdown writer.go:29: 2021-01-29T19:33:36.744Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server: shutting down server writer.go:29: 2021-01-29T19:33:36.745Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.746Z [WARN] TestDNS_CaseInsensitiveServiceLookup.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:36.745Z [ERROR] TestDNS_CaseInsensitiveServiceLookup.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:36.746Z [DEBUG] TestDNS_CaseInsensitiveServiceLookup.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:36.746Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server.serf.lan: serf: EventMemberJoin: Node-fa63907d-666a-c9ed-9aea-de78bcd9aedd 127.0.0.1 === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.748Z [WARN] TestDNS_CaseInsensitiveServiceLookup.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:36.753Z [INFO] TestDNS_CaseInsensitiveServiceLookup.server.router.manager: shutting down === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:36.754Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server: Adding LAN server: server="Node-fa63907d-666a-c9ed-9aea-de78bcd9aedd (Addr: tcp/127.0.0.1:30305) (DC: dc1)" === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:36.754Z [INFO] TestDNS_CaseInsensitiveServiceLookup: consul server down writer.go:29: 2021-01-29T19:33:36.755Z [INFO] TestDNS_CaseInsensitiveServiceLookup: shutdown complete writer.go:29: 2021-01-29T19:33:36.756Z [INFO] TestDNS_CaseInsensitiveServiceLookup: Stopping server: protocol=DNS address=127.0.0.1:30294 network=tcp writer.go:29: 2021-01-29T19:33:36.757Z [INFO] TestDNS_CaseInsensitiveServiceLookup: Stopping server: protocol=DNS address=127.0.0.1:30294 network=udp writer.go:29: 2021-01-29T19:33:36.758Z [INFO] TestDNS_CaseInsensitiveServiceLookup: Stopping server: protocol=HTTP address=127.0.0.1:30295 network=tcp === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:36.760Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server: Handled event for server in area: event=member-join server=Node-fa63907d-666a-c9ed-9aea-de78bcd9aedd.dc1 area=wan writer.go:29: 2021-01-29T19:33:36.765Z [INFO] TestDNS_Lookup_TaggedIPAddresses: Started DNS server: address=127.0.0.1:30300 network=tcp writer.go:29: 2021-01-29T19:33:36.766Z [INFO] TestDNS_Lookup_TaggedIPAddresses: Started DNS server: address=127.0.0.1:30300 network=udp writer.go:29: 2021-01-29T19:33:36.771Z [INFO] TestDNS_Lookup_TaggedIPAddresses: Started HTTP server: address=127.0.0.1:30301 network=tcp writer.go:29: 2021-01-29T19:33:36.772Z [INFO] TestDNS_Lookup_TaggedIPAddresses: started state syncer writer.go:29: 2021-01-29T19:33:36.799Z [WARN] TestDNS_Lookup_TaggedIPAddresses.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:36.800Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server.raft: entering candidate state: node="Node at 127.0.0.1:30305 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:36.802Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:36.803Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.server.raft: vote granted: from=fa63907d-666a-c9ed-9aea-de78bcd9aedd term=2 tally=1 writer.go:29: 2021-01-29T19:33:36.804Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:36.805Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server.raft: entering leader state: leader="Node at 127.0.0.1:30305 [Leader]" writer.go:29: 2021-01-29T19:33:36.806Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:36.806Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server: New leader elected: payload=Node-fa63907d-666a-c9ed-9aea-de78bcd9aedd writer.go:29: 2021-01-29T19:33:36.812Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30305 writer.go:29: 2021-01-29T19:33:36.817Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:36.822Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:36.823Z [INFO] TestDNS_Lookup_TaggedIPAddresses.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.824Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.server: Skipping self join check for node since the cluster is too small: node=Node-fa63907d-666a-c9ed-9aea-de78bcd9aedd writer.go:29: 2021-01-29T19:33:36.825Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server: member joined, marking health alive: member=Node-fa63907d-666a-c9ed-9aea-de78bcd9aedd writer.go:29: 2021-01-29T19:33:36.978Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:36.982Z [INFO] TestDNS_Lookup_TaggedIPAddresses: Synced node info writer.go:29: 2021-01-29T19:33:36.984Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses: Node info in sync === CONT TestDNS_ServiceLookup_TagPeriod writer.go:29: 2021-01-29T19:33:37.122Z [INFO] TestDNS_ServiceLookup_TagPeriod: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:37.123Z [INFO] TestDNS_ServiceLookup_TagPeriod: Endpoints down --- PASS: TestDNS_ServiceLookup_TagPeriod (0.84s) === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.138Z [WARN] TestDNS_ServiceLookup_WanTranslation: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:37.140Z [WARN] TestDNS_ServiceLookup_WanTranslation: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:37.141Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.tlsutil: Update: version=1 === RUN TestDNS_Lookup_TaggedIPAddresses/simple-ipv4 === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.149Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=db.service.consul. type=A class=IN latency=186.698µs client=127.0.0.1:56314 client_network=udp writer.go:29: 2021-01-29T19:33:37.154Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=db.service.consul. type=AAAA class=IN latency=98.756µs client=127.0.0.1:42584 client_network=udp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.148Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.157Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=74a14cc2-bf0b-94b9-91a3-60c2841a72ba.query.consul. type=A class=IN latency=185.04µs client=127.0.0.1:45606 client_network=udp writer.go:29: 2021-01-29T19:33:37.161Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=74a14cc2-bf0b-94b9-91a3-60c2841a72ba.query.consul. type=AAAA class=IN latency=90.411µs client=127.0.0.1:47540 client_network=udp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.161Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee Address:127.0.0.1:30311}]" writer.go:29: 2021-01-29T19:33:37.164Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.raft: entering follower state: follower="Node at 127.0.0.1:30311 [Follower]" leader= writer.go:29: 2021-01-29T19:33:37.165Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.serf.wan: serf: EventMemberJoin: Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee.dc1 127.0.0.1 === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.168Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=foo.node.consul. type=A class=IN latency=118.729µs client=127.0.0.1:38052 client_network=udp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.169Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.serf.lan: serf: EventMemberJoin: Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee 127.0.0.1 writer.go:29: 2021-01-29T19:33:37.172Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Handled event for server in area: event=member-join server=Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee.dc1 area=wan === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.173Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=foo.node.consul. type=AAAA class=IN latency=72.395µs client=127.0.0.1:32796 client_network=udp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.174Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Adding LAN server: server="Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee (Addr: tcp/127.0.0.1:30311) (DC: dc1)" === RUN TestDNS_Lookup_TaggedIPAddresses/simple-ipv6 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.174Z [INFO] TestDNS_ServiceLookup_WanTranslation: Started DNS server: address=127.0.0.1:30306 network=udp writer.go:29: 2021-01-29T19:33:37.177Z [INFO] TestDNS_ServiceLookup_WanTranslation: Started DNS server: address=127.0.0.1:30306 network=tcp === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.179Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=db.service.consul. type=A class=IN latency=62.098µs client=127.0.0.1:38387 client_network=udp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.181Z [INFO] TestDNS_ServiceLookup_WanTranslation: Started HTTP server: address=127.0.0.1:30307 network=tcp === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.181Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=db.service.consul. type=AAAA class=IN latency=118.516µs client=127.0.0.1:33698 client_network=udp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.183Z [INFO] TestDNS_ServiceLookup_WanTranslation: started state syncer === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.185Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=74a14cc2-bf0b-94b9-91a3-60c2841a72ba.query.consul. type=A class=IN latency=97.997µs client=127.0.0.1:48695 client_network=udp writer.go:29: 2021-01-29T19:33:37.188Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=74a14cc2-bf0b-94b9-91a3-60c2841a72ba.query.consul. type=AAAA class=IN latency=668.262µs client=127.0.0.1:59323 client_network=udp writer.go:29: 2021-01-29T19:33:37.194Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=foo.node.consul. type=A class=IN latency=54.853µs client=127.0.0.1:49877 client_network=udp writer.go:29: 2021-01-29T19:33:37.196Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=foo.node.consul. type=AAAA class=IN latency=40.528µs client=127.0.0.1:56811 client_network=udp === RUN TestDNS_Lookup_TaggedIPAddresses/ipv4-with-tagged-ipv6 === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.204Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=db.service.consul. type=A class=IN latency=83.462µs client=127.0.0.1:50246 client_network=udp writer.go:29: 2021-01-29T19:33:37.206Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=db.service.consul. type=AAAA class=IN latency=69.097µs client=127.0.0.1:60428 client_network=udp writer.go:29: 2021-01-29T19:33:37.209Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=74a14cc2-bf0b-94b9-91a3-60c2841a72ba.query.consul. type=A class=IN latency=165.294µs client=127.0.0.1:60509 client_network=udp writer.go:29: 2021-01-29T19:33:37.213Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=74a14cc2-bf0b-94b9-91a3-60c2841a72ba.query.consul. type=AAAA class=IN latency=137.707µs client=127.0.0.1:57854 client_network=udp writer.go:29: 2021-01-29T19:33:37.215Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=foo.node.consul. type=A class=IN latency=57.405µs client=127.0.0.1:48617 client_network=udp writer.go:29: 2021-01-29T19:33:37.218Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=foo.node.consul. type=AAAA class=IN latency=75.739µs client=127.0.0.1:44279 client_network=udp === RUN TestDNS_Lookup_TaggedIPAddresses/ipv6-with-tagged-ipv4 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.224Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:33:37.224Z [WARN] TestDNS_ServiceLookup_WanTranslation.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:37.225Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.raft: entering candidate state: node="Node at 127.0.0.1:30311 [Candidate]" term=2 === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.226Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=db.service.consul. type=A class=IN latency=163.579µs client=127.0.0.1:38560 client_network=udp writer.go:29: 2021-01-29T19:33:37.231Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=db.service.consul. type=AAAA class=IN latency=82.688µs client=127.0.0.1:58132 client_network=udp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.236Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:37.237Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server.raft: vote granted: from=f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee term=2 tally=1 writer.go:29: 2021-01-29T19:33:37.238Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:37.238Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.raft: entering leader state: leader="Node at 127.0.0.1:30311 [Leader]" writer.go:29: 2021-01-29T19:33:37.239Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:37.241Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: initializing acls writer.go:29: 2021-01-29T19:33:37.241Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: New leader elected: payload=Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.238Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=74a14cc2-bf0b-94b9-91a3-60c2841a72ba.query.consul. type=A class=IN latency=127.863µs client=127.0.0.1:37282 client_network=udp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.245Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Created ACL 'global-management' policy === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.245Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=74a14cc2-bf0b-94b9-91a3-60c2841a72ba.query.consul. type=AAAA class=IN latency=101.057µs client=127.0.0.1:40121 client_network=udp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.247Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:37.248Z [INFO] TestDNS_ServiceLookup_WanTranslation.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:37.249Z [INFO] TestDNS_ServiceLookup_WanTranslation.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:37.250Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.serf.lan: serf: EventMemberUpdate: Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee writer.go:29: 2021-01-29T19:33:37.251Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.serf.wan: serf: EventMemberUpdate: Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee.dc1 writer.go:29: 2021-01-29T19:33:37.251Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Updating LAN server: server="Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee (Addr: tcp/127.0.0.1:30311) (DC: dc1)" writer.go:29: 2021-01-29T19:33:37.253Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Handled event for server in area: event=member-update server=Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee.dc1 area=wan === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.251Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=foo.node.consul. type=A class=IN latency=75.696µs client=127.0.0.1:49271 client_network=udp writer.go:29: 2021-01-29T19:33:37.257Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.dns: request served from client: name=foo.node.consul. type=AAAA class=IN latency=133.295µs client=127.0.0.1:40820 client_network=udp === CONT TestDNS_CaseInsensitiveServiceLookup writer.go:29: 2021-01-29T19:33:37.259Z [INFO] TestDNS_CaseInsensitiveServiceLookup: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:37.261Z [INFO] TestDNS_CaseInsensitiveServiceLookup: Endpoints down --- PASS: TestDNS_CaseInsensitiveServiceLookup (0.77s) === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.268Z [INFO] TestDNS_Lookup_TaggedIPAddresses: Requesting shutdown writer.go:29: 2021-01-29T19:33:37.270Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server: shutting down server writer.go:29: 2021-01-29T19:33:37.271Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:37.272Z [WARN] TestDNS_Lookup_TaggedIPAddresses.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:37.272Z [DEBUG] TestDNS_Lookup_TaggedIPAddresses.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.280Z [WARN] TestDNS_ServiceLookup_ServiceAddressIPV6: bootstrap = true: do not enable unless necessary === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.276Z [WARN] TestDNS_Lookup_TaggedIPAddresses.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:37.284Z [INFO] TestDNS_Lookup_TaggedIPAddresses.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:37.285Z [INFO] TestDNS_Lookup_TaggedIPAddresses: consul server down writer.go:29: 2021-01-29T19:33:37.286Z [INFO] TestDNS_Lookup_TaggedIPAddresses: shutdown complete writer.go:29: 2021-01-29T19:33:37.287Z [INFO] TestDNS_Lookup_TaggedIPAddresses: Stopping server: protocol=DNS address=127.0.0.1:30300 network=tcp writer.go:29: 2021-01-29T19:33:37.289Z [INFO] TestDNS_Lookup_TaggedIPAddresses: Stopping server: protocol=DNS address=127.0.0.1:30300 network=udp writer.go:29: 2021-01-29T19:33:37.290Z [INFO] TestDNS_Lookup_TaggedIPAddresses: Stopping server: protocol=HTTP address=127.0.0.1:30301 network=tcp === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.290Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:37.296Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.302Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.320Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6c26091b-223b-e6b3-0a73-5922cf2155b8 Address:127.0.0.1:30293}]" === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.326Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.327Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server.serf.wan: serf: EventMemberJoin: Node-6c26091b-223b-e6b3-0a73-5922cf2155b8.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.329Z [INFO] TestDNS_ServiceLookup_WanTranslation.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.330Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server.raft: entering follower state: follower="Node at 127.0.0.1:30293 [Follower]" leader= === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.332Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server: Skipping self join check for node since the cluster is too small: node=Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee writer.go:29: 2021-01-29T19:33:37.336Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: member joined, marking health alive: member=Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.336Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server.serf.lan: serf: EventMemberJoin: Node-6c26091b-223b-e6b3-0a73-5922cf2155b8 127.0.0.1 writer.go:29: 2021-01-29T19:33:37.340Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server: Handled event for server in area: event=member-join server=Node-6c26091b-223b-e6b3-0a73-5922cf2155b8.dc1 area=wan writer.go:29: 2021-01-29T19:33:37.344Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server: Adding LAN server: server="Node-6c26091b-223b-e6b3-0a73-5922cf2155b8 (Addr: tcp/127.0.0.1:30293) (DC: dc1)" === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:37.348Z [DEBUG] TestDNS_RecursorTimeout: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:37.350Z [DEBUG] TestDNS_RecursorTimeout: Node info in sync writer.go:29: 2021-01-29T19:33:37.351Z [DEBUG] TestDNS_RecursorTimeout: Node info in sync === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.353Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: Started DNS server: address=127.0.0.1:30288 network=udp writer.go:29: 2021-01-29T19:33:37.355Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: Started DNS server: address=127.0.0.1:30288 network=tcp writer.go:29: 2021-01-29T19:33:37.361Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: Started HTTP server: address=127.0.0.1:30289 network=tcp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.361Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server: Skipping self join check for node since the cluster is too small: node=Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.363Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: started state syncer === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.386Z [DEBUG] TestDNS_ServiceLookup_WanTranslation: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.398Z [WARN] TestDNS_ServiceLookup_ServiceAddressIPV6.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:37.402Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server.raft: entering candidate state: node="Node at 127.0.0.1:30293 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.407Z [INFO] TestDNS_ServiceLookup_WanTranslation: Synced node info === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.417Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:37.418Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6.server.raft: vote granted: from=6c26091b-223b-e6b3-0a73-5922cf2155b8 term=2 tally=1 writer.go:29: 2021-01-29T19:33:37.419Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:37.420Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server.raft: entering leader state: leader="Node at 127.0.0.1:30293 [Leader]" writer.go:29: 2021-01-29T19:33:37.421Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:37.423Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server: New leader elected: payload=Node-6c26091b-223b-e6b3-0a73-5922cf2155b8 writer.go:29: 2021-01-29T19:33:37.445Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30293 writer.go:29: 2021-01-29T19:33:37.451Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:37.481Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:37.483Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:37.485Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6.server: Skipping self join check for node since the cluster is too small: node=Node-6c26091b-223b-e6b3-0a73-5922cf2155b8 writer.go:29: 2021-01-29T19:33:37.487Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server: member joined, marking health alive: member=Node-6c26091b-223b-e6b3-0a73-5922cf2155b8 writer.go:29: 2021-01-29T19:33:37.533Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=136.09µs client=127.0.0.1:55954 client_network=udp writer.go:29: 2021-01-29T19:33:37.535Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:37.534Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6.dns: request served from client: name=a51422e8-d85a-ad23-40d1-8d1942a20226.query.consul. type=SRV class=IN latency=81.564µs client=127.0.0.1:45713 client_network=udp writer.go:29: 2021-01-29T19:33:37.534Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: Requesting shutdown === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:37.541Z [ERROR] TestDNS_RecursorTimeout.dns: recurse failed: error="read udp 127.0.0.1:49893->127.0.0.1:32806: i/o timeout" === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.541Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server: shutting down server === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:37.542Z [ERROR] TestDNS_RecursorTimeout.dns: all resolvers failed for question from client: question="{apple.com. 255 1}" client=127.0.0.1:42733 client_network=udp === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.542Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:37.543Z [DEBUG] TestDNS_RecursorTimeout.dns: request served from client: question="{apple.com. 255 1}" network=udp latency=3.008288521s client=127.0.0.1:42733 client_network=udp writer.go:29: 2021-01-29T19:33:37.545Z [INFO] TestDNS_RecursorTimeout: Requesting shutdown writer.go:29: 2021-01-29T19:33:37.546Z [INFO] TestDNS_RecursorTimeout.server: shutting down server === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.544Z [WARN] TestDNS_ServiceLookup_ServiceAddressIPV6.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:37.548Z [DEBUG] TestDNS_RecursorTimeout.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.549Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:37.550Z [WARN] TestDNS_RecursorTimeout.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:37.553Z [DEBUG] TestDNS_RecursorTimeout.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:37.557Z [WARN] TestDNS_RecursorTimeout.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.555Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: Synced node info writer.go:29: 2021-01-29T19:33:37.559Z [DEBUG] TestDNS_ServiceLookup_ServiceAddressIPV6: Node info in sync writer.go:29: 2021-01-29T19:33:37.556Z [WARN] TestDNS_ServiceLookup_ServiceAddressIPV6.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:37.566Z [INFO] TestDNS_RecursorTimeout.server.router.manager: shutting down === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.566Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6.server.router.manager: shutting down === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:37.567Z [INFO] TestDNS_RecursorTimeout: consul server down writer.go:29: 2021-01-29T19:33:37.568Z [INFO] TestDNS_RecursorTimeout: shutdown complete === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.568Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: consul server down === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:37.568Z [INFO] TestDNS_RecursorTimeout: Stopping server: protocol=DNS address=127.0.0.1:30216 network=tcp writer.go:29: 2021-01-29T19:33:37.569Z [INFO] TestDNS_RecursorTimeout: Stopping server: protocol=DNS address=127.0.0.1:30216 network=udp writer.go:29: 2021-01-29T19:33:37.570Z [INFO] TestDNS_RecursorTimeout: Stopping server: protocol=HTTP address=127.0.0.1:30217 network=tcp === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:37.569Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: shutdown complete writer.go:29: 2021-01-29T19:33:37.581Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: Stopping server: protocol=DNS address=127.0.0.1:30288 network=tcp writer.go:29: 2021-01-29T19:33:37.582Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: Stopping server: protocol=DNS address=127.0.0.1:30288 network=udp writer.go:29: 2021-01-29T19:33:37.583Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: Stopping server: protocol=HTTP address=127.0.0.1:30289 network=tcp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.608Z [WARN] TestDNS_ServiceLookup_WanTranslation: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:37.610Z [WARN] TestDNS_ServiceLookup_WanTranslation: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:37.611Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:37.616Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:37.630Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cdc78bd3-612a-d321-8a0e-3bab76a53f94 Address:127.0.0.1:30317}]" writer.go:29: 2021-01-29T19:33:37.634Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.raft: entering follower state: follower="Node at 127.0.0.1:30317 [Follower]" leader= writer.go:29: 2021-01-29T19:33:37.634Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.serf.wan: serf: EventMemberJoin: Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94.dc2 127.0.0.1 writer.go:29: 2021-01-29T19:33:37.640Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.serf.lan: serf: EventMemberJoin: Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94 127.0.0.1 writer.go:29: 2021-01-29T19:33:37.648Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Adding LAN server: server="Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94 (Addr: tcp/127.0.0.1:30317) (DC: dc2)" writer.go:29: 2021-01-29T19:33:37.649Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Handled event for server in area: event=member-join server=Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94.dc2 area=wan writer.go:29: 2021-01-29T19:33:37.653Z [INFO] TestDNS_ServiceLookup_WanTranslation: Started DNS server: address=127.0.0.1:30312 network=tcp writer.go:29: 2021-01-29T19:33:37.655Z [INFO] TestDNS_ServiceLookup_WanTranslation: Started DNS server: address=127.0.0.1:30312 network=udp writer.go:29: 2021-01-29T19:33:37.657Z [INFO] TestDNS_ServiceLookup_WanTranslation: Started HTTP server: address=127.0.0.1:30313 network=tcp writer.go:29: 2021-01-29T19:33:37.658Z [INFO] TestDNS_ServiceLookup_WanTranslation: started state syncer writer.go:29: 2021-01-29T19:33:37.694Z [WARN] TestDNS_ServiceLookup_WanTranslation.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:37.696Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.raft: entering candidate state: node="Node at 127.0.0.1:30317 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:37.699Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:37.700Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server.raft: vote granted: from=cdc78bd3-612a-d321-8a0e-3bab76a53f94 term=2 tally=1 writer.go:29: 2021-01-29T19:33:37.702Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:37.703Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.raft: entering leader state: leader="Node at 127.0.0.1:30317 [Leader]" writer.go:29: 2021-01-29T19:33:37.701Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:33:37.706Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:37.707Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: New leader elected: payload=Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94 writer.go:29: 2021-01-29T19:33:37.708Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: initializing acls writer.go:29: 2021-01-29T19:33:37.713Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:37.717Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:37.719Z [INFO] TestDNS_ServiceLookup_WanTranslation.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:37.721Z [INFO] TestDNS_ServiceLookup_WanTranslation.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:37.723Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.serf.lan: serf: EventMemberUpdate: Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94 writer.go:29: 2021-01-29T19:33:37.725Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.serf.wan: serf: EventMemberUpdate: Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94.dc2 writer.go:29: 2021-01-29T19:33:37.725Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Updating LAN server: server="Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94 (Addr: tcp/127.0.0.1:30317) (DC: dc2)" writer.go:29: 2021-01-29T19:33:37.728Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Handled event for server in area: event=member-update server=Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94.dc2 area=wan writer.go:29: 2021-01-29T19:33:37.733Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:37.741Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:37.743Z [INFO] TestDNS_ServiceLookup_WanTranslation.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:37.744Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server: Skipping self join check for node since the cluster is too small: node=Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94 writer.go:29: 2021-01-29T19:33:37.746Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: member joined, marking health alive: member=Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94 writer.go:29: 2021-01-29T19:33:37.749Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server: Skipping self join check for node since the cluster is too small: node=Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94 === CONT TestDNS_Lookup_TaggedIPAddresses writer.go:29: 2021-01-29T19:33:37.792Z [INFO] TestDNS_Lookup_TaggedIPAddresses: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:37.793Z [INFO] TestDNS_Lookup_TaggedIPAddresses: Endpoints down --- PASS: TestDNS_Lookup_TaggedIPAddresses (1.09s) --- PASS: TestDNS_Lookup_TaggedIPAddresses/simple-ipv4 (0.03s) --- PASS: TestDNS_Lookup_TaggedIPAddresses/simple-ipv6 (0.02s) --- PASS: TestDNS_Lookup_TaggedIPAddresses/ipv4-with-tagged-ipv6 (0.02s) --- PASS: TestDNS_Lookup_TaggedIPAddresses/ipv6-with-tagged-ipv4 (0.05s) === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:37.804Z [WARN] TestDNS_ServiceLookup_ServiceAddress_SRV: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:37.806Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:37.810Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:37.830Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:69b23946-7282-4055-ccb9-351535f45d2d Address:127.0.0.1:30323}]" writer.go:29: 2021-01-29T19:33:37.833Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server.raft: entering follower state: follower="Node at 127.0.0.1:30323 [Follower]" leader= writer.go:29: 2021-01-29T19:33:37.833Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server.serf.wan: serf: EventMemberJoin: Node-69b23946-7282-4055-ccb9-351535f45d2d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:37.836Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server.serf.lan: serf: EventMemberJoin: Node-69b23946-7282-4055-ccb9-351535f45d2d 127.0.0.1 writer.go:29: 2021-01-29T19:33:37.838Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.dns: recursor enabled writer.go:29: 2021-01-29T19:33:37.840Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: Started DNS server: address=127.0.0.1:30318 network=udp writer.go:29: 2021-01-29T19:33:37.841Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server: Handled event for server in area: event=member-join server=Node-69b23946-7282-4055-ccb9-351535f45d2d.dc1 area=wan writer.go:29: 2021-01-29T19:33:37.841Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server: Adding LAN server: server="Node-69b23946-7282-4055-ccb9-351535f45d2d (Addr: tcp/127.0.0.1:30323) (DC: dc1)" writer.go:29: 2021-01-29T19:33:37.842Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.dns: recursor enabled writer.go:29: 2021-01-29T19:33:37.844Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: Started DNS server: address=127.0.0.1:30318 network=tcp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.843Z [DEBUG] TestDNS_ServiceLookup_WanTranslation: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:37.844Z [INFO] TestDNS_ServiceLookup_WanTranslation: (WAN) joining: wan_addresses=[127.0.0.1:30310] === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:37.846Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: Started HTTP server: address=127.0.0.1:30319 network=tcp writer.go:29: 2021-01-29T19:33:37.847Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: started state syncer === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.846Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server.memberlist.wan: memberlist: Initiating push/pull sync with: 127.0.0.1:30310 writer.go:29: 2021-01-29T19:33:37.849Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server.memberlist.wan: memberlist: Stream connection from=127.0.0.1:38346 writer.go:29: 2021-01-29T19:33:37.850Z [INFO] TestDNS_ServiceLookup_WanTranslation: Synced node info writer.go:29: 2021-01-29T19:33:37.851Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.serf.wan: serf: EventMemberJoin: Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94.dc2 127.0.0.1 writer.go:29: 2021-01-29T19:33:37.854Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Handled event for server in area: event=member-join server=Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94.dc2 area=wan writer.go:29: 2021-01-29T19:33:37.852Z [DEBUG] TestDNS_ServiceLookup_WanTranslation: Node info in sync writer.go:29: 2021-01-29T19:33:37.852Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.serf.wan: serf: EventMemberJoin: Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:37.861Z [INFO] TestDNS_ServiceLookup_WanTranslation: (WAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:33:37.862Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: Handled event for server in area: event=member-join server=Node-f9a1fd0e-5f75-e219-83ac-2e4d3d3692ee.dc1 area=wan === RUN TestDNS_ServiceLookup_WanTranslation/node-wan-from-dc2 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.880Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=SRV class=IN latency=117.199µs client=127.0.0.1:40949 client_network=udp writer.go:29: 2021-01-29T19:33:37.883Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=SRV class=IN latency=90.214µs client=127.0.0.1:48162 client_network=udp === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:37.884Z [WARN] TestDNS_ServiceLookup_ServiceAddress_SRV.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:37.886Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server.raft: entering candidate state: node="Node at 127.0.0.1:30323 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.887Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=A class=IN latency=83.016µs client=127.0.0.1:43936 client_network=udp === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:37.888Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.891Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=A class=IN latency=171.612µs client=127.0.0.1:54310 client_network=udp === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:37.890Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.server.raft: vote granted: from=69b23946-7282-4055-ccb9-351535f45d2d term=2 tally=1 writer.go:29: 2021-01-29T19:33:37.894Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:37.895Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server.raft: entering leader state: leader="Node at 127.0.0.1:30323 [Leader]" writer.go:29: 2021-01-29T19:33:37.897Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:37.897Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server: New leader elected: payload=Node-69b23946-7282-4055-ccb9-351535f45d2d === RUN TestDNS_ServiceLookup_WanTranslation/service-addr-from-dc2 === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:37.908Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30323 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.915Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=SRV class=IN latency=135.437µs client=127.0.0.1:34845 client_network=udp === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:37.913Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.921Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=SRV class=IN latency=140.599µs client=127.0.0.1:60748 client_network=udp === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:37.924Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:37.926Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:37.927Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.server: Skipping self join check for node since the cluster is too small: node=Node-69b23946-7282-4055-ccb9-351535f45d2d writer.go:29: 2021-01-29T19:33:37.928Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server: member joined, marking health alive: member=Node-69b23946-7282-4055-ccb9-351535f45d2d === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.934Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=A class=IN latency=96.676µs client=127.0.0.1:42556 client_network=udp writer.go:29: 2021-01-29T19:33:37.938Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=A class=IN latency=68.615µs client=127.0.0.1:59990 client_network=udp === RUN TestDNS_ServiceLookup_WanTranslation/service-wan-from-dc2 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.947Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=SRV class=IN latency=65.411µs client=127.0.0.1:54906 client_network=udp writer.go:29: 2021-01-29T19:33:37.952Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=SRV class=IN latency=73.346µs client=127.0.0.1:33516 client_network=udp writer.go:29: 2021-01-29T19:33:37.958Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=A class=IN latency=139.979µs client=127.0.0.1:56237 client_network=udp writer.go:29: 2021-01-29T19:33:37.966Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=A class=IN latency=1.252139ms client=127.0.0.1:35184 client_network=udp === RUN TestDNS_ServiceLookup_WanTranslation/node-addr-from-dc1 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:37.976Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:37.989Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=SRV class=IN latency=13.046104ms client=127.0.0.1:38190 client_network=udp writer.go:29: 2021-01-29T19:33:37.999Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=SRV class=IN latency=2.741374ms client=127.0.0.1:47492 client_network=udp writer.go:29: 2021-01-29T19:33:38.006Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=A class=IN latency=2.263001ms client=127.0.0.1:54607 client_network=udp writer.go:29: 2021-01-29T19:33:38.016Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=A class=IN latency=4.11875ms client=127.0.0.1:43764 client_network=udp === RUN TestDNS_ServiceLookup_WanTranslation/node-wan-from-dc1 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.037Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=SRV class=IN latency=8.092205ms client=127.0.0.1:59937 client_network=udp writer.go:29: 2021-01-29T19:33:38.044Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=SRV class=IN latency=2.888125ms client=127.0.0.1:59498 client_network=udp writer.go:29: 2021-01-29T19:33:38.054Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=A class=IN latency=2.816575ms client=127.0.0.1:36737 client_network=udp writer.go:29: 2021-01-29T19:33:38.059Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=A class=IN latency=2.629374ms client=127.0.0.1:39103 client_network=udp === RUN TestDNS_ServiceLookup_WanTranslation/service-addr-from-dc1 === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:38.080Z [INFO] TestDNS_RecursorTimeout: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:38.080Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_RecursorTimeout writer.go:29: 2021-01-29T19:33:38.081Z [INFO] TestDNS_RecursorTimeout: Endpoints down --- PASS: TestDNS_RecursorTimeout (3.66s) === CONT TestDNS_ServiceLookup_ServiceAddress_A === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.083Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=SRV class=IN latency=2.163485ms client=127.0.0.1:54802 client_network=udp === CONT TestDNS_ServiceLookup_ServiceAddressIPV6 writer.go:29: 2021-01-29T19:33:38.084Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:38.086Z [INFO] TestDNS_ServiceLookup_ServiceAddressIPV6: Endpoints down --- PASS: TestDNS_ServiceLookup_ServiceAddressIPV6 (0.82s) === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:38.089Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: Synced node info === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.097Z [WARN] TestDNS_ExternalServiceToConsulCNAMENestedLookup: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.094Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=SRV class=IN latency=4.971612ms client=127.0.0.1:32829 client_network=udp === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.097Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMENestedLookup.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_ServiceAddress_A writer.go:29: 2021-01-29T19:33:38.091Z [WARN] TestDNS_ServiceLookup_ServiceAddress_A: bootstrap = true: do not enable unless necessary === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.099Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMENestedLookup.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_ServiceAddress_A writer.go:29: 2021-01-29T19:33:38.099Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:38.100Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:38.105Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:33a9e827-f68d-000f-a0fe-9419011e1368 Address:127.0.0.1:30341}]" writer.go:29: 2021-01-29T19:33:38.106Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server.raft: entering follower state: follower="Node at 127.0.0.1:30341 [Follower]" leader= writer.go:29: 2021-01-29T19:33:38.107Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server.serf.wan: serf: EventMemberJoin: Node-33a9e827-f68d-000f-a0fe-9419011e1368.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.110Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=A class=IN latency=2.012534ms client=127.0.0.1:53967 client_network=udp === CONT TestDNS_ServiceLookup_ServiceAddress_A writer.go:29: 2021-01-29T19:33:38.110Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server.serf.lan: serf: EventMemberJoin: Node-33a9e827-f68d-000f-a0fe-9419011e1368 127.0.0.1 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.114Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=A class=IN latency=1.750136ms client=127.0.0.1:34446 client_network=udp === RUN TestDNS_ServiceLookup_WanTranslation/service-wan-from-dc1 === CONT TestDNS_ServiceLookup_ServiceAddress_A writer.go:29: 2021-01-29T19:33:38.113Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: Started DNS server: address=127.0.0.1:30336 network=udp writer.go:29: 2021-01-29T19:33:38.117Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: Started DNS server: address=127.0.0.1:30336 network=tcp writer.go:29: 2021-01-29T19:33:38.114Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server: Adding LAN server: server="Node-33a9e827-f68d-000f-a0fe-9419011e1368 (Addr: tcp/127.0.0.1:30341) (DC: dc1)" writer.go:29: 2021-01-29T19:33:38.114Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server: Handled event for server in area: event=member-join server=Node-33a9e827-f68d-000f-a0fe-9419011e1368.dc1 area=wan writer.go:29: 2021-01-29T19:33:38.119Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: Started HTTP server: address=127.0.0.1:30337 network=tcp writer.go:29: 2021-01-29T19:33:38.122Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: started state syncer === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.124Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:08dbb034-1271-0a7b-0059-4a271c9e5d1a Address:127.0.0.1:30329}]" writer.go:29: 2021-01-29T19:33:38.125Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.raft: entering follower state: follower="Node at 127.0.0.1:30329 [Follower]" leader= writer.go:29: 2021-01-29T19:33:38.126Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:38.129Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:33:38.132Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:33:38.132Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:30329) (DC: dc1)" === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.124Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=SRV class=IN latency=2.144905ms client=127.0.0.1:39858 client_network=udp writer.go:29: 2021-01-29T19:33:38.140Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server.serf.wan: serf: messageJoinType: Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94.dc2 === CONT TestDNS_ServiceLookup_ServiceAddress_A writer.go:29: 2021-01-29T19:33:38.145Z [WARN] TestDNS_ServiceLookup_ServiceAddress_A.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:38.149Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server.raft: entering candidate state: node="Node at 127.0.0.1:30341 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:38.153Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:38.154Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A.server.raft: vote granted: from=33a9e827-f68d-000f-a0fe-9419011e1368 term=2 tally=1 writer.go:29: 2021-01-29T19:33:38.155Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:38.156Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server.raft: entering leader state: leader="Node at 127.0.0.1:30341 [Leader]" writer.go:29: 2021-01-29T19:33:38.157Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:38.157Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server: New leader elected: payload=Node-33a9e827-f68d-000f-a0fe-9419011e1368 writer.go:29: 2021-01-29T19:33:38.161Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30341 writer.go:29: 2021-01-29T19:33:38.165Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:38.170Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.169Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.server.serf.wan: serf: messageJoinType: Node-cdc78bd3-612a-d321-8a0e-3bab76a53f94.dc2 === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.172Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Started DNS server: address=127.0.0.1:30324 network=udp writer.go:29: 2021-01-29T19:33:38.174Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Started DNS server: address=127.0.0.1:30324 network=tcp === CONT TestDNS_ServiceLookup_ServiceAddress_A writer.go:29: 2021-01-29T19:33:38.171Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.leader: started routine: routine="CA root pruning" === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.175Z [WARN] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_ServiceAddress_A writer.go:29: 2021-01-29T19:33:38.176Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A.server: Skipping self join check for node since the cluster is too small: node=Node-33a9e827-f68d-000f-a0fe-9419011e1368 writer.go:29: 2021-01-29T19:33:38.177Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server: member joined, marking health alive: member=Node-33a9e827-f68d-000f-a0fe-9419011e1368 === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.177Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.raft: entering candidate state: node="Node at 127.0.0.1:30329 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:38.176Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Started HTTP server: address=127.0.0.1:30325 network=tcp writer.go:29: 2021-01-29T19:33:38.179Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: started state syncer writer.go:29: 2021-01-29T19:33:38.180Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:38.181Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.raft: vote granted: from=08dbb034-1271-0a7b-0059-4a271c9e5d1a term=2 tally=1 writer.go:29: 2021-01-29T19:33:38.181Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:38.182Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.raft: entering leader state: leader="Node at 127.0.0.1:30329 [Leader]" writer.go:29: 2021-01-29T19:33:38.183Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:38.184Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:33:38.185Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30329 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.186Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=SRV class=IN latency=5.07552ms client=127.0.0.1:57061 client_network=udp writer.go:29: 2021-01-29T19:33:38.192Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=A class=IN latency=1.815415ms client=127.0.0.1:35641 client_network=udp writer.go:29: 2021-01-29T19:33:38.198Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=A class=IN latency=2.130614ms client=127.0.0.1:39041 client_network=udp === RUN TestDNS_ServiceLookup_WanTranslation/node-addr-from-dc2 === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.202Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.209Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=SRV class=IN latency=128.64µs client=127.0.0.1:46900 client_network=udp === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.211Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:38.212Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.214Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.216Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=SRV class=IN latency=81.866µs client=127.0.0.1:52167 client_network=udp === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.216Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.221Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=db.service.dc2.consul. type=A class=IN latency=74.262µs client=127.0.0.1:57418 client_network=udp writer.go:29: 2021-01-29T19:33:38.225Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.dns: request served from client: name=73d8e07c-a4ae-cb1d-bce9-324aee32edd3.query.dc2.consul. type=A class=IN latency=122.909µs client=127.0.0.1:55678 client_network=udp === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.228Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.229Z [INFO] TestDNS_ServiceLookup_WanTranslation: Requesting shutdown writer.go:29: 2021-01-29T19:33:38.230Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: shutting down server writer.go:29: 2021-01-29T19:33:38.231Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:38.232Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:38.232Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.233Z [WARN] TestDNS_ServiceLookup_WanTranslation.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.234Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:38.234Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:38.234Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.235Z [WARN] TestDNS_ServiceLookup_WanTranslation.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.241Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.router.manager: shutting down === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.241Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Synced node info === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.242Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:38.244Z [INFO] TestDNS_ServiceLookup_WanTranslation: consul server down writer.go:29: 2021-01-29T19:33:38.245Z [INFO] TestDNS_ServiceLookup_WanTranslation: shutdown complete writer.go:29: 2021-01-29T19:33:38.246Z [INFO] TestDNS_ServiceLookup_WanTranslation: Stopping server: protocol=DNS address=127.0.0.1:30312 network=tcp writer.go:29: 2021-01-29T19:33:38.247Z [INFO] TestDNS_ServiceLookup_WanTranslation: Stopping server: protocol=DNS address=127.0.0.1:30312 network=udp writer.go:29: 2021-01-29T19:33:38.248Z [INFO] TestDNS_ServiceLookup_WanTranslation: Stopping server: protocol=HTTP address=127.0.0.1:30313 network=tcp === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.248Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMENestedLookup.dns: request served from client: name=alias2.service.consul. type=SRV class=IN latency=2.057421ms client=127.0.0.1:59605 client_network=udp writer.go:29: 2021-01-29T19:33:38.248Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Requesting shutdown writer.go:29: 2021-01-29T19:33:38.251Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server: shutting down server writer.go:29: 2021-01-29T19:33:38.253Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMENestedLookup.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.255Z [WARN] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.255Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMENestedLookup.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.258Z [WARN] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.261Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:38.261Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: consul server down writer.go:29: 2021-01-29T19:33:38.263Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: shutdown complete writer.go:29: 2021-01-29T19:33:38.264Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Stopping server: protocol=DNS address=127.0.0.1:30324 network=tcp writer.go:29: 2021-01-29T19:33:38.265Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Stopping server: protocol=DNS address=127.0.0.1:30324 network=udp writer.go:29: 2021-01-29T19:33:38.266Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Stopping server: protocol=HTTP address=127.0.0.1:30325 network=tcp === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:38.278Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.dns: cname recurse RTT for name: name=www.google.com. rtt=493.319µs writer.go:29: 2021-01-29T19:33:38.279Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=1.601644ms client=127.0.0.1:59786 client_network=udp writer.go:29: 2021-01-29T19:33:38.280Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.dns: cname recurse RTT for name: name=www.google.com. rtt=699.518µs writer.go:29: 2021-01-29T19:33:38.281Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.dns: request served from client: name=30e02226-8f57-9198-a3d8-43cb861b1d9d.query.consul. type=SRV class=IN latency=1.917928ms client=127.0.0.1:58946 client_network=udp writer.go:29: 2021-01-29T19:33:38.281Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: Requesting shutdown writer.go:29: 2021-01-29T19:33:38.284Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server: shutting down server writer.go:29: 2021-01-29T19:33:38.284Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.285Z [WARN] TestDNS_ServiceLookup_ServiceAddress_SRV.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.287Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_SRV.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.287Z [WARN] TestDNS_ServiceLookup_ServiceAddress_SRV.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.290Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:38.291Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: consul server down writer.go:29: 2021-01-29T19:33:38.292Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: shutdown complete writer.go:29: 2021-01-29T19:33:38.293Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: Stopping server: protocol=DNS address=127.0.0.1:30318 network=tcp writer.go:29: 2021-01-29T19:33:38.294Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: Stopping server: protocol=DNS address=127.0.0.1:30318 network=udp writer.go:29: 2021-01-29T19:33:38.295Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: Stopping server: protocol=HTTP address=127.0.0.1:30319 network=tcp === CONT TestDNS_ServiceLookup_ServiceAddress_A writer.go:29: 2021-01-29T19:33:38.512Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=128.309µs client=127.0.0.1:43612 client_network=udp writer.go:29: 2021-01-29T19:33:38.513Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A.dns: request served from client: name=6c3c9e9d-00f3-6ae2-7a80-ff0460c662aa.query.consul. type=SRV class=IN latency=92.102µs client=127.0.0.1:34410 client_network=udp writer.go:29: 2021-01-29T19:33:38.513Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: Requesting shutdown writer.go:29: 2021-01-29T19:33:38.518Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server: shutting down server writer.go:29: 2021-01-29T19:33:38.520Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.514Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:38.522Z [WARN] TestDNS_ServiceLookup_ServiceAddress_A.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.522Z [DEBUG] TestDNS_ServiceLookup_ServiceAddress_A.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.525Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: Synced node info writer.go:29: 2021-01-29T19:33:38.528Z [WARN] TestDNS_ServiceLookup_ServiceAddress_A.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.542Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:38.543Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: consul server down writer.go:29: 2021-01-29T19:33:38.545Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: shutdown complete writer.go:29: 2021-01-29T19:33:38.546Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: Stopping server: protocol=DNS address=127.0.0.1:30336 network=tcp writer.go:29: 2021-01-29T19:33:38.548Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: Stopping server: protocol=DNS address=127.0.0.1:30336 network=udp writer.go:29: 2021-01-29T19:33:38.550Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: Stopping server: protocol=HTTP address=127.0.0.1:30337 network=tcp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.749Z [INFO] TestDNS_ServiceLookup_WanTranslation: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:38.751Z [INFO] TestDNS_ServiceLookup_WanTranslation: Endpoints down writer.go:29: 2021-01-29T19:33:38.752Z [INFO] TestDNS_ServiceLookup_WanTranslation: Requesting shutdown writer.go:29: 2021-01-29T19:33:38.754Z [INFO] TestDNS_ServiceLookup_WanTranslation.server: shutting down server writer.go:29: 2021-01-29T19:33:38.755Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:38.756Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.758Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:38.760Z [WARN] TestDNS_ServiceLookup_WanTranslation.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.758Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:38.759Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.760Z [DEBUG] TestDNS_ServiceLookup_WanTranslation.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.767Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.764Z [WARN] TestDNS_ServiceLookup_WanTranslation.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ExternalServiceToConsulCNAMENestedLookup writer.go:29: 2021-01-29T19:33:38.769Z [INFO] TestDNS_ExternalServiceToConsulCNAMENestedLookup: Endpoints down --- PASS: TestDNS_ExternalServiceToConsulCNAMENestedLookup (0.68s) === CONT TestDNS_NSRecords_IPV6 === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:38.771Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:38.772Z [INFO] TestDNS_ServiceLookup_WanTranslation.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:38.772Z [INFO] TestDNS_ServiceLookup_WanTranslation: consul server down writer.go:29: 2021-01-29T19:33:38.777Z [INFO] TestDNS_ServiceLookup_WanTranslation: shutdown complete writer.go:29: 2021-01-29T19:33:38.778Z [INFO] TestDNS_ServiceLookup_WanTranslation: Stopping server: protocol=DNS address=127.0.0.1:30306 network=tcp writer.go:29: 2021-01-29T19:33:38.780Z [INFO] TestDNS_ServiceLookup_WanTranslation: Stopping server: protocol=DNS address=127.0.0.1:30306 network=udp writer.go:29: 2021-01-29T19:33:38.782Z [INFO] TestDNS_ServiceLookup_WanTranslation: Stopping server: protocol=HTTP address=127.0.0.1:30307 network=tcp === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.790Z [WARN] TestDNS_NSRecords_IPV6: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:38.792Z [DEBUG] TestDNS_NSRecords_IPV6.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_ServiceAddress_SRV writer.go:29: 2021-01-29T19:33:38.796Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:38.798Z [INFO] TestDNS_ServiceLookup_ServiceAddress_SRV: Endpoints down --- PASS: TestDNS_ServiceLookup_ServiceAddress_SRV (1.01s) === CONT TestDNS_ExternalServiceToConsulCNAMELookup === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.794Z [DEBUG] TestDNS_NSRecords_IPV6.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:38.802Z [INFO] TestDNS_NSRecords_IPV6.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e3b7608f-df45-1618-a3a5-ae530a3ffe8a Address:[::1]:30335}]" writer.go:29: 2021-01-29T19:33:38.805Z [INFO] TestDNS_NSRecords_IPV6.server.raft: entering follower state: follower="Node at [::1]:30335 [Follower]" leader= === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:38.809Z [WARN] TestDNS_ExternalServiceToConsulCNAMELookup: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:38.811Z [WARN] TestDNS_ExternalServiceToConsulCNAMELookup: Node name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: node_name="test node" === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.810Z [INFO] TestDNS_NSRecords_IPV6.server.serf.wan: serf: EventMemberJoin: server1.dc1 ::1 === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:38.812Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:38.815Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.816Z [INFO] TestDNS_NSRecords_IPV6.server.serf.lan: serf: EventMemberJoin: server1 ::1 writer.go:29: 2021-01-29T19:33:38.821Z [INFO] TestDNS_NSRecords_IPV6.server: Adding LAN server: server="server1 (Addr: tcp/[::1]:30335) (DC: dc1)" writer.go:29: 2021-01-29T19:33:38.823Z [INFO] TestDNS_NSRecords_IPV6.server: Handled event for server in area: event=member-join server=server1.dc1 area=wan === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:38.822Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5e86ea68-eaee-0950-dfd6-af7ed2297020 Address:127.0.0.1:30347}]" writer.go:29: 2021-01-29T19:33:38.826Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server.raft: entering follower state: follower="Node at 127.0.0.1:30347 [Follower]" leader= === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.828Z [INFO] TestDNS_NSRecords_IPV6: Started DNS server: address=127.0.0.1:30330 network=tcp writer.go:29: 2021-01-29T19:33:38.831Z [INFO] TestDNS_NSRecords_IPV6: Started DNS server: address=127.0.0.1:30330 network=udp writer.go:29: 2021-01-29T19:33:38.834Z [INFO] TestDNS_NSRecords_IPV6: Started HTTP server: address=127.0.0.1:30331 network=tcp writer.go:29: 2021-01-29T19:33:38.835Z [INFO] TestDNS_NSRecords_IPV6: started state syncer === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:38.831Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server.serf.wan: serf: EventMemberJoin: test node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:38.840Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server.serf.lan: serf: EventMemberJoin: test node 127.0.0.1 writer.go:29: 2021-01-29T19:33:38.845Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server: Adding LAN server: server="test node (Addr: tcp/127.0.0.1:30347) (DC: dc1)" writer.go:29: 2021-01-29T19:33:38.846Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server: Handled event for server in area: event=member-join server="test node.dc1" area=wan writer.go:29: 2021-01-29T19:33:38.851Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: Started DNS server: address=127.0.0.1:30342 network=tcp writer.go:29: 2021-01-29T19:33:38.852Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: Started DNS server: address=127.0.0.1:30342 network=udp === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.855Z [WARN] TestDNS_NSRecords_IPV6.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:38.856Z [INFO] TestDNS_NSRecords_IPV6.server.raft: entering candidate state: node="Node at [::1]:30335 [Candidate]" term=2 === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:38.858Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: Started HTTP server: address=127.0.0.1:30343 network=tcp === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.858Z [DEBUG] TestDNS_NSRecords_IPV6.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:38.860Z [DEBUG] TestDNS_NSRecords_IPV6.server.raft: vote granted: from=e3b7608f-df45-1618-a3a5-ae530a3ffe8a term=2 tally=1 writer.go:29: 2021-01-29T19:33:38.861Z [INFO] TestDNS_NSRecords_IPV6.server.raft: election won: tally=1 === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:38.861Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: started state syncer === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.863Z [INFO] TestDNS_NSRecords_IPV6.server.raft: entering leader state: leader="Node at [::1]:30335 [Leader]" writer.go:29: 2021-01-29T19:33:38.865Z [INFO] TestDNS_NSRecords_IPV6.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:38.865Z [INFO] TestDNS_NSRecords_IPV6.server: New leader elected: payload=server1 writer.go:29: 2021-01-29T19:33:38.866Z [DEBUG] TestDNS_NSRecords_IPV6.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=[::1]:30335 === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:38.877Z [WARN] TestDNS_ExternalServiceToConsulCNAMELookup.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:38.880Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server.raft: entering candidate state: node="Node at 127.0.0.1:30347 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:38.884Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup.server.raft: votes: needed=1 === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.876Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:38.887Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup.server.raft: vote granted: from=5e86ea68-eaee-0950-dfd6-af7ed2297020 term=2 tally=1 === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.894Z [INFO] TestDNS_NSRecords_IPV6.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:38.896Z [INFO] TestDNS_NSRecords_IPV6.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.898Z [DEBUG] TestDNS_NSRecords_IPV6.server: Skipping self join check for node since the cluster is too small: node=server1 writer.go:29: 2021-01-29T19:33:38.900Z [INFO] TestDNS_NSRecords_IPV6.server: member joined, marking health alive: member=server1 === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:38.894Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:38.904Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server.raft: entering leader state: leader="Node at 127.0.0.1:30347 [Leader]" writer.go:29: 2021-01-29T19:33:38.906Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:38.906Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server: New leader elected: payload="test node" writer.go:29: 2021-01-29T19:33:38.913Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30347 writer.go:29: 2021-01-29T19:33:38.919Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:38.925Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:38.927Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.928Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup.server: Skipping self join check for node since the cluster is too small: node="test node" writer.go:29: 2021-01-29T19:33:38.929Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server: member joined, marking health alive: member="test node" === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:38.937Z [DEBUG] TestDNS_NSRecords_IPV6.dns: request served from client: name=server1.node.dc1.consul. type=NS class=IN latency=92.02µs client=127.0.0.1:35380 client_network=udp writer.go:29: 2021-01-29T19:33:38.939Z [INFO] TestDNS_NSRecords_IPV6: Requesting shutdown writer.go:29: 2021-01-29T19:33:38.941Z [INFO] TestDNS_NSRecords_IPV6.server: shutting down server writer.go:29: 2021-01-29T19:33:38.942Z [DEBUG] TestDNS_NSRecords_IPV6.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.944Z [WARN] TestDNS_NSRecords_IPV6.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.942Z [ERROR] TestDNS_NSRecords_IPV6.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:38.944Z [DEBUG] TestDNS_NSRecords_IPV6.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:38.947Z [WARN] TestDNS_NSRecords_IPV6.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:38.954Z [INFO] TestDNS_NSRecords_IPV6.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:38.956Z [INFO] TestDNS_NSRecords_IPV6: consul server down writer.go:29: 2021-01-29T19:33:38.958Z [INFO] TestDNS_NSRecords_IPV6: shutdown complete writer.go:29: 2021-01-29T19:33:38.960Z [INFO] TestDNS_NSRecords_IPV6: Stopping server: protocol=DNS address=127.0.0.1:30330 network=tcp writer.go:29: 2021-01-29T19:33:38.962Z [INFO] TestDNS_NSRecords_IPV6: Stopping server: protocol=DNS address=127.0.0.1:30330 network=udp writer.go:29: 2021-01-29T19:33:38.964Z [INFO] TestDNS_NSRecords_IPV6: Stopping server: protocol=HTTP address=127.0.0.1:30331 network=tcp === CONT TestDNS_ServiceLookup_ServiceAddress_A writer.go:29: 2021-01-29T19:33:39.059Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:39.060Z [INFO] TestDNS_ServiceLookup_ServiceAddress_A: Endpoints down --- PASS: TestDNS_ServiceLookup_ServiceAddress_A (0.98s) === CONT TestDNS_InifiniteRecursion writer.go:29: 2021-01-29T19:33:39.071Z [WARN] TestDNS_InifiniteRecursion: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:39.072Z [WARN] TestDNS_InifiniteRecursion: Node name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: node_name="test node" writer.go:29: 2021-01-29T19:33:39.073Z [DEBUG] TestDNS_InifiniteRecursion.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:39.075Z [DEBUG] TestDNS_InifiniteRecursion.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:39.102Z [INFO] TestDNS_InifiniteRecursion.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5532f2e5-2069-3e9f-1c38-f4e7838ec83f Address:127.0.0.1:30353}]" writer.go:29: 2021-01-29T19:33:39.106Z [INFO] TestDNS_InifiniteRecursion.server.raft: entering follower state: follower="Node at 127.0.0.1:30353 [Follower]" leader= writer.go:29: 2021-01-29T19:33:39.107Z [INFO] TestDNS_InifiniteRecursion.server.serf.wan: serf: EventMemberJoin: test node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.115Z [INFO] TestDNS_InifiniteRecursion.server.serf.lan: serf: EventMemberJoin: test node 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.121Z [INFO] TestDNS_InifiniteRecursion: Started DNS server: address=127.0.0.1:30348 network=udp writer.go:29: 2021-01-29T19:33:39.125Z [INFO] TestDNS_InifiniteRecursion.server: Adding LAN server: server="test node (Addr: tcp/127.0.0.1:30353) (DC: dc1)" writer.go:29: 2021-01-29T19:33:39.127Z [INFO] TestDNS_InifiniteRecursion: Started DNS server: address=127.0.0.1:30348 network=tcp writer.go:29: 2021-01-29T19:33:39.125Z [INFO] TestDNS_InifiniteRecursion.server: Handled event for server in area: event=member-join server="test node.dc1" area=wan writer.go:29: 2021-01-29T19:33:39.131Z [INFO] TestDNS_InifiniteRecursion: Started HTTP server: address=127.0.0.1:30349 network=tcp writer.go:29: 2021-01-29T19:33:39.134Z [INFO] TestDNS_InifiniteRecursion: started state syncer writer.go:29: 2021-01-29T19:33:39.172Z [WARN] TestDNS_InifiniteRecursion.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:39.173Z [INFO] TestDNS_InifiniteRecursion.server.raft: entering candidate state: node="Node at 127.0.0.1:30353 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:39.176Z [DEBUG] TestDNS_InifiniteRecursion.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:39.177Z [DEBUG] TestDNS_InifiniteRecursion.server.raft: vote granted: from=5532f2e5-2069-3e9f-1c38-f4e7838ec83f term=2 tally=1 writer.go:29: 2021-01-29T19:33:39.178Z [INFO] TestDNS_InifiniteRecursion.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:39.179Z [INFO] TestDNS_InifiniteRecursion.server.raft: entering leader state: leader="Node at 127.0.0.1:30353 [Leader]" writer.go:29: 2021-01-29T19:33:39.180Z [INFO] TestDNS_InifiniteRecursion.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:39.180Z [INFO] TestDNS_InifiniteRecursion.server: New leader elected: payload="test node" writer.go:29: 2021-01-29T19:33:39.181Z [DEBUG] TestDNS_InifiniteRecursion.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30353 writer.go:29: 2021-01-29T19:33:39.189Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:39.193Z [INFO] TestDNS_InifiniteRecursion.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:39.194Z [INFO] TestDNS_InifiniteRecursion.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.195Z [DEBUG] TestDNS_InifiniteRecursion.server: Skipping self join check for node since the cluster is too small: node="test node" writer.go:29: 2021-01-29T19:33:39.196Z [INFO] TestDNS_InifiniteRecursion.server: member joined, marking health alive: member="test node" === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:39.221Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:39.224Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup.dns: request served from client: name=alias.service.consul. type=SRV class=IN latency=1.218655ms client=127.0.0.1:36802 client_network=udp writer.go:29: 2021-01-29T19:33:39.226Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: Synced node info writer.go:29: 2021-01-29T19:33:39.227Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup.dns: request served from client: name=alias.service.CoNsUl. type=SRV class=IN latency=1.643103ms client=127.0.0.1:47965 client_network=udp writer.go:29: 2021-01-29T19:33:39.227Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: Requesting shutdown writer.go:29: 2021-01-29T19:33:39.230Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server: shutting down server writer.go:29: 2021-01-29T19:33:39.231Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.233Z [WARN] TestDNS_ExternalServiceToConsulCNAMELookup.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:39.235Z [DEBUG] TestDNS_ExternalServiceToConsulCNAMELookup.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.236Z [WARN] TestDNS_ExternalServiceToConsulCNAMELookup.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:39.257Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:39.258Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: consul server down writer.go:29: 2021-01-29T19:33:39.261Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: shutdown complete writer.go:29: 2021-01-29T19:33:39.263Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: Stopping server: protocol=DNS address=127.0.0.1:30342 network=tcp writer.go:29: 2021-01-29T19:33:39.265Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: Stopping server: protocol=DNS address=127.0.0.1:30342 network=udp writer.go:29: 2021-01-29T19:33:39.267Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: Stopping server: protocol=HTTP address=127.0.0.1:30343 network=tcp === CONT TestDNS_ServiceLookup_WanTranslation writer.go:29: 2021-01-29T19:33:39.283Z [INFO] TestDNS_ServiceLookup_WanTranslation: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:39.285Z [INFO] TestDNS_ServiceLookup_WanTranslation: Endpoints down --- PASS: TestDNS_ServiceLookup_WanTranslation (2.16s) --- PASS: TestDNS_ServiceLookup_WanTranslation/node-wan-from-dc2 (0.02s) --- PASS: TestDNS_ServiceLookup_WanTranslation/service-addr-from-dc2 (0.04s) --- PASS: TestDNS_ServiceLookup_WanTranslation/service-wan-from-dc2 (0.03s) --- PASS: TestDNS_ServiceLookup_WanTranslation/node-addr-from-dc1 (0.05s) --- PASS: TestDNS_ServiceLookup_WanTranslation/node-wan-from-dc1 (0.04s) --- PASS: TestDNS_ServiceLookup_WanTranslation/service-addr-from-dc1 (0.05s) --- PASS: TestDNS_ServiceLookup_WanTranslation/service-wan-from-dc1 (0.08s) --- PASS: TestDNS_ServiceLookup_WanTranslation/node-addr-from-dc2 (0.03s) === CONT TestDNS_ExternalServiceLookup writer.go:29: 2021-01-29T19:33:39.307Z [WARN] TestDNS_ExternalServiceLookup: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:39.309Z [DEBUG] TestDNS_ExternalServiceLookup.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:39.310Z [DEBUG] TestDNS_ExternalServiceLookup.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:39.321Z [INFO] TestDNS_ExternalServiceLookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bdbb4f25-d716-2cfb-3a1d-6b5cf9c733c9 Address:127.0.0.1:30365}]" writer.go:29: 2021-01-29T19:33:39.323Z [INFO] TestDNS_ExternalServiceLookup.server.raft: entering follower state: follower="Node at 127.0.0.1:30365 [Follower]" leader= writer.go:29: 2021-01-29T19:33:39.324Z [INFO] TestDNS_ExternalServiceLookup.server.serf.wan: serf: EventMemberJoin: Node-bdbb4f25-d716-2cfb-3a1d-6b5cf9c733c9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.329Z [INFO] TestDNS_ExternalServiceLookup.server.serf.lan: serf: EventMemberJoin: Node-bdbb4f25-d716-2cfb-3a1d-6b5cf9c733c9 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.332Z [INFO] TestDNS_ExternalServiceLookup.server: Adding LAN server: server="Node-bdbb4f25-d716-2cfb-3a1d-6b5cf9c733c9 (Addr: tcp/127.0.0.1:30365) (DC: dc1)" writer.go:29: 2021-01-29T19:33:39.333Z [INFO] TestDNS_ExternalServiceLookup.server: Handled event for server in area: event=member-join server=Node-bdbb4f25-d716-2cfb-3a1d-6b5cf9c733c9.dc1 area=wan writer.go:29: 2021-01-29T19:33:39.334Z [INFO] TestDNS_ExternalServiceLookup: Started DNS server: address=127.0.0.1:30360 network=udp === CONT TestDNS_InifiniteRecursion writer.go:29: 2021-01-29T19:33:39.333Z [ERROR] TestDNS_InifiniteRecursion.dns: Infinite recursion detected for name, won't perform any CNAME resolution.: name=web.service.consul. === CONT TestDNS_ExternalServiceLookup writer.go:29: 2021-01-29T19:33:39.341Z [INFO] TestDNS_ExternalServiceLookup: Started DNS server: address=127.0.0.1:30360 network=tcp === CONT TestDNS_InifiniteRecursion writer.go:29: 2021-01-29T19:33:39.342Z [DEBUG] TestDNS_InifiniteRecursion.dns: request served from client: name=web.service.consul. type=A class=IN latency=12.198844ms client=127.0.0.1:54764 client_network=udp === CONT TestDNS_ExternalServiceLookup writer.go:29: 2021-01-29T19:33:39.344Z [INFO] TestDNS_ExternalServiceLookup: Started HTTP server: address=127.0.0.1:30361 network=tcp === CONT TestDNS_InifiniteRecursion writer.go:29: 2021-01-29T19:33:39.342Z [INFO] TestDNS_InifiniteRecursion: Requesting shutdown === CONT TestDNS_ExternalServiceLookup writer.go:29: 2021-01-29T19:33:39.345Z [INFO] TestDNS_ExternalServiceLookup: started state syncer === CONT TestDNS_InifiniteRecursion writer.go:29: 2021-01-29T19:33:39.345Z [INFO] TestDNS_InifiniteRecursion.server: shutting down server writer.go:29: 2021-01-29T19:33:39.347Z [DEBUG] TestDNS_InifiniteRecursion.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.348Z [ERROR] TestDNS_InifiniteRecursion.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:39.349Z [WARN] TestDNS_InifiniteRecursion.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:39.349Z [DEBUG] TestDNS_InifiniteRecursion.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.359Z [WARN] TestDNS_InifiniteRecursion.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:39.372Z [INFO] TestDNS_InifiniteRecursion.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:39.373Z [INFO] TestDNS_InifiniteRecursion: consul server down writer.go:29: 2021-01-29T19:33:39.375Z [INFO] TestDNS_InifiniteRecursion: shutdown complete writer.go:29: 2021-01-29T19:33:39.377Z [INFO] TestDNS_InifiniteRecursion: Stopping server: protocol=DNS address=127.0.0.1:30348 network=tcp writer.go:29: 2021-01-29T19:33:39.379Z [INFO] TestDNS_InifiniteRecursion: Stopping server: protocol=DNS address=127.0.0.1:30348 network=udp === CONT TestDNS_ExternalServiceLookup writer.go:29: 2021-01-29T19:33:39.381Z [WARN] TestDNS_ExternalServiceLookup.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:39.382Z [INFO] TestDNS_ExternalServiceLookup.server.raft: entering candidate state: node="Node at 127.0.0.1:30365 [Candidate]" term=2 === CONT TestDNS_InifiniteRecursion writer.go:29: 2021-01-29T19:33:39.381Z [INFO] TestDNS_InifiniteRecursion: Stopping server: protocol=HTTP address=127.0.0.1:30349 network=tcp === CONT TestDNS_ExternalServiceLookup writer.go:29: 2021-01-29T19:33:39.384Z [DEBUG] TestDNS_ExternalServiceLookup.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:39.385Z [DEBUG] TestDNS_ExternalServiceLookup.server.raft: vote granted: from=bdbb4f25-d716-2cfb-3a1d-6b5cf9c733c9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:39.386Z [INFO] TestDNS_ExternalServiceLookup.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:39.387Z [INFO] TestDNS_ExternalServiceLookup.server.raft: entering leader state: leader="Node at 127.0.0.1:30365 [Leader]" writer.go:29: 2021-01-29T19:33:39.389Z [INFO] TestDNS_ExternalServiceLookup.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:39.390Z [INFO] TestDNS_ExternalServiceLookup.server: New leader elected: payload=Node-bdbb4f25-d716-2cfb-3a1d-6b5cf9c733c9 writer.go:29: 2021-01-29T19:33:39.392Z [DEBUG] TestDNS_ExternalServiceLookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30365 writer.go:29: 2021-01-29T19:33:39.400Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:39.408Z [INFO] TestDNS_ExternalServiceLookup.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:39.410Z [INFO] TestDNS_ExternalServiceLookup.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.412Z [DEBUG] TestDNS_ExternalServiceLookup.server: Skipping self join check for node since the cluster is too small: node=Node-bdbb4f25-d716-2cfb-3a1d-6b5cf9c733c9 writer.go:29: 2021-01-29T19:33:39.413Z [INFO] TestDNS_ExternalServiceLookup.server: member joined, marking health alive: member=Node-bdbb4f25-d716-2cfb-3a1d-6b5cf9c733c9 writer.go:29: 2021-01-29T19:33:39.458Z [DEBUG] TestDNS_ExternalServiceLookup.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=97.218µs client=127.0.0.1:58377 client_network=udp writer.go:29: 2021-01-29T19:33:39.458Z [INFO] TestDNS_ExternalServiceLookup: Requesting shutdown writer.go:29: 2021-01-29T19:33:39.461Z [INFO] TestDNS_ExternalServiceLookup.server: shutting down server writer.go:29: 2021-01-29T19:33:39.463Z [DEBUG] TestDNS_ExternalServiceLookup.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.464Z [WARN] TestDNS_ExternalServiceLookup.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:39.466Z [ERROR] TestDNS_ExternalServiceLookup.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:39.467Z [INFO] TestDNS_NSRecords_IPV6: Waiting for endpoints to shut down === CONT TestDNS_ExternalServiceLookup writer.go:29: 2021-01-29T19:33:39.467Z [DEBUG] TestDNS_ExternalServiceLookup.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_NSRecords_IPV6 writer.go:29: 2021-01-29T19:33:39.469Z [INFO] TestDNS_NSRecords_IPV6: Endpoints down --- PASS: TestDNS_NSRecords_IPV6 (0.70s) === CONT TestDNS_ConnectServiceLookup === CONT TestDNS_ExternalServiceLookup writer.go:29: 2021-01-29T19:33:39.481Z [WARN] TestDNS_ExternalServiceLookup.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:39.483Z [INFO] TestDNS_ExternalServiceLookup.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:39.484Z [INFO] TestDNS_ExternalServiceLookup: consul server down writer.go:29: 2021-01-29T19:33:39.485Z [INFO] TestDNS_ExternalServiceLookup: shutdown complete writer.go:29: 2021-01-29T19:33:39.492Z [INFO] TestDNS_ExternalServiceLookup: Stopping server: protocol=DNS address=127.0.0.1:30360 network=tcp writer.go:29: 2021-01-29T19:33:39.494Z [INFO] TestDNS_ExternalServiceLookup: Stopping server: protocol=DNS address=127.0.0.1:30360 network=udp writer.go:29: 2021-01-29T19:33:39.497Z [INFO] TestDNS_ExternalServiceLookup: Stopping server: protocol=HTTP address=127.0.0.1:30361 network=tcp === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.596Z [WARN] TestDNS_ConnectServiceLookup: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:39.599Z [DEBUG] TestDNS_ConnectServiceLookup.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:39.606Z [DEBUG] TestDNS_ConnectServiceLookup.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:39.642Z [INFO] TestDNS_ConnectServiceLookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c7d333ec-3c53-4857-4b0e-ea87eda3ead9 Address:127.0.0.1:30377}]" writer.go:29: 2021-01-29T19:33:39.648Z [INFO] TestDNS_ConnectServiceLookup.server.serf.wan: serf: EventMemberJoin: Node-c7d333ec-3c53-4857-4b0e-ea87eda3ead9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.653Z [INFO] TestDNS_ConnectServiceLookup.server.raft: entering follower state: follower="Node at 127.0.0.1:30377 [Follower]" leader= writer.go:29: 2021-01-29T19:33:39.656Z [INFO] TestDNS_ConnectServiceLookup.server.serf.lan: serf: EventMemberJoin: Node-c7d333ec-3c53-4857-4b0e-ea87eda3ead9 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.664Z [INFO] TestDNS_ConnectServiceLookup.server: Adding LAN server: server="Node-c7d333ec-3c53-4857-4b0e-ea87eda3ead9 (Addr: tcp/127.0.0.1:30377) (DC: dc1)" writer.go:29: 2021-01-29T19:33:39.664Z [INFO] TestDNS_ConnectServiceLookup: Started DNS server: address=127.0.0.1:30372 network=udp writer.go:29: 2021-01-29T19:33:39.666Z [INFO] TestDNS_ConnectServiceLookup.server: Handled event for server in area: event=member-join server=Node-c7d333ec-3c53-4857-4b0e-ea87eda3ead9.dc1 area=wan writer.go:29: 2021-01-29T19:33:39.666Z [INFO] TestDNS_ConnectServiceLookup: Started DNS server: address=127.0.0.1:30372 network=tcp writer.go:29: 2021-01-29T19:33:39.670Z [INFO] TestDNS_ConnectServiceLookup: Started HTTP server: address=127.0.0.1:30373 network=tcp writer.go:29: 2021-01-29T19:33:39.674Z [INFO] TestDNS_ConnectServiceLookup: started state syncer writer.go:29: 2021-01-29T19:33:39.700Z [WARN] TestDNS_ConnectServiceLookup.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:39.704Z [INFO] TestDNS_ConnectServiceLookup.server.raft: entering candidate state: node="Node at 127.0.0.1:30377 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:39.725Z [DEBUG] TestDNS_ConnectServiceLookup.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:39.727Z [DEBUG] TestDNS_ConnectServiceLookup.server.raft: vote granted: from=c7d333ec-3c53-4857-4b0e-ea87eda3ead9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:39.729Z [INFO] TestDNS_ConnectServiceLookup.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:39.731Z [INFO] TestDNS_ConnectServiceLookup.server.raft: entering leader state: leader="Node at 127.0.0.1:30377 [Leader]" writer.go:29: 2021-01-29T19:33:39.736Z [INFO] TestDNS_ConnectServiceLookup.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:39.738Z [INFO] TestDNS_ConnectServiceLookup.server: New leader elected: payload=Node-c7d333ec-3c53-4857-4b0e-ea87eda3ead9 writer.go:29: 2021-01-29T19:33:39.750Z [DEBUG] TestDNS_ConnectServiceLookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30377 === CONT TestDNS_ExternalServiceToConsulCNAMELookup writer.go:29: 2021-01-29T19:33:39.771Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:39.774Z [INFO] TestDNS_ExternalServiceToConsulCNAMELookup: Endpoints down --- PASS: TestDNS_ExternalServiceToConsulCNAMELookup (0.98s) === CONT TestDNS_ServiceLookupWithInternalServiceAddress === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.823Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:39.835Z [INFO] TestDNS_ConnectServiceLookup.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:39.838Z [INFO] TestDNS_ConnectServiceLookup.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.847Z [DEBUG] TestDNS_ConnectServiceLookup.server: Skipping self join check for node since the cluster is too small: node=Node-c7d333ec-3c53-4857-4b0e-ea87eda3ead9 writer.go:29: 2021-01-29T19:33:39.850Z [INFO] TestDNS_ConnectServiceLookup.server: member joined, marking health alive: member=Node-c7d333ec-3c53-4857-4b0e-ea87eda3ead9 === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:39.865Z [WARN] TestDNS_ServiceLookupWithInternalServiceAddress: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:39.877Z [WARN] TestDNS_ServiceLookupWithInternalServiceAddress: Node name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: node_name=my.test-node writer.go:29: 2021-01-29T19:33:39.881Z [DEBUG] TestDNS_ServiceLookupWithInternalServiceAddress.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:39.884Z [DEBUG] TestDNS_ServiceLookupWithInternalServiceAddress.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_InifiniteRecursion writer.go:29: 2021-01-29T19:33:39.887Z [INFO] TestDNS_InifiniteRecursion: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:39.897Z [INFO] TestDNS_InifiniteRecursion: Endpoints down --- PASS: TestDNS_InifiniteRecursion (0.84s) === CONT TestDNS_ServiceLookup === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.890Z [DEBUG] TestDNS_ConnectServiceLookup.dns: request served from client: name=db.connect.consul. type=SRV class=IN latency=115.485µs client=127.0.0.1:33331 client_network=udp === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:39.904Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:951fe3a0-d231-c875-43b2-f58c75d8b910 Address:127.0.0.1:30383}]" writer.go:29: 2021-01-29T19:33:39.909Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server.raft: entering follower state: follower="Node at 127.0.0.1:30383 [Follower]" leader= writer.go:29: 2021-01-29T19:33:39.918Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server.serf.wan: serf: EventMemberJoin: my.test-node.dc1 127.0.0.1 === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.933Z [INFO] TestDNS_ConnectServiceLookup: Requesting shutdown writer.go:29: 2021-01-29T19:33:39.934Z [INFO] TestDNS_ConnectServiceLookup.server: shutting down server === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:39.933Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server.serf.lan: serf: EventMemberJoin: my.test-node 127.0.0.1 === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.935Z [DEBUG] TestDNS_ConnectServiceLookup.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:39.937Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server: Adding LAN server: server="my.test-node (Addr: tcp/127.0.0.1:30383) (DC: dc1)" === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.937Z [WARN] TestDNS_ConnectServiceLookup.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:39.938Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server: Handled event for server in area: event=member-join server=my.test-node.dc1 area=wan === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.936Z [ERROR] TestDNS_ConnectServiceLookup.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:39.940Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: Started DNS server: address=127.0.0.1:30378 network=tcp writer.go:29: 2021-01-29T19:33:39.943Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: Started DNS server: address=127.0.0.1:30378 network=udp === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.937Z [DEBUG] TestDNS_ConnectServiceLookup.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:39.945Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: Started HTTP server: address=127.0.0.1:30379 network=tcp writer.go:29: 2021-01-29T19:33:39.946Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: started state syncer === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.945Z [WARN] TestDNS_ConnectServiceLookup.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:39.949Z [WARN] TestDNS_ServiceLookup: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:39.950Z [DEBUG] TestDNS_ServiceLookup.tlsutil: Update: version=1 === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.957Z [INFO] TestDNS_ConnectServiceLookup.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:39.959Z [INFO] TestDNS_ConnectServiceLookup: consul server down writer.go:29: 2021-01-29T19:33:39.961Z [INFO] TestDNS_ConnectServiceLookup: shutdown complete writer.go:29: 2021-01-29T19:33:39.973Z [INFO] TestDNS_ConnectServiceLookup: Stopping server: protocol=DNS address=127.0.0.1:30372 network=tcp === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:39.959Z [DEBUG] TestDNS_ServiceLookup.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:39.975Z [INFO] TestDNS_ConnectServiceLookup: Stopping server: protocol=DNS address=127.0.0.1:30372 network=udp writer.go:29: 2021-01-29T19:33:39.978Z [INFO] TestDNS_ConnectServiceLookup: Stopping server: protocol=HTTP address=127.0.0.1:30373 network=tcp === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:39.982Z [INFO] TestDNS_ServiceLookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cfa9ee81-a9f3-3e1b-20d0-9f35c62b07bb Address:127.0.0.1:30389}]" === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:39.983Z [WARN] TestDNS_ServiceLookupWithInternalServiceAddress.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:39.985Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server.raft: entering candidate state: node="Node at 127.0.0.1:30383 [Candidate]" term=2 === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:39.984Z [INFO] TestDNS_ServiceLookup.server.raft: entering follower state: follower="Node at 127.0.0.1:30389 [Follower]" leader= writer.go:29: 2021-01-29T19:33:39.988Z [INFO] TestDNS_ServiceLookup.server.serf.wan: serf: EventMemberJoin: Node-cfa9ee81-a9f3-3e1b-20d0-9f35c62b07bb.dc1 127.0.0.1 === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:39.988Z [DEBUG] TestDNS_ServiceLookupWithInternalServiceAddress.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:39.991Z [INFO] TestDNS_ServiceLookup.server.serf.lan: serf: EventMemberJoin: Node-cfa9ee81-a9f3-3e1b-20d0-9f35c62b07bb 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.994Z [INFO] TestDNS_ServiceLookup: Started DNS server: address=127.0.0.1:30384 network=udp writer.go:29: 2021-01-29T19:33:39.995Z [INFO] TestDNS_ServiceLookup.server: Adding LAN server: server="Node-cfa9ee81-a9f3-3e1b-20d0-9f35c62b07bb (Addr: tcp/127.0.0.1:30389) (DC: dc1)" writer.go:29: 2021-01-29T19:33:39.996Z [INFO] TestDNS_ServiceLookup.server: Handled event for server in area: event=member-join server=Node-cfa9ee81-a9f3-3e1b-20d0-9f35c62b07bb.dc1 area=wan writer.go:29: 2021-01-29T19:33:39.998Z [INFO] TestDNS_ServiceLookup: Started DNS server: address=127.0.0.1:30384 network=tcp writer.go:29: 2021-01-29T19:33:40.000Z [INFO] TestDNS_ServiceLookup: Started HTTP server: address=127.0.0.1:30385 network=tcp writer.go:29: 2021-01-29T19:33:40.000Z [INFO] TestDNS_ServiceLookup: started state syncer === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:39.989Z [DEBUG] TestDNS_ServiceLookupWithInternalServiceAddress.server.raft: vote granted: from=951fe3a0-d231-c875-43b2-f58c75d8b910 term=2 tally=1 === CONT TestDNS_ExternalServiceLookup writer.go:29: 2021-01-29T19:33:40.002Z [INFO] TestDNS_ExternalServiceLookup: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:40.002Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server.raft: election won: tally=1 === CONT TestDNS_ExternalServiceLookup writer.go:29: 2021-01-29T19:33:40.003Z [INFO] TestDNS_ExternalServiceLookup: Endpoints down --- PASS: TestDNS_ExternalServiceLookup (0.72s) === CONT TestDNS_ServiceLookupMultiAddrNoCNAME === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:40.003Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server.raft: entering leader state: leader="Node at 127.0.0.1:30383 [Leader]" writer.go:29: 2021-01-29T19:33:40.004Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:40.005Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server: New leader elected: payload=my.test-node === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.012Z [WARN] TestDNS_ServiceLookupMultiAddrNoCNAME: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:40.013Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:40.005Z [DEBUG] TestDNS_ServiceLookupWithInternalServiceAddress.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30383 === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.014Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:40.020Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a7d06913-3033-d4e2-4ae7-f8920653eed8 Address:127.0.0.1:30395}]" === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:40.018Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.022Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server.raft: entering follower state: follower="Node at 127.0.0.1:30395 [Follower]" leader= writer.go:29: 2021-01-29T19:33:40.022Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server.serf.wan: serf: EventMemberJoin: Node-a7d06913-3033-d4e2-4ae7-f8920653eed8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.027Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server.serf.lan: serf: EventMemberJoin: Node-a7d06913-3033-d4e2-4ae7-f8920653eed8 127.0.0.1 === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:40.027Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:40.029Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.030Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: Started DNS server: address=127.0.0.1:30390 network=udp writer.go:29: 2021-01-29T19:33:40.031Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server: Adding LAN server: server="Node-a7d06913-3033-d4e2-4ae7-f8920653eed8 (Addr: tcp/127.0.0.1:30395) (DC: dc1)" writer.go:29: 2021-01-29T19:33:40.031Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server: Handled event for server in area: event=member-join server=Node-a7d06913-3033-d4e2-4ae7-f8920653eed8.dc1 area=wan === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:40.030Z [DEBUG] TestDNS_ServiceLookupWithInternalServiceAddress.server: Skipping self join check for node since the cluster is too small: node=my.test-node === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.031Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: Started DNS server: address=127.0.0.1:30390 network=tcp === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:40.032Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server: member joined, marking health alive: member=my.test-node === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.033Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: Started HTTP server: address=127.0.0.1:30391 network=tcp writer.go:29: 2021-01-29T19:33:40.035Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: started state syncer === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.044Z [WARN] TestDNS_ServiceLookup.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:40.051Z [INFO] TestDNS_ServiceLookup.server.raft: entering candidate state: node="Node at 127.0.0.1:30389 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:40.054Z [DEBUG] TestDNS_ServiceLookup.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:40.055Z [DEBUG] TestDNS_ServiceLookup.server.raft: vote granted: from=cfa9ee81-a9f3-3e1b-20d0-9f35c62b07bb term=2 tally=1 writer.go:29: 2021-01-29T19:33:40.057Z [INFO] TestDNS_ServiceLookup.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:40.058Z [INFO] TestDNS_ServiceLookup.server.raft: entering leader state: leader="Node at 127.0.0.1:30389 [Leader]" writer.go:29: 2021-01-29T19:33:40.060Z [INFO] TestDNS_ServiceLookup.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:40.062Z [INFO] TestDNS_ServiceLookup.server: New leader elected: payload=Node-cfa9ee81-a9f3-3e1b-20d0-9f35c62b07bb writer.go:29: 2021-01-29T19:33:40.063Z [DEBUG] TestDNS_ServiceLookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30389 writer.go:29: 2021-01-29T19:33:40.072Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.085Z [WARN] TestDNS_ServiceLookupMultiAddrNoCNAME.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:40.087Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server.raft: entering candidate state: node="Node at 127.0.0.1:30395 [Candidate]" term=2 === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.083Z [INFO] TestDNS_ServiceLookup.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.099Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.100Z [INFO] TestDNS_ServiceLookup.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.101Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME.server.raft: vote granted: from=a7d06913-3033-d4e2-4ae7-f8920653eed8 term=2 tally=1 === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.102Z [DEBUG] TestDNS_ServiceLookup.server: Skipping self join check for node since the cluster is too small: node=Node-cfa9ee81-a9f3-3e1b-20d0-9f35c62b07bb === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.103Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.104Z [INFO] TestDNS_ServiceLookup.server: member joined, marking health alive: member=Node-cfa9ee81-a9f3-3e1b-20d0-9f35c62b07bb === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.105Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server.raft: entering leader state: leader="Node at 127.0.0.1:30395 [Leader]" writer.go:29: 2021-01-29T19:33:40.107Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:40.107Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server: New leader elected: payload=Node-a7d06913-3033-d4e2-4ae7-f8920653eed8 writer.go:29: 2021-01-29T19:33:40.109Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30395 writer.go:29: 2021-01-29T19:33:40.119Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:40.130Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:40.131Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.133Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME.server: Skipping self join check for node since the cluster is too small: node=Node-a7d06913-3033-d4e2-4ae7-f8920653eed8 writer.go:29: 2021-01-29T19:33:40.135Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server: member joined, marking health alive: member=Node-a7d06913-3033-d4e2-4ae7-f8920653eed8 writer.go:29: 2021-01-29T19:33:40.261Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:40.265Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: Synced node info writer.go:29: 2021-01-29T19:33:40.266Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME: Node info in sync writer.go:29: 2021-01-29T19:33:40.290Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME.dns: request served from client: name=db.service.consul. type=ANY class=IN latency=164.245µs client=127.0.0.1:51880 client_network=udp writer.go:29: 2021-01-29T19:33:40.295Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: Requesting shutdown writer.go:29: 2021-01-29T19:33:40.300Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server: shutting down server writer.go:29: 2021-01-29T19:33:40.302Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.303Z [WARN] TestDNS_ServiceLookupMultiAddrNoCNAME.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.303Z [DEBUG] TestDNS_ServiceLookupMultiAddrNoCNAME.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.306Z [WARN] TestDNS_ServiceLookupMultiAddrNoCNAME.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.305Z [DEBUG] TestDNS_ServiceLookup.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=389.491µs client=127.0.0.1:55789 client_network=udp === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.311Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME.server.router.manager: shutting down === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.307Z [DEBUG] TestDNS_ServiceLookup.dns: request served from client: name=3065efcc-cf13-47bb-a572-7197024ef0b9.query.consul. type=SRV class=IN latency=73.906µs client=127.0.0.1:46417 client_network=udp writer.go:29: 2021-01-29T19:33:40.308Z [DEBUG] TestDNS_ServiceLookup.dns: request served from client: name=nodb.service.consul. type=SRV class=IN latency=66.671µs client=127.0.0.1:46460 client_network=udp === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.312Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: consul server down === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.309Z [DEBUG] TestDNS_ServiceLookup.dns: request served from client: name=nope.query.consul. type=SRV class=IN latency=36.486µs client=127.0.0.1:57697 client_network=udp === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.316Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: shutdown complete === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.309Z [INFO] TestDNS_ServiceLookup: Requesting shutdown === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.318Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: Stopping server: protocol=DNS address=127.0.0.1:30390 network=tcp writer.go:29: 2021-01-29T19:33:40.320Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: Stopping server: protocol=DNS address=127.0.0.1:30390 network=udp writer.go:29: 2021-01-29T19:33:40.321Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: Stopping server: protocol=HTTP address=127.0.0.1:30391 network=tcp === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.319Z [INFO] TestDNS_ServiceLookup.server: shutting down server writer.go:29: 2021-01-29T19:33:40.327Z [DEBUG] TestDNS_ServiceLookup.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.329Z [WARN] TestDNS_ServiceLookup.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.327Z [ERROR] TestDNS_ServiceLookup.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:40.329Z [DEBUG] TestDNS_ServiceLookup.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.331Z [WARN] TestDNS_ServiceLookup.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.335Z [INFO] TestDNS_ServiceLookup.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:40.336Z [INFO] TestDNS_ServiceLookup: consul server down writer.go:29: 2021-01-29T19:33:40.339Z [INFO] TestDNS_ServiceLookup: shutdown complete writer.go:29: 2021-01-29T19:33:40.340Z [INFO] TestDNS_ServiceLookup: Stopping server: protocol=DNS address=127.0.0.1:30384 network=tcp writer.go:29: 2021-01-29T19:33:40.342Z [INFO] TestDNS_ServiceLookup: Stopping server: protocol=DNS address=127.0.0.1:30384 network=udp writer.go:29: 2021-01-29T19:33:40.344Z [INFO] TestDNS_ServiceLookup: Stopping server: protocol=HTTP address=127.0.0.1:30385 network=tcp === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:40.385Z [DEBUG] TestDNS_ServiceLookupWithInternalServiceAddress.dns: request served from client: name=db.service.consul. type=SRV class=IN latency=127.82µs client=127.0.0.1:50358 client_network=udp writer.go:29: 2021-01-29T19:33:40.385Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: Requesting shutdown writer.go:29: 2021-01-29T19:33:40.389Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server: shutting down server writer.go:29: 2021-01-29T19:33:40.391Z [DEBUG] TestDNS_ServiceLookupWithInternalServiceAddress.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.393Z [WARN] TestDNS_ServiceLookupWithInternalServiceAddress.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.392Z [ERROR] TestDNS_ServiceLookupWithInternalServiceAddress.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:40.393Z [DEBUG] TestDNS_ServiceLookupWithInternalServiceAddress.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.408Z [WARN] TestDNS_ServiceLookupWithInternalServiceAddress.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.429Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:40.430Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: consul server down writer.go:29: 2021-01-29T19:33:40.435Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: shutdown complete writer.go:29: 2021-01-29T19:33:40.436Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: Stopping server: protocol=DNS address=127.0.0.1:30378 network=tcp writer.go:29: 2021-01-29T19:33:40.438Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: Stopping server: protocol=DNS address=127.0.0.1:30378 network=udp writer.go:29: 2021-01-29T19:33:40.440Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: Stopping server: protocol=HTTP address=127.0.0.1:30379 network=tcp === CONT TestDNS_ConnectServiceLookup writer.go:29: 2021-01-29T19:33:40.480Z [INFO] TestDNS_ConnectServiceLookup: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:40.482Z [INFO] TestDNS_ConnectServiceLookup: Endpoints down --- PASS: TestDNS_ConnectServiceLookup (1.01s) === CONT TestDNS_ServiceLookupPreferNoCNAME writer.go:29: 2021-01-29T19:33:40.504Z [WARN] TestDNS_ServiceLookupPreferNoCNAME: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:40.506Z [DEBUG] TestDNS_ServiceLookupPreferNoCNAME.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:40.516Z [DEBUG] TestDNS_ServiceLookupPreferNoCNAME.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:40.535Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:de5f1734-5615-c2f2-95d7-191f92f1de55 Address:127.0.0.1:30401}]" writer.go:29: 2021-01-29T19:33:40.538Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server.raft: entering follower state: follower="Node at 127.0.0.1:30401 [Follower]" leader= writer.go:29: 2021-01-29T19:33:40.539Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server.serf.wan: serf: EventMemberJoin: Node-de5f1734-5615-c2f2-95d7-191f92f1de55.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.544Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server.serf.lan: serf: EventMemberJoin: Node-de5f1734-5615-c2f2-95d7-191f92f1de55 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.550Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: Started DNS server: address=127.0.0.1:30396 network=udp writer.go:29: 2021-01-29T19:33:40.553Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: Started DNS server: address=127.0.0.1:30396 network=tcp writer.go:29: 2021-01-29T19:33:40.551Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server: Adding LAN server: server="Node-de5f1734-5615-c2f2-95d7-191f92f1de55 (Addr: tcp/127.0.0.1:30401) (DC: dc1)" writer.go:29: 2021-01-29T19:33:40.552Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server: Handled event for server in area: event=member-join server=Node-de5f1734-5615-c2f2-95d7-191f92f1de55.dc1 area=wan writer.go:29: 2021-01-29T19:33:40.556Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: Started HTTP server: address=127.0.0.1:30397 network=tcp writer.go:29: 2021-01-29T19:33:40.562Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: started state syncer writer.go:29: 2021-01-29T19:33:40.583Z [WARN] TestDNS_ServiceLookupPreferNoCNAME.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:40.585Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server.raft: entering candidate state: node="Node at 127.0.0.1:30401 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:40.591Z [DEBUG] TestDNS_ServiceLookupPreferNoCNAME.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:40.593Z [DEBUG] TestDNS_ServiceLookupPreferNoCNAME.server.raft: vote granted: from=de5f1734-5615-c2f2-95d7-191f92f1de55 term=2 tally=1 writer.go:29: 2021-01-29T19:33:40.597Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:40.601Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server.raft: entering leader state: leader="Node at 127.0.0.1:30401 [Leader]" writer.go:29: 2021-01-29T19:33:40.606Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:40.607Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server: New leader elected: payload=Node-de5f1734-5615-c2f2-95d7-191f92f1de55 writer.go:29: 2021-01-29T19:33:40.612Z [DEBUG] TestDNS_ServiceLookupPreferNoCNAME.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30401 writer.go:29: 2021-01-29T19:33:40.629Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:40.644Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:40.646Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.648Z [DEBUG] TestDNS_ServiceLookupPreferNoCNAME.server: Skipping self join check for node since the cluster is too small: node=Node-de5f1734-5615-c2f2-95d7-191f92f1de55 writer.go:29: 2021-01-29T19:33:40.649Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server: member joined, marking health alive: member=Node-de5f1734-5615-c2f2-95d7-191f92f1de55 writer.go:29: 2021-01-29T19:33:40.668Z [DEBUG] TestDNS_ServiceLookupPreferNoCNAME.dns: request served from client: name=db.service.consul. type=ANY class=IN latency=168.195µs client=127.0.0.1:55884 client_network=udp writer.go:29: 2021-01-29T19:33:40.671Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: Requesting shutdown writer.go:29: 2021-01-29T19:33:40.672Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server: shutting down server writer.go:29: 2021-01-29T19:33:40.673Z [DEBUG] TestDNS_ServiceLookupPreferNoCNAME.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.674Z [WARN] TestDNS_ServiceLookupPreferNoCNAME.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.675Z [ERROR] TestDNS_ServiceLookupPreferNoCNAME.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:40.676Z [DEBUG] TestDNS_ServiceLookupPreferNoCNAME.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.676Z [WARN] TestDNS_ServiceLookupPreferNoCNAME.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.679Z [INFO] TestDNS_ServiceLookupPreferNoCNAME.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:40.681Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: consul server down writer.go:29: 2021-01-29T19:33:40.683Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: shutdown complete writer.go:29: 2021-01-29T19:33:40.685Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: Stopping server: protocol=DNS address=127.0.0.1:30396 network=tcp writer.go:29: 2021-01-29T19:33:40.687Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: Stopping server: protocol=DNS address=127.0.0.1:30396 network=udp writer.go:29: 2021-01-29T19:33:40.689Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: Stopping server: protocol=HTTP address=127.0.0.1:30397 network=tcp === CONT TestDNS_ServiceLookupMultiAddrNoCNAME writer.go:29: 2021-01-29T19:33:40.823Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:40.825Z [INFO] TestDNS_ServiceLookupMultiAddrNoCNAME: Endpoints down --- PASS: TestDNS_ServiceLookupMultiAddrNoCNAME (0.82s) === CONT TestDNS_ServiceReverseLookupNodeAddress === CONT TestDNS_ServiceLookup writer.go:29: 2021-01-29T19:33:40.846Z [INFO] TestDNS_ServiceLookup: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:40.847Z [INFO] TestDNS_ServiceLookup: Endpoints down --- PASS: TestDNS_ServiceLookup (0.95s) === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:40.867Z [WARN] TestDNS_SOA_Settings: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:40.868Z [DEBUG] TestDNS_SOA_Settings.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:40.870Z [DEBUG] TestDNS_SOA_Settings.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:40.874Z [INFO] TestDNS_SOA_Settings.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:81f9e480-2ee0-6468-b38d-687d08961103 Address:127.0.0.1:30413}]" writer.go:29: 2021-01-29T19:33:40.876Z [INFO] TestDNS_SOA_Settings.server.raft: entering follower state: follower="Node at 127.0.0.1:30413 [Follower]" leader= writer.go:29: 2021-01-29T19:33:40.876Z [INFO] TestDNS_SOA_Settings.server.serf.wan: serf: EventMemberJoin: Node-81f9e480-2ee0-6468-b38d-687d08961103.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.883Z [INFO] TestDNS_SOA_Settings.server.serf.lan: serf: EventMemberJoin: Node-81f9e480-2ee0-6468-b38d-687d08961103 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.890Z [INFO] TestDNS_SOA_Settings.server: Adding LAN server: server="Node-81f9e480-2ee0-6468-b38d-687d08961103 (Addr: tcp/127.0.0.1:30413) (DC: dc1)" writer.go:29: 2021-01-29T19:33:40.890Z [INFO] TestDNS_SOA_Settings.server: Handled event for server in area: event=member-join server=Node-81f9e480-2ee0-6468-b38d-687d08961103.dc1 area=wan writer.go:29: 2021-01-29T19:33:40.893Z [INFO] TestDNS_SOA_Settings: Started DNS server: address=127.0.0.1:30408 network=udp writer.go:29: 2021-01-29T19:33:40.894Z [INFO] TestDNS_SOA_Settings: Started DNS server: address=127.0.0.1:30408 network=tcp writer.go:29: 2021-01-29T19:33:40.896Z [INFO] TestDNS_SOA_Settings: Started HTTP server: address=127.0.0.1:30409 network=tcp writer.go:29: 2021-01-29T19:33:40.897Z [INFO] TestDNS_SOA_Settings: started state syncer === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:40.896Z [WARN] TestDNS_ServiceReverseLookupNodeAddress: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:40.898Z [DEBUG] TestDNS_ServiceReverseLookupNodeAddress.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:40.902Z [DEBUG] TestDNS_ServiceReverseLookupNodeAddress.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:40.911Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6de89fcd-4a19-a271-5928-034bf051ebef Address:127.0.0.1:30407}]" writer.go:29: 2021-01-29T19:33:40.913Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server.raft: entering follower state: follower="Node at 127.0.0.1:30407 [Follower]" leader= writer.go:29: 2021-01-29T19:33:40.914Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server.serf.wan: serf: EventMemberJoin: Node-6de89fcd-4a19-a271-5928-034bf051ebef.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.917Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server.serf.lan: serf: EventMemberJoin: Node-6de89fcd-4a19-a271-5928-034bf051ebef 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.919Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: Started DNS server: address=127.0.0.1:30402 network=udp writer.go:29: 2021-01-29T19:33:40.921Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server: Adding LAN server: server="Node-6de89fcd-4a19-a271-5928-034bf051ebef (Addr: tcp/127.0.0.1:30407) (DC: dc1)" writer.go:29: 2021-01-29T19:33:40.922Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server: Handled event for server in area: event=member-join server=Node-6de89fcd-4a19-a271-5928-034bf051ebef.dc1 area=wan writer.go:29: 2021-01-29T19:33:40.922Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: Started DNS server: address=127.0.0.1:30402 network=tcp === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:40.925Z [WARN] TestDNS_SOA_Settings.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:40.924Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: Started HTTP server: address=127.0.0.1:30403 network=tcp === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:40.925Z [INFO] TestDNS_SOA_Settings.server.raft: entering candidate state: node="Node at 127.0.0.1:30413 [Candidate]" term=2 === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:40.926Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: started state syncer === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:40.931Z [DEBUG] TestDNS_SOA_Settings.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:40.932Z [DEBUG] TestDNS_SOA_Settings.server.raft: vote granted: from=81f9e480-2ee0-6468-b38d-687d08961103 term=2 tally=1 writer.go:29: 2021-01-29T19:33:40.932Z [INFO] TestDNS_SOA_Settings.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:40.933Z [INFO] TestDNS_SOA_Settings.server.raft: entering leader state: leader="Node at 127.0.0.1:30413 [Leader]" writer.go:29: 2021-01-29T19:33:40.934Z [INFO] TestDNS_SOA_Settings.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:40.934Z [INFO] TestDNS_SOA_Settings.server: New leader elected: payload=Node-81f9e480-2ee0-6468-b38d-687d08961103 writer.go:29: 2021-01-29T19:33:40.939Z [DEBUG] TestDNS_SOA_Settings.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30413 === CONT TestDNS_ServiceLookupWithInternalServiceAddress writer.go:29: 2021-01-29T19:33:40.943Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:40.944Z [INFO] TestDNS_ServiceLookupWithInternalServiceAddress: Endpoints down --- PASS: TestDNS_ServiceLookupWithInternalServiceAddress (1.17s) === CONT TestDNS_ServiceReverseLookup_CustomDomain === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:40.947Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:40.958Z [WARN] TestDNS_ServiceReverseLookup_CustomDomain: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:40.960Z [DEBUG] TestDNS_ServiceReverseLookup_CustomDomain.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:40.962Z [DEBUG] TestDNS_ServiceReverseLookup_CustomDomain.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:40.966Z [INFO] TestDNS_SOA_Settings.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:40.967Z [INFO] TestDNS_SOA_Settings.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:40.966Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:abe76c84-a4e3-b962-f379-f3b187c15c3b Address:127.0.0.1:30419}]" === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:40.967Z [DEBUG] TestDNS_SOA_Settings.server: Skipping self join check for node since the cluster is too small: node=Node-81f9e480-2ee0-6468-b38d-687d08961103 writer.go:29: 2021-01-29T19:33:40.968Z [INFO] TestDNS_SOA_Settings.server: member joined, marking health alive: member=Node-81f9e480-2ee0-6468-b38d-687d08961103 === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:40.969Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server.raft: entering follower state: follower="Node at 127.0.0.1:30419 [Follower]" leader= === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:40.980Z [WARN] TestDNS_ServiceReverseLookupNodeAddress.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:40.981Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server.raft: entering candidate state: node="Node at 127.0.0.1:30407 [Candidate]" term=2 === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:40.971Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server.serf.wan: serf: EventMemberJoin: Node-abe76c84-a4e3-b962-f379-f3b187c15c3b.dc1 127.0.0.1 === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:40.987Z [DEBUG] TestDNS_ServiceReverseLookupNodeAddress.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:40.988Z [DEBUG] TestDNS_ServiceReverseLookupNodeAddress.server.raft: vote granted: from=6de89fcd-4a19-a271-5928-034bf051ebef term=2 tally=1 writer.go:29: 2021-01-29T19:33:40.990Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server.raft: election won: tally=1 === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:40.989Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server.serf.lan: serf: EventMemberJoin: Node-abe76c84-a4e3-b962-f379-f3b187c15c3b 127.0.0.1 === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:40.991Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server.raft: entering leader state: leader="Node at 127.0.0.1:30407 [Leader]" === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:40.992Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server: Adding LAN server: server="Node-abe76c84-a4e3-b962-f379-f3b187c15c3b (Addr: tcp/127.0.0.1:30419) (DC: dc1)" writer.go:29: 2021-01-29T19:33:40.993Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server: Handled event for server in area: event=member-join server=Node-abe76c84-a4e3-b962-f379-f3b187c15c3b.dc1 area=wan === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:40.992Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server: cluster leadership acquired === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:40.993Z [DEBUG] TestDNS_SOA_Settings: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:40.995Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: Started DNS server: address=127.0.0.1:30414 network=tcp === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:40.995Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server: New leader elected: payload=Node-6de89fcd-4a19-a271-5928-034bf051ebef === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:40.996Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: Started DNS server: address=127.0.0.1:30414 network=udp writer.go:29: 2021-01-29T19:33:40.998Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: Started HTTP server: address=127.0.0.1:30415 network=tcp writer.go:29: 2021-01-29T19:33:41.000Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: started state syncer === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.004Z [INFO] TestDNS_SOA_Settings: Synced node info === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:41.004Z [DEBUG] TestDNS_ServiceReverseLookupNodeAddress.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30407 writer.go:29: 2021-01-29T19:33:41.011Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.013Z [DEBUG] TestDNS_SOA_Settings.dns: request served from client: name=nofoo.node.dc1.consul. type=ANY class=IN latency=104.477µs client=127.0.0.1:36163 client_network=udp writer.go:29: 2021-01-29T19:33:41.018Z [INFO] TestDNS_SOA_Settings: Requesting shutdown writer.go:29: 2021-01-29T19:33:41.020Z [INFO] TestDNS_SOA_Settings.server: shutting down server writer.go:29: 2021-01-29T19:33:41.021Z [DEBUG] TestDNS_SOA_Settings.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:41.023Z [WARN] TestDNS_SOA_Settings.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:41.023Z [DEBUG] TestDNS_SOA_Settings.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:41.028Z [WARN] TestDNS_SOA_Settings.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.029Z [WARN] TestDNS_ServiceReverseLookup_CustomDomain.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:41.029Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.030Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server.raft: entering candidate state: node="Node at 127.0.0.1:30419 [Candidate]" term=2 === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:41.031Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:41.032Z [DEBUG] TestDNS_ServiceReverseLookupNodeAddress.server: Skipping self join check for node since the cluster is too small: node=Node-6de89fcd-4a19-a271-5928-034bf051ebef writer.go:29: 2021-01-29T19:33:41.034Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server: member joined, marking health alive: member=Node-6de89fcd-4a19-a271-5928-034bf051ebef === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.037Z [INFO] TestDNS_SOA_Settings.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:41.037Z [INFO] TestDNS_SOA_Settings: consul server down writer.go:29: 2021-01-29T19:33:41.039Z [INFO] TestDNS_SOA_Settings: shutdown complete === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.040Z [DEBUG] TestDNS_ServiceReverseLookup_CustomDomain.server.raft: votes: needed=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.040Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=DNS address=127.0.0.1:30408 network=tcp writer.go:29: 2021-01-29T19:33:41.041Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=DNS address=127.0.0.1:30408 network=udp writer.go:29: 2021-01-29T19:33:41.042Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=HTTP address=127.0.0.1:30409 network=tcp === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.041Z [DEBUG] TestDNS_ServiceReverseLookup_CustomDomain.server.raft: vote granted: from=abe76c84-a4e3-b962-f379-f3b187c15c3b term=2 tally=1 writer.go:29: 2021-01-29T19:33:41.043Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:41.044Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server.raft: entering leader state: leader="Node at 127.0.0.1:30419 [Leader]" writer.go:29: 2021-01-29T19:33:41.045Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:41.046Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server: New leader elected: payload=Node-abe76c84-a4e3-b962-f379-f3b187c15c3b writer.go:29: 2021-01-29T19:33:41.048Z [DEBUG] TestDNS_ServiceReverseLookup_CustomDomain.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30419 writer.go:29: 2021-01-29T19:33:41.062Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:41.070Z [DEBUG] TestDNS_ServiceReverseLookupNodeAddress.dns: request served from client: question="{1.0.0.127.in-addr.arpa. 255 1}" latency=100.549µs client=127.0.0.1:34473 client_network=udp writer.go:29: 2021-01-29T19:33:41.070Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: Requesting shutdown === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.072Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:41.073Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server: shutting down server === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.074Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:41.075Z [DEBUG] TestDNS_ServiceReverseLookupNodeAddress.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:41.075Z [ERROR] TestDNS_ServiceReverseLookupNodeAddress.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:41.076Z [WARN] TestDNS_ServiceReverseLookupNodeAddress.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.075Z [DEBUG] TestDNS_ServiceReverseLookup_CustomDomain.server: Skipping self join check for node since the cluster is too small: node=Node-abe76c84-a4e3-b962-f379-f3b187c15c3b === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:41.076Z [DEBUG] TestDNS_ServiceReverseLookupNodeAddress.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.080Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server: member joined, marking health alive: member=Node-abe76c84-a4e3-b962-f379-f3b187c15c3b === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:41.079Z [WARN] TestDNS_ServiceReverseLookupNodeAddress.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:41.085Z [INFO] TestDNS_ServiceReverseLookupNodeAddress.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:41.086Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: consul server down writer.go:29: 2021-01-29T19:33:41.089Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: shutdown complete writer.go:29: 2021-01-29T19:33:41.090Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: Stopping server: protocol=DNS address=127.0.0.1:30402 network=tcp writer.go:29: 2021-01-29T19:33:41.091Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: Stopping server: protocol=DNS address=127.0.0.1:30402 network=udp writer.go:29: 2021-01-29T19:33:41.092Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: Stopping server: protocol=HTTP address=127.0.0.1:30403 network=tcp === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.110Z [DEBUG] TestDNS_ServiceReverseLookup_CustomDomain.dns: request served from client: question="{2.0.0.127.in-addr.arpa. 255 1}" latency=136.584µs client=127.0.0.1:48250 client_network=udp writer.go:29: 2021-01-29T19:33:41.110Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: Requesting shutdown writer.go:29: 2021-01-29T19:33:41.111Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server: shutting down server writer.go:29: 2021-01-29T19:33:41.112Z [DEBUG] TestDNS_ServiceReverseLookup_CustomDomain.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:41.113Z [WARN] TestDNS_ServiceReverseLookup_CustomDomain.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:41.113Z [DEBUG] TestDNS_ServiceReverseLookup_CustomDomain.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:41.112Z [ERROR] TestDNS_ServiceReverseLookup_CustomDomain.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:41.115Z [WARN] TestDNS_ServiceReverseLookup_CustomDomain.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:41.119Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:41.121Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: consul server down writer.go:29: 2021-01-29T19:33:41.122Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: shutdown complete writer.go:29: 2021-01-29T19:33:41.123Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: Stopping server: protocol=DNS address=127.0.0.1:30414 network=tcp writer.go:29: 2021-01-29T19:33:41.124Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: Stopping server: protocol=DNS address=127.0.0.1:30414 network=udp writer.go:29: 2021-01-29T19:33:41.125Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: Stopping server: protocol=HTTP address=127.0.0.1:30415 network=tcp === CONT TestDNS_ServiceLookupPreferNoCNAME writer.go:29: 2021-01-29T19:33:41.190Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:41.191Z [INFO] TestDNS_ServiceLookupPreferNoCNAME: Endpoints down --- PASS: TestDNS_ServiceLookupPreferNoCNAME (0.71s) === CONT TestDNS_ServiceReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.201Z [WARN] TestDNS_ServiceReverseLookup_IPV6: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:41.202Z [DEBUG] TestDNS_ServiceReverseLookup_IPV6.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:41.204Z [DEBUG] TestDNS_ServiceReverseLookup_IPV6.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:41.208Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8d051f83-00ce-b326-3241-1cec01d701d1 Address:127.0.0.1:29411}]" writer.go:29: 2021-01-29T19:33:41.209Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server.raft: entering follower state: follower="Node at 127.0.0.1:29411 [Follower]" leader= writer.go:29: 2021-01-29T19:33:41.212Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server.serf.wan: serf: EventMemberJoin: Node-8d051f83-00ce-b326-3241-1cec01d701d1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:41.214Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server.serf.lan: serf: EventMemberJoin: Node-8d051f83-00ce-b326-3241-1cec01d701d1 127.0.0.1 writer.go:29: 2021-01-29T19:33:41.217Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server: Adding LAN server: server="Node-8d051f83-00ce-b326-3241-1cec01d701d1 (Addr: tcp/127.0.0.1:29411) (DC: dc1)" writer.go:29: 2021-01-29T19:33:41.218Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server: Handled event for server in area: event=member-join server=Node-8d051f83-00ce-b326-3241-1cec01d701d1.dc1 area=wan writer.go:29: 2021-01-29T19:33:41.220Z [INFO] TestDNS_ServiceReverseLookup_IPV6: Started DNS server: address=127.0.0.1:29406 network=tcp writer.go:29: 2021-01-29T19:33:41.222Z [INFO] TestDNS_ServiceReverseLookup_IPV6: Started DNS server: address=127.0.0.1:29406 network=udp writer.go:29: 2021-01-29T19:33:41.223Z [INFO] TestDNS_ServiceReverseLookup_IPV6: Started HTTP server: address=127.0.0.1:29407 network=tcp writer.go:29: 2021-01-29T19:33:41.224Z [INFO] TestDNS_ServiceReverseLookup_IPV6: started state syncer writer.go:29: 2021-01-29T19:33:41.271Z [WARN] TestDNS_ServiceReverseLookup_IPV6.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:41.273Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server.raft: entering candidate state: node="Node at 127.0.0.1:29411 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:41.275Z [DEBUG] TestDNS_ServiceReverseLookup_IPV6.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:41.277Z [DEBUG] TestDNS_ServiceReverseLookup_IPV6.server.raft: vote granted: from=8d051f83-00ce-b326-3241-1cec01d701d1 term=2 tally=1 writer.go:29: 2021-01-29T19:33:41.278Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:41.279Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server.raft: entering leader state: leader="Node at 127.0.0.1:29411 [Leader]" writer.go:29: 2021-01-29T19:33:41.281Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:41.282Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server: New leader elected: payload=Node-8d051f83-00ce-b326-3241-1cec01d701d1 writer.go:29: 2021-01-29T19:33:41.283Z [DEBUG] TestDNS_ServiceReverseLookup_IPV6.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29411 writer.go:29: 2021-01-29T19:33:41.290Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:41.307Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:41.309Z [INFO] TestDNS_ServiceReverseLookup_IPV6.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:41.310Z [DEBUG] TestDNS_ServiceReverseLookup_IPV6.server: Skipping self join check for node since the cluster is too small: node=Node-8d051f83-00ce-b326-3241-1cec01d701d1 writer.go:29: 2021-01-29T19:33:41.312Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server: member joined, marking health alive: member=Node-8d051f83-00ce-b326-3241-1cec01d701d1 writer.go:29: 2021-01-29T19:33:41.443Z [DEBUG] TestDNS_ServiceReverseLookup_IPV6: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:41.457Z [INFO] TestDNS_ServiceReverseLookup_IPV6: Synced node info === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.543Z [INFO] TestDNS_SOA_Settings: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:41.544Z [INFO] TestDNS_SOA_Settings: Endpoints down writer.go:29: 2021-01-29T19:33:41.560Z [WARN] TestDNS_SOA_Settings: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:41.564Z [DEBUG] TestDNS_SOA_Settings.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:41.566Z [DEBUG] TestDNS_SOA_Settings.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:41.571Z [INFO] TestDNS_SOA_Settings.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6f1f76e2-b643-a045-8515-89033e551969 Address:127.0.0.1:29405}]" writer.go:29: 2021-01-29T19:33:41.573Z [INFO] TestDNS_SOA_Settings.server.raft: entering follower state: follower="Node at 127.0.0.1:29405 [Follower]" leader= writer.go:29: 2021-01-29T19:33:41.575Z [INFO] TestDNS_SOA_Settings.server.serf.wan: serf: EventMemberJoin: Node-6f1f76e2-b643-a045-8515-89033e551969.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:41.581Z [INFO] TestDNS_SOA_Settings.server.serf.lan: serf: EventMemberJoin: Node-6f1f76e2-b643-a045-8515-89033e551969 127.0.0.1 writer.go:29: 2021-01-29T19:33:41.586Z [INFO] TestDNS_SOA_Settings.server: Adding LAN server: server="Node-6f1f76e2-b643-a045-8515-89033e551969 (Addr: tcp/127.0.0.1:29405) (DC: dc1)" writer.go:29: 2021-01-29T19:33:41.587Z [INFO] TestDNS_SOA_Settings.server: Handled event for server in area: event=member-join server=Node-6f1f76e2-b643-a045-8515-89033e551969.dc1 area=wan === CONT TestDNS_ServiceReverseLookupNodeAddress writer.go:29: 2021-01-29T19:33:41.593Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:41.595Z [INFO] TestDNS_ServiceReverseLookupNodeAddress: Endpoints down --- PASS: TestDNS_ServiceReverseLookupNodeAddress (0.77s) === CONT TestDNS_ReverseLookup_IPV6 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.591Z [INFO] TestDNS_SOA_Settings: Started DNS server: address=127.0.0.1:29400 network=tcp writer.go:29: 2021-01-29T19:33:41.613Z [INFO] TestDNS_SOA_Settings: Started DNS server: address=127.0.0.1:29400 network=udp writer.go:29: 2021-01-29T19:33:41.618Z [INFO] TestDNS_SOA_Settings: Started HTTP server: address=127.0.0.1:29401 network=tcp writer.go:29: 2021-01-29T19:33:41.620Z [INFO] TestDNS_SOA_Settings: started state syncer === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.623Z [WARN] TestDNS_ReverseLookup_IPV6: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:41.625Z [DEBUG] TestDNS_ReverseLookup_IPV6.tlsutil: Update: version=1 === CONT TestDNS_ServiceReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.626Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:41.627Z [INFO] TestDNS_ServiceReverseLookup_CustomDomain: Endpoints down --- PASS: TestDNS_ServiceReverseLookup_CustomDomain (0.68s) === CONT TestDNS_ReverseLookup_CustomDomain === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.627Z [DEBUG] TestDNS_ReverseLookup_IPV6.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:41.634Z [INFO] TestDNS_ReverseLookup_IPV6.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cdb32a9e-a1c6-da57-b785-b17b42d57af9 Address:127.0.0.1:30371}]" writer.go:29: 2021-01-29T19:33:41.636Z [INFO] TestDNS_ReverseLookup_IPV6.server.raft: entering follower state: follower="Node at 127.0.0.1:30371 [Follower]" leader= writer.go:29: 2021-01-29T19:33:41.640Z [INFO] TestDNS_ReverseLookup_IPV6.server.serf.wan: serf: EventMemberJoin: Node-cdb32a9e-a1c6-da57-b785-b17b42d57af9.dc1 127.0.0.1 === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.649Z [WARN] TestDNS_ReverseLookup_CustomDomain: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:41.650Z [DEBUG] TestDNS_ReverseLookup_CustomDomain.tlsutil: Update: version=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.654Z [WARN] TestDNS_SOA_Settings.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.662Z [DEBUG] TestDNS_ReverseLookup_CustomDomain.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.661Z [INFO] TestDNS_SOA_Settings.server.raft: entering candidate state: node="Node at 127.0.0.1:29405 [Candidate]" term=2 === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.663Z [INFO] TestDNS_ReverseLookup_IPV6.server.serf.lan: serf: EventMemberJoin: Node-cdb32a9e-a1c6-da57-b785-b17b42d57af9 127.0.0.1 writer.go:29: 2021-01-29T19:33:41.670Z [INFO] TestDNS_ReverseLookup_IPV6.server: Handled event for server in area: event=member-join server=Node-cdb32a9e-a1c6-da57-b785-b17b42d57af9.dc1 area=wan writer.go:29: 2021-01-29T19:33:41.671Z [INFO] TestDNS_ReverseLookup_IPV6: Started DNS server: address=127.0.0.1:30366 network=udp writer.go:29: 2021-01-29T19:33:41.680Z [INFO] TestDNS_ReverseLookup_IPV6.server: Adding LAN server: server="Node-cdb32a9e-a1c6-da57-b785-b17b42d57af9 (Addr: tcp/127.0.0.1:30371) (DC: dc1)" === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.691Z [DEBUG] TestDNS_SOA_Settings.server.raft: votes: needed=1 === CONT TestDNS_ServiceReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.681Z [DEBUG] TestDNS_ServiceReverseLookup_IPV6.dns: request served from client: question="{9.2.3.8.2.4.0.0.0.0.f.f.0.0.0.0.0.0.0.0.0.0.0.0.8.b.d.0.1.0.0.2.ip6.arpa. 255 1}" latency=136.228µs client=127.0.0.1:41611 client_network=udp === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.692Z [DEBUG] TestDNS_SOA_Settings.server.raft: vote granted: from=6f1f76e2-b643-a045-8515-89033e551969 term=2 tally=1 === CONT TestDNS_ServiceReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.681Z [INFO] TestDNS_ServiceReverseLookup_IPV6: Requesting shutdown === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.693Z [INFO] TestDNS_SOA_Settings.server.raft: election won: tally=1 === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.681Z [WARN] TestDNS_ReverseLookup_IPV6.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.694Z [INFO] TestDNS_SOA_Settings.server.raft: entering leader state: leader="Node at 127.0.0.1:29405 [Leader]" === CONT TestDNS_ServiceReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.693Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server: shutting down server writer.go:29: 2021-01-29T19:33:41.695Z [DEBUG] TestDNS_ServiceReverseLookup_IPV6.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.694Z [INFO] TestDNS_ReverseLookup_IPV6.server.raft: entering candidate state: node="Node at 127.0.0.1:30371 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:41.689Z [INFO] TestDNS_ReverseLookup_IPV6: Started DNS server: address=127.0.0.1:30366 network=tcp === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.695Z [INFO] TestDNS_ReverseLookup_CustomDomain.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:94b28cd7-92fe-900c-0f40-f967f6967ecc Address:127.0.0.1:29423}]" === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.704Z [INFO] TestDNS_ReverseLookup_IPV6: Started HTTP server: address=127.0.0.1:30367 network=tcp === CONT TestDNS_ServiceReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.696Z [WARN] TestDNS_ServiceReverseLookup_IPV6.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:41.708Z [DEBUG] TestDNS_ServiceReverseLookup_IPV6.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.706Z [INFO] TestDNS_ReverseLookup_IPV6: started state syncer === CONT TestDNS_ServiceReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.710Z [WARN] TestDNS_ServiceReverseLookup_IPV6.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.712Z [INFO] TestDNS_SOA_Settings.server: cluster leadership acquired === CONT TestDNS_ServiceReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.719Z [INFO] TestDNS_ServiceReverseLookup_IPV6.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:41.721Z [INFO] TestDNS_ServiceReverseLookup_IPV6: consul server down writer.go:29: 2021-01-29T19:33:41.722Z [INFO] TestDNS_ServiceReverseLookup_IPV6: shutdown complete writer.go:29: 2021-01-29T19:33:41.723Z [INFO] TestDNS_ServiceReverseLookup_IPV6: Stopping server: protocol=DNS address=127.0.0.1:29406 network=tcp writer.go:29: 2021-01-29T19:33:41.725Z [INFO] TestDNS_ServiceReverseLookup_IPV6: Stopping server: protocol=DNS address=127.0.0.1:29406 network=udp === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.717Z [DEBUG] TestDNS_ReverseLookup_IPV6.server.raft: votes: needed=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.738Z [INFO] TestDNS_SOA_Settings.server: New leader elected: payload=Node-6f1f76e2-b643-a045-8515-89033e551969 === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.742Z [DEBUG] TestDNS_ReverseLookup_IPV6.server.raft: vote granted: from=cdb32a9e-a1c6-da57-b785-b17b42d57af9 term=2 tally=1 === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.710Z [INFO] TestDNS_ReverseLookup_CustomDomain.server.serf.wan: serf: EventMemberJoin: Node-94b28cd7-92fe-900c-0f40-f967f6967ecc.dc1 127.0.0.1 === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.744Z [INFO] TestDNS_ReverseLookup_IPV6.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:41.745Z [INFO] TestDNS_ReverseLookup_IPV6.server.raft: entering leader state: leader="Node at 127.0.0.1:30371 [Leader]" === CONT TestDNS_ServiceReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.742Z [INFO] TestDNS_ServiceReverseLookup_IPV6: Stopping server: protocol=HTTP address=127.0.0.1:29407 network=tcp === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.747Z [INFO] TestDNS_ReverseLookup_IPV6.server: cluster leadership acquired === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.749Z [INFO] TestDNS_ReverseLookup_CustomDomain.server.serf.lan: serf: EventMemberJoin: Node-94b28cd7-92fe-900c-0f40-f967f6967ecc 127.0.0.1 === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.752Z [DEBUG] TestDNS_ReverseLookup_IPV6.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30371 writer.go:29: 2021-01-29T19:33:41.755Z [INFO] TestDNS_ReverseLookup_IPV6.server: New leader elected: payload=Node-cdb32a9e-a1c6-da57-b785-b17b42d57af9 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.754Z [DEBUG] TestDNS_SOA_Settings.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29405 === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.734Z [INFO] TestDNS_ReverseLookup_CustomDomain.server.raft: entering follower state: follower="Node at 127.0.0.1:29423 [Follower]" leader= writer.go:29: 2021-01-29T19:33:41.766Z [INFO] TestDNS_ReverseLookup_CustomDomain: Started DNS server: address=127.0.0.1:29418 network=udp writer.go:29: 2021-01-29T19:33:41.783Z [INFO] TestDNS_ReverseLookup_CustomDomain: Started DNS server: address=127.0.0.1:29418 network=tcp === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.792Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.811Z [INFO] TestDNS_ReverseLookup_CustomDomain: Started HTTP server: address=127.0.0.1:29419 network=tcp writer.go:29: 2021-01-29T19:33:41.812Z [INFO] TestDNS_ReverseLookup_CustomDomain: started state syncer writer.go:29: 2021-01-29T19:33:41.814Z [INFO] TestDNS_ReverseLookup_CustomDomain.server: Handled event for server in area: event=member-join server=Node-94b28cd7-92fe-900c-0f40-f967f6967ecc.dc1 area=wan === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.820Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:41.831Z [INFO] TestDNS_ReverseLookup_IPV6.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:41.833Z [INFO] TestDNS_ReverseLookup_IPV6.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:41.836Z [DEBUG] TestDNS_ReverseLookup_IPV6.server: Skipping self join check for node since the cluster is too small: node=Node-cdb32a9e-a1c6-da57-b785-b17b42d57af9 writer.go:29: 2021-01-29T19:33:41.838Z [INFO] TestDNS_ReverseLookup_IPV6.server: member joined, marking health alive: member=Node-cdb32a9e-a1c6-da57-b785-b17b42d57af9 === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.840Z [WARN] TestDNS_ReverseLookup_CustomDomain.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:41.842Z [INFO] TestDNS_ReverseLookup_CustomDomain.server.raft: entering candidate state: node="Node at 127.0.0.1:29423 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:41.846Z [WARN] TestDNS_ReverseLookup_CustomDomain.server.raft: unable to get address for sever, using fallback address: id=94b28cd7-92fe-900c-0f40-f967f6967ecc fallback=127.0.0.1:29423 error="Could not find address for server id 94b28cd7-92fe-900c-0f40-f967f6967ecc" writer.go:29: 2021-01-29T19:33:41.842Z [INFO] TestDNS_ReverseLookup_CustomDomain.server: Adding LAN server: server="Node-94b28cd7-92fe-900c-0f40-f967f6967ecc (Addr: tcp/127.0.0.1:29423) (DC: dc1)" writer.go:29: 2021-01-29T19:33:41.891Z [DEBUG] TestDNS_ReverseLookup_CustomDomain.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:41.903Z [DEBUG] TestDNS_ReverseLookup_CustomDomain.server.raft: vote granted: from=94b28cd7-92fe-900c-0f40-f967f6967ecc term=2 tally=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.906Z [INFO] TestDNS_SOA_Settings.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.908Z [INFO] TestDNS_ReverseLookup_CustomDomain.server.raft: election won: tally=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.910Z [INFO] TestDNS_SOA_Settings.leader: started routine: routine="CA root pruning" === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.912Z [INFO] TestDNS_ReverseLookup_CustomDomain.server.raft: entering leader state: leader="Node at 127.0.0.1:29423 [Leader]" === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.914Z [DEBUG] TestDNS_SOA_Settings.server: Skipping self join check for node since the cluster is too small: node=Node-6f1f76e2-b643-a045-8515-89033e551969 writer.go:29: 2021-01-29T19:33:41.915Z [INFO] TestDNS_SOA_Settings.server: member joined, marking health alive: member=Node-6f1f76e2-b643-a045-8515-89033e551969 === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.915Z [INFO] TestDNS_ReverseLookup_CustomDomain.server: cluster leadership acquired === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.924Z [DEBUG] TestDNS_SOA_Settings: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.918Z [INFO] TestDNS_ReverseLookup_CustomDomain.server: New leader elected: payload=Node-94b28cd7-92fe-900c-0f40-f967f6967ecc writer.go:29: 2021-01-29T19:33:41.926Z [DEBUG] TestDNS_ReverseLookup_CustomDomain.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29423 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.936Z [INFO] TestDNS_SOA_Settings: Synced node info writer.go:29: 2021-01-29T19:33:41.937Z [DEBUG] TestDNS_SOA_Settings: Node info in sync === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.949Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.957Z [DEBUG] TestDNS_ReverseLookup_IPV6: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.961Z [DEBUG] TestDNS_SOA_Settings.dns: request served from client: name=nofoo.node.dc1.consul. type=ANY class=IN latency=281.915µs client=127.0.0.1:58071 client_network=udp writer.go:29: 2021-01-29T19:33:41.978Z [INFO] TestDNS_SOA_Settings: Requesting shutdown writer.go:29: 2021-01-29T19:33:41.981Z [INFO] TestDNS_SOA_Settings.server: shutting down server === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.984Z [INFO] TestDNS_ReverseLookup_IPV6: Synced node info === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.984Z [DEBUG] TestDNS_SOA_Settings.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.985Z [DEBUG] TestDNS_ReverseLookup_IPV6.dns: request served from client: question="{2.4.2.4.2.4.2.4.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.0.ip6.arpa. 255 1}" latency=372.01µs client=127.0.0.1:40744 client_network=udp writer.go:29: 2021-01-29T19:33:41.989Z [INFO] TestDNS_ReverseLookup_IPV6: Requesting shutdown === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.987Z [WARN] TestDNS_SOA_Settings.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.992Z [INFO] TestDNS_ReverseLookup_IPV6.server: shutting down server === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:41.989Z [DEBUG] TestDNS_SOA_Settings.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:41.997Z [INFO] TestDNS_ReverseLookup_CustomDomain.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.001Z [WARN] TestDNS_SOA_Settings.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:41.998Z [DEBUG] TestDNS_ReverseLookup_IPV6.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:42.002Z [INFO] TestDNS_ReverseLookup_CustomDomain.leader: started routine: routine="CA root pruning" === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.005Z [INFO] TestDNS_SOA_Settings.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:42.006Z [INFO] TestDNS_SOA_Settings: consul server down === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:42.005Z [WARN] TestDNS_ReverseLookup_IPV6.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:42.005Z [DEBUG] TestDNS_ReverseLookup_CustomDomain.server: Skipping self join check for node since the cluster is too small: node=Node-94b28cd7-92fe-900c-0f40-f967f6967ecc === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:42.008Z [DEBUG] TestDNS_ReverseLookup_IPV6.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:42.012Z [WARN] TestDNS_ReverseLookup_IPV6.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:42.009Z [INFO] TestDNS_ReverseLookup_CustomDomain.server: member joined, marking health alive: member=Node-94b28cd7-92fe-900c-0f40-f967f6967ecc === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.007Z [INFO] TestDNS_SOA_Settings: shutdown complete writer.go:29: 2021-01-29T19:33:42.019Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=DNS address=127.0.0.1:29400 network=tcp writer.go:29: 2021-01-29T19:33:42.020Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=DNS address=127.0.0.1:29400 network=udp writer.go:29: 2021-01-29T19:33:42.021Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=HTTP address=127.0.0.1:29401 network=tcp === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:42.012Z [INFO] TestDNS_ReverseLookup_CustomDomain: Synced node info writer.go:29: 2021-01-29T19:33:42.023Z [DEBUG] TestDNS_ReverseLookup_CustomDomain: Node info in sync === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:42.022Z [INFO] TestDNS_ReverseLookup_IPV6.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:42.024Z [INFO] TestDNS_ReverseLookup_IPV6: consul server down writer.go:29: 2021-01-29T19:33:42.032Z [INFO] TestDNS_ReverseLookup_IPV6: shutdown complete writer.go:29: 2021-01-29T19:33:42.034Z [INFO] TestDNS_ReverseLookup_IPV6: Stopping server: protocol=DNS address=127.0.0.1:30366 network=tcp writer.go:29: 2021-01-29T19:33:42.036Z [INFO] TestDNS_ReverseLookup_IPV6: Stopping server: protocol=DNS address=127.0.0.1:30366 network=udp writer.go:29: 2021-01-29T19:33:42.039Z [INFO] TestDNS_ReverseLookup_IPV6: Stopping server: protocol=HTTP address=127.0.0.1:30367 network=tcp === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:42.040Z [DEBUG] TestDNS_ReverseLookup_CustomDomain.dns: request served from client: question="{2.0.0.127.in-addr.arpa. 255 1}" latency=52.574µs client=127.0.0.1:53451 client_network=udp writer.go:29: 2021-01-29T19:33:42.041Z [INFO] TestDNS_ReverseLookup_CustomDomain: Requesting shutdown writer.go:29: 2021-01-29T19:33:42.042Z [INFO] TestDNS_ReverseLookup_CustomDomain.server: shutting down server writer.go:29: 2021-01-29T19:33:42.043Z [DEBUG] TestDNS_ReverseLookup_CustomDomain.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:42.044Z [WARN] TestDNS_ReverseLookup_CustomDomain.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:42.045Z [DEBUG] TestDNS_ReverseLookup_CustomDomain.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:42.062Z [WARN] TestDNS_ReverseLookup_CustomDomain.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:42.079Z [INFO] TestDNS_ReverseLookup_CustomDomain.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:42.081Z [INFO] TestDNS_ReverseLookup_CustomDomain: consul server down writer.go:29: 2021-01-29T19:33:42.093Z [INFO] TestDNS_ReverseLookup_CustomDomain: shutdown complete writer.go:29: 2021-01-29T19:33:42.098Z [INFO] TestDNS_ReverseLookup_CustomDomain: Stopping server: protocol=DNS address=127.0.0.1:29418 network=tcp writer.go:29: 2021-01-29T19:33:42.105Z [INFO] TestDNS_ReverseLookup_CustomDomain: Stopping server: protocol=DNS address=127.0.0.1:29418 network=udp writer.go:29: 2021-01-29T19:33:42.107Z [INFO] TestDNS_ReverseLookup_CustomDomain: Stopping server: protocol=HTTP address=127.0.0.1:29419 network=tcp === CONT TestDNS_ServiceReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:42.250Z [INFO] TestDNS_ServiceReverseLookup_IPV6: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:42.252Z [INFO] TestDNS_ServiceReverseLookup_IPV6: Endpoints down --- PASS: TestDNS_ServiceReverseLookup_IPV6 (1.06s) === CONT TestDNS_ReverseLookup writer.go:29: 2021-01-29T19:33:42.277Z [WARN] TestDNS_ReverseLookup: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:42.279Z [DEBUG] TestDNS_ReverseLookup.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:42.282Z [DEBUG] TestDNS_ReverseLookup.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:42.291Z [INFO] TestDNS_ReverseLookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:28ec4cb7-c8c3-d7d8-d3b6-f3a01a6ec7f9 Address:127.0.0.1:29417}]" writer.go:29: 2021-01-29T19:33:42.294Z [INFO] TestDNS_ReverseLookup.server.raft: entering follower state: follower="Node at 127.0.0.1:29417 [Follower]" leader= writer.go:29: 2021-01-29T19:33:42.295Z [INFO] TestDNS_ReverseLookup.server.serf.wan: serf: EventMemberJoin: Node-28ec4cb7-c8c3-d7d8-d3b6-f3a01a6ec7f9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:42.297Z [INFO] TestDNS_ReverseLookup.server.serf.lan: serf: EventMemberJoin: Node-28ec4cb7-c8c3-d7d8-d3b6-f3a01a6ec7f9 127.0.0.1 writer.go:29: 2021-01-29T19:33:42.299Z [INFO] TestDNS_ReverseLookup.server: Handled event for server in area: event=member-join server=Node-28ec4cb7-c8c3-d7d8-d3b6-f3a01a6ec7f9.dc1 area=wan writer.go:29: 2021-01-29T19:33:42.300Z [INFO] TestDNS_ReverseLookup.server: Adding LAN server: server="Node-28ec4cb7-c8c3-d7d8-d3b6-f3a01a6ec7f9 (Addr: tcp/127.0.0.1:29417) (DC: dc1)" writer.go:29: 2021-01-29T19:33:42.300Z [INFO] TestDNS_ReverseLookup: Started DNS server: address=127.0.0.1:29412 network=udp writer.go:29: 2021-01-29T19:33:42.304Z [INFO] TestDNS_ReverseLookup: Started DNS server: address=127.0.0.1:29412 network=tcp writer.go:29: 2021-01-29T19:33:42.306Z [INFO] TestDNS_ReverseLookup: Started HTTP server: address=127.0.0.1:29413 network=tcp writer.go:29: 2021-01-29T19:33:42.308Z [INFO] TestDNS_ReverseLookup: started state syncer writer.go:29: 2021-01-29T19:33:42.341Z [WARN] TestDNS_ReverseLookup.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:42.343Z [INFO] TestDNS_ReverseLookup.server.raft: entering candidate state: node="Node at 127.0.0.1:29417 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:42.345Z [DEBUG] TestDNS_ReverseLookup.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:42.347Z [DEBUG] TestDNS_ReverseLookup.server.raft: vote granted: from=28ec4cb7-c8c3-d7d8-d3b6-f3a01a6ec7f9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:42.349Z [INFO] TestDNS_ReverseLookup.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:42.350Z [INFO] TestDNS_ReverseLookup.server.raft: entering leader state: leader="Node at 127.0.0.1:29417 [Leader]" writer.go:29: 2021-01-29T19:33:42.352Z [INFO] TestDNS_ReverseLookup.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:42.353Z [INFO] TestDNS_ReverseLookup.server: New leader elected: payload=Node-28ec4cb7-c8c3-d7d8-d3b6-f3a01a6ec7f9 writer.go:29: 2021-01-29T19:33:42.354Z [DEBUG] TestDNS_ReverseLookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29417 writer.go:29: 2021-01-29T19:33:42.369Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:42.376Z [INFO] TestDNS_ReverseLookup.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:42.378Z [INFO] TestDNS_ReverseLookup.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:42.380Z [DEBUG] TestDNS_ReverseLookup.server: Skipping self join check for node since the cluster is too small: node=Node-28ec4cb7-c8c3-d7d8-d3b6-f3a01a6ec7f9 writer.go:29: 2021-01-29T19:33:42.383Z [INFO] TestDNS_ReverseLookup.server: member joined, marking health alive: member=Node-28ec4cb7-c8c3-d7d8-d3b6-f3a01a6ec7f9 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.522Z [INFO] TestDNS_SOA_Settings: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:42.523Z [INFO] TestDNS_SOA_Settings: Endpoints down === CONT TestDNS_ReverseLookup_IPV6 writer.go:29: 2021-01-29T19:33:42.541Z [INFO] TestDNS_ReverseLookup_IPV6: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:42.542Z [INFO] TestDNS_ReverseLookup_IPV6: Endpoints down --- PASS: TestDNS_ReverseLookup_IPV6 (0.95s) === CONT TestDNS_EDNS0 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.555Z [WARN] TestDNS_SOA_Settings: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:42.557Z [DEBUG] TestDNS_SOA_Settings.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:42.559Z [DEBUG] TestDNS_SOA_Settings.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.559Z [WARN] TestDNS_EDNS0: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:42.561Z [DEBUG] TestDNS_EDNS0.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:42.563Z [DEBUG] TestDNS_EDNS0.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.563Z [INFO] TestDNS_SOA_Settings.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:af898af0-be25-ca7a-5901-ebcad7c2c2ec Address:127.0.0.1:29429}]" writer.go:29: 2021-01-29T19:33:42.566Z [INFO] TestDNS_SOA_Settings.server.raft: entering follower state: follower="Node at 127.0.0.1:29429 [Follower]" leader= writer.go:29: 2021-01-29T19:33:42.566Z [INFO] TestDNS_SOA_Settings.server.serf.wan: serf: EventMemberJoin: Node-af898af0-be25-ca7a-5901-ebcad7c2c2ec.dc1 127.0.0.1 === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.571Z [INFO] TestDNS_EDNS0.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3fc55f85-6c28-c7d6-aaa6-46b756212482 Address:127.0.0.1:30359}]" writer.go:29: 2021-01-29T19:33:42.574Z [INFO] TestDNS_EDNS0.server.raft: entering follower state: follower="Node at 127.0.0.1:30359 [Follower]" leader= === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.573Z [INFO] TestDNS_SOA_Settings.server.serf.lan: serf: EventMemberJoin: Node-af898af0-be25-ca7a-5901-ebcad7c2c2ec 127.0.0.1 writer.go:29: 2021-01-29T19:33:42.580Z [INFO] TestDNS_SOA_Settings.server: Adding LAN server: server="Node-af898af0-be25-ca7a-5901-ebcad7c2c2ec (Addr: tcp/127.0.0.1:29429) (DC: dc1)" writer.go:29: 2021-01-29T19:33:42.581Z [INFO] TestDNS_SOA_Settings.server: Handled event for server in area: event=member-join server=Node-af898af0-be25-ca7a-5901-ebcad7c2c2ec.dc1 area=wan writer.go:29: 2021-01-29T19:33:42.583Z [INFO] TestDNS_SOA_Settings: Started DNS server: address=127.0.0.1:29424 network=udp writer.go:29: 2021-01-29T19:33:42.585Z [INFO] TestDNS_SOA_Settings: Started DNS server: address=127.0.0.1:29424 network=tcp === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.588Z [INFO] TestDNS_EDNS0.server.serf.wan: serf: EventMemberJoin: Node-3fc55f85-6c28-c7d6-aaa6-46b756212482.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:42.592Z [INFO] TestDNS_EDNS0.server.serf.lan: serf: EventMemberJoin: Node-3fc55f85-6c28-c7d6-aaa6-46b756212482 127.0.0.1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.596Z [INFO] TestDNS_SOA_Settings: Started HTTP server: address=127.0.0.1:29425 network=tcp writer.go:29: 2021-01-29T19:33:42.604Z [INFO] TestDNS_SOA_Settings: started state syncer === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.602Z [INFO] TestDNS_EDNS0: Started DNS server: address=127.0.0.1:30354 network=udp writer.go:29: 2021-01-29T19:33:42.605Z [INFO] TestDNS_EDNS0: Started DNS server: address=127.0.0.1:30354 network=tcp writer.go:29: 2021-01-29T19:33:42.602Z [INFO] TestDNS_EDNS0.server: Handled event for server in area: event=member-join server=Node-3fc55f85-6c28-c7d6-aaa6-46b756212482.dc1 area=wan writer.go:29: 2021-01-29T19:33:42.603Z [INFO] TestDNS_EDNS0.server: Adding LAN server: server="Node-3fc55f85-6c28-c7d6-aaa6-46b756212482 (Addr: tcp/127.0.0.1:30359) (DC: dc1)" === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:42.608Z [INFO] TestDNS_ReverseLookup_CustomDomain: Waiting for endpoints to shut down === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.607Z [INFO] TestDNS_EDNS0: Started HTTP server: address=127.0.0.1:30355 network=tcp === CONT TestDNS_ReverseLookup_CustomDomain writer.go:29: 2021-01-29T19:33:42.610Z [INFO] TestDNS_ReverseLookup_CustomDomain: Endpoints down --- PASS: TestDNS_ReverseLookup_CustomDomain (0.98s) === CONT TestDNS_NodeLookup_CNAME === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.611Z [INFO] TestDNS_EDNS0: started state syncer === CONT TestDNS_ReverseLookup writer.go:29: 2021-01-29T19:33:42.619Z [DEBUG] TestDNS_ReverseLookup.dns: request served from client: question="{2.0.0.127.in-addr.arpa. 255 1}" latency=57.928µs client=127.0.0.1:60399 client_network=udp writer.go:29: 2021-01-29T19:33:42.620Z [INFO] TestDNS_ReverseLookup: Requesting shutdown writer.go:29: 2021-01-29T19:33:42.621Z [INFO] TestDNS_ReverseLookup.server: shutting down server writer.go:29: 2021-01-29T19:33:42.622Z [DEBUG] TestDNS_ReverseLookup.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:42.623Z [WARN] TestDNS_ReverseLookup.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:42.624Z [ERROR] TestDNS_ReverseLookup.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:42.625Z [DEBUG] TestDNS_ReverseLookup.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.629Z [WARN] TestDNS_EDNS0.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:42.630Z [INFO] TestDNS_EDNS0.server.raft: entering candidate state: node="Node at 127.0.0.1:30359 [Candidate]" term=2 === CONT TestDNS_ReverseLookup writer.go:29: 2021-01-29T19:33:42.629Z [WARN] TestDNS_ReverseLookup.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.631Z [WARN] TestDNS_SOA_Settings.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:42.633Z [INFO] TestDNS_SOA_Settings.server.raft: entering candidate state: node="Node at 127.0.0.1:29429 [Candidate]" term=2 === CONT TestDNS_ReverseLookup writer.go:29: 2021-01-29T19:33:42.632Z [INFO] TestDNS_ReverseLookup.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:42.635Z [INFO] TestDNS_ReverseLookup: consul server down === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.635Z [DEBUG] TestDNS_SOA_Settings.server.raft: votes: needed=1 === CONT TestDNS_ReverseLookup writer.go:29: 2021-01-29T19:33:42.636Z [INFO] TestDNS_ReverseLookup: shutdown complete === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.637Z [DEBUG] TestDNS_SOA_Settings.server.raft: vote granted: from=af898af0-be25-ca7a-5901-ebcad7c2c2ec term=2 tally=1 === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.637Z [DEBUG] TestDNS_EDNS0.server.raft: votes: needed=1 === CONT TestDNS_ReverseLookup writer.go:29: 2021-01-29T19:33:42.638Z [INFO] TestDNS_ReverseLookup: Stopping server: protocol=DNS address=127.0.0.1:29412 network=tcp === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.638Z [INFO] TestDNS_SOA_Settings.server.raft: election won: tally=1 === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.638Z [DEBUG] TestDNS_EDNS0.server.raft: vote granted: from=3fc55f85-6c28-c7d6-aaa6-46b756212482 term=2 tally=1 === CONT TestDNS_ReverseLookup writer.go:29: 2021-01-29T19:33:42.639Z [INFO] TestDNS_ReverseLookup: Stopping server: protocol=DNS address=127.0.0.1:29412 network=udp === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.639Z [INFO] TestDNS_SOA_Settings.server.raft: entering leader state: leader="Node at 127.0.0.1:29429 [Leader]" === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.640Z [INFO] TestDNS_EDNS0.server.raft: election won: tally=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.641Z [INFO] TestDNS_SOA_Settings.server: cluster leadership acquired === CONT TestDNS_ReverseLookup writer.go:29: 2021-01-29T19:33:42.641Z [INFO] TestDNS_ReverseLookup: Stopping server: protocol=HTTP address=127.0.0.1:29413 network=tcp === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.643Z [WARN] TestDNS_NodeLookup_CNAME: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:42.652Z [DEBUG] TestDNS_NodeLookup_CNAME.tlsutil: Update: version=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.643Z [INFO] TestDNS_SOA_Settings.server: New leader elected: payload=Node-af898af0-be25-ca7a-5901-ebcad7c2c2ec === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.655Z [DEBUG] TestDNS_NodeLookup_CNAME.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.651Z [INFO] TestDNS_EDNS0.server.raft: entering leader state: leader="Node at 127.0.0.1:30359 [Leader]" === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.644Z [DEBUG] TestDNS_SOA_Settings.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29429 === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.660Z [INFO] TestDNS_EDNS0.server: cluster leadership acquired === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.665Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.667Z [INFO] TestDNS_NodeLookup_CNAME.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1e374545-0a0a-0b74-9273-70d234b29d75 Address:127.0.0.1:29441}]" === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.661Z [INFO] TestDNS_EDNS0.server: New leader elected: payload=Node-3fc55f85-6c28-c7d6-aaa6-46b756212482 === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.671Z [INFO] TestDNS_NodeLookup_CNAME.server.serf.wan: serf: EventMemberJoin: Node-1e374545-0a0a-0b74-9273-70d234b29d75.dc1 127.0.0.1 === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.667Z [DEBUG] TestDNS_EDNS0.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30359 === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.677Z [INFO] TestDNS_NodeLookup_CNAME.server.serf.lan: serf: EventMemberJoin: Node-1e374545-0a0a-0b74-9273-70d234b29d75 127.0.0.1 === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.683Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.692Z [INFO] TestDNS_SOA_Settings.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:42.693Z [INFO] TestDNS_SOA_Settings.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:42.694Z [DEBUG] TestDNS_SOA_Settings.server: Skipping self join check for node since the cluster is too small: node=Node-af898af0-be25-ca7a-5901-ebcad7c2c2ec writer.go:29: 2021-01-29T19:33:42.695Z [INFO] TestDNS_SOA_Settings.server: member joined, marking health alive: member=Node-af898af0-be25-ca7a-5901-ebcad7c2c2ec === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.672Z [INFO] TestDNS_NodeLookup_CNAME.server.raft: entering follower state: follower="Node at 127.0.0.1:29441 [Follower]" leader= writer.go:29: 2021-01-29T19:33:42.694Z [INFO] TestDNS_NodeLookup_CNAME.server: Adding LAN server: server="Node-1e374545-0a0a-0b74-9273-70d234b29d75 (Addr: tcp/127.0.0.1:29441) (DC: dc1)" === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.700Z [INFO] TestDNS_EDNS0.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.694Z [INFO] TestDNS_NodeLookup_CNAME.server: Handled event for server in area: event=member-join server=Node-1e374545-0a0a-0b74-9273-70d234b29d75.dc1 area=wan === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.703Z [INFO] TestDNS_EDNS0.leader: started routine: routine="CA root pruning" === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.695Z [DEBUG] TestDNS_NodeLookup_CNAME.dns: recursor enabled === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.705Z [DEBUG] TestDNS_EDNS0.server: Skipping self join check for node since the cluster is too small: node=Node-3fc55f85-6c28-c7d6-aaa6-46b756212482 === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.695Z [DEBUG] TestDNS_NodeLookup_CNAME.dns: recursor enabled === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.707Z [INFO] TestDNS_EDNS0.server: member joined, marking health alive: member=Node-3fc55f85-6c28-c7d6-aaa6-46b756212482 === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.706Z [INFO] TestDNS_NodeLookup_CNAME: Started DNS server: address=127.0.0.1:29436 network=udp writer.go:29: 2021-01-29T19:33:42.710Z [INFO] TestDNS_NodeLookup_CNAME: Started DNS server: address=127.0.0.1:29436 network=tcp writer.go:29: 2021-01-29T19:33:42.712Z [INFO] TestDNS_NodeLookup_CNAME: Started HTTP server: address=127.0.0.1:29437 network=tcp writer.go:29: 2021-01-29T19:33:42.713Z [INFO] TestDNS_NodeLookup_CNAME: started state syncer === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:42.713Z [DEBUG] TestDNS_SOA_Settings.dns: request served from client: name=nofoo.node.dc1.consul. type=ANY class=IN latency=119.459µs client=127.0.0.1:51810 client_network=udp writer.go:29: 2021-01-29T19:33:42.724Z [INFO] TestDNS_SOA_Settings: Requesting shutdown writer.go:29: 2021-01-29T19:33:42.725Z [INFO] TestDNS_SOA_Settings.server: shutting down server writer.go:29: 2021-01-29T19:33:42.727Z [DEBUG] TestDNS_SOA_Settings.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:42.727Z [WARN] TestDNS_SOA_Settings.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:42.727Z [ERROR] TestDNS_SOA_Settings.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:42.727Z [DEBUG] TestDNS_SOA_Settings.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:42.729Z [WARN] TestDNS_SOA_Settings.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:42.735Z [INFO] TestDNS_SOA_Settings.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:42.735Z [INFO] TestDNS_SOA_Settings: consul server down writer.go:29: 2021-01-29T19:33:42.738Z [INFO] TestDNS_SOA_Settings: shutdown complete writer.go:29: 2021-01-29T19:33:42.739Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=DNS address=127.0.0.1:29424 network=tcp writer.go:29: 2021-01-29T19:33:42.741Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=DNS address=127.0.0.1:29424 network=udp writer.go:29: 2021-01-29T19:33:42.743Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=HTTP address=127.0.0.1:29425 network=tcp === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.763Z [WARN] TestDNS_NodeLookup_CNAME.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:42.765Z [INFO] TestDNS_NodeLookup_CNAME.server.raft: entering candidate state: node="Node at 127.0.0.1:29441 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:42.782Z [DEBUG] TestDNS_NodeLookup_CNAME.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:42.785Z [DEBUG] TestDNS_NodeLookup_CNAME.server.raft: vote granted: from=1e374545-0a0a-0b74-9273-70d234b29d75 term=2 tally=1 writer.go:29: 2021-01-29T19:33:42.786Z [INFO] TestDNS_NodeLookup_CNAME.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:42.788Z [INFO] TestDNS_NodeLookup_CNAME.server.raft: entering leader state: leader="Node at 127.0.0.1:29441 [Leader]" writer.go:29: 2021-01-29T19:33:42.789Z [INFO] TestDNS_NodeLookup_CNAME.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:42.790Z [INFO] TestDNS_NodeLookup_CNAME.server: New leader elected: payload=Node-1e374545-0a0a-0b74-9273-70d234b29d75 writer.go:29: 2021-01-29T19:33:42.805Z [DEBUG] TestDNS_NodeLookup_CNAME.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29441 === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.810Z [DEBUG] TestDNS_EDNS0: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:42.814Z [INFO] TestDNS_EDNS0: Synced node info writer.go:29: 2021-01-29T19:33:42.816Z [DEBUG] TestDNS_EDNS0: Node info in sync === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.820Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:42.833Z [INFO] TestDNS_NodeLookup_CNAME.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:42.835Z [INFO] TestDNS_NodeLookup_CNAME.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:42.836Z [DEBUG] TestDNS_NodeLookup_CNAME.server: Skipping self join check for node since the cluster is too small: node=Node-1e374545-0a0a-0b74-9273-70d234b29d75 writer.go:29: 2021-01-29T19:33:42.838Z [INFO] TestDNS_NodeLookup_CNAME.server: member joined, marking health alive: member=Node-1e374545-0a0a-0b74-9273-70d234b29d75 writer.go:29: 2021-01-29T19:33:42.862Z [DEBUG] TestDNS_NodeLookup_CNAME.dns: cname recurse RTT for name: name=www.google.com. rtt=664.607µs writer.go:29: 2021-01-29T19:33:42.864Z [DEBUG] TestDNS_NodeLookup_CNAME.dns: request served from client: name=google.node.consul. type=ANY class=IN latency=2.48817ms client=127.0.0.1:46516 client_network=udp writer.go:29: 2021-01-29T19:33:42.864Z [INFO] TestDNS_NodeLookup_CNAME: Requesting shutdown writer.go:29: 2021-01-29T19:33:42.867Z [INFO] TestDNS_NodeLookup_CNAME.server: shutting down server writer.go:29: 2021-01-29T19:33:42.869Z [DEBUG] TestDNS_NodeLookup_CNAME.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:42.871Z [WARN] TestDNS_NodeLookup_CNAME.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.874Z [DEBUG] TestDNS_EDNS0.dns: request served from client: name=foo.node.dc1.consul. type=ANY class=IN latency=86.214µs client=127.0.0.1:43967 client_network=udp === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.871Z [ERROR] TestDNS_NodeLookup_CNAME.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.874Z [INFO] TestDNS_EDNS0: Requesting shutdown === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.873Z [DEBUG] TestDNS_NodeLookup_CNAME.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.879Z [INFO] TestDNS_EDNS0.server: shutting down server writer.go:29: 2021-01-29T19:33:42.880Z [DEBUG] TestDNS_EDNS0.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.878Z [WARN] TestDNS_NodeLookup_CNAME.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.882Z [WARN] TestDNS_EDNS0.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.884Z [INFO] TestDNS_NodeLookup_CNAME.server.router.manager: shutting down === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.882Z [DEBUG] TestDNS_EDNS0.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.885Z [INFO] TestDNS_NodeLookup_CNAME: consul server down === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.884Z [WARN] TestDNS_EDNS0.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.887Z [INFO] TestDNS_NodeLookup_CNAME: shutdown complete writer.go:29: 2021-01-29T19:33:42.888Z [INFO] TestDNS_NodeLookup_CNAME: Stopping server: protocol=DNS address=127.0.0.1:29436 network=tcp === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.889Z [INFO] TestDNS_EDNS0.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:42.889Z [INFO] TestDNS_EDNS0: consul server down === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.891Z [INFO] TestDNS_NodeLookup_CNAME: Stopping server: protocol=DNS address=127.0.0.1:29436 network=udp === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.892Z [INFO] TestDNS_EDNS0: shutdown complete === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:42.893Z [INFO] TestDNS_NodeLookup_CNAME: Stopping server: protocol=HTTP address=127.0.0.1:29437 network=tcp === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:42.894Z [INFO] TestDNS_EDNS0: Stopping server: protocol=DNS address=127.0.0.1:30354 network=tcp writer.go:29: 2021-01-29T19:33:42.895Z [INFO] TestDNS_EDNS0: Stopping server: protocol=DNS address=127.0.0.1:30354 network=udp writer.go:29: 2021-01-29T19:33:42.897Z [INFO] TestDNS_EDNS0: Stopping server: protocol=HTTP address=127.0.0.1:30355 network=tcp === CONT TestDNS_ReverseLookup writer.go:29: 2021-01-29T19:33:43.151Z [INFO] TestDNS_ReverseLookup: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:43.154Z [INFO] TestDNS_ReverseLookup: Endpoints down --- PASS: TestDNS_ReverseLookup (0.90s) === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.180Z [WARN] TestDNSCycleRecursorCheck: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:43.191Z [DEBUG] TestDNSCycleRecursorCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:43.194Z [DEBUG] TestDNSCycleRecursorCheck.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:43.198Z [INFO] TestDNSCycleRecursorCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2d7e3faa-c8e1-dfe6-9ef4-fa2a31278ee1 Address:127.0.0.1:29447}]" writer.go:29: 2021-01-29T19:33:43.200Z [INFO] TestDNSCycleRecursorCheck.server.serf.wan: serf: EventMemberJoin: Node-2d7e3faa-c8e1-dfe6-9ef4-fa2a31278ee1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:43.200Z [INFO] TestDNSCycleRecursorCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:29447 [Follower]" leader= writer.go:29: 2021-01-29T19:33:43.203Z [INFO] TestDNSCycleRecursorCheck.server.serf.lan: serf: EventMemberJoin: Node-2d7e3faa-c8e1-dfe6-9ef4-fa2a31278ee1 127.0.0.1 writer.go:29: 2021-01-29T19:33:43.209Z [DEBUG] TestDNSCycleRecursorCheck.dns: recursor enabled writer.go:29: 2021-01-29T19:33:43.211Z [INFO] TestDNSCycleRecursorCheck: Started DNS server: address=127.0.0.1:29442 network=udp writer.go:29: 2021-01-29T19:33:43.211Z [INFO] TestDNSCycleRecursorCheck.server: Adding LAN server: server="Node-2d7e3faa-c8e1-dfe6-9ef4-fa2a31278ee1 (Addr: tcp/127.0.0.1:29447) (DC: dc1)" writer.go:29: 2021-01-29T19:33:43.212Z [DEBUG] TestDNSCycleRecursorCheck.dns: recursor enabled writer.go:29: 2021-01-29T19:33:43.213Z [INFO] TestDNSCycleRecursorCheck.server: Handled event for server in area: event=member-join server=Node-2d7e3faa-c8e1-dfe6-9ef4-fa2a31278ee1.dc1 area=wan writer.go:29: 2021-01-29T19:33:43.222Z [INFO] TestDNSCycleRecursorCheck: Started DNS server: address=127.0.0.1:29442 network=tcp writer.go:29: 2021-01-29T19:33:43.226Z [INFO] TestDNSCycleRecursorCheck: Started HTTP server: address=127.0.0.1:29443 network=tcp writer.go:29: 2021-01-29T19:33:43.228Z [INFO] TestDNSCycleRecursorCheck: started state syncer === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.245Z [INFO] TestDNS_SOA_Settings: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:43.247Z [INFO] TestDNS_SOA_Settings: Endpoints down writer.go:29: 2021-01-29T19:33:43.270Z [WARN] TestDNS_SOA_Settings: bootstrap = true: do not enable unless necessary === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.271Z [WARN] TestDNSCycleRecursorCheck.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.273Z [DEBUG] TestDNS_SOA_Settings.tlsutil: Update: version=1 === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.273Z [INFO] TestDNSCycleRecursorCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:29447 [Candidate]" term=2 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.276Z [DEBUG] TestDNS_SOA_Settings.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:43.284Z [INFO] TestDNS_SOA_Settings.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:77ae9329-c3ff-a9fb-d5cb-b02fb6bd72ce Address:127.0.0.1:29435}]" writer.go:29: 2021-01-29T19:33:43.290Z [INFO] TestDNS_SOA_Settings.server.raft: entering follower state: follower="Node at 127.0.0.1:29435 [Follower]" leader= === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.293Z [DEBUG] TestDNSCycleRecursorCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:43.294Z [DEBUG] TestDNSCycleRecursorCheck.server.raft: vote granted: from=2d7e3faa-c8e1-dfe6-9ef4-fa2a31278ee1 term=2 tally=1 writer.go:29: 2021-01-29T19:33:43.295Z [INFO] TestDNSCycleRecursorCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:43.296Z [INFO] TestDNSCycleRecursorCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:29447 [Leader]" writer.go:29: 2021-01-29T19:33:43.296Z [INFO] TestDNSCycleRecursorCheck.server: cluster leadership acquired === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.294Z [INFO] TestDNS_SOA_Settings.server.serf.wan: serf: EventMemberJoin: Node-77ae9329-c3ff-a9fb-d5cb-b02fb6bd72ce.dc1 127.0.0.1 === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.298Z [INFO] TestDNSCycleRecursorCheck.server: New leader elected: payload=Node-2d7e3faa-c8e1-dfe6-9ef4-fa2a31278ee1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.299Z [INFO] TestDNS_SOA_Settings.server.serf.lan: serf: EventMemberJoin: Node-77ae9329-c3ff-a9fb-d5cb-b02fb6bd72ce 127.0.0.1 writer.go:29: 2021-01-29T19:33:43.302Z [INFO] TestDNS_SOA_Settings.server: Adding LAN server: server="Node-77ae9329-c3ff-a9fb-d5cb-b02fb6bd72ce (Addr: tcp/127.0.0.1:29435) (DC: dc1)" writer.go:29: 2021-01-29T19:33:43.302Z [INFO] TestDNS_SOA_Settings: Started DNS server: address=127.0.0.1:29430 network=udp writer.go:29: 2021-01-29T19:33:43.305Z [INFO] TestDNS_SOA_Settings: Started DNS server: address=127.0.0.1:29430 network=tcp writer.go:29: 2021-01-29T19:33:43.302Z [INFO] TestDNS_SOA_Settings.server: Handled event for server in area: event=member-join server=Node-77ae9329-c3ff-a9fb-d5cb-b02fb6bd72ce.dc1 area=wan === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.308Z [DEBUG] TestDNSCycleRecursorCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29447 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.307Z [INFO] TestDNS_SOA_Settings: Started HTTP server: address=127.0.0.1:29431 network=tcp writer.go:29: 2021-01-29T19:33:43.315Z [INFO] TestDNS_SOA_Settings: started state syncer writer.go:29: 2021-01-29T19:33:43.332Z [WARN] TestDNS_SOA_Settings.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:43.334Z [INFO] TestDNS_SOA_Settings.server.raft: entering candidate state: node="Node at 127.0.0.1:29435 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:43.336Z [DEBUG] TestDNS_SOA_Settings.server.raft: votes: needed=1 === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.336Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.338Z [DEBUG] TestDNS_SOA_Settings.server.raft: vote granted: from=77ae9329-c3ff-a9fb-d5cb-b02fb6bd72ce term=2 tally=1 writer.go:29: 2021-01-29T19:33:43.339Z [INFO] TestDNS_SOA_Settings.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:43.341Z [INFO] TestDNS_SOA_Settings.server.raft: entering leader state: leader="Node at 127.0.0.1:29435 [Leader]" writer.go:29: 2021-01-29T19:33:43.344Z [INFO] TestDNS_SOA_Settings.server: cluster leadership acquired === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.345Z [INFO] TestDNSCycleRecursorCheck.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.346Z [INFO] TestDNS_SOA_Settings.server: New leader elected: payload=Node-77ae9329-c3ff-a9fb-d5cb-b02fb6bd72ce === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.346Z [INFO] TestDNSCycleRecursorCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:43.347Z [DEBUG] TestDNSCycleRecursorCheck.server: Skipping self join check for node since the cluster is too small: node=Node-2d7e3faa-c8e1-dfe6-9ef4-fa2a31278ee1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.352Z [DEBUG] TestDNS_SOA_Settings.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29435 === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.352Z [INFO] TestDNSCycleRecursorCheck.server: member joined, marking health alive: member=Node-2d7e3faa-c8e1-dfe6-9ef4-fa2a31278ee1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.365Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:43.375Z [INFO] TestDNS_SOA_Settings.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:43.376Z [INFO] TestDNS_SOA_Settings.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:43.377Z [DEBUG] TestDNS_SOA_Settings.server: Skipping self join check for node since the cluster is too small: node=Node-77ae9329-c3ff-a9fb-d5cb-b02fb6bd72ce writer.go:29: 2021-01-29T19:33:43.378Z [INFO] TestDNS_SOA_Settings.server: member joined, marking health alive: member=Node-77ae9329-c3ff-a9fb-d5cb-b02fb6bd72ce === CONT TestDNS_NodeLookup_CNAME writer.go:29: 2021-01-29T19:33:43.395Z [INFO] TestDNS_NodeLookup_CNAME: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:43.396Z [INFO] TestDNS_NodeLookup_CNAME: Endpoints down --- PASS: TestDNS_NodeLookup_CNAME (0.79s) === CONT TestDNS_NodeLookup_AAAA === CONT TestDNS_EDNS0 writer.go:29: 2021-01-29T19:33:43.399Z [INFO] TestDNS_EDNS0: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:43.400Z [INFO] TestDNS_EDNS0: Endpoints down --- PASS: TestDNS_EDNS0 (0.86s) === CONT TestDNS_NodeLookup_PeriodName === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.429Z [WARN] TestDNS_NodeLookup_AAAA: bootstrap = true: do not enable unless necessary === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.431Z [WARN] TestDNS_NodeLookup_PeriodName: bootstrap = true: do not enable unless necessary === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.431Z [DEBUG] TestDNS_NodeLookup_AAAA.tlsutil: Update: version=1 === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.432Z [DEBUG] TestDNS_NodeLookup_PeriodName.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:43.442Z [DEBUG] TestDNS_NodeLookup_PeriodName.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.443Z [DEBUG] TestDNS_NodeLookup_AAAA.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.453Z [INFO] TestDNS_NodeLookup_PeriodName.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b4d88d24-dcce-0214-e12c-18b0dd071b62 Address:127.0.0.1:29471}]" writer.go:29: 2021-01-29T19:33:43.455Z [INFO] TestDNS_NodeLookup_PeriodName.server.raft: entering follower state: follower="Node at 127.0.0.1:29471 [Follower]" leader= writer.go:29: 2021-01-29T19:33:43.456Z [INFO] TestDNS_NodeLookup_PeriodName.server.serf.wan: serf: EventMemberJoin: Node-b4d88d24-dcce-0214-e12c-18b0dd071b62.dc1 127.0.0.1 === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.461Z [INFO] TestDNS_NodeLookup_AAAA.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2dd195d7-bd5a-7d35-f242-b2932e5a69e3 Address:127.0.0.1:29453}]" writer.go:29: 2021-01-29T19:33:43.465Z [INFO] TestDNS_NodeLookup_AAAA.server.serf.wan: serf: EventMemberJoin: Node-2dd195d7-bd5a-7d35-f242-b2932e5a69e3.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:43.468Z [INFO] TestDNS_NodeLookup_AAAA.server.serf.lan: serf: EventMemberJoin: Node-2dd195d7-bd5a-7d35-f242-b2932e5a69e3 127.0.0.1 === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.473Z [INFO] TestDNS_NodeLookup_PeriodName.server.serf.lan: serf: EventMemberJoin: Node-b4d88d24-dcce-0214-e12c-18b0dd071b62 127.0.0.1 === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.471Z [INFO] TestDNS_NodeLookup_AAAA.server.raft: entering follower state: follower="Node at 127.0.0.1:29453 [Follower]" leader= writer.go:29: 2021-01-29T19:33:43.477Z [INFO] TestDNS_NodeLookup_AAAA.server: Handled event for server in area: event=member-join server=Node-2dd195d7-bd5a-7d35-f242-b2932e5a69e3.dc1 area=wan writer.go:29: 2021-01-29T19:33:43.478Z [INFO] TestDNS_NodeLookup_AAAA.server: Adding LAN server: server="Node-2dd195d7-bd5a-7d35-f242-b2932e5a69e3 (Addr: tcp/127.0.0.1:29453) (DC: dc1)" writer.go:29: 2021-01-29T19:33:43.478Z [INFO] TestDNS_NodeLookup_AAAA: Started DNS server: address=127.0.0.1:29448 network=udp === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.482Z [INFO] TestDNS_NodeLookup_PeriodName.server: Adding LAN server: server="Node-b4d88d24-dcce-0214-e12c-18b0dd071b62 (Addr: tcp/127.0.0.1:29471) (DC: dc1)" writer.go:29: 2021-01-29T19:33:43.483Z [INFO] TestDNS_NodeLookup_PeriodName: Started DNS server: address=127.0.0.1:29466 network=udp === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.484Z [DEBUG] TestDNS_SOA_Settings.dns: request served from client: name=nofoo.node.dc1.consul. type=ANY class=IN latency=52.262µs client=127.0.0.1:46169 client_network=udp === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.481Z [INFO] TestDNS_NodeLookup_AAAA: Started DNS server: address=127.0.0.1:29448 network=tcp === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.489Z [INFO] TestDNS_NodeLookup_PeriodName: Started DNS server: address=127.0.0.1:29466 network=tcp === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.490Z [INFO] TestDNS_NodeLookup_AAAA: Started HTTP server: address=127.0.0.1:29449 network=tcp writer.go:29: 2021-01-29T19:33:43.491Z [INFO] TestDNS_NodeLookup_AAAA: started state syncer === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.485Z [INFO] TestDNS_NodeLookup_PeriodName.server: Handled event for server in area: event=member-join server=Node-b4d88d24-dcce-0214-e12c-18b0dd071b62.dc1 area=wan writer.go:29: 2021-01-29T19:33:43.493Z [INFO] TestDNS_NodeLookup_PeriodName: Started HTTP server: address=127.0.0.1:29467 network=tcp writer.go:29: 2021-01-29T19:33:43.494Z [INFO] TestDNS_NodeLookup_PeriodName: started state syncer === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.498Z [INFO] TestDNS_SOA_Settings: Requesting shutdown writer.go:29: 2021-01-29T19:33:43.499Z [INFO] TestDNS_SOA_Settings.server: shutting down server writer.go:29: 2021-01-29T19:33:43.500Z [DEBUG] TestDNS_SOA_Settings.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.500Z [WARN] TestDNS_NodeLookup_PeriodName.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.501Z [WARN] TestDNS_SOA_Settings.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.501Z [INFO] TestDNS_NodeLookup_PeriodName.server.raft: entering candidate state: node="Node at 127.0.0.1:29471 [Candidate]" term=2 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.502Z [ERROR] TestDNS_SOA_Settings.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:43.502Z [DEBUG] TestDNS_SOA_Settings.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.523Z [DEBUG] TestDNS_NodeLookup_PeriodName.server.raft: votes: needed=1 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.506Z [WARN] TestDNS_SOA_Settings.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.525Z [DEBUG] TestDNS_NodeLookup_PeriodName.server.raft: vote granted: from=b4d88d24-dcce-0214-e12c-18b0dd071b62 term=2 tally=1 writer.go:29: 2021-01-29T19:33:43.527Z [INFO] TestDNS_NodeLookup_PeriodName.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:43.529Z [INFO] TestDNS_NodeLookup_PeriodName.server.raft: entering leader state: leader="Node at 127.0.0.1:29471 [Leader]" === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.529Z [INFO] TestDNS_SOA_Settings.server.router.manager: shutting down === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.531Z [INFO] TestDNS_NodeLookup_PeriodName.server: cluster leadership acquired === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.529Z [INFO] TestDNS_SOA_Settings: consul server down writer.go:29: 2021-01-29T19:33:43.536Z [INFO] TestDNS_SOA_Settings: shutdown complete writer.go:29: 2021-01-29T19:33:43.537Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=DNS address=127.0.0.1:29430 network=tcp === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.531Z [INFO] TestDNS_NodeLookup_PeriodName.server: New leader elected: payload=Node-b4d88d24-dcce-0214-e12c-18b0dd071b62 === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:43.538Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=DNS address=127.0.0.1:29430 network=udp writer.go:29: 2021-01-29T19:33:43.540Z [INFO] TestDNS_SOA_Settings: Stopping server: protocol=HTTP address=127.0.0.1:29431 network=tcp === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.534Z [DEBUG] TestDNS_NodeLookup_PeriodName.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29471 === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.547Z [WARN] TestDNS_NodeLookup_AAAA.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:43.551Z [INFO] TestDNS_NodeLookup_AAAA.server.raft: entering candidate state: node="Node at 127.0.0.1:29453 [Candidate]" term=2 === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.558Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.558Z [DEBUG] TestDNS_NodeLookup_AAAA.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:43.561Z [DEBUG] TestDNS_NodeLookup_AAAA.server.raft: vote granted: from=2dd195d7-bd5a-7d35-f242-b2932e5a69e3 term=2 tally=1 writer.go:29: 2021-01-29T19:33:43.564Z [INFO] TestDNS_NodeLookup_AAAA.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:43.566Z [INFO] TestDNS_NodeLookup_AAAA.server.raft: entering leader state: leader="Node at 127.0.0.1:29453 [Leader]" === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.568Z [INFO] TestDNS_NodeLookup_PeriodName.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.568Z [INFO] TestDNS_NodeLookup_AAAA.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:43.569Z [INFO] TestDNS_NodeLookup_AAAA.server: New leader elected: payload=Node-2dd195d7-bd5a-7d35-f242-b2932e5a69e3 === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.570Z [INFO] TestDNS_NodeLookup_PeriodName.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:43.571Z [DEBUG] TestDNS_NodeLookup_PeriodName.server: Skipping self join check for node since the cluster is too small: node=Node-b4d88d24-dcce-0214-e12c-18b0dd071b62 writer.go:29: 2021-01-29T19:33:43.573Z [INFO] TestDNS_NodeLookup_PeriodName.server: member joined, marking health alive: member=Node-b4d88d24-dcce-0214-e12c-18b0dd071b62 === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.572Z [DEBUG] TestDNS_NodeLookup_AAAA.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29453 writer.go:29: 2021-01-29T19:33:43.594Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:43.597Z [INFO] TestDNS_NodeLookup_AAAA.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:43.598Z [INFO] TestDNS_NodeLookup_AAAA.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:43.598Z [DEBUG] TestDNS_NodeLookup_AAAA.server: Skipping self join check for node since the cluster is too small: node=Node-2dd195d7-bd5a-7d35-f242-b2932e5a69e3 writer.go:29: 2021-01-29T19:33:43.599Z [INFO] TestDNS_NodeLookup_AAAA.server: member joined, marking health alive: member=Node-2dd195d7-bd5a-7d35-f242-b2932e5a69e3 === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.600Z [DEBUG] TestDNS_NodeLookup_PeriodName.dns: request served from client: name=foo.bar.node.consul. type=ANY class=IN latency=114.769µs client=127.0.0.1:54483 client_network=udp writer.go:29: 2021-01-29T19:33:43.600Z [INFO] TestDNS_NodeLookup_PeriodName: Requesting shutdown writer.go:29: 2021-01-29T19:33:43.605Z [INFO] TestDNS_NodeLookup_PeriodName.server: shutting down server writer.go:29: 2021-01-29T19:33:43.606Z [DEBUG] TestDNS_NodeLookup_PeriodName.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:43.608Z [WARN] TestDNS_NodeLookup_PeriodName.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.609Z [DEBUG] TestDNSCycleRecursorCheck.dns: recurse failed for question: question="{google.com. 1 1}" rtt=475.007µs recursor=127.0.0.1:59117 rcode=SERVFAIL === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.607Z [ERROR] TestDNS_NodeLookup_PeriodName.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:43.608Z [DEBUG] TestDNS_NodeLookup_PeriodName.leader: stopped routine: routine="CA root pruning" === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.613Z [DEBUG] TestDNSCycleRecursorCheck.dns: recurse succeeded for question: question="{google.com. 1 1}" rtt=663.834µs recursor=127.0.0.1:59819 writer.go:29: 2021-01-29T19:33:43.616Z [INFO] TestDNSCycleRecursorCheck: Requesting shutdown === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.611Z [WARN] TestDNS_NodeLookup_PeriodName.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.616Z [DEBUG] TestDNSCycleRecursorCheck.dns: request served from client: question="{google.com. 1 1}" network=udp latency=7.837207ms client=127.0.0.1:46322 client_network=udp writer.go:29: 2021-01-29T19:33:43.618Z [INFO] TestDNSCycleRecursorCheck.server: shutting down server writer.go:29: 2021-01-29T19:33:43.629Z [DEBUG] TestDNSCycleRecursorCheck.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.629Z [INFO] TestDNS_NodeLookup_PeriodName.server.router.manager: shutting down === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.629Z [DEBUG] TestDNSCycleRecursorCheck: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.630Z [INFO] TestDNS_NodeLookup_PeriodName: consul server down === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.631Z [WARN] TestDNSCycleRecursorCheck.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.634Z [INFO] TestDNS_NodeLookup_PeriodName: shutdown complete === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.631Z [DEBUG] TestDNSCycleRecursorCheck.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.635Z [INFO] TestDNS_NodeLookup_PeriodName: Stopping server: protocol=DNS address=127.0.0.1:29466 network=tcp === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.633Z [INFO] TestDNSCycleRecursorCheck: Synced node info === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.637Z [INFO] TestDNS_NodeLookup_PeriodName: Stopping server: protocol=DNS address=127.0.0.1:29466 network=udp === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.638Z [DEBUG] TestDNSCycleRecursorCheck: Node info in sync === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:43.639Z [INFO] TestDNS_NodeLookup_PeriodName: Stopping server: protocol=HTTP address=127.0.0.1:29467 network=tcp === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.641Z [WARN] TestDNSCycleRecursorCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:43.644Z [INFO] TestDNSCycleRecursorCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:43.645Z [INFO] TestDNSCycleRecursorCheck: consul server down writer.go:29: 2021-01-29T19:33:43.646Z [INFO] TestDNSCycleRecursorCheck: shutdown complete writer.go:29: 2021-01-29T19:33:43.646Z [INFO] TestDNSCycleRecursorCheck: Stopping server: protocol=DNS address=127.0.0.1:29442 network=tcp === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.645Z [DEBUG] TestDNS_NodeLookup_AAAA: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:43.647Z [INFO] TestDNSCycleRecursorCheck: Stopping server: protocol=DNS address=127.0.0.1:29442 network=udp writer.go:29: 2021-01-29T19:33:43.649Z [INFO] TestDNSCycleRecursorCheck: Stopping server: protocol=HTTP address=127.0.0.1:29443 network=tcp === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:43.651Z [INFO] TestDNS_NodeLookup_AAAA: Synced node info writer.go:29: 2021-01-29T19:33:43.653Z [DEBUG] TestDNS_NodeLookup_AAAA: Node info in sync writer.go:29: 2021-01-29T19:33:43.740Z [DEBUG] TestDNS_NodeLookup_AAAA.dns: request served from client: name=bar.node.consul. type=AAAA class=IN latency=99.49µs client=127.0.0.1:35261 client_network=udp writer.go:29: 2021-01-29T19:33:43.740Z [INFO] TestDNS_NodeLookup_AAAA: Requesting shutdown writer.go:29: 2021-01-29T19:33:43.742Z [INFO] TestDNS_NodeLookup_AAAA.server: shutting down server writer.go:29: 2021-01-29T19:33:43.743Z [DEBUG] TestDNS_NodeLookup_AAAA.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:43.744Z [WARN] TestDNS_NodeLookup_AAAA.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:43.744Z [DEBUG] TestDNS_NodeLookup_AAAA.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:43.746Z [WARN] TestDNS_NodeLookup_AAAA.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:43.750Z [INFO] TestDNS_NodeLookup_AAAA.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:43.750Z [INFO] TestDNS_NodeLookup_AAAA: consul server down writer.go:29: 2021-01-29T19:33:43.752Z [INFO] TestDNS_NodeLookup_AAAA: shutdown complete writer.go:29: 2021-01-29T19:33:43.753Z [INFO] TestDNS_NodeLookup_AAAA: Stopping server: protocol=DNS address=127.0.0.1:29448 network=tcp writer.go:29: 2021-01-29T19:33:43.753Z [INFO] TestDNS_NodeLookup_AAAA: Stopping server: protocol=DNS address=127.0.0.1:29448 network=udp writer.go:29: 2021-01-29T19:33:43.755Z [INFO] TestDNS_NodeLookup_AAAA: Stopping server: protocol=HTTP address=127.0.0.1:29449 network=tcp === CONT TestDNS_SOA_Settings writer.go:29: 2021-01-29T19:33:44.042Z [INFO] TestDNS_SOA_Settings: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:44.043Z [INFO] TestDNS_SOA_Settings: Endpoints down --- PASS: TestDNS_SOA_Settings (3.20s) === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.133Z [WARN] TestDNS_CaseInsensitiveNodeLookup: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:44.152Z [DEBUG] TestDNS_CaseInsensitiveNodeLookup.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:44.155Z [DEBUG] TestDNS_CaseInsensitiveNodeLookup.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_NodeLookup_PeriodName writer.go:29: 2021-01-29T19:33:44.156Z [INFO] TestDNS_NodeLookup_PeriodName: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:44.157Z [INFO] TestDNS_NodeLookup_PeriodName: Endpoints down --- PASS: TestDNS_NodeLookup_PeriodName (0.76s) === CONT TestDNS_EmptyAltDomain === CONT TestDNSCycleRecursorCheck writer.go:29: 2021-01-29T19:33:44.162Z [INFO] TestDNSCycleRecursorCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:44.164Z [INFO] TestDNSCycleRecursorCheck: Endpoints down --- PASS: TestDNSCycleRecursorCheck (1.01s) === CONT TestDNS_Over_TCP === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.211Z [WARN] TestDNS_EmptyAltDomain: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:44.223Z [DEBUG] TestDNS_EmptyAltDomain.tlsutil: Update: version=1 === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.225Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e2ed11f6-f84b-992d-6e47-dff24a3c0cab Address:127.0.0.1:29459}]" === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.226Z [WARN] TestDNS_Over_TCP: bootstrap = true: do not enable unless necessary === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.228Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server.serf.wan: serf: EventMemberJoin: Node-e2ed11f6-f84b-992d-6e47-dff24a3c0cab.dc1 127.0.0.1 === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.226Z [DEBUG] TestDNS_EmptyAltDomain.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.228Z [DEBUG] TestDNS_Over_TCP.tlsutil: Update: version=1 === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.242Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server.serf.lan: serf: EventMemberJoin: Node-e2ed11f6-f84b-992d-6e47-dff24a3c0cab 127.0.0.1 writer.go:29: 2021-01-29T19:33:44.252Z [INFO] TestDNS_CaseInsensitiveNodeLookup: Started DNS server: address=127.0.0.1:29454 network=udp === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.270Z [DEBUG] TestDNS_Over_TCP.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.277Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server: Handled event for server in area: event=member-join server=Node-e2ed11f6-f84b-992d-6e47-dff24a3c0cab.dc1 area=wan === CONT TestDNS_NodeLookup_AAAA writer.go:29: 2021-01-29T19:33:44.264Z [INFO] TestDNS_NodeLookup_AAAA: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:44.296Z [INFO] TestDNS_NodeLookup_AAAA: Endpoints down --- PASS: TestDNS_NodeLookup_AAAA (0.90s) === CONT TestRecursorAddr --- PASS: TestRecursorAddr (0.00s) === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.279Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server.raft: entering follower state: follower="Node at 127.0.0.1:29459 [Follower]" leader= writer.go:29: 2021-01-29T19:33:44.338Z [INFO] TestDNS_CaseInsensitiveNodeLookup: Started DNS server: address=127.0.0.1:29454 network=tcp === CONT TestDiscoveryChainRead === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.363Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server: Adding LAN server: server="Node-e2ed11f6-f84b-992d-6e47-dff24a3c0cab (Addr: tcp/127.0.0.1:29459) (DC: dc1)" writer.go:29: 2021-01-29T19:33:44.395Z [INFO] TestDNS_CaseInsensitiveNodeLookup: Started HTTP server: address=127.0.0.1:29455 network=tcp writer.go:29: 2021-01-29T19:33:44.396Z [INFO] TestDNS_CaseInsensitiveNodeLookup: started state syncer === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.396Z [INFO] TestDNS_Over_TCP.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fcdd8dc5-0e47-ad55-a8b6-5278ae53a67b Address:127.0.0.1:29477}]" writer.go:29: 2021-01-29T19:33:44.400Z [INFO] TestDNS_Over_TCP.server.raft: entering follower state: follower="Node at 127.0.0.1:29477 [Follower]" leader= writer.go:29: 2021-01-29T19:33:44.404Z [INFO] TestDNS_Over_TCP.server.serf.wan: serf: EventMemberJoin: Node-fcdd8dc5-0e47-ad55-a8b6-5278ae53a67b.dc1 127.0.0.1 === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.406Z [WARN] TestDNS_CaseInsensitiveNodeLookup.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:44.407Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server.raft: entering candidate state: node="Node at 127.0.0.1:29459 [Candidate]" term=2 === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.410Z [INFO] TestDNS_EmptyAltDomain.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:95bb7162-b2b9-b576-400d-92b79ebd5f82 Address:127.0.0.1:29465}]" === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.414Z [INFO] TestDNS_Over_TCP.server.serf.lan: serf: EventMemberJoin: Node-fcdd8dc5-0e47-ad55-a8b6-5278ae53a67b 127.0.0.1 === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.414Z [INFO] TestDNS_EmptyAltDomain.server.serf.wan: serf: EventMemberJoin: Node-95bb7162-b2b9-b576-400d-92b79ebd5f82.dc1 127.0.0.1 === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.417Z [DEBUG] TestDNS_CaseInsensitiveNodeLookup.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:44.418Z [DEBUG] TestDNS_CaseInsensitiveNodeLookup.server.raft: vote granted: from=e2ed11f6-f84b-992d-6e47-dff24a3c0cab term=2 tally=1 writer.go:29: 2021-01-29T19:33:44.419Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server.raft: election won: tally=1 === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.418Z [INFO] TestDNS_EmptyAltDomain.server.serf.lan: serf: EventMemberJoin: Node-95bb7162-b2b9-b576-400d-92b79ebd5f82 127.0.0.1 === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.420Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server.raft: entering leader state: leader="Node at 127.0.0.1:29459 [Leader]" === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.422Z [WARN] TestDiscoveryChainRead: bootstrap = true: do not enable unless necessary === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.422Z [INFO] TestDNS_Over_TCP: Started DNS server: address=127.0.0.1:29472 network=udp writer.go:29: 2021-01-29T19:33:44.424Z [INFO] TestDNS_Over_TCP.server: Adding LAN server: server="Node-fcdd8dc5-0e47-ad55-a8b6-5278ae53a67b (Addr: tcp/127.0.0.1:29477) (DC: dc1)" writer.go:29: 2021-01-29T19:33:44.425Z [INFO] TestDNS_Over_TCP.server: Handled event for server in area: event=member-join server=Node-fcdd8dc5-0e47-ad55-a8b6-5278ae53a67b.dc1 area=wan === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.427Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:44.429Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server: New leader elected: payload=Node-e2ed11f6-f84b-992d-6e47-dff24a3c0cab === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.429Z [DEBUG] TestDiscoveryChainRead.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:44.430Z [DEBUG] TestDiscoveryChainRead.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.431Z [INFO] TestDNS_EmptyAltDomain.server.raft: entering follower state: follower="Node at 127.0.0.1:29465 [Follower]" leader= writer.go:29: 2021-01-29T19:33:44.433Z [INFO] TestDNS_EmptyAltDomain.server: Adding LAN server: server="Node-95bb7162-b2b9-b576-400d-92b79ebd5f82 (Addr: tcp/127.0.0.1:29465) (DC: dc1)" === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.427Z [INFO] TestDNS_Over_TCP: Started DNS server: address=127.0.0.1:29472 network=tcp === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.435Z [DEBUG] TestDNS_CaseInsensitiveNodeLookup.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29459 === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.433Z [INFO] TestDNS_EmptyAltDomain: Started DNS server: address=127.0.0.1:29460 network=tcp === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.436Z [INFO] TestDNS_Over_TCP: Started HTTP server: address=127.0.0.1:29473 network=tcp === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.436Z [INFO] TestDNS_EmptyAltDomain: Started DNS server: address=127.0.0.1:29460 network=udp === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.437Z [INFO] TestDNS_Over_TCP: started state syncer === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.433Z [INFO] TestDNS_EmptyAltDomain.server: Handled event for server in area: event=member-join server=Node-95bb7162-b2b9-b576-400d-92b79ebd5f82.dc1 area=wan writer.go:29: 2021-01-29T19:33:44.440Z [INFO] TestDNS_EmptyAltDomain: Started HTTP server: address=127.0.0.1:29461 network=tcp === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.440Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.441Z [INFO] TestDNS_EmptyAltDomain: started state syncer === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.442Z [INFO] TestDiscoveryChainRead.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:307ef1d3-e649-181c-ff8e-956d595d1c93 Address:127.0.0.1:29489}]" writer.go:29: 2021-01-29T19:33:44.443Z [INFO] TestDiscoveryChainRead.server.serf.wan: serf: EventMemberJoin: Node-307ef1d3-e649-181c-ff8e-956d595d1c93.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:44.443Z [INFO] TestDiscoveryChainRead.server.raft: entering follower state: follower="Node at 127.0.0.1:29489 [Follower]" leader= writer.go:29: 2021-01-29T19:33:44.447Z [INFO] TestDiscoveryChainRead.server.serf.lan: serf: EventMemberJoin: Node-307ef1d3-e649-181c-ff8e-956d595d1c93 127.0.0.1 writer.go:29: 2021-01-29T19:33:44.450Z [INFO] TestDiscoveryChainRead.server: Handled event for server in area: event=member-join server=Node-307ef1d3-e649-181c-ff8e-956d595d1c93.dc1 area=wan === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.450Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.451Z [INFO] TestDiscoveryChainRead.server: Adding LAN server: server="Node-307ef1d3-e649-181c-ff8e-956d595d1c93 (Addr: tcp/127.0.0.1:29489) (DC: dc1)" === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.451Z [INFO] TestDNS_CaseInsensitiveNodeLookup.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.452Z [DEBUG] TestDNS_CaseInsensitiveNodeLookup.server: Skipping self join check for node since the cluster is too small: node=Node-e2ed11f6-f84b-992d-6e47-dff24a3c0cab === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.454Z [WARN] TestDNS_Over_TCP.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.454Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server: member joined, marking health alive: member=Node-e2ed11f6-f84b-992d-6e47-dff24a3c0cab === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.455Z [INFO] TestDNS_Over_TCP.server.raft: entering candidate state: node="Node at 127.0.0.1:29477 [Candidate]" term=2 === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.453Z [INFO] TestDiscoveryChainRead: Started DNS server: address=127.0.0.1:29484 network=tcp writer.go:29: 2021-01-29T19:33:44.456Z [INFO] TestDiscoveryChainRead: Started DNS server: address=127.0.0.1:29484 network=udp writer.go:29: 2021-01-29T19:33:44.458Z [INFO] TestDiscoveryChainRead: Started HTTP server: address=127.0.0.1:29485 network=tcp === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.460Z [DEBUG] TestDNS_Over_TCP.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:44.461Z [DEBUG] TestDNS_Over_TCP.server.raft: vote granted: from=fcdd8dc5-0e47-ad55-a8b6-5278ae53a67b term=2 tally=1 writer.go:29: 2021-01-29T19:33:44.462Z [INFO] TestDNS_Over_TCP.server.raft: election won: tally=1 === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.459Z [INFO] TestDiscoveryChainRead: started state syncer === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.465Z [INFO] TestDNS_Over_TCP.server.raft: entering leader state: leader="Node at 127.0.0.1:29477 [Leader]" writer.go:29: 2021-01-29T19:33:44.467Z [INFO] TestDNS_Over_TCP.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:44.468Z [INFO] TestDNS_Over_TCP.server: New leader elected: payload=Node-fcdd8dc5-0e47-ad55-a8b6-5278ae53a67b writer.go:29: 2021-01-29T19:33:44.476Z [DEBUG] TestDNS_Over_TCP.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29477 writer.go:29: 2021-01-29T19:33:44.479Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:44.484Z [INFO] TestDNS_Over_TCP.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:44.485Z [INFO] TestDNS_Over_TCP.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.486Z [DEBUG] TestDNS_Over_TCP.server: Skipping self join check for node since the cluster is too small: node=Node-fcdd8dc5-0e47-ad55-a8b6-5278ae53a67b writer.go:29: 2021-01-29T19:33:44.487Z [INFO] TestDNS_Over_TCP.server: member joined, marking health alive: member=Node-fcdd8dc5-0e47-ad55-a8b6-5278ae53a67b === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.492Z [WARN] TestDNS_EmptyAltDomain.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.498Z [DEBUG] TestDNS_Over_TCP.dns: request served from client: name=foo.node.dc1.consul. type=ANY class=IN latency=879.364µs client=127.0.0.1:41988 client_network=tcp writer.go:29: 2021-01-29T19:33:44.499Z [INFO] TestDNS_Over_TCP: Requesting shutdown writer.go:29: 2021-01-29T19:33:44.501Z [INFO] TestDNS_Over_TCP.server: shutting down server writer.go:29: 2021-01-29T19:33:44.502Z [DEBUG] TestDNS_Over_TCP.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.503Z [WARN] TestDNS_Over_TCP.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:44.502Z [ERROR] TestDNS_Over_TCP.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.493Z [INFO] TestDNS_EmptyAltDomain.server.raft: entering candidate state: node="Node at 127.0.0.1:29465 [Candidate]" term=2 === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.503Z [DEBUG] TestDNS_Over_TCP.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.505Z [WARN] TestDNS_Over_TCP.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.509Z [DEBUG] TestDNS_EmptyAltDomain.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:44.510Z [DEBUG] TestDNS_EmptyAltDomain.server.raft: vote granted: from=95bb7162-b2b9-b576-400d-92b79ebd5f82 term=2 tally=1 === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.511Z [INFO] TestDNS_Over_TCP.server.router.manager: shutting down === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.512Z [INFO] TestDNS_EmptyAltDomain.server.raft: election won: tally=1 === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.511Z [INFO] TestDNS_Over_TCP: consul server down writer.go:29: 2021-01-29T19:33:44.513Z [INFO] TestDNS_Over_TCP: shutdown complete === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.513Z [INFO] TestDNS_EmptyAltDomain.server.raft: entering leader state: leader="Node at 127.0.0.1:29465 [Leader]" === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.513Z [INFO] TestDNS_Over_TCP: Stopping server: protocol=DNS address=127.0.0.1:29472 network=tcp === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.514Z [INFO] TestDNS_EmptyAltDomain.server: cluster leadership acquired === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.514Z [INFO] TestDNS_Over_TCP: Stopping server: protocol=DNS address=127.0.0.1:29472 network=udp === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.515Z [INFO] TestDNS_EmptyAltDomain.server: New leader elected: payload=Node-95bb7162-b2b9-b576-400d-92b79ebd5f82 === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:44.515Z [INFO] TestDNS_Over_TCP: Stopping server: protocol=HTTP address=127.0.0.1:29473 network=tcp === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.515Z [WARN] TestDiscoveryChainRead.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:44.517Z [INFO] TestDiscoveryChainRead.server.raft: entering candidate state: node="Node at 127.0.0.1:29489 [Candidate]" term=2 === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.516Z [DEBUG] TestDNS_EmptyAltDomain.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29465 === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.528Z [DEBUG] TestDiscoveryChainRead.server.raft: votes: needed=1 === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.529Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.530Z [DEBUG] TestDiscoveryChainRead.server.raft: vote granted: from=307ef1d3-e649-181c-ff8e-956d595d1c93 term=2 tally=1 writer.go:29: 2021-01-29T19:33:44.533Z [INFO] TestDiscoveryChainRead.server.raft: election won: tally=1 === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.534Z [INFO] TestDNS_EmptyAltDomain.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:44.535Z [INFO] TestDNS_EmptyAltDomain.leader: started routine: routine="CA root pruning" === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.535Z [INFO] TestDiscoveryChainRead.server.raft: entering leader state: leader="Node at 127.0.0.1:29489 [Leader]" === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.536Z [DEBUG] TestDNS_EmptyAltDomain.server: Skipping self join check for node since the cluster is too small: node=Node-95bb7162-b2b9-b576-400d-92b79ebd5f82 === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.537Z [INFO] TestDiscoveryChainRead.server: cluster leadership acquired === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.537Z [INFO] TestDNS_EmptyAltDomain.server: member joined, marking health alive: member=Node-95bb7162-b2b9-b576-400d-92b79ebd5f82 === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.537Z [INFO] TestDiscoveryChainRead.server: New leader elected: payload=Node-307ef1d3-e649-181c-ff8e-956d595d1c93 writer.go:29: 2021-01-29T19:33:44.539Z [DEBUG] TestDiscoveryChainRead.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29489 writer.go:29: 2021-01-29T19:33:44.550Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:44.562Z [INFO] TestDiscoveryChainRead.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:44.564Z [INFO] TestDiscoveryChainRead.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.565Z [DEBUG] TestDiscoveryChainRead.server: Skipping self join check for node since the cluster is too small: node=Node-307ef1d3-e649-181c-ff8e-956d595d1c93 writer.go:29: 2021-01-29T19:33:44.566Z [INFO] TestDiscoveryChainRead.server: member joined, marking health alive: member=Node-307ef1d3-e649-181c-ff8e-956d595d1c93 === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:44.573Z [INFO] TestDNS_EmptyAltDomain: Requesting shutdown writer.go:29: 2021-01-29T19:33:44.574Z [INFO] TestDNS_EmptyAltDomain.server: shutting down server writer.go:29: 2021-01-29T19:33:44.575Z [DEBUG] TestDNS_EmptyAltDomain.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.575Z [ERROR] TestDNS_EmptyAltDomain.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:44.576Z [WARN] TestDNS_EmptyAltDomain.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:44.576Z [DEBUG] TestDNS_EmptyAltDomain.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.580Z [WARN] TestDNS_EmptyAltDomain.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:44.589Z [INFO] TestDNS_EmptyAltDomain.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:44.589Z [INFO] TestDNS_EmptyAltDomain: consul server down writer.go:29: 2021-01-29T19:33:44.592Z [INFO] TestDNS_EmptyAltDomain: shutdown complete writer.go:29: 2021-01-29T19:33:44.594Z [INFO] TestDNS_EmptyAltDomain: Stopping server: protocol=DNS address=127.0.0.1:29460 network=tcp writer.go:29: 2021-01-29T19:33:44.596Z [INFO] TestDNS_EmptyAltDomain: Stopping server: protocol=DNS address=127.0.0.1:29460 network=udp writer.go:29: 2021-01-29T19:33:44.599Z [INFO] TestDNS_EmptyAltDomain: Stopping server: protocol=HTTP address=127.0.0.1:29461 network=tcp === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.613Z [DEBUG] TestDiscoveryChainRead: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:44.617Z [INFO] TestDiscoveryChainRead: Synced node info === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:44.640Z [DEBUG] TestDNS_CaseInsensitiveNodeLookup.dns: request served from client: name=fOO.node.dc1.consul. type=ANY class=IN latency=96.637µs client=127.0.0.1:57956 client_network=udp writer.go:29: 2021-01-29T19:33:44.640Z [INFO] TestDNS_CaseInsensitiveNodeLookup: Requesting shutdown writer.go:29: 2021-01-29T19:33:44.645Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server: shutting down server writer.go:29: 2021-01-29T19:33:44.646Z [DEBUG] TestDNS_CaseInsensitiveNodeLookup.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.648Z [WARN] TestDNS_CaseInsensitiveNodeLookup.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:44.646Z [ERROR] TestDNS_CaseInsensitiveNodeLookup.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:44.648Z [DEBUG] TestDNS_CaseInsensitiveNodeLookup.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.651Z [WARN] TestDNS_CaseInsensitiveNodeLookup.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:44.665Z [INFO] TestDNS_CaseInsensitiveNodeLookup.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:44.666Z [INFO] TestDNS_CaseInsensitiveNodeLookup: consul server down writer.go:29: 2021-01-29T19:33:44.667Z [INFO] TestDNS_CaseInsensitiveNodeLookup: shutdown complete writer.go:29: 2021-01-29T19:33:44.668Z [INFO] TestDNS_CaseInsensitiveNodeLookup: Stopping server: protocol=DNS address=127.0.0.1:29454 network=tcp writer.go:29: 2021-01-29T19:33:44.669Z [INFO] TestDNS_CaseInsensitiveNodeLookup: Stopping server: protocol=DNS address=127.0.0.1:29454 network=udp writer.go:29: 2021-01-29T19:33:44.670Z [INFO] TestDNS_CaseInsensitiveNodeLookup: Stopping server: protocol=HTTP address=127.0.0.1:29455 network=tcp === RUN TestDiscoveryChainRead/GET:_error_on_no_service_name === RUN TestDiscoveryChainRead/GET:_read_default_chain === RUN TestDiscoveryChainRead/GET:_read_default_chain;_evaluate_in_dc2 === RUN TestDiscoveryChainRead/GET:_read_default_chain_with_cache === RUN TestDiscoveryChainRead/POST:_error_on_no_service_name === RUN TestDiscoveryChainRead/POST:_read_default_chain === RUN TestDiscoveryChainRead/POST:_read_default_chain;_evaluate_in_dc2 === RUN TestDiscoveryChainRead/POST:_read_default_chain_with_cache === RUN TestDiscoveryChainRead/GET:_read_modified_chain === RUN TestDiscoveryChainRead/POST:_read_modified_chain_with_overrides_(camel_case) === RUN TestDiscoveryChainRead/POST:_read_modified_chain_with_overrides_(snake_case) === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:44.815Z [INFO] TestDiscoveryChainRead: Requesting shutdown writer.go:29: 2021-01-29T19:33:44.816Z [INFO] TestDiscoveryChainRead.server: shutting down server writer.go:29: 2021-01-29T19:33:44.816Z [DEBUG] TestDiscoveryChainRead.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.817Z [WARN] TestDiscoveryChainRead.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:44.817Z [DEBUG] TestDiscoveryChainRead.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:44.821Z [WARN] TestDiscoveryChainRead.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:44.836Z [INFO] TestDiscoveryChainRead.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:44.838Z [INFO] TestDiscoveryChainRead: consul server down writer.go:29: 2021-01-29T19:33:44.840Z [INFO] TestDiscoveryChainRead: shutdown complete writer.go:29: 2021-01-29T19:33:44.842Z [INFO] TestDiscoveryChainRead: Stopping server: protocol=DNS address=127.0.0.1:29484 network=tcp writer.go:29: 2021-01-29T19:33:44.845Z [INFO] TestDiscoveryChainRead: Stopping server: protocol=DNS address=127.0.0.1:29484 network=udp writer.go:29: 2021-01-29T19:33:44.846Z [INFO] TestDiscoveryChainRead: Stopping server: protocol=HTTP address=127.0.0.1:29485 network=tcp === CONT TestDNS_Over_TCP writer.go:29: 2021-01-29T19:33:45.016Z [INFO] TestDNS_Over_TCP: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:45.018Z [INFO] TestDNS_Over_TCP: Endpoints down --- PASS: TestDNS_Over_TCP (0.85s) === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.038Z [WARN] TestCoordinate_Update_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:33:45.042Z [WARN] TestCoordinate_Update_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:45.049Z [DEBUG] TestCoordinate_Update_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:45.054Z [DEBUG] TestCoordinate_Update_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:45.062Z [INFO] TestCoordinate_Update_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb Address:127.0.0.1:29483}]" writer.go:29: 2021-01-29T19:33:45.065Z [INFO] TestCoordinate_Update_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:29483 [Follower]" leader= writer.go:29: 2021-01-29T19:33:45.065Z [INFO] TestCoordinate_Update_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:45.074Z [INFO] TestCoordinate_Update_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb 127.0.0.1 writer.go:29: 2021-01-29T19:33:45.077Z [INFO] TestCoordinate_Update_ACLDeny.server: Handled event for server in area: event=member-join server=Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb.dc1 area=wan writer.go:29: 2021-01-29T19:33:45.078Z [INFO] TestCoordinate_Update_ACLDeny.server: Adding LAN server: server="Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb (Addr: tcp/127.0.0.1:29483) (DC: dc1)" writer.go:29: 2021-01-29T19:33:45.080Z [INFO] TestCoordinate_Update_ACLDeny: Started DNS server: address=127.0.0.1:29478 network=udp writer.go:29: 2021-01-29T19:33:45.081Z [INFO] TestCoordinate_Update_ACLDeny: Started DNS server: address=127.0.0.1:29478 network=tcp writer.go:29: 2021-01-29T19:33:45.082Z [INFO] TestCoordinate_Update_ACLDeny: Started HTTP server: address=127.0.0.1:29479 network=tcp writer.go:29: 2021-01-29T19:33:45.085Z [INFO] TestCoordinate_Update_ACLDeny: started state syncer === CONT TestDNS_EmptyAltDomain writer.go:29: 2021-01-29T19:33:45.101Z [INFO] TestDNS_EmptyAltDomain: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:45.103Z [INFO] TestDNS_EmptyAltDomain: Endpoints down --- PASS: TestDNS_EmptyAltDomain (0.95s) === CONT TestCoordinate_Update === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.109Z [WARN] TestCoordinate_Update_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:45.111Z [INFO] TestCoordinate_Update_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:29483 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:45.113Z [DEBUG] TestCoordinate_Update_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:45.115Z [DEBUG] TestCoordinate_Update_ACLDeny.server.raft: vote granted: from=7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb term=2 tally=1 === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.117Z [WARN] TestCoordinate_Update: bootstrap = true: do not enable unless necessary === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.117Z [INFO] TestCoordinate_Update_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:45.119Z [INFO] TestCoordinate_Update_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:29483 [Leader]" writer.go:29: 2021-01-29T19:33:45.121Z [INFO] TestCoordinate_Update_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:45.121Z [INFO] TestCoordinate_Update_ACLDeny.server: New leader elected: payload=Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb writer.go:29: 2021-01-29T19:33:45.122Z [INFO] TestCoordinate_Update_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:33:45.131Z [INFO] TestCoordinate_Update_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:33:45.153Z [INFO] TestCoordinate_Update_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:45.154Z [WARN] TestCoordinate_Update_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:33:45.153Z [INFO] TestCoordinate_Update_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:33:45.157Z [WARN] TestCoordinate_Update_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:33:45.157Z [INFO] TestCoordinate_Update_ACLDeny.server: Bootstrapped ACL master token from configuration === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.119Z [DEBUG] TestCoordinate_Update.tlsutil: Update: version=1 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.165Z [INFO] TestCoordinate_Update_ACLDeny.server: Created ACL anonymous token from configuration === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.162Z [DEBUG] TestCoordinate_Update.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.166Z [INFO] TestCoordinate_Update_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:45.165Z [INFO] TestCoordinate_Update_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:33:45.168Z [INFO] TestCoordinate_Update_ACLDeny.leader: started routine: routine="acl token reaping" === CONT TestDNS_CaseInsensitiveNodeLookup writer.go:29: 2021-01-29T19:33:45.171Z [INFO] TestDNS_CaseInsensitiveNodeLookup: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:45.172Z [INFO] TestDNS_CaseInsensitiveNodeLookup: Endpoints down --- PASS: TestDNS_CaseInsensitiveNodeLookup (1.13s) === CONT TestCoordinate_Node === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.182Z [INFO] TestCoordinate_Update.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:560f7dcd-33a2-73c3-de5d-7d2ed99598f7 Address:127.0.0.1:29495}]" === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.170Z [INFO] TestCoordinate_Update_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.183Z [WARN] TestCoordinate_Node: bootstrap = true: do not enable unless necessary === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.171Z [DEBUG] TestCoordinate_Update_ACLDeny.server: transitioning out of legacy ACL mode === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.185Z [DEBUG] TestCoordinate_Node.tlsutil: Update: version=1 === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.184Z [INFO] TestCoordinate_Update.server.raft: entering follower state: follower="Node at 127.0.0.1:29495 [Follower]" leader= === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.183Z [INFO] TestCoordinate_Update_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb.dc1 writer.go:29: 2021-01-29T19:33:45.183Z [INFO] TestCoordinate_Update_ACLDeny.server: Updating LAN server: server="Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb (Addr: tcp/127.0.0.1:29483) (DC: dc1)" writer.go:29: 2021-01-29T19:33:45.185Z [INFO] TestCoordinate_Update_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb writer.go:29: 2021-01-29T19:33:45.191Z [INFO] TestCoordinate_Update_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb.dc1 writer.go:29: 2021-01-29T19:33:45.192Z [INFO] TestCoordinate_Update_ACLDeny.server: Handled event for server in area: event=member-update server=Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb.dc1 area=wan === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.194Z [DEBUG] TestCoordinate_Node.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.195Z [INFO] TestCoordinate_Update_ACLDeny.server: Handled event for server in area: event=member-update server=Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb.dc1 area=wan === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.185Z [INFO] TestCoordinate_Update.server.serf.wan: serf: EventMemberJoin: Node-560f7dcd-33a2-73c3-de5d-7d2ed99598f7.dc1 127.0.0.1 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.197Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.201Z [INFO] TestCoordinate_Update.server.serf.lan: serf: EventMemberJoin: Node-560f7dcd-33a2-73c3-de5d-7d2ed99598f7 127.0.0.1 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.193Z [INFO] TestCoordinate_Update_ACLDeny.server: Updating LAN server: server="Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb (Addr: tcp/127.0.0.1:29483) (DC: dc1)" === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.207Z [INFO] TestCoordinate_Update.server: Adding LAN server: server="Node-560f7dcd-33a2-73c3-de5d-7d2ed99598f7 (Addr: tcp/127.0.0.1:29495) (DC: dc1)" writer.go:29: 2021-01-29T19:33:45.207Z [INFO] TestCoordinate_Update: Started DNS server: address=127.0.0.1:29490 network=udp === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.213Z [INFO] TestCoordinate_Update_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:45.214Z [INFO] TestCoordinate_Update_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.215Z [DEBUG] TestCoordinate_Update_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.217Z [INFO] TestCoordinate_Node.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4b4e4343-5e79-b2d8-16b0-75c450e93ac8 Address:127.0.0.1:29519}]" === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.217Z [INFO] TestCoordinate_Update_ACLDeny.server: member joined, marking health alive: member=Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb writer.go:29: 2021-01-29T19:33:45.219Z [DEBUG] TestCoordinate_Update_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.211Z [INFO] TestCoordinate_Update.server: Handled event for server in area: event=member-join server=Node-560f7dcd-33a2-73c3-de5d-7d2ed99598f7.dc1 area=wan === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.221Z [DEBUG] TestCoordinate_Update_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.213Z [INFO] TestCoordinate_Update: Started DNS server: address=127.0.0.1:29490 network=tcp writer.go:29: 2021-01-29T19:33:45.225Z [INFO] TestCoordinate_Update: Started HTTP server: address=127.0.0.1:29491 network=tcp writer.go:29: 2021-01-29T19:33:45.226Z [INFO] TestCoordinate_Update: started state syncer === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.219Z [INFO] TestCoordinate_Node.server.raft: entering follower state: follower="Node at 127.0.0.1:29519 [Follower]" leader= writer.go:29: 2021-01-29T19:33:45.219Z [INFO] TestCoordinate_Node.server.serf.wan: serf: EventMemberJoin: Node-4b4e4343-5e79-b2d8-16b0-75c450e93ac8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:45.237Z [INFO] TestCoordinate_Node.server.serf.lan: serf: EventMemberJoin: Node-4b4e4343-5e79-b2d8-16b0-75c450e93ac8 127.0.0.1 === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.238Z [WARN] TestCoordinate_Update.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.240Z [INFO] TestCoordinate_Node.server: Handled event for server in area: event=member-join server=Node-4b4e4343-5e79-b2d8-16b0-75c450e93ac8.dc1 area=wan === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.241Z [INFO] TestCoordinate_Update.server.raft: entering candidate state: node="Node at 127.0.0.1:29495 [Candidate]" term=2 === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.240Z [INFO] TestCoordinate_Node.server: Adding LAN server: server="Node-4b4e4343-5e79-b2d8-16b0-75c450e93ac8 (Addr: tcp/127.0.0.1:29519) (DC: dc1)" === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.244Z [DEBUG] TestCoordinate_Update.server.raft: votes: needed=1 === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.242Z [INFO] TestCoordinate_Node: Started DNS server: address=127.0.0.1:29514 network=udp === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.245Z [DEBUG] TestCoordinate_Update.server.raft: vote granted: from=560f7dcd-33a2-73c3-de5d-7d2ed99598f7 term=2 tally=1 === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.245Z [INFO] TestCoordinate_Node: Started DNS server: address=127.0.0.1:29514 network=tcp writer.go:29: 2021-01-29T19:33:45.248Z [INFO] TestCoordinate_Node: Started HTTP server: address=127.0.0.1:29515 network=tcp === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.246Z [INFO] TestCoordinate_Update.server.raft: election won: tally=1 === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.250Z [INFO] TestCoordinate_Node: started state syncer === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.252Z [INFO] TestCoordinate_Update.server.raft: entering leader state: leader="Node at 127.0.0.1:29495 [Leader]" writer.go:29: 2021-01-29T19:33:45.258Z [INFO] TestCoordinate_Update.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:45.259Z [INFO] TestCoordinate_Update.server: New leader elected: payload=Node-560f7dcd-33a2-73c3-de5d-7d2ed99598f7 writer.go:29: 2021-01-29T19:33:45.261Z [DEBUG] TestCoordinate_Update.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29495 === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.268Z [WARN] TestCoordinate_Node.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:45.270Z [INFO] TestCoordinate_Node.server.raft: entering candidate state: node="Node at 127.0.0.1:29519 [Candidate]" term=2 === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.272Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.273Z [DEBUG] TestCoordinate_Node.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:45.274Z [DEBUG] TestCoordinate_Node.server.raft: vote granted: from=4b4e4343-5e79-b2d8-16b0-75c450e93ac8 term=2 tally=1 writer.go:29: 2021-01-29T19:33:45.276Z [INFO] TestCoordinate_Node.server.raft: election won: tally=1 === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.277Z [INFO] TestCoordinate_Update.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.278Z [INFO] TestCoordinate_Node.server.raft: entering leader state: leader="Node at 127.0.0.1:29519 [Leader]" === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.279Z [INFO] TestCoordinate_Update.leader: started routine: routine="CA root pruning" === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.280Z [INFO] TestCoordinate_Node.server: cluster leadership acquired === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.280Z [DEBUG] TestCoordinate_Update.server: Skipping self join check for node since the cluster is too small: node=Node-560f7dcd-33a2-73c3-de5d-7d2ed99598f7 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.277Z [DEBUG] TestCoordinate_Update_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.282Z [INFO] TestCoordinate_Node.server: New leader elected: payload=Node-4b4e4343-5e79-b2d8-16b0-75c450e93ac8 === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.282Z [INFO] TestCoordinate_Update.server: member joined, marking health alive: member=Node-560f7dcd-33a2-73c3-de5d-7d2ed99598f7 === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.284Z [DEBUG] TestCoordinate_Node.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29519 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.284Z [INFO] TestCoordinate_Update_ACLDeny: Synced node info === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.291Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:45.303Z [INFO] TestCoordinate_Node.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:45.308Z [INFO] TestCoordinate_Node.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.314Z [DEBUG] TestCoordinate_Node.server: Skipping self join check for node since the cluster is too small: node=Node-4b4e4343-5e79-b2d8-16b0-75c450e93ac8 writer.go:29: 2021-01-29T19:33:45.321Z [INFO] TestCoordinate_Node.server: member joined, marking health alive: member=Node-4b4e4343-5e79-b2d8-16b0-75c450e93ac8 === CONT TestDiscoveryChainRead writer.go:29: 2021-01-29T19:33:45.350Z [INFO] TestDiscoveryChainRead: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:45.352Z [INFO] TestDiscoveryChainRead: Endpoints down --- PASS: TestDiscoveryChainRead (1.01s) --- PASS: TestDiscoveryChainRead/GET:_error_on_no_service_name (0.00s) --- PASS: TestDiscoveryChainRead/GET:_read_default_chain (0.00s) --- PASS: TestDiscoveryChainRead/GET:_read_default_chain;_evaluate_in_dc2 (0.00s) --- PASS: TestDiscoveryChainRead/GET:_read_default_chain_with_cache (0.00s) --- PASS: TestDiscoveryChainRead/POST:_error_on_no_service_name (0.00s) --- PASS: TestDiscoveryChainRead/POST:_read_default_chain (0.01s) --- PASS: TestDiscoveryChainRead/POST:_read_default_chain;_evaluate_in_dc2 (0.01s) --- PASS: TestDiscoveryChainRead/POST:_read_default_chain_with_cache (0.00s) --- PASS: TestDiscoveryChainRead/GET:_read_modified_chain (0.00s) --- PASS: TestDiscoveryChainRead/POST:_read_modified_chain_with_overrides_(camel_case) (0.00s) --- PASS: TestDiscoveryChainRead/POST:_read_modified_chain_with_overrides_(snake_case) (0.00s) === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.362Z [WARN] TestCoordinate_Disabled_Response: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:45.363Z [DEBUG] TestCoordinate_Disabled_Response.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:45.365Z [DEBUG] TestCoordinate_Disabled_Response.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:45.371Z [INFO] TestCoordinate_Disabled_Response.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:dba8c3bc-2ddd-f111-88ca-0d26e75c0b54 Address:127.0.0.1:29501}]" writer.go:29: 2021-01-29T19:33:45.372Z [INFO] TestCoordinate_Disabled_Response.server.raft: entering follower state: follower="Node at 127.0.0.1:29501 [Follower]" leader= writer.go:29: 2021-01-29T19:33:45.378Z [INFO] TestCoordinate_Disabled_Response.server.serf.wan: serf: EventMemberJoin: Node-dba8c3bc-2ddd-f111-88ca-0d26e75c0b54.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:45.381Z [INFO] TestCoordinate_Disabled_Response.server.serf.lan: serf: EventMemberJoin: Node-dba8c3bc-2ddd-f111-88ca-0d26e75c0b54 127.0.0.1 writer.go:29: 2021-01-29T19:33:45.383Z [INFO] TestCoordinate_Disabled_Response.server: Adding LAN server: server="Node-dba8c3bc-2ddd-f111-88ca-0d26e75c0b54 (Addr: tcp/127.0.0.1:29501) (DC: dc1)" writer.go:29: 2021-01-29T19:33:45.384Z [INFO] TestCoordinate_Disabled_Response.server: Handled event for server in area: event=member-join server=Node-dba8c3bc-2ddd-f111-88ca-0d26e75c0b54.dc1 area=wan writer.go:29: 2021-01-29T19:33:45.387Z [INFO] TestCoordinate_Disabled_Response: Started DNS server: address=127.0.0.1:29496 network=tcp writer.go:29: 2021-01-29T19:33:45.388Z [INFO] TestCoordinate_Disabled_Response: Started DNS server: address=127.0.0.1:29496 network=udp writer.go:29: 2021-01-29T19:33:45.390Z [INFO] TestCoordinate_Disabled_Response: Started HTTP server: address=127.0.0.1:29497 network=tcp writer.go:29: 2021-01-29T19:33:45.391Z [INFO] TestCoordinate_Disabled_Response: started state syncer === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.418Z [DEBUG] TestCoordinate_Node: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:45.421Z [INFO] TestCoordinate_Node: Synced node info === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.427Z [WARN] TestCoordinate_Disabled_Response.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:45.430Z [INFO] TestCoordinate_Disabled_Response.server.raft: entering candidate state: node="Node at 127.0.0.1:29501 [Candidate]" term=2 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.441Z [DEBUG] TestCoordinate_Update_ACLDeny.acl: dropping node from result due to ACLs: node=Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb writer.go:29: 2021-01-29T19:33:45.444Z [DEBUG] TestCoordinate_Update_ACLDeny.acl: dropping node from result due to ACLs: node=Node-7cfde0ef-ae57-6b65-2ffd-c3a51b3904bb === RUN TestCoordinate_Update_ACLDeny/no_token === RUN TestCoordinate_Update_ACLDeny/valid_token === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.448Z [DEBUG] TestCoordinate_Disabled_Response.server.raft: votes: needed=1 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.450Z [INFO] TestCoordinate_Update_ACLDeny: Requesting shutdown === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.451Z [DEBUG] TestCoordinate_Disabled_Response.server.raft: vote granted: from=dba8c3bc-2ddd-f111-88ca-0d26e75c0b54 term=2 tally=1 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.451Z [INFO] TestCoordinate_Update_ACLDeny.server: shutting down server === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.456Z [INFO] TestCoordinate_Disabled_Response.server.raft: election won: tally=1 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.457Z [DEBUG] TestCoordinate_Update_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.462Z [INFO] TestCoordinate_Disabled_Response.server.raft: entering leader state: leader="Node at 127.0.0.1:29501 [Leader]" writer.go:29: 2021-01-29T19:33:45.465Z [INFO] TestCoordinate_Disabled_Response.server: cluster leadership acquired === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.462Z [DEBUG] TestCoordinate_Update_ACLDeny.leader: stopping routine: routine="acl token reaping" === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.466Z [INFO] TestCoordinate_Disabled_Response.server: New leader elected: payload=Node-dba8c3bc-2ddd-f111-88ca-0d26e75c0b54 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.466Z [DEBUG] TestCoordinate_Update_ACLDeny.leader: stopping routine: routine="CA root pruning" === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.469Z [DEBUG] TestCoordinate_Disabled_Response.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29501 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.462Z [DEBUG] TestCoordinate_Update_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:45.466Z [DEBUG] TestCoordinate_Update_ACLDeny.leader: stopped routine: routine="acl token reaping" === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.479Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:45.485Z [INFO] TestCoordinate_Disabled_Response.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:45.486Z [INFO] TestCoordinate_Disabled_Response.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.488Z [DEBUG] TestCoordinate_Disabled_Response.server: Skipping self join check for node since the cluster is too small: node=Node-dba8c3bc-2ddd-f111-88ca-0d26e75c0b54 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.468Z [WARN] TestCoordinate_Update_ACLDeny.server.serf.lan: serf: Shutdown without a Leave === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.490Z [INFO] TestCoordinate_Disabled_Response.server: member joined, marking health alive: member=Node-dba8c3bc-2ddd-f111-88ca-0d26e75c0b54 === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:45.468Z [DEBUG] TestCoordinate_Update_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.494Z [WARN] TestCoordinate_Update_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:45.502Z [INFO] TestCoordinate_Update_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:45.503Z [INFO] TestCoordinate_Update_ACLDeny: consul server down writer.go:29: 2021-01-29T19:33:45.508Z [INFO] TestCoordinate_Update_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:33:45.509Z [INFO] TestCoordinate_Update_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29478 network=tcp writer.go:29: 2021-01-29T19:33:45.511Z [INFO] TestCoordinate_Update_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29478 network=udp writer.go:29: 2021-01-29T19:33:45.513Z [INFO] TestCoordinate_Update_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:29479 network=tcp === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.542Z [DEBUG] TestCoordinate_Update: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:45.544Z [INFO] TestCoordinate_Update: Synced node info writer.go:29: 2021-01-29T19:33:45.544Z [DEBUG] TestCoordinate_Update: Node info in sync === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.572Z [DEBUG] TestCoordinate_Disabled_Response: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:45.575Z [INFO] TestCoordinate_Disabled_Response: Synced node info === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:45.749Z [INFO] TestCoordinate_Update: Requesting shutdown writer.go:29: 2021-01-29T19:33:45.750Z [INFO] TestCoordinate_Update.server: shutting down server writer.go:29: 2021-01-29T19:33:45.751Z [DEBUG] TestCoordinate_Update.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.753Z [WARN] TestCoordinate_Update.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:45.755Z [DEBUG] TestCoordinate_Update.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.757Z [WARN] TestCoordinate_Update.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:45.760Z [INFO] TestCoordinate_Update.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:45.762Z [INFO] TestCoordinate_Update: consul server down writer.go:29: 2021-01-29T19:33:45.764Z [INFO] TestCoordinate_Update: shutdown complete writer.go:29: 2021-01-29T19:33:45.766Z [INFO] TestCoordinate_Update: Stopping server: protocol=DNS address=127.0.0.1:29490 network=tcp writer.go:29: 2021-01-29T19:33:45.768Z [INFO] TestCoordinate_Update: Stopping server: protocol=DNS address=127.0.0.1:29490 network=udp writer.go:29: 2021-01-29T19:33:45.769Z [INFO] TestCoordinate_Update: Stopping server: protocol=HTTP address=127.0.0.1:29491 network=tcp === RUN TestCoordinate_Disabled_Response/0 === RUN TestCoordinate_Disabled_Response/1 === RUN TestCoordinate_Disabled_Response/2 === RUN TestCoordinate_Disabled_Response/3 === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:45.817Z [INFO] TestCoordinate_Disabled_Response: Requesting shutdown writer.go:29: 2021-01-29T19:33:45.819Z [INFO] TestCoordinate_Disabled_Response.server: shutting down server writer.go:29: 2021-01-29T19:33:45.820Z [DEBUG] TestCoordinate_Disabled_Response.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.821Z [WARN] TestCoordinate_Disabled_Response.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:45.822Z [DEBUG] TestCoordinate_Disabled_Response.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.826Z [WARN] TestCoordinate_Disabled_Response.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:45.828Z [INFO] TestCoordinate_Disabled_Response.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:45.830Z [INFO] TestCoordinate_Disabled_Response: consul server down writer.go:29: 2021-01-29T19:33:45.832Z [INFO] TestCoordinate_Disabled_Response: shutdown complete writer.go:29: 2021-01-29T19:33:45.833Z [INFO] TestCoordinate_Disabled_Response: Stopping server: protocol=DNS address=127.0.0.1:29496 network=tcp writer.go:29: 2021-01-29T19:33:45.835Z [INFO] TestCoordinate_Disabled_Response: Stopping server: protocol=DNS address=127.0.0.1:29496 network=udp writer.go:29: 2021-01-29T19:33:45.837Z [INFO] TestCoordinate_Disabled_Response: Stopping server: protocol=HTTP address=127.0.0.1:29497 network=tcp === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:45.891Z [INFO] TestCoordinate_Node: Requesting shutdown writer.go:29: 2021-01-29T19:33:45.893Z [INFO] TestCoordinate_Node.server: shutting down server writer.go:29: 2021-01-29T19:33:45.896Z [DEBUG] TestCoordinate_Node.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.900Z [WARN] TestCoordinate_Node.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:45.900Z [DEBUG] TestCoordinate_Node.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.905Z [WARN] TestCoordinate_Node.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:45.922Z [INFO] TestCoordinate_Node.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:45.922Z [INFO] TestCoordinate_Node: consul server down writer.go:29: 2021-01-29T19:33:45.925Z [INFO] TestCoordinate_Node: shutdown complete writer.go:29: 2021-01-29T19:33:45.926Z [INFO] TestCoordinate_Node: Stopping server: protocol=DNS address=127.0.0.1:29514 network=tcp writer.go:29: 2021-01-29T19:33:45.928Z [INFO] TestCoordinate_Node: Stopping server: protocol=DNS address=127.0.0.1:29514 network=udp writer.go:29: 2021-01-29T19:33:45.929Z [INFO] TestCoordinate_Node: Stopping server: protocol=HTTP address=127.0.0.1:29515 network=tcp === CONT TestCoordinate_Update_ACLDeny writer.go:29: 2021-01-29T19:33:46.015Z [INFO] TestCoordinate_Update_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:46.016Z [INFO] TestCoordinate_Update_ACLDeny: Endpoints down --- PASS: TestCoordinate_Update_ACLDeny (1.00s) --- PASS: TestCoordinate_Update_ACLDeny/no_token (0.00s) --- PASS: TestCoordinate_Update_ACLDeny/valid_token (0.00s) === CONT TestConnectCAConfig === RUN TestConnectCAConfig/basic writer.go:29: 2021-01-29T19:33:46.038Z [WARN] TestConnectCAConfig/basic: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:46.040Z [DEBUG] TestConnectCAConfig/basic.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:46.043Z [DEBUG] TestConnectCAConfig/basic.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:46.050Z [INFO] TestConnectCAConfig/basic.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5eab8920-b32d-e6bb-19b2-7a5d1c4aec05 Address:127.0.0.1:29531}]" writer.go:29: 2021-01-29T19:33:46.053Z [INFO] TestConnectCAConfig/basic.server.raft: entering follower state: follower="Node at 127.0.0.1:29531 [Follower]" leader= writer.go:29: 2021-01-29T19:33:46.059Z [INFO] TestConnectCAConfig/basic.server.serf.wan: serf: EventMemberJoin: Node-5eab8920-b32d-e6bb-19b2-7a5d1c4aec05.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:46.083Z [INFO] TestConnectCAConfig/basic.server.serf.lan: serf: EventMemberJoin: Node-5eab8920-b32d-e6bb-19b2-7a5d1c4aec05 127.0.0.1 writer.go:29: 2021-01-29T19:33:46.094Z [INFO] TestConnectCAConfig/basic: Started DNS server: address=127.0.0.1:29526 network=udp writer.go:29: 2021-01-29T19:33:46.095Z [INFO] TestConnectCAConfig/basic.server: Adding LAN server: server="Node-5eab8920-b32d-e6bb-19b2-7a5d1c4aec05 (Addr: tcp/127.0.0.1:29531) (DC: dc1)" writer.go:29: 2021-01-29T19:33:46.095Z [INFO] TestConnectCAConfig/basic.server: Handled event for server in area: event=member-join server=Node-5eab8920-b32d-e6bb-19b2-7a5d1c4aec05.dc1 area=wan writer.go:29: 2021-01-29T19:33:46.096Z [INFO] TestConnectCAConfig/basic: Started DNS server: address=127.0.0.1:29526 network=tcp writer.go:29: 2021-01-29T19:33:46.099Z [INFO] TestConnectCAConfig/basic: Started HTTP server: address=127.0.0.1:29527 network=tcp writer.go:29: 2021-01-29T19:33:46.100Z [INFO] TestConnectCAConfig/basic: started state syncer writer.go:29: 2021-01-29T19:33:46.115Z [WARN] TestConnectCAConfig/basic.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:46.117Z [INFO] TestConnectCAConfig/basic.server.raft: entering candidate state: node="Node at 127.0.0.1:29531 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:46.142Z [DEBUG] TestConnectCAConfig/basic.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:46.144Z [DEBUG] TestConnectCAConfig/basic.server.raft: vote granted: from=5eab8920-b32d-e6bb-19b2-7a5d1c4aec05 term=2 tally=1 writer.go:29: 2021-01-29T19:33:46.145Z [INFO] TestConnectCAConfig/basic.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:46.147Z [INFO] TestConnectCAConfig/basic.server.raft: entering leader state: leader="Node at 127.0.0.1:29531 [Leader]" writer.go:29: 2021-01-29T19:33:46.149Z [INFO] TestConnectCAConfig/basic.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:46.150Z [INFO] TestConnectCAConfig/basic.server: New leader elected: payload=Node-5eab8920-b32d-e6bb-19b2-7a5d1c4aec05 writer.go:29: 2021-01-29T19:33:46.151Z [DEBUG] TestConnectCAConfig/basic.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29531 writer.go:29: 2021-01-29T19:33:46.163Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:46.169Z [INFO] TestConnectCAConfig/basic.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:46.171Z [INFO] TestConnectCAConfig/basic.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.172Z [DEBUG] TestConnectCAConfig/basic.server: Skipping self join check for node since the cluster is too small: node=Node-5eab8920-b32d-e6bb-19b2-7a5d1c4aec05 writer.go:29: 2021-01-29T19:33:46.174Z [INFO] TestConnectCAConfig/basic.server: member joined, marking health alive: member=Node-5eab8920-b32d-e6bb-19b2-7a5d1c4aec05 writer.go:29: 2021-01-29T19:33:46.216Z [DEBUG] TestConnectCAConfig/basic: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:46.218Z [INFO] TestConnectCAConfig/basic: Synced node info === CONT TestCoordinate_Update writer.go:29: 2021-01-29T19:33:46.271Z [INFO] TestCoordinate_Update: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:46.273Z [INFO] TestCoordinate_Update: Endpoints down --- PASS: TestCoordinate_Update (1.17s) === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.295Z [WARN] TestConnectCARoots_list: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:46.297Z [DEBUG] TestConnectCARoots_list.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:46.299Z [DEBUG] TestConnectCARoots_list.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:46.316Z [INFO] TestConnectCARoots_list.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5888476f-b9ff-2072-0d2d-f1bd7d53dccc Address:127.0.0.1:29525}]" writer.go:29: 2021-01-29T19:33:46.319Z [INFO] TestConnectCARoots_list.server.raft: entering follower state: follower="Node at 127.0.0.1:29525 [Follower]" leader= writer.go:29: 2021-01-29T19:33:46.323Z [INFO] TestConnectCARoots_list.server.serf.wan: serf: EventMemberJoin: Node-5888476f-b9ff-2072-0d2d-f1bd7d53dccc.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:46.326Z [INFO] TestConnectCARoots_list.server.serf.lan: serf: EventMemberJoin: Node-5888476f-b9ff-2072-0d2d-f1bd7d53dccc 127.0.0.1 writer.go:29: 2021-01-29T19:33:46.331Z [INFO] TestConnectCARoots_list: Started DNS server: address=127.0.0.1:29520 network=udp writer.go:29: 2021-01-29T19:33:46.332Z [INFO] TestConnectCARoots_list.server: Adding LAN server: server="Node-5888476f-b9ff-2072-0d2d-f1bd7d53dccc (Addr: tcp/127.0.0.1:29525) (DC: dc1)" writer.go:29: 2021-01-29T19:33:46.333Z [INFO] TestConnectCARoots_list.server: Handled event for server in area: event=member-join server=Node-5888476f-b9ff-2072-0d2d-f1bd7d53dccc.dc1 area=wan writer.go:29: 2021-01-29T19:33:46.334Z [INFO] TestConnectCARoots_list: Started DNS server: address=127.0.0.1:29520 network=tcp === CONT TestCoordinate_Disabled_Response writer.go:29: 2021-01-29T19:33:46.339Z [INFO] TestCoordinate_Disabled_Response: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:46.340Z [INFO] TestCoordinate_Disabled_Response: Endpoints down --- PASS: TestCoordinate_Disabled_Response (0.99s) --- PASS: TestCoordinate_Disabled_Response/0 (0.00s) --- PASS: TestCoordinate_Disabled_Response/1 (0.00s) --- PASS: TestCoordinate_Disabled_Response/2 (0.00s) --- PASS: TestCoordinate_Disabled_Response/3 (0.00s) === CONT TestConnectCARoots_empty === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.341Z [INFO] TestConnectCARoots_list: Started HTTP server: address=127.0.0.1:29521 network=tcp writer.go:29: 2021-01-29T19:33:46.343Z [INFO] TestConnectCARoots_list: started state syncer === CONT TestConnectCAConfig/basic writer.go:29: 2021-01-29T19:33:46.351Z [INFO] TestConnectCAConfig/basic.server.connect: CA provider config updated writer.go:29: 2021-01-29T19:33:46.358Z [INFO] TestConnectCAConfig/basic: Requesting shutdown writer.go:29: 2021-01-29T19:33:46.360Z [INFO] TestConnectCAConfig/basic.server: shutting down server writer.go:29: 2021-01-29T19:33:46.365Z [DEBUG] TestConnectCAConfig/basic.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.368Z [WARN] TestConnectCAConfig/basic.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:46.374Z [DEBUG] TestConnectCAConfig/basic.leader: stopped routine: routine="CA root pruning" === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.376Z [WARN] TestConnectCARoots_list.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:46.377Z [INFO] TestConnectCARoots_list.server.raft: entering candidate state: node="Node at 127.0.0.1:29525 [Candidate]" term=2 === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.386Z [WARN] TestConnectCARoots_empty: bootstrap = true: do not enable unless necessary === CONT TestConnectCAConfig/basic writer.go:29: 2021-01-29T19:33:46.387Z [WARN] TestConnectCAConfig/basic.server.serf.wan: serf: Shutdown without a Leave === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.388Z [DEBUG] TestConnectCARoots_empty.tlsutil: Update: version=1 === CONT TestConnectCAConfig/basic writer.go:29: 2021-01-29T19:33:46.391Z [INFO] TestConnectCAConfig/basic.server.router.manager: shutting down === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.391Z [DEBUG] TestConnectCARoots_empty.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestConnectCAConfig/basic writer.go:29: 2021-01-29T19:33:46.393Z [INFO] TestConnectCAConfig/basic: consul server down === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.392Z [DEBUG] TestConnectCARoots_list.server.raft: votes: needed=1 === CONT TestConnectCAConfig/basic writer.go:29: 2021-01-29T19:33:46.394Z [INFO] TestConnectCAConfig/basic: shutdown complete === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.395Z [DEBUG] TestConnectCARoots_list.server.raft: vote granted: from=5888476f-b9ff-2072-0d2d-f1bd7d53dccc term=2 tally=1 === CONT TestConnectCAConfig/basic writer.go:29: 2021-01-29T19:33:46.396Z [INFO] TestConnectCAConfig/basic: Stopping server: protocol=DNS address=127.0.0.1:29526 network=tcp writer.go:29: 2021-01-29T19:33:46.398Z [INFO] TestConnectCAConfig/basic: Stopping server: protocol=DNS address=127.0.0.1:29526 network=udp === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.397Z [INFO] TestConnectCARoots_list.server.raft: election won: tally=1 === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.405Z [INFO] TestConnectCARoots_empty.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:486dd7cf-cd79-5a9a-fcf6-6661ecc732f9 Address:127.0.0.1:29507}]" === CONT TestConnectCAConfig/basic writer.go:29: 2021-01-29T19:33:46.400Z [INFO] TestConnectCAConfig/basic: Stopping server: protocol=HTTP address=127.0.0.1:29527 network=tcp === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.400Z [INFO] TestConnectCARoots_list.server.raft: entering leader state: leader="Node at 127.0.0.1:29525 [Leader]" writer.go:29: 2021-01-29T19:33:46.409Z [INFO] TestConnectCARoots_list.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:46.413Z [INFO] TestConnectCARoots_list.server: New leader elected: payload=Node-5888476f-b9ff-2072-0d2d-f1bd7d53dccc === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.408Z [INFO] TestConnectCARoots_empty.server.raft: entering follower state: follower="Node at 127.0.0.1:29507 [Follower]" leader= writer.go:29: 2021-01-29T19:33:46.411Z [INFO] TestConnectCARoots_empty.server.serf.wan: serf: EventMemberJoin: Node-486dd7cf-cd79-5a9a-fcf6-6661ecc732f9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:46.423Z [INFO] TestConnectCARoots_empty.server.serf.lan: serf: EventMemberJoin: Node-486dd7cf-cd79-5a9a-fcf6-6661ecc732f9 127.0.0.1 writer.go:29: 2021-01-29T19:33:46.430Z [INFO] TestConnectCARoots_empty.server: Adding LAN server: server="Node-486dd7cf-cd79-5a9a-fcf6-6661ecc732f9 (Addr: tcp/127.0.0.1:29507) (DC: dc1)" writer.go:29: 2021-01-29T19:33:46.437Z [INFO] TestConnectCARoots_empty: Started DNS server: address=127.0.0.1:29502 network=udp === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.440Z [DEBUG] TestConnectCARoots_list.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29525 === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:46.435Z [INFO] TestCoordinate_Node: Waiting for endpoints to shut down === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.431Z [INFO] TestConnectCARoots_empty.server: Handled event for server in area: event=member-join server=Node-486dd7cf-cd79-5a9a-fcf6-6661ecc732f9.dc1 area=wan === CONT TestCoordinate_Node writer.go:29: 2021-01-29T19:33:46.444Z [INFO] TestCoordinate_Node: Endpoints down --- PASS: TestCoordinate_Node (1.27s) === CONT TestConfig_Apply_ProxyDefaultsExpose === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.444Z [INFO] TestConnectCARoots_empty: Started DNS server: address=127.0.0.1:29502 network=tcp writer.go:29: 2021-01-29T19:33:46.455Z [INFO] TestConnectCARoots_empty: Started HTTP server: address=127.0.0.1:29503 network=tcp writer.go:29: 2021-01-29T19:33:46.457Z [INFO] TestConnectCARoots_empty: started state syncer === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.459Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.470Z [WARN] TestConnectCARoots_empty.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:46.471Z [INFO] TestConnectCARoots_empty.server.raft: entering candidate state: node="Node at 127.0.0.1:29507 [Candidate]" term=2 === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.481Z [INFO] TestConnectCARoots_list.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:46.485Z [INFO] TestConnectCARoots_list.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.489Z [DEBUG] TestConnectCARoots_list.server: Skipping self join check for node since the cluster is too small: node=Node-5888476f-b9ff-2072-0d2d-f1bd7d53dccc === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.493Z [WARN] TestConfig_Apply_ProxyDefaultsExpose: bootstrap = true: do not enable unless necessary === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.492Z [INFO] TestConnectCARoots_list.server: member joined, marking health alive: member=Node-5888476f-b9ff-2072-0d2d-f1bd7d53dccc === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.494Z [DEBUG] TestConfig_Apply_ProxyDefaultsExpose.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:46.496Z [DEBUG] TestConfig_Apply_ProxyDefaultsExpose.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.514Z [DEBUG] TestConnectCARoots_empty.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:46.517Z [DEBUG] TestConnectCARoots_empty.server.raft: vote granted: from=486dd7cf-cd79-5a9a-fcf6-6661ecc732f9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:46.521Z [INFO] TestConnectCARoots_empty.server.raft: election won: tally=1 === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.524Z [DEBUG] connect.ca.consul: consul CA provider configured: id=29:14:9e:00:b9:7e:70:cf:4b:3b:82:ba:e7:8f:c8:9a:15:d0:bb:60:ac:7e:c4:d7:cd:dc:77:e4:65:d1:84:0c is_primary=true === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.524Z [INFO] TestConnectCARoots_empty.server.raft: entering leader state: leader="Node at 127.0.0.1:29507 [Leader]" writer.go:29: 2021-01-29T19:33:46.526Z [INFO] TestConnectCARoots_empty.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:46.528Z [INFO] TestConnectCARoots_empty.server: New leader elected: payload=Node-486dd7cf-cd79-5a9a-fcf6-6661ecc732f9 === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.531Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9810119e-880c-476e-9ec5-820f4bd746b9 Address:127.0.0.1:29537}]" === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.532Z [DEBUG] TestConnectCARoots_empty.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29507 === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.535Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server.raft: entering follower state: follower="Node at 127.0.0.1:29537 [Follower]" leader= === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.536Z [INFO] TestConnectCARoots_empty.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.539Z [DEBUG] TestConnectCARoots_empty.server: Skipping self join check for node since the cluster is too small: node=Node-486dd7cf-cd79-5a9a-fcf6-6661ecc732f9 writer.go:29: 2021-01-29T19:33:46.543Z [INFO] TestConnectCARoots_empty.server: member joined, marking health alive: member=Node-486dd7cf-cd79-5a9a-fcf6-6661ecc732f9 === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.540Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server.serf.wan: serf: EventMemberJoin: Node-9810119e-880c-476e-9ec5-820f4bd746b9.dc1 127.0.0.1 === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.550Z [INFO] TestConnectCARoots_list.server.connect: CA rotated to new root under provider: provider=consul writer.go:29: 2021-01-29T19:33:46.554Z [INFO] TestConnectCARoots_list: Requesting shutdown writer.go:29: 2021-01-29T19:33:46.555Z [INFO] TestConnectCARoots_list.server: shutting down server writer.go:29: 2021-01-29T19:33:46.556Z [DEBUG] TestConnectCARoots_list.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.557Z [WARN] TestConnectCARoots_list.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:46.559Z [ERROR] TestConnectCARoots_list.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:46.560Z [DEBUG] TestConnectCARoots_list.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.564Z [WARN] TestConnectCARoots_list.server.serf.wan: serf: Shutdown without a Leave === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.568Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server.serf.lan: serf: EventMemberJoin: Node-9810119e-880c-476e-9ec5-820f4bd746b9 127.0.0.1 === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:46.572Z [INFO] TestConnectCARoots_list.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:46.576Z [INFO] TestConnectCARoots_list: consul server down writer.go:29: 2021-01-29T19:33:46.578Z [INFO] TestConnectCARoots_list: shutdown complete writer.go:29: 2021-01-29T19:33:46.579Z [INFO] TestConnectCARoots_list: Stopping server: protocol=DNS address=127.0.0.1:29520 network=tcp writer.go:29: 2021-01-29T19:33:46.581Z [INFO] TestConnectCARoots_list: Stopping server: protocol=DNS address=127.0.0.1:29520 network=udp writer.go:29: 2021-01-29T19:33:46.583Z [INFO] TestConnectCARoots_list: Stopping server: protocol=HTTP address=127.0.0.1:29521 network=tcp === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.585Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server: Handled event for server in area: event=member-join server=Node-9810119e-880c-476e-9ec5-820f4bd746b9.dc1 area=wan writer.go:29: 2021-01-29T19:33:46.585Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server: Adding LAN server: server="Node-9810119e-880c-476e-9ec5-820f4bd746b9 (Addr: tcp/127.0.0.1:29537) (DC: dc1)" writer.go:29: 2021-01-29T19:33:46.594Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: Started DNS server: address=127.0.0.1:29532 network=udp writer.go:29: 2021-01-29T19:33:46.597Z [WARN] TestConfig_Apply_ProxyDefaultsExpose.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:46.598Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server.raft: entering candidate state: node="Node at 127.0.0.1:29537 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:46.599Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: Started DNS server: address=127.0.0.1:29532 network=tcp writer.go:29: 2021-01-29T19:33:46.606Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: Started HTTP server: address=127.0.0.1:29533 network=tcp writer.go:29: 2021-01-29T19:33:46.607Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: started state syncer writer.go:29: 2021-01-29T19:33:46.619Z [DEBUG] TestConfig_Apply_ProxyDefaultsExpose.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:46.622Z [DEBUG] TestConfig_Apply_ProxyDefaultsExpose.server.raft: vote granted: from=9810119e-880c-476e-9ec5-820f4bd746b9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:46.625Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:46.629Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server.raft: entering leader state: leader="Node at 127.0.0.1:29537 [Leader]" writer.go:29: 2021-01-29T19:33:46.632Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:46.633Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server: New leader elected: payload=Node-9810119e-880c-476e-9ec5-820f4bd746b9 writer.go:29: 2021-01-29T19:33:46.637Z [DEBUG] TestConfig_Apply_ProxyDefaultsExpose.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29537 writer.go:29: 2021-01-29T19:33:46.677Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:46.697Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:46.699Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.701Z [DEBUG] TestConfig_Apply_ProxyDefaultsExpose.server: Skipping self join check for node since the cluster is too small: node=Node-9810119e-880c-476e-9ec5-820f4bd746b9 writer.go:29: 2021-01-29T19:33:46.703Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server: member joined, marking health alive: member=Node-9810119e-880c-476e-9ec5-820f4bd746b9 === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:46.821Z [INFO] TestConnectCARoots_empty: Requesting shutdown writer.go:29: 2021-01-29T19:33:46.823Z [INFO] TestConnectCARoots_empty.server: shutting down server writer.go:29: 2021-01-29T19:33:46.825Z [DEBUG] TestConnectCARoots_empty.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.827Z [WARN] TestConnectCARoots_empty.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:46.825Z [ERROR] TestConnectCARoots_empty.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:46.827Z [DEBUG] TestConnectCARoots_empty.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.830Z [WARN] TestConnectCARoots_empty.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:46.837Z [INFO] TestConnectCARoots_empty.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:46.838Z [INFO] TestConnectCARoots_empty: consul server down writer.go:29: 2021-01-29T19:33:46.841Z [INFO] TestConnectCARoots_empty: shutdown complete writer.go:29: 2021-01-29T19:33:46.843Z [INFO] TestConnectCARoots_empty: Stopping server: protocol=DNS address=127.0.0.1:29502 network=tcp writer.go:29: 2021-01-29T19:33:46.844Z [INFO] TestConnectCARoots_empty: Stopping server: protocol=DNS address=127.0.0.1:29502 network=udp writer.go:29: 2021-01-29T19:33:46.846Z [INFO] TestConnectCARoots_empty: Stopping server: protocol=HTTP address=127.0.0.1:29503 network=tcp === CONT TestConnectCAConfig/basic writer.go:29: 2021-01-29T19:33:46.907Z [INFO] TestConnectCAConfig/basic: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:46.908Z [INFO] TestConnectCAConfig/basic: Endpoints down === RUN TestConnectCAConfig/basic_with_IntermediateCertTTL === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.916Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: Requesting shutdown writer.go:29: 2021-01-29T19:33:46.918Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server: shutting down server writer.go:29: 2021-01-29T19:33:46.919Z [DEBUG] TestConfig_Apply_ProxyDefaultsExpose.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.921Z [WARN] TestConfig_Apply_ProxyDefaultsExpose.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:46.919Z [ERROR] TestConfig_Apply_ProxyDefaultsExpose.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:46.921Z [DEBUG] TestConfig_Apply_ProxyDefaultsExpose.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:46.924Z [WARN] TestConfig_Apply_ProxyDefaultsExpose.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:46.931Z [INFO] TestConfig_Apply_ProxyDefaultsExpose.server.router.manager: shutting down === CONT TestConnectCAConfig/basic_with_IntermediateCertTTL writer.go:29: 2021-01-29T19:33:46.934Z [WARN] TestConnectCAConfig/basic_with_IntermediateCertTTL: bootstrap = true: do not enable unless necessary === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.931Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: consul server down === CONT TestConnectCAConfig/basic_with_IntermediateCertTTL writer.go:29: 2021-01-29T19:33:46.937Z [DEBUG] TestConnectCAConfig/basic_with_IntermediateCertTTL.tlsutil: Update: version=1 === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.937Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: shutdown complete === CONT TestConnectCAConfig/basic_with_IntermediateCertTTL writer.go:29: 2021-01-29T19:33:46.939Z [DEBUG] TestConnectCAConfig/basic_with_IntermediateCertTTL.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:46.938Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: Stopping server: protocol=DNS address=127.0.0.1:29532 network=tcp writer.go:29: 2021-01-29T19:33:46.941Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: Stopping server: protocol=DNS address=127.0.0.1:29532 network=udp writer.go:29: 2021-01-29T19:33:46.942Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: Stopping server: protocol=HTTP address=127.0.0.1:29533 network=tcp === CONT TestConnectCAConfig/basic_with_IntermediateCertTTL writer.go:29: 2021-01-29T19:33:46.945Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:50dace0a-37b6-d8ac-f67f-12b2501b0b08 Address:127.0.0.1:29543}]" writer.go:29: 2021-01-29T19:33:46.948Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.raft: entering follower state: follower="Node at 127.0.0.1:29543 [Follower]" leader= writer.go:29: 2021-01-29T19:33:46.951Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.serf.wan: serf: EventMemberJoin: Node-50dace0a-37b6-d8ac-f67f-12b2501b0b08.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:46.961Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.serf.lan: serf: EventMemberJoin: Node-50dace0a-37b6-d8ac-f67f-12b2501b0b08 127.0.0.1 writer.go:29: 2021-01-29T19:33:46.978Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server: Adding LAN server: server="Node-50dace0a-37b6-d8ac-f67f-12b2501b0b08 (Addr: tcp/127.0.0.1:29543) (DC: dc1)" writer.go:29: 2021-01-29T19:33:46.984Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server: Handled event for server in area: event=member-join server=Node-50dace0a-37b6-d8ac-f67f-12b2501b0b08.dc1 area=wan writer.go:29: 2021-01-29T19:33:46.988Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: Started DNS server: address=127.0.0.1:29538 network=tcp writer.go:29: 2021-01-29T19:33:46.990Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: Started DNS server: address=127.0.0.1:29538 network=udp writer.go:29: 2021-01-29T19:33:46.992Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: Started HTTP server: address=127.0.0.1:29539 network=tcp writer.go:29: 2021-01-29T19:33:46.994Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: started state syncer writer.go:29: 2021-01-29T19:33:47.002Z [WARN] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:47.004Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.raft: entering candidate state: node="Node at 127.0.0.1:29543 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:47.012Z [DEBUG] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:47.018Z [DEBUG] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.raft: vote granted: from=50dace0a-37b6-d8ac-f67f-12b2501b0b08 term=2 tally=1 writer.go:29: 2021-01-29T19:33:47.021Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:47.024Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.raft: entering leader state: leader="Node at 127.0.0.1:29543 [Leader]" writer.go:29: 2021-01-29T19:33:47.027Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:47.028Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server: New leader elected: payload=Node-50dace0a-37b6-d8ac-f67f-12b2501b0b08 writer.go:29: 2021-01-29T19:33:47.031Z [DEBUG] TestConnectCAConfig/basic_with_IntermediateCertTTL.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29543 writer.go:29: 2021-01-29T19:33:47.044Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:47.053Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:47.054Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.056Z [DEBUG] TestConnectCAConfig/basic_with_IntermediateCertTTL.server: Skipping self join check for node since the cluster is too small: node=Node-50dace0a-37b6-d8ac-f67f-12b2501b0b08 writer.go:29: 2021-01-29T19:33:47.057Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server: member joined, marking health alive: member=Node-50dace0a-37b6-d8ac-f67f-12b2501b0b08 === CONT TestConnectCARoots_list writer.go:29: 2021-01-29T19:33:47.085Z [INFO] TestConnectCARoots_list: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:47.087Z [INFO] TestConnectCARoots_list: Endpoints down --- PASS: TestConnectCARoots_list (0.81s) === CONT TestConfig_Apply_Decoding writer.go:29: 2021-01-29T19:33:47.109Z [WARN] TestConfig_Apply_Decoding: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:47.114Z [DEBUG] TestConfig_Apply_Decoding.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:47.121Z [DEBUG] TestConfig_Apply_Decoding.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:47.134Z [INFO] TestConfig_Apply_Decoding.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7db32d65-80c4-d0ce-64b1-69adae2e1da2 Address:127.0.0.1:29513}]" writer.go:29: 2021-01-29T19:33:47.137Z [INFO] TestConfig_Apply_Decoding.server.raft: entering follower state: follower="Node at 127.0.0.1:29513 [Follower]" leader= writer.go:29: 2021-01-29T19:33:47.138Z [INFO] TestConfig_Apply_Decoding.server.serf.wan: serf: EventMemberJoin: Node-7db32d65-80c4-d0ce-64b1-69adae2e1da2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:47.146Z [INFO] TestConfig_Apply_Decoding.server.serf.lan: serf: EventMemberJoin: Node-7db32d65-80c4-d0ce-64b1-69adae2e1da2 127.0.0.1 writer.go:29: 2021-01-29T19:33:47.155Z [INFO] TestConfig_Apply_Decoding.server: Handled event for server in area: event=member-join server=Node-7db32d65-80c4-d0ce-64b1-69adae2e1da2.dc1 area=wan writer.go:29: 2021-01-29T19:33:47.155Z [INFO] TestConfig_Apply_Decoding.server: Adding LAN server: server="Node-7db32d65-80c4-d0ce-64b1-69adae2e1da2 (Addr: tcp/127.0.0.1:29513) (DC: dc1)" === CONT TestConnectCAConfig/basic_with_IntermediateCertTTL writer.go:29: 2021-01-29T19:33:47.157Z [DEBUG] TestConnectCAConfig/basic_with_IntermediateCertTTL: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:47.160Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: Synced node info writer.go:29: 2021-01-29T19:33:47.162Z [DEBUG] TestConnectCAConfig/basic_with_IntermediateCertTTL: Node info in sync === CONT TestConfig_Apply_Decoding writer.go:29: 2021-01-29T19:33:47.160Z [INFO] TestConfig_Apply_Decoding: Started DNS server: address=127.0.0.1:29508 network=tcp writer.go:29: 2021-01-29T19:33:47.169Z [INFO] TestConfig_Apply_Decoding: Started DNS server: address=127.0.0.1:29508 network=udp writer.go:29: 2021-01-29T19:33:47.175Z [INFO] TestConfig_Apply_Decoding: Started HTTP server: address=127.0.0.1:29509 network=tcp writer.go:29: 2021-01-29T19:33:47.177Z [INFO] TestConfig_Apply_Decoding: started state syncer writer.go:29: 2021-01-29T19:33:47.205Z [WARN] TestConfig_Apply_Decoding.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:47.207Z [INFO] TestConfig_Apply_Decoding.server.raft: entering candidate state: node="Node at 127.0.0.1:29513 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:47.209Z [DEBUG] TestConfig_Apply_Decoding.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:47.211Z [DEBUG] TestConfig_Apply_Decoding.server.raft: vote granted: from=7db32d65-80c4-d0ce-64b1-69adae2e1da2 term=2 tally=1 writer.go:29: 2021-01-29T19:33:47.212Z [INFO] TestConfig_Apply_Decoding.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:47.213Z [INFO] TestConfig_Apply_Decoding.server.raft: entering leader state: leader="Node at 127.0.0.1:29513 [Leader]" writer.go:29: 2021-01-29T19:33:47.217Z [INFO] TestConfig_Apply_Decoding.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:47.217Z [INFO] TestConfig_Apply_Decoding.server: New leader elected: payload=Node-7db32d65-80c4-d0ce-64b1-69adae2e1da2 writer.go:29: 2021-01-29T19:33:47.220Z [DEBUG] TestConfig_Apply_Decoding.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29513 === CONT TestConnectCAConfig/basic_with_IntermediateCertTTL writer.go:29: 2021-01-29T19:33:47.232Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.connect: CA provider config updated writer.go:29: 2021-01-29T19:33:47.236Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: Requesting shutdown writer.go:29: 2021-01-29T19:33:47.237Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server: shutting down server writer.go:29: 2021-01-29T19:33:47.239Z [DEBUG] TestConnectCAConfig/basic_with_IntermediateCertTTL.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.241Z [DEBUG] TestConnectCAConfig/basic_with_IntermediateCertTTL.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.241Z [WARN] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:47.245Z [WARN] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:47.248Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:47.250Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: consul server down writer.go:29: 2021-01-29T19:33:47.252Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: shutdown complete writer.go:29: 2021-01-29T19:33:47.255Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: Stopping server: protocol=DNS address=127.0.0.1:29538 network=tcp writer.go:29: 2021-01-29T19:33:47.265Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: Stopping server: protocol=DNS address=127.0.0.1:29538 network=udp writer.go:29: 2021-01-29T19:33:47.267Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: Stopping server: protocol=HTTP address=127.0.0.1:29539 network=tcp === CONT TestConfig_Apply_Decoding writer.go:29: 2021-01-29T19:33:47.261Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:47.281Z [INFO] TestConfig_Apply_Decoding.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:47.289Z [INFO] TestConfig_Apply_Decoding: Synced node info writer.go:29: 2021-01-29T19:33:47.294Z [DEBUG] TestConfig_Apply_Decoding: Node info in sync writer.go:29: 2021-01-29T19:33:47.296Z [INFO] TestConfig_Apply_Decoding.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.301Z [DEBUG] TestConfig_Apply_Decoding.server: Skipping self join check for node since the cluster is too small: node=Node-7db32d65-80c4-d0ce-64b1-69adae2e1da2 writer.go:29: 2021-01-29T19:33:47.307Z [INFO] TestConfig_Apply_Decoding.server: member joined, marking health alive: member=Node-7db32d65-80c4-d0ce-64b1-69adae2e1da2 === RUN TestConfig_Apply_Decoding/No_Kind === RUN TestConfig_Apply_Decoding/Kind_Not_String === RUN TestConfig_Apply_Decoding/Lowercase_kind === CONT TestConnectCARoots_empty writer.go:29: 2021-01-29T19:33:47.347Z [INFO] TestConnectCARoots_empty: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:47.349Z [INFO] TestConnectCARoots_empty: Endpoints down --- PASS: TestConnectCARoots_empty (1.01s) === CONT TestConfig_Apply_CAS === CONT TestConfig_Apply_Decoding writer.go:29: 2021-01-29T19:33:47.362Z [INFO] TestConfig_Apply_Decoding: Requesting shutdown writer.go:29: 2021-01-29T19:33:47.365Z [INFO] TestConfig_Apply_Decoding.server: shutting down server writer.go:29: 2021-01-29T19:33:47.366Z [DEBUG] TestConfig_Apply_Decoding.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.368Z [WARN] TestConfig_Apply_Decoding.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:47.370Z [DEBUG] TestConfig_Apply_Decoding.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.376Z [WARN] TestConfig_Apply_Decoding.server.serf.wan: serf: Shutdown without a Leave === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.381Z [WARN] TestConfig_Apply_CAS: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:47.383Z [DEBUG] TestConfig_Apply_CAS.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:47.384Z [DEBUG] TestConfig_Apply_CAS.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestConfig_Apply_Decoding writer.go:29: 2021-01-29T19:33:47.382Z [INFO] TestConfig_Apply_Decoding.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:47.382Z [INFO] TestConfig_Apply_Decoding: consul server down writer.go:29: 2021-01-29T19:33:47.387Z [INFO] TestConfig_Apply_Decoding: shutdown complete writer.go:29: 2021-01-29T19:33:47.389Z [INFO] TestConfig_Apply_Decoding: Stopping server: protocol=DNS address=127.0.0.1:29508 network=tcp writer.go:29: 2021-01-29T19:33:47.390Z [INFO] TestConfig_Apply_Decoding: Stopping server: protocol=DNS address=127.0.0.1:29508 network=udp === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.390Z [INFO] TestConfig_Apply_CAS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fca6258c-fe06-c347-2ccf-0b1e5aaa46f6 Address:127.0.0.1:29549}]" === CONT TestConfig_Apply_Decoding writer.go:29: 2021-01-29T19:33:47.391Z [INFO] TestConfig_Apply_Decoding: Stopping server: protocol=HTTP address=127.0.0.1:29509 network=tcp === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.393Z [INFO] TestConfig_Apply_CAS.server.serf.wan: serf: EventMemberJoin: Node-fca6258c-fe06-c347-2ccf-0b1e5aaa46f6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:47.393Z [INFO] TestConfig_Apply_CAS.server.raft: entering follower state: follower="Node at 127.0.0.1:29549 [Follower]" leader= writer.go:29: 2021-01-29T19:33:47.396Z [INFO] TestConfig_Apply_CAS.server.serf.lan: serf: EventMemberJoin: Node-fca6258c-fe06-c347-2ccf-0b1e5aaa46f6 127.0.0.1 writer.go:29: 2021-01-29T19:33:47.405Z [INFO] TestConfig_Apply_CAS.server: Adding LAN server: server="Node-fca6258c-fe06-c347-2ccf-0b1e5aaa46f6 (Addr: tcp/127.0.0.1:29549) (DC: dc1)" writer.go:29: 2021-01-29T19:33:47.405Z [INFO] TestConfig_Apply_CAS.server: Handled event for server in area: event=member-join server=Node-fca6258c-fe06-c347-2ccf-0b1e5aaa46f6.dc1 area=wan writer.go:29: 2021-01-29T19:33:47.406Z [INFO] TestConfig_Apply_CAS: Started DNS server: address=127.0.0.1:29544 network=tcp writer.go:29: 2021-01-29T19:33:47.409Z [INFO] TestConfig_Apply_CAS: Started DNS server: address=127.0.0.1:29544 network=udp writer.go:29: 2021-01-29T19:33:47.415Z [INFO] TestConfig_Apply_CAS: Started HTTP server: address=127.0.0.1:29545 network=tcp writer.go:29: 2021-01-29T19:33:47.416Z [INFO] TestConfig_Apply_CAS: started state syncer === CONT TestConfig_Apply_ProxyDefaultsExpose writer.go:29: 2021-01-29T19:33:47.443Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:47.445Z [INFO] TestConfig_Apply_ProxyDefaultsExpose: Endpoints down --- PASS: TestConfig_Apply_ProxyDefaultsExpose (1.00s) === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.462Z [WARN] TestConfig_Apply_ProxyDefaultsMeshGateway: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:47.464Z [DEBUG] TestConfig_Apply_ProxyDefaultsMeshGateway.tlsutil: Update: version=1 === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.464Z [WARN] TestConfig_Apply_CAS.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:47.467Z [INFO] TestConfig_Apply_CAS.server.raft: entering candidate state: node="Node at 127.0.0.1:29549 [Candidate]" term=2 === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.467Z [DEBUG] TestConfig_Apply_ProxyDefaultsMeshGateway.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.470Z [DEBUG] TestConfig_Apply_CAS.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:47.472Z [DEBUG] TestConfig_Apply_CAS.server.raft: vote granted: from=fca6258c-fe06-c347-2ccf-0b1e5aaa46f6 term=2 tally=1 writer.go:29: 2021-01-29T19:33:47.475Z [INFO] TestConfig_Apply_CAS.server.raft: election won: tally=1 === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.473Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6d498615-5f7d-a3d1-79e4-fc62927cc50f Address:127.0.0.1:29567}]" === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.476Z [INFO] TestConfig_Apply_CAS.server.raft: entering leader state: leader="Node at 127.0.0.1:29549 [Leader]" === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.479Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server.raft: entering follower state: follower="Node at 127.0.0.1:29567 [Follower]" leader= === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.481Z [INFO] TestConfig_Apply_CAS.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:47.482Z [INFO] TestConfig_Apply_CAS.server: New leader elected: payload=Node-fca6258c-fe06-c347-2ccf-0b1e5aaa46f6 === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.480Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server.serf.wan: serf: EventMemberJoin: Node-6d498615-5f7d-a3d1-79e4-fc62927cc50f.dc1 127.0.0.1 === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.483Z [DEBUG] TestConfig_Apply_CAS.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29549 === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.486Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server.serf.lan: serf: EventMemberJoin: Node-6d498615-5f7d-a3d1-79e4-fc62927cc50f 127.0.0.1 === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.490Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.489Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server: Handled event for server in area: event=member-join server=Node-6d498615-5f7d-a3d1-79e4-fc62927cc50f.dc1 area=wan writer.go:29: 2021-01-29T19:33:47.490Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: Started DNS server: address=127.0.0.1:29562 network=udp writer.go:29: 2021-01-29T19:33:47.495Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: Started DNS server: address=127.0.0.1:29562 network=tcp writer.go:29: 2021-01-29T19:33:47.498Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: Started HTTP server: address=127.0.0.1:29563 network=tcp writer.go:29: 2021-01-29T19:33:47.500Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: started state syncer === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.502Z [INFO] TestConfig_Apply_CAS.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.504Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server: Adding LAN server: server="Node-6d498615-5f7d-a3d1-79e4-fc62927cc50f (Addr: tcp/127.0.0.1:29567) (DC: dc1)" === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.504Z [INFO] TestConfig_Apply_CAS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.508Z [DEBUG] TestConfig_Apply_CAS.server: Skipping self join check for node since the cluster is too small: node=Node-fca6258c-fe06-c347-2ccf-0b1e5aaa46f6 writer.go:29: 2021-01-29T19:33:47.511Z [INFO] TestConfig_Apply_CAS.server: member joined, marking health alive: member=Node-fca6258c-fe06-c347-2ccf-0b1e5aaa46f6 === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.542Z [WARN] TestConfig_Apply_ProxyDefaultsMeshGateway.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:47.543Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server.raft: entering candidate state: node="Node at 127.0.0.1:29567 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:47.548Z [DEBUG] TestConfig_Apply_ProxyDefaultsMeshGateway.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:47.549Z [DEBUG] TestConfig_Apply_ProxyDefaultsMeshGateway.server.raft: vote granted: from=6d498615-5f7d-a3d1-79e4-fc62927cc50f term=2 tally=1 writer.go:29: 2021-01-29T19:33:47.549Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:47.550Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server.raft: entering leader state: leader="Node at 127.0.0.1:29567 [Leader]" writer.go:29: 2021-01-29T19:33:47.552Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:47.553Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server: New leader elected: payload=Node-6d498615-5f7d-a3d1-79e4-fc62927cc50f writer.go:29: 2021-01-29T19:33:47.559Z [DEBUG] TestConfig_Apply_ProxyDefaultsMeshGateway.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29567 writer.go:29: 2021-01-29T19:33:47.581Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:47.590Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:47.592Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.593Z [DEBUG] TestConfig_Apply_ProxyDefaultsMeshGateway.server: Skipping self join check for node since the cluster is too small: node=Node-6d498615-5f7d-a3d1-79e4-fc62927cc50f writer.go:29: 2021-01-29T19:33:47.594Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server: member joined, marking health alive: member=Node-6d498615-5f7d-a3d1-79e4-fc62927cc50f === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.627Z [DEBUG] TestConfig_Apply_CAS: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:47.629Z [INFO] TestConfig_Apply_CAS: Synced node info writer.go:29: 2021-01-29T19:33:47.631Z [DEBUG] TestConfig_Apply_CAS: Node info in sync === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.665Z [DEBUG] TestConfig_Apply_ProxyDefaultsMeshGateway: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:47.669Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: Synced node info writer.go:29: 2021-01-29T19:33:47.670Z [DEBUG] TestConfig_Apply_ProxyDefaultsMeshGateway: Node info in sync === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:47.698Z [INFO] TestConfig_Apply_CAS: Requesting shutdown writer.go:29: 2021-01-29T19:33:47.699Z [INFO] TestConfig_Apply_CAS.server: shutting down server writer.go:29: 2021-01-29T19:33:47.700Z [DEBUG] TestConfig_Apply_CAS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.701Z [WARN] TestConfig_Apply_CAS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:47.701Z [DEBUG] TestConfig_Apply_CAS.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.704Z [WARN] TestConfig_Apply_CAS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:47.706Z [INFO] TestConfig_Apply_CAS.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:47.707Z [INFO] TestConfig_Apply_CAS: consul server down writer.go:29: 2021-01-29T19:33:47.709Z [INFO] TestConfig_Apply_CAS: shutdown complete writer.go:29: 2021-01-29T19:33:47.711Z [INFO] TestConfig_Apply_CAS: Stopping server: protocol=DNS address=127.0.0.1:29544 network=tcp writer.go:29: 2021-01-29T19:33:47.713Z [INFO] TestConfig_Apply_CAS: Stopping server: protocol=DNS address=127.0.0.1:29544 network=udp writer.go:29: 2021-01-29T19:33:47.714Z [INFO] TestConfig_Apply_CAS: Stopping server: protocol=HTTP address=127.0.0.1:29545 network=tcp === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.776Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: Requesting shutdown writer.go:29: 2021-01-29T19:33:47.777Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server: shutting down server writer.go:29: 2021-01-29T19:33:47.778Z [DEBUG] TestConfig_Apply_ProxyDefaultsMeshGateway.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.780Z [WARN] TestConfig_Apply_ProxyDefaultsMeshGateway.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:47.780Z [DEBUG] TestConfig_Apply_ProxyDefaultsMeshGateway.leader: stopped routine: routine="CA root pruning" === CONT TestConnectCAConfig/basic_with_IntermediateCertTTL writer.go:29: 2021-01-29T19:33:47.771Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: Waiting for endpoints to shut down === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.783Z [WARN] TestConfig_Apply_ProxyDefaultsMeshGateway.server.serf.wan: serf: Shutdown without a Leave === CONT TestConnectCAConfig/basic_with_IntermediateCertTTL writer.go:29: 2021-01-29T19:33:47.784Z [INFO] TestConnectCAConfig/basic_with_IntermediateCertTTL: Endpoints down === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.786Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway.server.router.manager: shutting down === RUN TestConnectCAConfig/force_without_cross_sign_CamelCase === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:47.787Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: consul server down writer.go:29: 2021-01-29T19:33:47.794Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: shutdown complete writer.go:29: 2021-01-29T19:33:47.798Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: Stopping server: protocol=DNS address=127.0.0.1:29562 network=tcp writer.go:29: 2021-01-29T19:33:47.802Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: Stopping server: protocol=DNS address=127.0.0.1:29562 network=udp writer.go:29: 2021-01-29T19:33:47.804Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: Stopping server: protocol=HTTP address=127.0.0.1:29563 network=tcp === CONT TestConnectCAConfig/force_without_cross_sign_CamelCase writer.go:29: 2021-01-29T19:33:47.808Z [WARN] TestConnectCAConfig/force_without_cross_sign_CamelCase: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:47.810Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:47.812Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:47.820Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3aecbdc4-67ef-0a00-783c-1bcfb185075c Address:127.0.0.1:29555}]" writer.go:29: 2021-01-29T19:33:47.823Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.raft: entering follower state: follower="Node at 127.0.0.1:29555 [Follower]" leader= writer.go:29: 2021-01-29T19:33:47.825Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.serf.wan: serf: EventMemberJoin: Node-3aecbdc4-67ef-0a00-783c-1bcfb185075c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:47.857Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.serf.lan: serf: EventMemberJoin: Node-3aecbdc4-67ef-0a00-783c-1bcfb185075c 127.0.0.1 writer.go:29: 2021-01-29T19:33:47.869Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server: Adding LAN server: server="Node-3aecbdc4-67ef-0a00-783c-1bcfb185075c (Addr: tcp/127.0.0.1:29555) (DC: dc1)" writer.go:29: 2021-01-29T19:33:47.871Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server: Handled event for server in area: event=member-join server=Node-3aecbdc4-67ef-0a00-783c-1bcfb185075c.dc1 area=wan writer.go:29: 2021-01-29T19:33:47.880Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: Started DNS server: address=127.0.0.1:29550 network=tcp writer.go:29: 2021-01-29T19:33:47.882Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: Started DNS server: address=127.0.0.1:29550 network=udp writer.go:29: 2021-01-29T19:33:47.885Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: Started HTTP server: address=127.0.0.1:29551 network=tcp === CONT TestConfig_Apply_Decoding writer.go:29: 2021-01-29T19:33:47.893Z [INFO] TestConfig_Apply_Decoding: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:47.894Z [INFO] TestConfig_Apply_Decoding: Endpoints down --- PASS: TestConfig_Apply_Decoding (0.81s) --- PASS: TestConfig_Apply_Decoding/No_Kind (0.00s) --- PASS: TestConfig_Apply_Decoding/Kind_Not_String (0.00s) --- PASS: TestConfig_Apply_Decoding/Lowercase_kind (0.02s) === CONT TestConfig_Apply === CONT TestConnectCAConfig/force_without_cross_sign_CamelCase writer.go:29: 2021-01-29T19:33:47.891Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: started state syncer writer.go:29: 2021-01-29T19:33:47.915Z [WARN] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:47.918Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.raft: entering candidate state: node="Node at 127.0.0.1:29555 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:47.923Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:47.925Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.raft: vote granted: from=3aecbdc4-67ef-0a00-783c-1bcfb185075c term=2 tally=1 writer.go:29: 2021-01-29T19:33:47.927Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:47.929Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.raft: entering leader state: leader="Node at 127.0.0.1:29555 [Leader]" writer.go:29: 2021-01-29T19:33:47.931Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:47.933Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server: New leader elected: payload=Node-3aecbdc4-67ef-0a00-783c-1bcfb185075c === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:47.936Z [WARN] TestConfig_Apply: bootstrap = true: do not enable unless necessary === CONT TestConnectCAConfig/force_without_cross_sign_CamelCase writer.go:29: 2021-01-29T19:33:47.936Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29555 === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:47.939Z [DEBUG] TestConfig_Apply.tlsutil: Update: version=1 === CONT TestConnectCAConfig/force_without_cross_sign_CamelCase writer.go:29: 2021-01-29T19:33:47.959Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:47.961Z [DEBUG] TestConfig_Apply.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestConnectCAConfig/force_without_cross_sign_CamelCase writer.go:29: 2021-01-29T19:33:47.965Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:47.968Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.leader: started routine: routine="CA root pruning" === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:47.974Z [INFO] TestConfig_Apply.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:db4ce188-d99e-520c-4a91-5906e34c3c56 Address:127.0.0.1:29561}]" === CONT TestConnectCAConfig/force_without_cross_sign_CamelCase writer.go:29: 2021-01-29T19:33:47.971Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase.server: Skipping self join check for node since the cluster is too small: node=Node-3aecbdc4-67ef-0a00-783c-1bcfb185075c writer.go:29: 2021-01-29T19:33:47.975Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server: member joined, marking health alive: member=Node-3aecbdc4-67ef-0a00-783c-1bcfb185075c === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:47.975Z [INFO] TestConfig_Apply.server.serf.wan: serf: EventMemberJoin: Node-db4ce188-d99e-520c-4a91-5906e34c3c56.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:47.976Z [INFO] TestConfig_Apply.server.raft: entering follower state: follower="Node at 127.0.0.1:29561 [Follower]" leader= writer.go:29: 2021-01-29T19:33:47.985Z [INFO] TestConfig_Apply.server.serf.lan: serf: EventMemberJoin: Node-db4ce188-d99e-520c-4a91-5906e34c3c56 127.0.0.1 writer.go:29: 2021-01-29T19:33:47.991Z [INFO] TestConfig_Apply.server: Handled event for server in area: event=member-join server=Node-db4ce188-d99e-520c-4a91-5906e34c3c56.dc1 area=wan writer.go:29: 2021-01-29T19:33:47.991Z [INFO] TestConfig_Apply.server: Adding LAN server: server="Node-db4ce188-d99e-520c-4a91-5906e34c3c56 (Addr: tcp/127.0.0.1:29561) (DC: dc1)" writer.go:29: 2021-01-29T19:33:47.993Z [INFO] TestConfig_Apply: Started DNS server: address=127.0.0.1:29556 network=tcp writer.go:29: 2021-01-29T19:33:47.994Z [INFO] TestConfig_Apply: Started DNS server: address=127.0.0.1:29556 network=udp writer.go:29: 2021-01-29T19:33:47.997Z [INFO] TestConfig_Apply: Started HTTP server: address=127.0.0.1:29557 network=tcp writer.go:29: 2021-01-29T19:33:47.999Z [INFO] TestConfig_Apply: started state syncer writer.go:29: 2021-01-29T19:33:48.047Z [WARN] TestConfig_Apply.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:48.049Z [INFO] TestConfig_Apply.server.raft: entering candidate state: node="Node at 127.0.0.1:29561 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:48.052Z [DEBUG] TestConfig_Apply.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:48.055Z [DEBUG] TestConfig_Apply.server.raft: vote granted: from=db4ce188-d99e-520c-4a91-5906e34c3c56 term=2 tally=1 writer.go:29: 2021-01-29T19:33:48.057Z [INFO] TestConfig_Apply.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:48.059Z [INFO] TestConfig_Apply.server.raft: entering leader state: leader="Node at 127.0.0.1:29561 [Leader]" writer.go:29: 2021-01-29T19:33:48.062Z [INFO] TestConfig_Apply.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:48.063Z [INFO] TestConfig_Apply.server: New leader elected: payload=Node-db4ce188-d99e-520c-4a91-5906e34c3c56 writer.go:29: 2021-01-29T19:33:48.073Z [DEBUG] TestConfig_Apply.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29561 writer.go:29: 2021-01-29T19:33:48.091Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:48.097Z [INFO] TestConfig_Apply.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:48.099Z [INFO] TestConfig_Apply.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.100Z [DEBUG] TestConfig_Apply.server: Skipping self join check for node since the cluster is too small: node=Node-db4ce188-d99e-520c-4a91-5906e34c3c56 writer.go:29: 2021-01-29T19:33:48.101Z [INFO] TestConfig_Apply.server: member joined, marking health alive: member=Node-db4ce188-d99e-520c-4a91-5906e34c3c56 === CONT TestConnectCAConfig/force_without_cross_sign_CamelCase writer.go:29: 2021-01-29T19:33:48.141Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:48.144Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: Synced node info writer.go:29: 2021-01-29T19:33:48.168Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:48.176Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase: Node info in sync writer.go:29: 2021-01-29T19:33:48.177Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase: Node info in sync === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:48.208Z [DEBUG] TestConfig_Apply: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:48.210Z [INFO] TestConfig_Apply: Synced node info === CONT TestConfig_Apply_CAS writer.go:29: 2021-01-29T19:33:48.215Z [INFO] TestConfig_Apply_CAS: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:48.216Z [INFO] TestConfig_Apply_CAS: Endpoints down --- PASS: TestConfig_Apply_CAS (0.87s) === CONT TestConfig_Delete writer.go:29: 2021-01-29T19:33:48.228Z [WARN] TestConfig_Delete: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:48.230Z [DEBUG] TestConfig_Delete.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:48.233Z [DEBUG] TestConfig_Delete.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:48.245Z [INFO] TestConfig_Delete.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9654863c-d898-46d7-277e-78d38158f306 Address:127.0.0.1:29573}]" writer.go:29: 2021-01-29T19:33:48.248Z [INFO] TestConfig_Delete.server.raft: entering follower state: follower="Node at 127.0.0.1:29573 [Follower]" leader= writer.go:29: 2021-01-29T19:33:48.248Z [INFO] TestConfig_Delete.server.serf.wan: serf: EventMemberJoin: Node-9654863c-d898-46d7-277e-78d38158f306.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:48.255Z [INFO] TestConfig_Delete.server.serf.lan: serf: EventMemberJoin: Node-9654863c-d898-46d7-277e-78d38158f306 127.0.0.1 writer.go:29: 2021-01-29T19:33:48.258Z [INFO] TestConfig_Delete.server: Handled event for server in area: event=member-join server=Node-9654863c-d898-46d7-277e-78d38158f306.dc1 area=wan writer.go:29: 2021-01-29T19:33:48.258Z [INFO] TestConfig_Delete.server: Adding LAN server: server="Node-9654863c-d898-46d7-277e-78d38158f306 (Addr: tcp/127.0.0.1:29573) (DC: dc1)" writer.go:29: 2021-01-29T19:33:48.259Z [INFO] TestConfig_Delete: Started DNS server: address=127.0.0.1:29568 network=udp writer.go:29: 2021-01-29T19:33:48.263Z [INFO] TestConfig_Delete: Started DNS server: address=127.0.0.1:29568 network=tcp writer.go:29: 2021-01-29T19:33:48.266Z [INFO] TestConfig_Delete: Started HTTP server: address=127.0.0.1:29569 network=tcp writer.go:29: 2021-01-29T19:33:48.268Z [INFO] TestConfig_Delete: started state syncer === CONT TestConnectCAConfig/force_without_cross_sign_CamelCase writer.go:29: 2021-01-29T19:33:48.274Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.connect: CA provider config updated writer.go:29: 2021-01-29T19:33:48.280Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: Requesting shutdown writer.go:29: 2021-01-29T19:33:48.281Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server: shutting down server writer.go:29: 2021-01-29T19:33:48.283Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.284Z [WARN] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:48.284Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_CamelCase.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.288Z [WARN] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:48.292Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:48.292Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: consul server down writer.go:29: 2021-01-29T19:33:48.295Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: shutdown complete writer.go:29: 2021-01-29T19:33:48.296Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: Stopping server: protocol=DNS address=127.0.0.1:29550 network=tcp writer.go:29: 2021-01-29T19:33:48.298Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: Stopping server: protocol=DNS address=127.0.0.1:29550 network=udp writer.go:29: 2021-01-29T19:33:48.300Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: Stopping server: protocol=HTTP address=127.0.0.1:29551 network=tcp === CONT TestConfig_Delete writer.go:29: 2021-01-29T19:33:48.304Z [WARN] TestConfig_Delete.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:48.306Z [INFO] TestConfig_Delete.server.raft: entering candidate state: node="Node at 127.0.0.1:29573 [Candidate]" term=2 === CONT TestConfig_Apply_ProxyDefaultsMeshGateway writer.go:29: 2021-01-29T19:33:48.306Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:48.307Z [INFO] TestConfig_Apply_ProxyDefaultsMeshGateway: Endpoints down --- PASS: TestConfig_Apply_ProxyDefaultsMeshGateway (0.86s) === CONT TestConfig_Get === CONT TestConfig_Delete writer.go:29: 2021-01-29T19:33:48.308Z [DEBUG] TestConfig_Delete.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:48.309Z [DEBUG] TestConfig_Delete.server.raft: vote granted: from=9654863c-d898-46d7-277e-78d38158f306 term=2 tally=1 writer.go:29: 2021-01-29T19:33:48.311Z [INFO] TestConfig_Delete.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:48.312Z [INFO] TestConfig_Delete.server.raft: entering leader state: leader="Node at 127.0.0.1:29573 [Leader]" === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:48.313Z [INFO] TestConfig_Apply: Requesting shutdown === CONT TestConfig_Delete writer.go:29: 2021-01-29T19:33:48.314Z [INFO] TestConfig_Delete.server: cluster leadership acquired === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:48.314Z [INFO] TestConfig_Apply.server: shutting down server === CONT TestConfig_Delete writer.go:29: 2021-01-29T19:33:48.315Z [INFO] TestConfig_Delete.server: New leader elected: payload=Node-9654863c-d898-46d7-277e-78d38158f306 === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:48.316Z [DEBUG] TestConfig_Apply.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.318Z [WARN] TestConfig_Apply.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:48.318Z [DEBUG] TestConfig_Apply.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.322Z [WARN] TestConfig_Apply.server.serf.wan: serf: Shutdown without a Leave === CONT TestConfig_Get writer.go:29: 2021-01-29T19:33:48.320Z [WARN] TestConfig_Get: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:48.328Z [DEBUG] TestConfig_Get.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:48.329Z [DEBUG] TestConfig_Get.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:48.327Z [INFO] TestConfig_Apply.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:48.330Z [INFO] TestConfig_Apply: consul server down === CONT TestConfig_Delete writer.go:29: 2021-01-29T19:33:48.319Z [DEBUG] TestConfig_Delete.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29573 === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:48.332Z [INFO] TestConfig_Apply: shutdown complete writer.go:29: 2021-01-29T19:33:48.334Z [INFO] TestConfig_Apply: Stopping server: protocol=DNS address=127.0.0.1:29556 network=tcp writer.go:29: 2021-01-29T19:33:48.336Z [INFO] TestConfig_Apply: Stopping server: protocol=DNS address=127.0.0.1:29556 network=udp === CONT TestConfig_Get writer.go:29: 2021-01-29T19:33:48.338Z [INFO] TestConfig_Get.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:572dfc4c-9c53-9f98-5632-d697932143dd Address:127.0.0.1:29579}]" === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:48.339Z [INFO] TestConfig_Apply: Stopping server: protocol=HTTP address=127.0.0.1:29557 network=tcp === CONT TestConfig_Get writer.go:29: 2021-01-29T19:33:48.340Z [INFO] TestConfig_Get.server.serf.wan: serf: EventMemberJoin: Node-572dfc4c-9c53-9f98-5632-d697932143dd.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:48.340Z [INFO] TestConfig_Get.server.raft: entering follower state: follower="Node at 127.0.0.1:29579 [Follower]" leader= === CONT TestConfig_Delete writer.go:29: 2021-01-29T19:33:48.343Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestConfig_Get writer.go:29: 2021-01-29T19:33:48.346Z [INFO] TestConfig_Get.server.serf.lan: serf: EventMemberJoin: Node-572dfc4c-9c53-9f98-5632-d697932143dd 127.0.0.1 writer.go:29: 2021-01-29T19:33:48.348Z [INFO] TestConfig_Get.server: Adding LAN server: server="Node-572dfc4c-9c53-9f98-5632-d697932143dd (Addr: tcp/127.0.0.1:29579) (DC: dc1)" writer.go:29: 2021-01-29T19:33:48.348Z [INFO] TestConfig_Get.server: Handled event for server in area: event=member-join server=Node-572dfc4c-9c53-9f98-5632-d697932143dd.dc1 area=wan writer.go:29: 2021-01-29T19:33:48.348Z [INFO] TestConfig_Get: Started DNS server: address=127.0.0.1:29574 network=udp writer.go:29: 2021-01-29T19:33:48.353Z [INFO] TestConfig_Get: Started DNS server: address=127.0.0.1:29574 network=tcp writer.go:29: 2021-01-29T19:33:48.356Z [INFO] TestConfig_Get: Started HTTP server: address=127.0.0.1:29575 network=tcp writer.go:29: 2021-01-29T19:33:48.357Z [INFO] TestConfig_Get: started state syncer === CONT TestConfig_Delete writer.go:29: 2021-01-29T19:33:48.355Z [INFO] TestConfig_Delete.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:48.358Z [INFO] TestConfig_Delete.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.361Z [DEBUG] TestConfig_Delete.server: Skipping self join check for node since the cluster is too small: node=Node-9654863c-d898-46d7-277e-78d38158f306 writer.go:29: 2021-01-29T19:33:48.364Z [INFO] TestConfig_Delete.server: member joined, marking health alive: member=Node-9654863c-d898-46d7-277e-78d38158f306 === CONT TestConfig_Get writer.go:29: 2021-01-29T19:33:48.394Z [WARN] TestConfig_Get.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:48.396Z [INFO] TestConfig_Get.server.raft: entering candidate state: node="Node at 127.0.0.1:29579 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:48.399Z [DEBUG] TestConfig_Get.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:48.400Z [DEBUG] TestConfig_Get.server.raft: vote granted: from=572dfc4c-9c53-9f98-5632-d697932143dd term=2 tally=1 writer.go:29: 2021-01-29T19:33:48.401Z [INFO] TestConfig_Get.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:48.402Z [INFO] TestConfig_Get.server.raft: entering leader state: leader="Node at 127.0.0.1:29579 [Leader]" writer.go:29: 2021-01-29T19:33:48.403Z [INFO] TestConfig_Get.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:48.403Z [INFO] TestConfig_Get.server: New leader elected: payload=Node-572dfc4c-9c53-9f98-5632-d697932143dd writer.go:29: 2021-01-29T19:33:48.404Z [DEBUG] TestConfig_Get.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29579 writer.go:29: 2021-01-29T19:33:48.413Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:48.421Z [INFO] TestConfig_Get.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:48.423Z [INFO] TestConfig_Get.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.424Z [DEBUG] TestConfig_Get.server: Skipping self join check for node since the cluster is too small: node=Node-572dfc4c-9c53-9f98-5632-d697932143dd writer.go:29: 2021-01-29T19:33:48.425Z [INFO] TestConfig_Get.server: member joined, marking health alive: member=Node-572dfc4c-9c53-9f98-5632-d697932143dd writer.go:29: 2021-01-29T19:33:48.441Z [DEBUG] TestConfig_Get: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:48.443Z [INFO] TestConfig_Get: Synced node info writer.go:29: 2021-01-29T19:33:48.444Z [DEBUG] TestConfig_Get: Node info in sync === RUN TestConfig_Get/get_a_single_service_entry === RUN TestConfig_Get/list_both_service_entries === RUN TestConfig_Get/get_global_proxy_config === RUN TestConfig_Get/error_on_no_arguments === CONT TestConfig_Get writer.go:29: 2021-01-29T19:33:48.512Z [INFO] TestConfig_Get: Requesting shutdown writer.go:29: 2021-01-29T19:33:48.514Z [INFO] TestConfig_Get.server: shutting down server writer.go:29: 2021-01-29T19:33:48.515Z [DEBUG] TestConfig_Get.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.516Z [WARN] TestConfig_Get.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:48.517Z [DEBUG] TestConfig_Get.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.518Z [WARN] TestConfig_Get.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:48.524Z [INFO] TestConfig_Get.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:48.524Z [INFO] TestConfig_Get: consul server down writer.go:29: 2021-01-29T19:33:48.526Z [INFO] TestConfig_Get: shutdown complete writer.go:29: 2021-01-29T19:33:48.527Z [INFO] TestConfig_Get: Stopping server: protocol=DNS address=127.0.0.1:29574 network=tcp writer.go:29: 2021-01-29T19:33:48.527Z [INFO] TestConfig_Get: Stopping server: protocol=DNS address=127.0.0.1:29574 network=udp writer.go:29: 2021-01-29T19:33:48.528Z [INFO] TestConfig_Get: Stopping server: protocol=HTTP address=127.0.0.1:29575 network=tcp === CONT TestConfig_Delete writer.go:29: 2021-01-29T19:33:48.685Z [INFO] TestConfig_Delete: Requesting shutdown writer.go:29: 2021-01-29T19:33:48.686Z [INFO] TestConfig_Delete.server: shutting down server writer.go:29: 2021-01-29T19:33:48.688Z [DEBUG] TestConfig_Delete.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.688Z [ERROR] TestConfig_Delete.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:48.689Z [WARN] TestConfig_Delete.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:48.689Z [DEBUG] TestConfig_Delete.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:48.692Z [WARN] TestConfig_Delete.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:48.697Z [INFO] TestConfig_Delete.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:48.697Z [INFO] TestConfig_Delete: consul server down writer.go:29: 2021-01-29T19:33:48.699Z [INFO] TestConfig_Delete: shutdown complete writer.go:29: 2021-01-29T19:33:48.700Z [INFO] TestConfig_Delete: Stopping server: protocol=DNS address=127.0.0.1:29568 network=tcp writer.go:29: 2021-01-29T19:33:48.702Z [INFO] TestConfig_Delete: Stopping server: protocol=DNS address=127.0.0.1:29568 network=udp writer.go:29: 2021-01-29T19:33:48.704Z [INFO] TestConfig_Delete: Stopping server: protocol=HTTP address=127.0.0.1:29569 network=tcp === CONT TestConnectCAConfig/force_without_cross_sign_CamelCase writer.go:29: 2021-01-29T19:33:48.802Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:48.804Z [INFO] TestConnectCAConfig/force_without_cross_sign_CamelCase: Endpoints down === RUN TestConnectCAConfig/force_without_cross_sign_snake_case === CONT TestConfig_Apply writer.go:29: 2021-01-29T19:33:48.842Z [INFO] TestConfig_Apply: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:48.844Z [INFO] TestConfig_Apply: Endpoints down --- PASS: TestConfig_Apply (0.95s) === CONT TestCatalogNodeServices_ConnectProxy === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:48.861Z [WARN] TestConnectCAConfig/force_without_cross_sign_snake_case: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:48.863Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_snake_case.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:48.865Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_snake_case.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:48.893Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:545f6dfb-ede4-e09e-12e2-bf19807d2b81 Address:127.0.0.1:29585}]" writer.go:29: 2021-01-29T19:33:48.896Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server.serf.wan: serf: EventMemberJoin: Node-545f6dfb-ede4-e09e-12e2-bf19807d2b81.dc1 127.0.0.1 === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:48.901Z [WARN] TestCatalogNodeServices_ConnectProxy: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:48.902Z [DEBUG] TestCatalogNodeServices_ConnectProxy.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:48.904Z [DEBUG] TestCatalogNodeServices_ConnectProxy.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:48.900Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server.serf.lan: serf: EventMemberJoin: Node-545f6dfb-ede4-e09e-12e2-bf19807d2b81 127.0.0.1 writer.go:29: 2021-01-29T19:33:48.896Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server.raft: entering follower state: follower="Node at 127.0.0.1:29585 [Follower]" leader= writer.go:29: 2021-01-29T19:33:48.916Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: Started DNS server: address=127.0.0.1:29580 network=udp writer.go:29: 2021-01-29T19:33:48.923Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: Started DNS server: address=127.0.0.1:29580 network=tcp writer.go:29: 2021-01-29T19:33:48.918Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server: Adding LAN server: server="Node-545f6dfb-ede4-e09e-12e2-bf19807d2b81 (Addr: tcp/127.0.0.1:29585) (DC: dc1)" writer.go:29: 2021-01-29T19:33:48.918Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server: Handled event for server in area: event=member-join server=Node-545f6dfb-ede4-e09e-12e2-bf19807d2b81.dc1 area=wan writer.go:29: 2021-01-29T19:33:48.936Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: Started HTTP server: address=127.0.0.1:29581 network=tcp writer.go:29: 2021-01-29T19:33:48.938Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: started state syncer === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:48.964Z [INFO] TestCatalogNodeServices_ConnectProxy.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:29a9670d-10a3-530e-7a23-4d2daa22af46 Address:127.0.0.1:29591}]" writer.go:29: 2021-01-29T19:33:48.967Z [INFO] TestCatalogNodeServices_ConnectProxy.server.serf.wan: serf: EventMemberJoin: Node-29a9670d-10a3-530e-7a23-4d2daa22af46.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:48.978Z [INFO] TestCatalogNodeServices_ConnectProxy.server.raft: entering follower state: follower="Node at 127.0.0.1:29591 [Follower]" leader= writer.go:29: 2021-01-29T19:33:48.992Z [INFO] TestCatalogNodeServices_ConnectProxy.server.serf.lan: serf: EventMemberJoin: Node-29a9670d-10a3-530e-7a23-4d2daa22af46 127.0.0.1 writer.go:29: 2021-01-29T19:33:48.996Z [INFO] TestCatalogNodeServices_ConnectProxy: Started DNS server: address=127.0.0.1:29586 network=udp writer.go:29: 2021-01-29T19:33:48.999Z [INFO] TestCatalogNodeServices_ConnectProxy.server: Adding LAN server: server="Node-29a9670d-10a3-530e-7a23-4d2daa22af46 (Addr: tcp/127.0.0.1:29591) (DC: dc1)" === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:48.974Z [WARN] TestConnectCAConfig/force_without_cross_sign_snake_case.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:49.000Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server.raft: entering candidate state: node="Node at 127.0.0.1:29585 [Candidate]" term=2 === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.022Z [INFO] TestCatalogNodeServices_ConnectProxy.server: Handled event for server in area: event=member-join server=Node-29a9670d-10a3-530e-7a23-4d2daa22af46.dc1 area=wan writer.go:29: 2021-01-29T19:33:49.030Z [INFO] TestCatalogNodeServices_ConnectProxy: Started DNS server: address=127.0.0.1:29586 network=tcp === CONT TestConfig_Get writer.go:29: 2021-01-29T19:33:49.030Z [INFO] TestConfig_Get: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:49.035Z [INFO] TestConfig_Get: Endpoints down --- PASS: TestConfig_Get (0.73s) --- PASS: TestConfig_Get/get_a_single_service_entry (0.01s) --- PASS: TestConfig_Get/list_both_service_entries (0.01s) --- PASS: TestConfig_Get/get_global_proxy_config (0.01s) --- PASS: TestConfig_Get/error_on_no_arguments (0.00s) === CONT TestCatalogNodeServices_Filter === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.034Z [WARN] TestCatalogNodeServices_ConnectProxy.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:49.040Z [INFO] TestCatalogNodeServices_ConnectProxy.server.raft: entering candidate state: node="Node at 127.0.0.1:29591 [Candidate]" term=2 === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.050Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_snake_case.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:49.051Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_snake_case.server.raft: vote granted: from=545f6dfb-ede4-e09e-12e2-bf19807d2b81 term=2 tally=1 writer.go:29: 2021-01-29T19:33:49.055Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:49.056Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server.raft: entering leader state: leader="Node at 127.0.0.1:29585 [Leader]" writer.go:29: 2021-01-29T19:33:49.057Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:49.058Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server: New leader elected: payload=Node-545f6dfb-ede4-e09e-12e2-bf19807d2b81 === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.039Z [INFO] TestCatalogNodeServices_ConnectProxy: Started HTTP server: address=127.0.0.1:29587 network=tcp writer.go:29: 2021-01-29T19:33:49.061Z [INFO] TestCatalogNodeServices_ConnectProxy: started state syncer writer.go:29: 2021-01-29T19:33:49.070Z [DEBUG] TestCatalogNodeServices_ConnectProxy.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:49.070Z [DEBUG] TestCatalogNodeServices_ConnectProxy.server.raft: vote granted: from=29a9670d-10a3-530e-7a23-4d2daa22af46 term=2 tally=1 writer.go:29: 2021-01-29T19:33:49.071Z [INFO] TestCatalogNodeServices_ConnectProxy.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:49.076Z [INFO] TestCatalogNodeServices_ConnectProxy.server.raft: entering leader state: leader="Node at 127.0.0.1:29591 [Leader]" writer.go:29: 2021-01-29T19:33:49.077Z [INFO] TestCatalogNodeServices_ConnectProxy.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:49.079Z [INFO] TestCatalogNodeServices_ConnectProxy.server: New leader elected: payload=Node-29a9670d-10a3-530e-7a23-4d2daa22af46 === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.079Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_snake_case.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29585 === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.083Z [DEBUG] TestCatalogNodeServices_ConnectProxy.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29591 === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.091Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.091Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.100Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:49.101Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.115Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_snake_case.server: Skipping self join check for node since the cluster is too small: node=Node-545f6dfb-ede4-e09e-12e2-bf19807d2b81 writer.go:29: 2021-01-29T19:33:49.117Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server: member joined, marking health alive: member=Node-545f6dfb-ede4-e09e-12e2-bf19807d2b81 === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.121Z [INFO] TestCatalogNodeServices_ConnectProxy.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:49.132Z [INFO] TestCatalogNodeServices_ConnectProxy.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.134Z [DEBUG] TestCatalogNodeServices_ConnectProxy.server: Skipping self join check for node since the cluster is too small: node=Node-29a9670d-10a3-530e-7a23-4d2daa22af46 writer.go:29: 2021-01-29T19:33:49.135Z [INFO] TestCatalogNodeServices_ConnectProxy.server: member joined, marking health alive: member=Node-29a9670d-10a3-530e-7a23-4d2daa22af46 === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.156Z [WARN] TestCatalogNodeServices_Filter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:49.160Z [DEBUG] TestCatalogNodeServices_Filter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:49.162Z [DEBUG] TestCatalogNodeServices_Filter.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:49.175Z [INFO] TestCatalogNodeServices_Filter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e611bdb1-a39c-3b42-6aa4-f8429b298bae Address:127.0.0.1:29597}]" writer.go:29: 2021-01-29T19:33:49.183Z [INFO] TestCatalogNodeServices_Filter.server.raft: entering follower state: follower="Node at 127.0.0.1:29597 [Follower]" leader= writer.go:29: 2021-01-29T19:33:49.188Z [INFO] TestCatalogNodeServices_Filter.server.serf.wan: serf: EventMemberJoin: Node-e611bdb1-a39c-3b42-6aa4-f8429b298bae.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:49.194Z [INFO] TestCatalogNodeServices_Filter.server.serf.lan: serf: EventMemberJoin: Node-e611bdb1-a39c-3b42-6aa4-f8429b298bae 127.0.0.1 writer.go:29: 2021-01-29T19:33:49.198Z [INFO] TestCatalogNodeServices_Filter.server: Handled event for server in area: event=member-join server=Node-e611bdb1-a39c-3b42-6aa4-f8429b298bae.dc1 area=wan writer.go:29: 2021-01-29T19:33:49.199Z [INFO] TestCatalogNodeServices_Filter.server: Adding LAN server: server="Node-e611bdb1-a39c-3b42-6aa4-f8429b298bae (Addr: tcp/127.0.0.1:29597) (DC: dc1)" writer.go:29: 2021-01-29T19:33:49.201Z [INFO] TestCatalogNodeServices_Filter: Started DNS server: address=127.0.0.1:29592 network=tcp writer.go:29: 2021-01-29T19:33:49.203Z [INFO] TestCatalogNodeServices_Filter: Started DNS server: address=127.0.0.1:29592 network=udp === CONT TestConfig_Delete writer.go:29: 2021-01-29T19:33:49.205Z [INFO] TestConfig_Delete: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:49.206Z [INFO] TestConfig_Delete: Endpoints down --- PASS: TestConfig_Delete (0.99s) === CONT TestCatalogNodeServiceList === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.206Z [INFO] TestCatalogNodeServices_Filter: Started HTTP server: address=127.0.0.1:29593 network=tcp writer.go:29: 2021-01-29T19:33:49.211Z [INFO] TestCatalogNodeServices_Filter: started state syncer === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.214Z [WARN] TestCatalogNodeServiceList: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:49.215Z [DEBUG] TestCatalogNodeServiceList.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:49.216Z [DEBUG] TestCatalogNodeServiceList.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.230Z [WARN] TestCatalogNodeServices_Filter.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.232Z [INFO] TestCatalogNodeServiceList.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:13e64a60-c625-2a4a-afb0-54717fef532c Address:127.0.0.1:29603}]" === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.232Z [INFO] TestCatalogNodeServices_Filter.server.raft: entering candidate state: node="Node at 127.0.0.1:29597 [Candidate]" term=2 === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.234Z [INFO] TestCatalogNodeServices_ConnectProxy: Requesting shutdown writer.go:29: 2021-01-29T19:33:49.235Z [INFO] TestCatalogNodeServices_ConnectProxy.server: shutting down server === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.235Z [DEBUG] TestCatalogNodeServices_Filter.server.raft: votes: needed=1 === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.235Z [INFO] TestCatalogNodeServiceList.server.raft: entering follower state: follower="Node at 127.0.0.1:29603 [Follower]" leader= === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.236Z [DEBUG] TestCatalogNodeServices_ConnectProxy.leader: stopping routine: routine="CA root pruning" === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.236Z [DEBUG] TestCatalogNodeServices_Filter.server.raft: vote granted: from=e611bdb1-a39c-3b42-6aa4-f8429b298bae term=2 tally=1 writer.go:29: 2021-01-29T19:33:49.238Z [INFO] TestCatalogNodeServices_Filter.server.raft: election won: tally=1 === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.235Z [INFO] TestCatalogNodeServiceList.server.serf.wan: serf: EventMemberJoin: Node-13e64a60-c625-2a4a-afb0-54717fef532c.dc1 127.0.0.1 === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.237Z [ERROR] TestCatalogNodeServices_ConnectProxy.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.239Z [INFO] TestCatalogNodeServices_Filter.server.raft: entering leader state: leader="Node at 127.0.0.1:29597 [Leader]" === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.240Z [INFO] TestCatalogNodeServiceList.server.serf.lan: serf: EventMemberJoin: Node-13e64a60-c625-2a4a-afb0-54717fef532c 127.0.0.1 === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.240Z [INFO] TestCatalogNodeServices_Filter.server: cluster leadership acquired === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.238Z [WARN] TestCatalogNodeServices_ConnectProxy.server.serf.lan: serf: Shutdown without a Leave === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.241Z [INFO] TestCatalogNodeServices_Filter.server: New leader elected: payload=Node-e611bdb1-a39c-3b42-6aa4-f8429b298bae === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.238Z [DEBUG] TestCatalogNodeServices_ConnectProxy.leader: stopped routine: routine="CA root pruning" === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.243Z [INFO] TestCatalogNodeServiceList: Started DNS server: address=127.0.0.1:29598 network=udp writer.go:29: 2021-01-29T19:33:49.245Z [INFO] TestCatalogNodeServiceList.server: Handled event for server in area: event=member-join server=Node-13e64a60-c625-2a4a-afb0-54717fef532c.dc1 area=wan === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.246Z [WARN] TestCatalogNodeServices_ConnectProxy.server.serf.wan: serf: Shutdown without a Leave === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.247Z [INFO] TestCatalogNodeServiceList.server: Adding LAN server: server="Node-13e64a60-c625-2a4a-afb0-54717fef532c (Addr: tcp/127.0.0.1:29603) (DC: dc1)" writer.go:29: 2021-01-29T19:33:49.249Z [INFO] TestCatalogNodeServiceList: Started DNS server: address=127.0.0.1:29598 network=tcp === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.249Z [INFO] TestCatalogNodeServices_ConnectProxy.server.router.manager: shutting down === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.252Z [INFO] TestCatalogNodeServiceList: Started HTTP server: address=127.0.0.1:29599 network=tcp === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.250Z [INFO] TestCatalogNodeServices_ConnectProxy: consul server down === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.253Z [INFO] TestCatalogNodeServiceList: started state syncer === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.254Z [INFO] TestCatalogNodeServices_ConnectProxy: shutdown complete writer.go:29: 2021-01-29T19:33:49.256Z [INFO] TestCatalogNodeServices_ConnectProxy: Stopping server: protocol=DNS address=127.0.0.1:29586 network=tcp writer.go:29: 2021-01-29T19:33:49.258Z [INFO] TestCatalogNodeServices_ConnectProxy: Stopping server: protocol=DNS address=127.0.0.1:29586 network=udp writer.go:29: 2021-01-29T19:33:49.259Z [INFO] TestCatalogNodeServices_ConnectProxy: Stopping server: protocol=HTTP address=127.0.0.1:29587 network=tcp === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.248Z [DEBUG] TestCatalogNodeServices_Filter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29597 writer.go:29: 2021-01-29T19:33:49.267Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:49.274Z [INFO] TestCatalogNodeServices_Filter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:49.275Z [INFO] TestCatalogNodeServices_Filter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.276Z [DEBUG] TestCatalogNodeServices_Filter.server: Skipping self join check for node since the cluster is too small: node=Node-e611bdb1-a39c-3b42-6aa4-f8429b298bae === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.279Z [WARN] TestCatalogNodeServiceList.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.278Z [INFO] TestCatalogNodeServices_Filter.server: member joined, marking health alive: member=Node-e611bdb1-a39c-3b42-6aa4-f8429b298bae === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.280Z [INFO] TestCatalogNodeServiceList.server.raft: entering candidate state: node="Node at 127.0.0.1:29603 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:49.283Z [DEBUG] TestCatalogNodeServiceList.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:49.284Z [DEBUG] TestCatalogNodeServiceList.server.raft: vote granted: from=13e64a60-c625-2a4a-afb0-54717fef532c term=2 tally=1 writer.go:29: 2021-01-29T19:33:49.286Z [INFO] TestCatalogNodeServiceList.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:49.287Z [INFO] TestCatalogNodeServiceList.server.raft: entering leader state: leader="Node at 127.0.0.1:29603 [Leader]" writer.go:29: 2021-01-29T19:33:49.288Z [INFO] TestCatalogNodeServiceList.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:49.289Z [DEBUG] TestCatalogNodeServiceList.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29603 writer.go:29: 2021-01-29T19:33:49.289Z [INFO] TestCatalogNodeServiceList.server: New leader elected: payload=Node-13e64a60-c625-2a4a-afb0-54717fef532c writer.go:29: 2021-01-29T19:33:49.297Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.301Z [INFO] TestCatalogNodeServices_Filter: Requesting shutdown writer.go:29: 2021-01-29T19:33:49.302Z [INFO] TestCatalogNodeServices_Filter.server: shutting down server writer.go:29: 2021-01-29T19:33:49.303Z [DEBUG] TestCatalogNodeServices_Filter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.303Z [WARN] TestCatalogNodeServices_Filter.server.serf.lan: serf: Shutdown without a Leave === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.304Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_snake_case: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.303Z [ERROR] TestCatalogNodeServices_Filter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:49.304Z [DEBUG] TestCatalogNodeServices_Filter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.306Z [WARN] TestCatalogNodeServices_Filter.server.serf.wan: serf: Shutdown without a Leave === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.306Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: Synced node info === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.308Z [INFO] TestCatalogNodeServices_Filter.server.router.manager: shutting down === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.308Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_snake_case: Node info in sync === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.308Z [INFO] TestCatalogNodeServices_Filter: consul server down === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.307Z [INFO] TestCatalogNodeServiceList.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.312Z [INFO] TestCatalogNodeServices_Filter: shutdown complete writer.go:29: 2021-01-29T19:33:49.313Z [INFO] TestCatalogNodeServices_Filter: Stopping server: protocol=DNS address=127.0.0.1:29592 network=tcp === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.312Z [INFO] TestCatalogNodeServiceList.leader: started routine: routine="CA root pruning" === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.316Z [INFO] TestCatalogNodeServices_Filter: Stopping server: protocol=DNS address=127.0.0.1:29592 network=udp writer.go:29: 2021-01-29T19:33:49.317Z [INFO] TestCatalogNodeServices_Filter: Stopping server: protocol=HTTP address=127.0.0.1:29593 network=tcp === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.317Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server.connect: CA provider config updated === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.317Z [DEBUG] TestCatalogNodeServiceList.server: Skipping self join check for node since the cluster is too small: node=Node-13e64a60-c625-2a4a-afb0-54717fef532c writer.go:29: 2021-01-29T19:33:49.323Z [INFO] TestCatalogNodeServiceList.server: member joined, marking health alive: member=Node-13e64a60-c625-2a4a-afb0-54717fef532c === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.324Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: Requesting shutdown writer.go:29: 2021-01-29T19:33:49.326Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server: shutting down server writer.go:29: 2021-01-29T19:33:49.327Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_snake_case.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.328Z [WARN] TestConnectCAConfig/force_without_cross_sign_snake_case.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:49.328Z [DEBUG] TestConnectCAConfig/force_without_cross_sign_snake_case.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.338Z [WARN] TestConnectCAConfig/force_without_cross_sign_snake_case.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:49.340Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:49.341Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: consul server down writer.go:29: 2021-01-29T19:33:49.344Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: shutdown complete writer.go:29: 2021-01-29T19:33:49.345Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: Stopping server: protocol=DNS address=127.0.0.1:29580 network=tcp writer.go:29: 2021-01-29T19:33:49.347Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: Stopping server: protocol=DNS address=127.0.0.1:29580 network=udp writer.go:29: 2021-01-29T19:33:49.348Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: Stopping server: protocol=HTTP address=127.0.0.1:29581 network=tcp === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.363Z [DEBUG] TestCatalogNodeServiceList: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:49.366Z [INFO] TestCatalogNodeServiceList: Synced node info writer.go:29: 2021-01-29T19:33:49.423Z [INFO] TestCatalogNodeServiceList: Requesting shutdown writer.go:29: 2021-01-29T19:33:49.424Z [INFO] TestCatalogNodeServiceList.server: shutting down server writer.go:29: 2021-01-29T19:33:49.426Z [DEBUG] TestCatalogNodeServiceList.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.427Z [WARN] TestCatalogNodeServiceList.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:49.427Z [DEBUG] TestCatalogNodeServiceList.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.431Z [WARN] TestCatalogNodeServiceList.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:49.434Z [INFO] TestCatalogNodeServiceList.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:49.434Z [INFO] TestCatalogNodeServiceList: consul server down writer.go:29: 2021-01-29T19:33:49.436Z [INFO] TestCatalogNodeServiceList: shutdown complete writer.go:29: 2021-01-29T19:33:49.438Z [INFO] TestCatalogNodeServiceList: Stopping server: protocol=DNS address=127.0.0.1:29598 network=tcp writer.go:29: 2021-01-29T19:33:49.439Z [INFO] TestCatalogNodeServiceList: Stopping server: protocol=DNS address=127.0.0.1:29598 network=udp writer.go:29: 2021-01-29T19:33:49.441Z [INFO] TestCatalogNodeServiceList: Stopping server: protocol=HTTP address=127.0.0.1:29599 network=tcp === CONT TestCatalogNodeServices_ConnectProxy writer.go:29: 2021-01-29T19:33:49.761Z [INFO] TestCatalogNodeServices_ConnectProxy: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:49.764Z [INFO] TestCatalogNodeServices_ConnectProxy: Endpoints down --- PASS: TestCatalogNodeServices_ConnectProxy (0.92s) === CONT TestCatalogNodeServices === CONT TestCatalogNodeServices_Filter writer.go:29: 2021-01-29T19:33:49.818Z [INFO] TestCatalogNodeServices_Filter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:49.819Z [INFO] TestCatalogNodeServices_Filter: Endpoints down --- PASS: TestCatalogNodeServices_Filter (0.78s) === CONT TestCatalogConnectServiceNodes_good === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.824Z [WARN] TestCatalogNodeServices: bootstrap = true: do not enable unless necessary === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.834Z [WARN] TestCatalogConnectServiceNodes_good: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:49.836Z [DEBUG] TestCatalogConnectServiceNodes_good.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:49.838Z [DEBUG] TestCatalogConnectServiceNodes_good.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:49.843Z [INFO] TestCatalogConnectServiceNodes_good.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:13fcc084-20fa-066b-903d-cd8c1b5c5842 Address:127.0.0.1:29609}]" === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.830Z [DEBUG] TestCatalogNodeServices.tlsutil: Update: version=1 === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.846Z [INFO] TestCatalogConnectServiceNodes_good.server.raft: entering follower state: follower="Node at 127.0.0.1:29609 [Follower]" leader= === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.850Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: Waiting for endpoints to shut down === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.850Z [DEBUG] TestCatalogNodeServices.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestConnectCAConfig/force_without_cross_sign_snake_case writer.go:29: 2021-01-29T19:33:49.858Z [INFO] TestConnectCAConfig/force_without_cross_sign_snake_case: Endpoints down === RUN TestConnectCAConfig/setting_state_fails === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.855Z [INFO] TestCatalogConnectServiceNodes_good.server.serf.wan: serf: EventMemberJoin: Node-13fcc084-20fa-066b-903d-cd8c1b5c5842.dc1 127.0.0.1 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.879Z [INFO] TestCatalogNodeServices.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cd4aed90-1cd2-6bba-d9ed-173ad718a4c1 Address:127.0.0.1:29615}]" === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.888Z [INFO] TestCatalogConnectServiceNodes_good.server.serf.lan: serf: EventMemberJoin: Node-13fcc084-20fa-066b-903d-cd8c1b5c5842 127.0.0.1 writer.go:29: 2021-01-29T19:33:49.892Z [INFO] TestCatalogConnectServiceNodes_good.server: Adding LAN server: server="Node-13fcc084-20fa-066b-903d-cd8c1b5c5842 (Addr: tcp/127.0.0.1:29609) (DC: dc1)" writer.go:29: 2021-01-29T19:33:49.894Z [INFO] TestCatalogConnectServiceNodes_good: Started DNS server: address=127.0.0.1:29604 network=udp === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:49.895Z [WARN] TestConnectCAConfig/setting_state_fails: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:49.897Z [DEBUG] TestConnectCAConfig/setting_state_fails.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:49.899Z [DEBUG] TestConnectCAConfig/setting_state_fails.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.896Z [INFO] TestCatalogConnectServiceNodes_good.server: Handled event for server in area: event=member-join server=Node-13fcc084-20fa-066b-903d-cd8c1b5c5842.dc1 area=wan === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.891Z [INFO] TestCatalogNodeServices.server.raft: entering follower state: follower="Node at 127.0.0.1:29615 [Follower]" leader= writer.go:29: 2021-01-29T19:33:49.894Z [INFO] TestCatalogNodeServices.server.serf.wan: serf: EventMemberJoin: Node-cd4aed90-1cd2-6bba-d9ed-173ad718a4c1.dc1 127.0.0.1 === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.899Z [WARN] TestCatalogConnectServiceNodes_good.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:49.910Z [INFO] TestCatalogConnectServiceNodes_good.server.raft: entering candidate state: node="Node at 127.0.0.1:29609 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:49.899Z [INFO] TestCatalogConnectServiceNodes_good: Started DNS server: address=127.0.0.1:29604 network=tcp writer.go:29: 2021-01-29T19:33:49.919Z [INFO] TestCatalogConnectServiceNodes_good: Started HTTP server: address=127.0.0.1:29605 network=tcp writer.go:29: 2021-01-29T19:33:49.920Z [INFO] TestCatalogConnectServiceNodes_good: started state syncer === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.915Z [INFO] TestCatalogNodeServices.server.serf.lan: serf: EventMemberJoin: Node-cd4aed90-1cd2-6bba-d9ed-173ad718a4c1 127.0.0.1 === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.925Z [DEBUG] TestCatalogConnectServiceNodes_good.server.raft: votes: needed=1 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.930Z [INFO] TestCatalogNodeServices.server: Adding LAN server: server="Node-cd4aed90-1cd2-6bba-d9ed-173ad718a4c1 (Addr: tcp/127.0.0.1:29615) (DC: dc1)" writer.go:29: 2021-01-29T19:33:49.930Z [INFO] TestCatalogNodeServices.server: Handled event for server in area: event=member-join server=Node-cd4aed90-1cd2-6bba-d9ed-173ad718a4c1.dc1 area=wan === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:49.934Z [INFO] TestConnectCAConfig/setting_state_fails.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c19b42df-2da5-e2a0-9ba1-c70c5374a68d Address:127.0.0.1:29621}]" === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.929Z [DEBUG] TestCatalogConnectServiceNodes_good.server.raft: vote granted: from=13fcc084-20fa-066b-903d-cd8c1b5c5842 term=2 tally=1 writer.go:29: 2021-01-29T19:33:49.940Z [INFO] TestCatalogConnectServiceNodes_good.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:49.942Z [INFO] TestCatalogConnectServiceNodes_good.server.raft: entering leader state: leader="Node at 127.0.0.1:29609 [Leader]" writer.go:29: 2021-01-29T19:33:49.944Z [INFO] TestCatalogConnectServiceNodes_good.server: cluster leadership acquired === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.943Z [INFO] TestCatalogNodeServices: Started DNS server: address=127.0.0.1:29610 network=udp === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:49.939Z [INFO] TestConnectCAConfig/setting_state_fails.server.serf.wan: serf: EventMemberJoin: Node-c19b42df-2da5-e2a0-9ba1-c70c5374a68d.dc1 127.0.0.1 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.951Z [INFO] TestCatalogNodeServices: Started DNS server: address=127.0.0.1:29610 network=tcp === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.954Z [INFO] TestCatalogNodeServiceList: Waiting for endpoints to shut down === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.944Z [WARN] TestCatalogNodeServices.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCatalogNodeServiceList writer.go:29: 2021-01-29T19:33:49.956Z [INFO] TestCatalogNodeServiceList: Endpoints down --- PASS: TestCatalogNodeServiceList (0.75s) === CONT TestCatalogServiceNodes_ConnectProxy === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.957Z [INFO] TestCatalogNodeServices.server.raft: entering candidate state: node="Node at 127.0.0.1:29615 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:49.959Z [DEBUG] TestCatalogNodeServices.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:49.962Z [DEBUG] TestCatalogNodeServices.server.raft: vote granted: from=cd4aed90-1cd2-6bba-d9ed-173ad718a4c1 term=2 tally=1 === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.956Z [DEBUG] TestCatalogConnectServiceNodes_good.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29609 === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:49.958Z [INFO] TestConnectCAConfig/setting_state_fails.server.serf.lan: serf: EventMemberJoin: Node-c19b42df-2da5-e2a0-9ba1-c70c5374a68d 127.0.0.1 === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.961Z [INFO] TestCatalogConnectServiceNodes_good.server: New leader elected: payload=Node-13fcc084-20fa-066b-903d-cd8c1b5c5842 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.967Z [INFO] TestCatalogNodeServices.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:49.964Z [INFO] TestCatalogNodeServices: Started HTTP server: address=127.0.0.1:29611 network=tcp === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.982Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:49.940Z [INFO] TestConnectCAConfig/setting_state_fails.server.raft: entering follower state: follower="Node at 127.0.0.1:29621 [Follower]" leader= writer.go:29: 2021-01-29T19:33:49.976Z [INFO] TestConnectCAConfig/setting_state_fails: Started DNS server: address=127.0.0.1:29616 network=udp writer.go:29: 2021-01-29T19:33:49.987Z [INFO] TestConnectCAConfig/setting_state_fails: Started DNS server: address=127.0.0.1:29616 network=tcp writer.go:29: 2021-01-29T19:33:49.977Z [INFO] TestConnectCAConfig/setting_state_fails.server: Adding LAN server: server="Node-c19b42df-2da5-e2a0-9ba1-c70c5374a68d (Addr: tcp/127.0.0.1:29621) (DC: dc1)" === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.981Z [INFO] TestCatalogNodeServices: started state syncer === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.996Z [INFO] TestCatalogConnectServiceNodes_good: Synced node info === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:49.977Z [INFO] TestConnectCAConfig/setting_state_fails.server: Handled event for server in area: event=member-join server=Node-c19b42df-2da5-e2a0-9ba1-c70c5374a68d.dc1 area=wan === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:49.998Z [DEBUG] TestCatalogConnectServiceNodes_good: Node info in sync === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:49.999Z [WARN] TestCatalogServiceNodes_ConnectProxy: bootstrap = true: do not enable unless necessary === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:49.974Z [INFO] TestCatalogNodeServices.server.raft: entering leader state: leader="Node at 127.0.0.1:29615 [Leader]" === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:49.992Z [INFO] TestConnectCAConfig/setting_state_fails: Started HTTP server: address=127.0.0.1:29617 network=tcp === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.002Z [DEBUG] TestCatalogServiceNodes_ConnectProxy.tlsutil: Update: version=1 === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:50.002Z [INFO] TestCatalogConnectServiceNodes_good.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.004Z [INFO] TestConnectCAConfig/setting_state_fails: started state syncer === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.007Z [DEBUG] TestCatalogServiceNodes_ConnectProxy.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.013Z [INFO] TestCatalogNodeServices.server: cluster leadership acquired === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:50.007Z [INFO] TestCatalogConnectServiceNodes_good.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.014Z [DEBUG] TestCatalogConnectServiceNodes_good.server: Skipping self join check for node since the cluster is too small: node=Node-13fcc084-20fa-066b-903d-cd8c1b5c5842 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.014Z [INFO] TestCatalogNodeServices.server: New leader elected: payload=Node-cd4aed90-1cd2-6bba-d9ed-173ad718a4c1 === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:50.015Z [INFO] TestCatalogConnectServiceNodes_good.server: member joined, marking health alive: member=Node-13fcc084-20fa-066b-903d-cd8c1b5c5842 === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.018Z [INFO] TestCatalogServiceNodes_ConnectProxy.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0869be48-978f-3e03-f93a-267339cece96 Address:127.0.0.1:29627}]" === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.018Z [INFO] TestCatalogNodeServices: Synced node info === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.020Z [INFO] TestCatalogServiceNodes_ConnectProxy.server.raft: entering follower state: follower="Node at 127.0.0.1:29627 [Follower]" leader= writer.go:29: 2021-01-29T19:33:50.020Z [INFO] TestCatalogServiceNodes_ConnectProxy.server.serf.wan: serf: EventMemberJoin: Node-0869be48-978f-3e03-f93a-267339cece96.dc1 127.0.0.1 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.020Z [DEBUG] TestCatalogNodeServices.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29615 === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.024Z [INFO] TestCatalogServiceNodes_ConnectProxy.server.serf.lan: serf: EventMemberJoin: Node-0869be48-978f-3e03-f93a-267339cece96 127.0.0.1 writer.go:29: 2021-01-29T19:33:50.027Z [INFO] TestCatalogServiceNodes_ConnectProxy.server: Adding LAN server: server="Node-0869be48-978f-3e03-f93a-267339cece96 (Addr: tcp/127.0.0.1:29627) (DC: dc1)" writer.go:29: 2021-01-29T19:33:50.027Z [INFO] TestCatalogServiceNodes_ConnectProxy.server: Handled event for server in area: event=member-join server=Node-0869be48-978f-3e03-f93a-267339cece96.dc1 area=wan writer.go:29: 2021-01-29T19:33:50.028Z [INFO] TestCatalogServiceNodes_ConnectProxy: Started DNS server: address=127.0.0.1:29622 network=udp writer.go:29: 2021-01-29T19:33:50.030Z [INFO] TestCatalogServiceNodes_ConnectProxy: Started DNS server: address=127.0.0.1:29622 network=tcp === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.032Z [WARN] TestConnectCAConfig/setting_state_fails.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.031Z [INFO] TestCatalogServiceNodes_ConnectProxy: Started HTTP server: address=127.0.0.1:29623 network=tcp === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.033Z [INFO] TestConnectCAConfig/setting_state_fails.server.raft: entering candidate state: node="Node at 127.0.0.1:29621 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:50.036Z [DEBUG] TestConnectCAConfig/setting_state_fails.server.raft: votes: needed=1 === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.034Z [INFO] TestCatalogServiceNodes_ConnectProxy: started state syncer === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.037Z [DEBUG] TestConnectCAConfig/setting_state_fails.server.raft: vote granted: from=c19b42df-2da5-e2a0-9ba1-c70c5374a68d term=2 tally=1 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.036Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.039Z [INFO] TestConnectCAConfig/setting_state_fails.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:50.040Z [INFO] TestConnectCAConfig/setting_state_fails.server.raft: entering leader state: leader="Node at 127.0.0.1:29621 [Leader]" writer.go:29: 2021-01-29T19:33:50.042Z [INFO] TestConnectCAConfig/setting_state_fails.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:50.043Z [INFO] TestConnectCAConfig/setting_state_fails.server: New leader elected: payload=Node-c19b42df-2da5-e2a0-9ba1-c70c5374a68d === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.048Z [INFO] TestCatalogNodeServices.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.046Z [DEBUG] TestConnectCAConfig/setting_state_fails.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29621 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.049Z [INFO] TestCatalogNodeServices.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.050Z [DEBUG] TestCatalogNodeServices.server: Skipping self join check for node since the cluster is too small: node=Node-cd4aed90-1cd2-6bba-d9ed-173ad718a4c1 writer.go:29: 2021-01-29T19:33:50.051Z [INFO] TestCatalogNodeServices.server: member joined, marking health alive: member=Node-cd4aed90-1cd2-6bba-d9ed-173ad718a4c1 === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.054Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.074Z [WARN] TestCatalogServiceNodes_ConnectProxy.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:50.076Z [INFO] TestCatalogServiceNodes_ConnectProxy.server.raft: entering candidate state: node="Node at 127.0.0.1:29627 [Candidate]" term=2 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.079Z [INFO] TestCatalogNodeServices: Requesting shutdown writer.go:29: 2021-01-29T19:33:50.081Z [INFO] TestCatalogNodeServices.server: shutting down server writer.go:29: 2021-01-29T19:33:50.084Z [DEBUG] TestCatalogNodeServices.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.086Z [WARN] TestCatalogNodeServices.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:50.088Z [DEBUG] TestCatalogNodeServices.leader: stopped routine: routine="CA root pruning" === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.098Z [DEBUG] TestCatalogServiceNodes_ConnectProxy.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:50.100Z [DEBUG] TestCatalogServiceNodes_ConnectProxy.server.raft: vote granted: from=0869be48-978f-3e03-f93a-267339cece96 term=2 tally=1 === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.101Z [INFO] TestConnectCAConfig/setting_state_fails.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.102Z [INFO] TestCatalogServiceNodes_ConnectProxy.server.raft: election won: tally=1 === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.103Z [INFO] TestConnectCAConfig/setting_state_fails.leader: started routine: routine="CA root pruning" === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.103Z [INFO] TestCatalogServiceNodes_ConnectProxy.server.raft: entering leader state: leader="Node at 127.0.0.1:29627 [Leader]" writer.go:29: 2021-01-29T19:33:50.105Z [INFO] TestCatalogServiceNodes_ConnectProxy.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:50.108Z [INFO] TestCatalogServiceNodes_ConnectProxy.server: New leader elected: payload=Node-0869be48-978f-3e03-f93a-267339cece96 === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.105Z [DEBUG] TestConnectCAConfig/setting_state_fails.server: Skipping self join check for node since the cluster is too small: node=Node-c19b42df-2da5-e2a0-9ba1-c70c5374a68d === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.099Z [WARN] TestCatalogNodeServices.server.serf.wan: serf: Shutdown without a Leave === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.111Z [INFO] TestConnectCAConfig/setting_state_fails.server: member joined, marking health alive: member=Node-c19b42df-2da5-e2a0-9ba1-c70c5374a68d === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.115Z [INFO] TestCatalogNodeServices.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:50.115Z [INFO] TestCatalogNodeServices: consul server down === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.118Z [INFO] TestCatalogServiceNodes_ConnectProxy: Synced node info === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.119Z [INFO] TestCatalogNodeServices: shutdown complete === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.119Z [DEBUG] TestCatalogServiceNodes_ConnectProxy.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29627 === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.122Z [INFO] TestCatalogNodeServices: Stopping server: protocol=DNS address=127.0.0.1:29610 network=tcp writer.go:29: 2021-01-29T19:33:50.125Z [INFO] TestCatalogNodeServices: Stopping server: protocol=DNS address=127.0.0.1:29610 network=udp writer.go:29: 2021-01-29T19:33:50.127Z [INFO] TestCatalogNodeServices: Stopping server: protocol=HTTP address=127.0.0.1:29611 network=tcp === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.135Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:50.149Z [INFO] TestCatalogServiceNodes_ConnectProxy.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:50.150Z [INFO] TestCatalogServiceNodes_ConnectProxy.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.152Z [DEBUG] TestCatalogServiceNodes_ConnectProxy.server: Skipping self join check for node since the cluster is too small: node=Node-0869be48-978f-3e03-f93a-267339cece96 writer.go:29: 2021-01-29T19:33:50.154Z [INFO] TestCatalogServiceNodes_ConnectProxy.server: member joined, marking health alive: member=Node-0869be48-978f-3e03-f93a-267339cece96 === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.257Z [DEBUG] TestConnectCAConfig/setting_state_fails: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:50.260Z [INFO] TestConnectCAConfig/setting_state_fails: Synced node info writer.go:29: 2021-01-29T19:33:50.262Z [DEBUG] TestConnectCAConfig/setting_state_fails: Node info in sync writer.go:29: 2021-01-29T19:33:50.346Z [INFO] TestConnectCAConfig/setting_state_fails: Requesting shutdown writer.go:29: 2021-01-29T19:33:50.346Z [INFO] TestConnectCAConfig/setting_state_fails.server: shutting down server writer.go:29: 2021-01-29T19:33:50.347Z [DEBUG] TestConnectCAConfig/setting_state_fails.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.348Z [WARN] TestConnectCAConfig/setting_state_fails.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:50.348Z [DEBUG] TestConnectCAConfig/setting_state_fails.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.351Z [WARN] TestConnectCAConfig/setting_state_fails.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:50.354Z [INFO] TestConnectCAConfig/setting_state_fails.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:50.356Z [INFO] TestConnectCAConfig/setting_state_fails: consul server down writer.go:29: 2021-01-29T19:33:50.358Z [INFO] TestConnectCAConfig/setting_state_fails: shutdown complete writer.go:29: 2021-01-29T19:33:50.360Z [INFO] TestConnectCAConfig/setting_state_fails: Stopping server: protocol=DNS address=127.0.0.1:29616 network=tcp writer.go:29: 2021-01-29T19:33:50.361Z [INFO] TestConnectCAConfig/setting_state_fails: Stopping server: protocol=DNS address=127.0.0.1:29616 network=udp === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:50.374Z [INFO] TestCatalogConnectServiceNodes_good: Requesting shutdown writer.go:29: 2021-01-29T19:33:50.375Z [INFO] TestCatalogConnectServiceNodes_good.server: shutting down server === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.376Z [INFO] TestConnectCAConfig/setting_state_fails: Stopping server: protocol=HTTP address=127.0.0.1:29617 network=tcp === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:50.377Z [DEBUG] TestCatalogConnectServiceNodes_good.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.378Z [WARN] TestCatalogConnectServiceNodes_good.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:50.379Z [DEBUG] TestCatalogConnectServiceNodes_good.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.386Z [WARN] TestCatalogConnectServiceNodes_good.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:50.388Z [INFO] TestCatalogConnectServiceNodes_good.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:50.389Z [INFO] TestCatalogConnectServiceNodes_good: consul server down writer.go:29: 2021-01-29T19:33:50.391Z [INFO] TestCatalogConnectServiceNodes_good: shutdown complete writer.go:29: 2021-01-29T19:33:50.393Z [INFO] TestCatalogConnectServiceNodes_good: Stopping server: protocol=DNS address=127.0.0.1:29604 network=tcp writer.go:29: 2021-01-29T19:33:50.395Z [INFO] TestCatalogConnectServiceNodes_good: Stopping server: protocol=DNS address=127.0.0.1:29604 network=udp writer.go:29: 2021-01-29T19:33:50.396Z [INFO] TestCatalogConnectServiceNodes_good: Stopping server: protocol=HTTP address=127.0.0.1:29605 network=tcp === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.400Z [INFO] TestCatalogServiceNodes_ConnectProxy: Requesting shutdown writer.go:29: 2021-01-29T19:33:50.401Z [INFO] TestCatalogServiceNodes_ConnectProxy.server: shutting down server writer.go:29: 2021-01-29T19:33:50.403Z [DEBUG] TestCatalogServiceNodes_ConnectProxy.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.404Z [WARN] TestCatalogServiceNodes_ConnectProxy.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:50.404Z [DEBUG] TestCatalogServiceNodes_ConnectProxy.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.408Z [WARN] TestCatalogServiceNodes_ConnectProxy.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:50.410Z [INFO] TestCatalogServiceNodes_ConnectProxy.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:50.410Z [INFO] TestCatalogServiceNodes_ConnectProxy: consul server down writer.go:29: 2021-01-29T19:33:50.413Z [INFO] TestCatalogServiceNodes_ConnectProxy: shutdown complete writer.go:29: 2021-01-29T19:33:50.414Z [INFO] TestCatalogServiceNodes_ConnectProxy: Stopping server: protocol=DNS address=127.0.0.1:29622 network=tcp writer.go:29: 2021-01-29T19:33:50.415Z [INFO] TestCatalogServiceNodes_ConnectProxy: Stopping server: protocol=DNS address=127.0.0.1:29622 network=udp writer.go:29: 2021-01-29T19:33:50.416Z [INFO] TestCatalogServiceNodes_ConnectProxy: Stopping server: protocol=HTTP address=127.0.0.1:29623 network=tcp === CONT TestCatalogNodeServices writer.go:29: 2021-01-29T19:33:50.629Z [INFO] TestCatalogNodeServices: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:50.631Z [INFO] TestCatalogNodeServices: Endpoints down --- PASS: TestCatalogNodeServices (0.87s) === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:50.657Z [WARN] TestCatalogServiceNodes_DistanceSort: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:50.659Z [DEBUG] TestCatalogServiceNodes_DistanceSort.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:50.661Z [DEBUG] TestCatalogServiceNodes_DistanceSort.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:50.677Z [INFO] TestCatalogServiceNodes_DistanceSort.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c363166f-af69-69be-1dcf-ad00e13abb56 Address:127.0.0.1:29633}]" writer.go:29: 2021-01-29T19:33:50.680Z [INFO] TestCatalogServiceNodes_DistanceSort.server.raft: entering follower state: follower="Node at 127.0.0.1:29633 [Follower]" leader= writer.go:29: 2021-01-29T19:33:50.680Z [INFO] TestCatalogServiceNodes_DistanceSort.server.serf.wan: serf: EventMemberJoin: Node-c363166f-af69-69be-1dcf-ad00e13abb56.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:50.689Z [INFO] TestCatalogServiceNodes_DistanceSort.server.serf.lan: serf: EventMemberJoin: Node-c363166f-af69-69be-1dcf-ad00e13abb56 127.0.0.1 writer.go:29: 2021-01-29T19:33:50.697Z [INFO] TestCatalogServiceNodes_DistanceSort.server: Adding LAN server: server="Node-c363166f-af69-69be-1dcf-ad00e13abb56 (Addr: tcp/127.0.0.1:29633) (DC: dc1)" writer.go:29: 2021-01-29T19:33:50.699Z [INFO] TestCatalogServiceNodes_DistanceSort.server: Handled event for server in area: event=member-join server=Node-c363166f-af69-69be-1dcf-ad00e13abb56.dc1 area=wan writer.go:29: 2021-01-29T19:33:50.703Z [INFO] TestCatalogServiceNodes_DistanceSort: Started DNS server: address=127.0.0.1:29628 network=tcp writer.go:29: 2021-01-29T19:33:50.706Z [INFO] TestCatalogServiceNodes_DistanceSort: Started DNS server: address=127.0.0.1:29628 network=udp writer.go:29: 2021-01-29T19:33:50.708Z [INFO] TestCatalogServiceNodes_DistanceSort: Started HTTP server: address=127.0.0.1:29629 network=tcp writer.go:29: 2021-01-29T19:33:50.711Z [INFO] TestCatalogServiceNodes_DistanceSort: started state syncer writer.go:29: 2021-01-29T19:33:50.743Z [WARN] TestCatalogServiceNodes_DistanceSort.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:50.746Z [INFO] TestCatalogServiceNodes_DistanceSort.server.raft: entering candidate state: node="Node at 127.0.0.1:29633 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:50.750Z [DEBUG] TestCatalogServiceNodes_DistanceSort.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:50.751Z [DEBUG] TestCatalogServiceNodes_DistanceSort.server.raft: vote granted: from=c363166f-af69-69be-1dcf-ad00e13abb56 term=2 tally=1 writer.go:29: 2021-01-29T19:33:50.753Z [INFO] TestCatalogServiceNodes_DistanceSort.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:50.754Z [INFO] TestCatalogServiceNodes_DistanceSort.server.raft: entering leader state: leader="Node at 127.0.0.1:29633 [Leader]" writer.go:29: 2021-01-29T19:33:50.756Z [INFO] TestCatalogServiceNodes_DistanceSort.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:50.757Z [INFO] TestCatalogServiceNodes_DistanceSort.server: New leader elected: payload=Node-c363166f-af69-69be-1dcf-ad00e13abb56 writer.go:29: 2021-01-29T19:33:50.760Z [DEBUG] TestCatalogServiceNodes_DistanceSort.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29633 writer.go:29: 2021-01-29T19:33:50.766Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:50.773Z [INFO] TestCatalogServiceNodes_DistanceSort.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:50.775Z [INFO] TestCatalogServiceNodes_DistanceSort.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:50.776Z [DEBUG] TestCatalogServiceNodes_DistanceSort.server: Skipping self join check for node since the cluster is too small: node=Node-c363166f-af69-69be-1dcf-ad00e13abb56 writer.go:29: 2021-01-29T19:33:50.778Z [INFO] TestCatalogServiceNodes_DistanceSort.server: member joined, marking health alive: member=Node-c363166f-af69-69be-1dcf-ad00e13abb56 === CONT TestConnectCAConfig/setting_state_fails writer.go:29: 2021-01-29T19:33:50.878Z [INFO] TestConnectCAConfig/setting_state_fails: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:50.880Z [INFO] TestConnectCAConfig/setting_state_fails: Endpoints down === RUN TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:50.895Z [WARN] TestConnectCAConfig/updating_config_with_same_state: bootstrap = true: do not enable unless necessary === CONT TestCatalogConnectServiceNodes_good writer.go:29: 2021-01-29T19:33:50.898Z [INFO] TestCatalogConnectServiceNodes_good: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:50.899Z [INFO] TestCatalogConnectServiceNodes_good: Endpoints down --- PASS: TestCatalogConnectServiceNodes_good (1.08s) === CONT TestCatalogServiceNodes_Filter === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:50.897Z [DEBUG] TestConnectCAConfig/updating_config_with_same_state.tlsutil: Update: version=1 === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:50.914Z [WARN] TestCatalogServiceNodes_Filter: bootstrap = true: do not enable unless necessary === CONT TestCatalogServiceNodes_ConnectProxy writer.go:29: 2021-01-29T19:33:50.917Z [INFO] TestCatalogServiceNodes_ConnectProxy: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:50.918Z [INFO] TestCatalogServiceNodes_ConnectProxy: Endpoints down --- PASS: TestCatalogServiceNodes_ConnectProxy (0.96s) === CONT TestCatalogServiceNodes_NodeMetaFilter === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:50.915Z [DEBUG] TestConnectCAConfig/updating_config_with_same_state.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:50.917Z [DEBUG] TestCatalogServiceNodes_Filter.tlsutil: Update: version=1 === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:50.936Z [WARN] TestCatalogServiceNodes_NodeMetaFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:50.938Z [DEBUG] TestCatalogServiceNodes_NodeMetaFilter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:50.940Z [DEBUG] TestCatalogServiceNodes_NodeMetaFilter.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:50.949Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:80c01b0b-0fc4-f19b-7b45-54544c117a05 Address:127.0.0.1:29651}]" === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:50.953Z [DEBUG] TestCatalogServiceNodes_Filter.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:50.955Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:29651 [Follower]" leader= writer.go:29: 2021-01-29T19:33:50.956Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server.serf.wan: serf: EventMemberJoin: Node-80c01b0b-0fc4-f19b-7b45-54544c117a05.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:50.970Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server.serf.lan: serf: EventMemberJoin: Node-80c01b0b-0fc4-f19b-7b45-54544c117a05 127.0.0.1 === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:50.996Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:62d63f73-ff97-c2fa-add2-faa9376dc04c Address:127.0.0.1:29639}]" writer.go:29: 2021-01-29T19:33:51.000Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server.serf.wan: serf: EventMemberJoin: Node-62d63f73-ff97-c2fa-add2-faa9376dc04c.dc1 127.0.0.1 === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.008Z [WARN] TestCatalogServiceNodes_NodeMetaFilter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:51.009Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:29651 [Candidate]" term=2 === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.013Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server.serf.lan: serf: EventMemberJoin: Node-62d63f73-ff97-c2fa-add2-faa9376dc04c 127.0.0.1 writer.go:29: 2021-01-29T19:33:51.008Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server.raft: entering follower state: follower="Node at 127.0.0.1:29639 [Follower]" leader= === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.026Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: Started DNS server: address=127.0.0.1:29646 network=udp writer.go:29: 2021-01-29T19:33:51.028Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: Started DNS server: address=127.0.0.1:29646 network=tcp === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.029Z [INFO] TestConnectCAConfig/updating_config_with_same_state: Started DNS server: address=127.0.0.1:29634 network=udp === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.037Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: Started HTTP server: address=127.0.0.1:29647 network=tcp === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.039Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server: Adding LAN server: server="Node-62d63f73-ff97-c2fa-add2-faa9376dc04c (Addr: tcp/127.0.0.1:29639) (DC: dc1)" === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.039Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: started state syncer === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.044Z [DEBUG] TestCatalogServiceNodes_DistanceSort: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.046Z [WARN] TestCatalogServiceNodes_NodeMetaFilter.server.raft: unable to get address for sever, using fallback address: id=80c01b0b-0fc4-f19b-7b45-54544c117a05 fallback=127.0.0.1:29651 error="Could not find address for server id 80c01b0b-0fc4-f19b-7b45-54544c117a05" === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.048Z [INFO] TestCatalogServiceNodes_DistanceSort: Synced node info === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.051Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server: Handled event for server in area: event=member-join server=Node-62d63f73-ff97-c2fa-add2-faa9376dc04c.dc1 area=wan writer.go:29: 2021-01-29T19:33:51.069Z [INFO] TestConnectCAConfig/updating_config_with_same_state: Started DNS server: address=127.0.0.1:29634 network=tcp === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.076Z [DEBUG] TestCatalogServiceNodes_NodeMetaFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:51.078Z [DEBUG] TestCatalogServiceNodes_NodeMetaFilter.server.raft: vote granted: from=80c01b0b-0fc4-f19b-7b45-54544c117a05 term=2 tally=1 writer.go:29: 2021-01-29T19:33:51.079Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server.raft: election won: tally=1 === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.076Z [INFO] TestCatalogServiceNodes_Filter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4fcf2587-db7b-9edb-8874-33b473858f6f Address:127.0.0.1:29645}]" === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.081Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:29651 [Leader]" writer.go:29: 2021-01-29T19:33:51.093Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server: cluster leadership acquired === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.093Z [WARN] TestConnectCAConfig/updating_config_with_same_state.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.083Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server: Adding LAN server: server="Node-80c01b0b-0fc4-f19b-7b45-54544c117a05 (Addr: tcp/127.0.0.1:29651) (DC: dc1)" === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.097Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server.raft: entering candidate state: node="Node at 127.0.0.1:29639 [Candidate]" term=2 === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.097Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server: New leader elected: payload=Node-80c01b0b-0fc4-f19b-7b45-54544c117a05 === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.095Z [INFO] TestConnectCAConfig/updating_config_with_same_state: Started HTTP server: address=127.0.0.1:29635 network=tcp === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.083Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server: Handled event for server in area: event=member-join server=Node-80c01b0b-0fc4-f19b-7b45-54544c117a05.dc1 area=wan === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.100Z [INFO] TestConnectCAConfig/updating_config_with_same_state: started state syncer === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.093Z [INFO] TestCatalogServiceNodes_Filter.server.serf.wan: serf: EventMemberJoin: Node-4fcf2587-db7b-9edb-8874-33b473858f6f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:51.112Z [INFO] TestCatalogServiceNodes_Filter.server.raft: entering follower state: follower="Node at 127.0.0.1:29645 [Follower]" leader= writer.go:29: 2021-01-29T19:33:51.115Z [INFO] TestCatalogServiceNodes_Filter.server.serf.lan: serf: EventMemberJoin: Node-4fcf2587-db7b-9edb-8874-33b473858f6f 127.0.0.1 === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.131Z [DEBUG] TestCatalogServiceNodes_NodeMetaFilter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29651 === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.133Z [INFO] TestCatalogServiceNodes_Filter: Started DNS server: address=127.0.0.1:29640 network=udp === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.138Z [DEBUG] TestConnectCAConfig/updating_config_with_same_state.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:51.140Z [DEBUG] TestConnectCAConfig/updating_config_with_same_state.server.raft: vote granted: from=62d63f73-ff97-c2fa-add2-faa9376dc04c term=2 tally=1 === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.134Z [INFO] TestCatalogServiceNodes_Filter.server: Handled event for server in area: event=member-join server=Node-4fcf2587-db7b-9edb-8874-33b473858f6f.dc1 area=wan === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.142Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:51.144Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server.raft: entering leader state: leader="Node at 127.0.0.1:29639 [Leader]" writer.go:29: 2021-01-29T19:33:51.146Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server: cluster leadership acquired === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.137Z [INFO] TestCatalogServiceNodes_Filter.server: Adding LAN server: server="Node-4fcf2587-db7b-9edb-8874-33b473858f6f (Addr: tcp/127.0.0.1:29645) (DC: dc1)" === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.149Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server: New leader elected: payload=Node-62d63f73-ff97-c2fa-add2-faa9376dc04c === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.140Z [INFO] TestCatalogServiceNodes_Filter: Started DNS server: address=127.0.0.1:29640 network=tcp writer.go:29: 2021-01-29T19:33:51.167Z [INFO] TestCatalogServiceNodes_Filter: Started HTTP server: address=127.0.0.1:29641 network=tcp === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.166Z [INFO] TestCatalogServiceNodes_DistanceSort: Requesting shutdown === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.168Z [INFO] TestCatalogServiceNodes_Filter: started state syncer === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.170Z [INFO] TestCatalogServiceNodes_DistanceSort.server: shutting down server writer.go:29: 2021-01-29T19:33:51.174Z [DEBUG] TestCatalogServiceNodes_DistanceSort.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:51.177Z [WARN] TestCatalogServiceNodes_DistanceSort.server.serf.lan: serf: Shutdown without a Leave === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.185Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: Synced node info writer.go:29: 2021-01-29T19:33:51.186Z [DEBUG] TestCatalogServiceNodes_NodeMetaFilter: Node info in sync writer.go:29: 2021-01-29T19:33:51.195Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.200Z [DEBUG] TestConnectCAConfig/updating_config_with_same_state.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29639 === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.204Z [WARN] TestCatalogServiceNodes_Filter.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.202Z [DEBUG] TestCatalogServiceNodes_DistanceSort.leader: stopped routine: routine="CA root pruning" === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.205Z [INFO] TestCatalogServiceNodes_Filter.server.raft: entering candidate state: node="Node at 127.0.0.1:29645 [Candidate]" term=2 === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.207Z [WARN] TestCatalogServiceNodes_DistanceSort.server.serf.wan: serf: Shutdown without a Leave === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.228Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.230Z [INFO] TestCatalogServiceNodes_DistanceSort.server.router.manager: shutting down === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.232Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.233Z [INFO] TestCatalogServiceNodes_DistanceSort: consul server down === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.244Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.leader: started routine: routine="CA root pruning" === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.246Z [INFO] TestCatalogServiceNodes_DistanceSort: shutdown complete writer.go:29: 2021-01-29T19:33:51.253Z [INFO] TestCatalogServiceNodes_DistanceSort: Stopping server: protocol=DNS address=127.0.0.1:29628 network=tcp writer.go:29: 2021-01-29T19:33:51.254Z [INFO] TestCatalogServiceNodes_DistanceSort: Stopping server: protocol=DNS address=127.0.0.1:29628 network=udp === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.250Z [DEBUG] TestCatalogServiceNodes_Filter.server.raft: votes: needed=1 === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.255Z [INFO] TestCatalogServiceNodes_DistanceSort: Stopping server: protocol=HTTP address=127.0.0.1:29629 network=tcp === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.251Z [DEBUG] TestCatalogServiceNodes_NodeMetaFilter.server: Skipping self join check for node since the cluster is too small: node=Node-80c01b0b-0fc4-f19b-7b45-54544c117a05 === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.261Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.256Z [DEBUG] TestCatalogServiceNodes_Filter.server.raft: vote granted: from=4fcf2587-db7b-9edb-8874-33b473858f6f term=2 tally=1 === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.259Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server: member joined, marking health alive: member=Node-80c01b0b-0fc4-f19b-7b45-54544c117a05 === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.262Z [INFO] TestConnectCAConfig/updating_config_with_same_state.leader: started routine: routine="CA root pruning" === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.263Z [INFO] TestCatalogServiceNodes_Filter.server.raft: election won: tally=1 === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.272Z [DEBUG] TestConnectCAConfig/updating_config_with_same_state.server: Skipping self join check for node since the cluster is too small: node=Node-62d63f73-ff97-c2fa-add2-faa9376dc04c === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.277Z [DEBUG] TestCatalogServiceNodes_NodeMetaFilter.server: Skipping self join check for node since the cluster is too small: node=Node-80c01b0b-0fc4-f19b-7b45-54544c117a05 === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.275Z [INFO] TestCatalogServiceNodes_Filter.server.raft: entering leader state: leader="Node at 127.0.0.1:29645 [Leader]" writer.go:29: 2021-01-29T19:33:51.280Z [INFO] TestCatalogServiceNodes_Filter.server: cluster leadership acquired === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.277Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server: member joined, marking health alive: member=Node-62d63f73-ff97-c2fa-add2-faa9376dc04c === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.288Z [INFO] TestCatalogServiceNodes_Filter.server: New leader elected: payload=Node-4fcf2587-db7b-9edb-8874-33b473858f6f === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.280Z [INFO] TestConnectCAConfig/updating_config_with_same_state: Synced node info === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.290Z [DEBUG] TestCatalogServiceNodes_Filter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29645 === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.291Z [DEBUG] TestConnectCAConfig/updating_config_with_same_state: Node info in sync === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.328Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:51.356Z [INFO] TestCatalogServiceNodes_Filter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:51.357Z [INFO] TestCatalogServiceNodes_Filter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:51.358Z [DEBUG] TestCatalogServiceNodes_Filter.server: Skipping self join check for node since the cluster is too small: node=Node-4fcf2587-db7b-9edb-8874-33b473858f6f writer.go:29: 2021-01-29T19:33:51.359Z [INFO] TestCatalogServiceNodes_Filter.server: member joined, marking health alive: member=Node-4fcf2587-db7b-9edb-8874-33b473858f6f === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.393Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: Requesting shutdown writer.go:29: 2021-01-29T19:33:51.412Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server: shutting down server writer.go:29: 2021-01-29T19:33:51.413Z [DEBUG] TestCatalogServiceNodes_NodeMetaFilter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:51.415Z [WARN] TestCatalogServiceNodes_NodeMetaFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:51.417Z [DEBUG] TestCatalogServiceNodes_NodeMetaFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:51.434Z [WARN] TestCatalogServiceNodes_NodeMetaFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:51.453Z [INFO] TestCatalogServiceNodes_NodeMetaFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:51.454Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: consul server down writer.go:29: 2021-01-29T19:33:51.455Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: shutdown complete writer.go:29: 2021-01-29T19:33:51.456Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: Stopping server: protocol=DNS address=127.0.0.1:29646 network=tcp writer.go:29: 2021-01-29T19:33:51.457Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: Stopping server: protocol=DNS address=127.0.0.1:29646 network=udp writer.go:29: 2021-01-29T19:33:51.458Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: Stopping server: protocol=HTTP address=127.0.0.1:29647 network=tcp === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:51.467Z [INFO] TestCatalogServiceNodes_Filter: Requesting shutdown writer.go:29: 2021-01-29T19:33:51.472Z [INFO] TestCatalogServiceNodes_Filter.server: shutting down server writer.go:29: 2021-01-29T19:33:51.477Z [DEBUG] TestCatalogServiceNodes_Filter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:51.480Z [WARN] TestCatalogServiceNodes_Filter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:51.479Z [ERROR] TestCatalogServiceNodes_Filter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:51.489Z [DEBUG] TestCatalogServiceNodes_Filter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:51.493Z [WARN] TestCatalogServiceNodes_Filter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:51.495Z [INFO] TestCatalogServiceNodes_Filter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:51.496Z [INFO] TestCatalogServiceNodes_Filter: consul server down writer.go:29: 2021-01-29T19:33:51.498Z [INFO] TestCatalogServiceNodes_Filter: shutdown complete writer.go:29: 2021-01-29T19:33:51.499Z [INFO] TestCatalogServiceNodes_Filter: Stopping server: protocol=DNS address=127.0.0.1:29640 network=tcp writer.go:29: 2021-01-29T19:33:51.500Z [INFO] TestCatalogServiceNodes_Filter: Stopping server: protocol=DNS address=127.0.0.1:29640 network=udp writer.go:29: 2021-01-29T19:33:51.502Z [INFO] TestCatalogServiceNodes_Filter: Stopping server: protocol=HTTP address=127.0.0.1:29641 network=tcp === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:51.539Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server.connect: CA provider config updated writer.go:29: 2021-01-29T19:33:51.544Z [INFO] TestConnectCAConfig/updating_config_with_same_state: Requesting shutdown writer.go:29: 2021-01-29T19:33:51.546Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server: shutting down server writer.go:29: 2021-01-29T19:33:51.547Z [DEBUG] TestConnectCAConfig/updating_config_with_same_state.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:51.548Z [WARN] TestConnectCAConfig/updating_config_with_same_state.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:51.548Z [DEBUG] TestConnectCAConfig/updating_config_with_same_state.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:51.551Z [WARN] TestConnectCAConfig/updating_config_with_same_state.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:51.553Z [INFO] TestConnectCAConfig/updating_config_with_same_state.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:51.553Z [INFO] TestConnectCAConfig/updating_config_with_same_state: consul server down writer.go:29: 2021-01-29T19:33:51.557Z [INFO] TestConnectCAConfig/updating_config_with_same_state: shutdown complete writer.go:29: 2021-01-29T19:33:51.558Z [INFO] TestConnectCAConfig/updating_config_with_same_state: Stopping server: protocol=DNS address=127.0.0.1:29634 network=tcp writer.go:29: 2021-01-29T19:33:51.560Z [INFO] TestConnectCAConfig/updating_config_with_same_state: Stopping server: protocol=DNS address=127.0.0.1:29634 network=udp writer.go:29: 2021-01-29T19:33:51.561Z [INFO] TestConnectCAConfig/updating_config_with_same_state: Stopping server: protocol=HTTP address=127.0.0.1:29635 network=tcp === CONT TestCatalogServiceNodes_DistanceSort writer.go:29: 2021-01-29T19:33:51.757Z [INFO] TestCatalogServiceNodes_DistanceSort: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:51.758Z [INFO] TestCatalogServiceNodes_DistanceSort: Endpoints down --- PASS: TestCatalogServiceNodes_DistanceSort (1.13s) === CONT TestCatalogServiceNodes writer.go:29: 2021-01-29T19:33:51.783Z [WARN] TestCatalogServiceNodes: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:51.784Z [DEBUG] TestCatalogServiceNodes.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:51.785Z [DEBUG] TestCatalogServiceNodes.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:51.789Z [INFO] TestCatalogServiceNodes.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2361578a-a4fd-34ed-bb80-212962471d33 Address:127.0.0.1:29657}]" writer.go:29: 2021-01-29T19:33:51.791Z [INFO] TestCatalogServiceNodes.server.raft: entering follower state: follower="Node at 127.0.0.1:29657 [Follower]" leader= writer.go:29: 2021-01-29T19:33:51.792Z [INFO] TestCatalogServiceNodes.server.serf.wan: serf: EventMemberJoin: Node-2361578a-a4fd-34ed-bb80-212962471d33.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:51.796Z [INFO] TestCatalogServiceNodes.server.serf.lan: serf: EventMemberJoin: Node-2361578a-a4fd-34ed-bb80-212962471d33 127.0.0.1 writer.go:29: 2021-01-29T19:33:51.798Z [INFO] TestCatalogServiceNodes.server: Handled event for server in area: event=member-join server=Node-2361578a-a4fd-34ed-bb80-212962471d33.dc1 area=wan writer.go:29: 2021-01-29T19:33:51.798Z [INFO] TestCatalogServiceNodes.server: Adding LAN server: server="Node-2361578a-a4fd-34ed-bb80-212962471d33 (Addr: tcp/127.0.0.1:29657) (DC: dc1)" writer.go:29: 2021-01-29T19:33:51.799Z [INFO] TestCatalogServiceNodes: Started DNS server: address=127.0.0.1:29652 network=udp writer.go:29: 2021-01-29T19:33:51.802Z [INFO] TestCatalogServiceNodes: Started DNS server: address=127.0.0.1:29652 network=tcp writer.go:29: 2021-01-29T19:33:51.804Z [INFO] TestCatalogServiceNodes: Started HTTP server: address=127.0.0.1:29653 network=tcp writer.go:29: 2021-01-29T19:33:51.805Z [INFO] TestCatalogServiceNodes: started state syncer writer.go:29: 2021-01-29T19:33:51.843Z [WARN] TestCatalogServiceNodes.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:51.844Z [INFO] TestCatalogServiceNodes.server.raft: entering candidate state: node="Node at 127.0.0.1:29657 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:51.847Z [DEBUG] TestCatalogServiceNodes.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:51.848Z [DEBUG] TestCatalogServiceNodes.server.raft: vote granted: from=2361578a-a4fd-34ed-bb80-212962471d33 term=2 tally=1 writer.go:29: 2021-01-29T19:33:51.849Z [INFO] TestCatalogServiceNodes.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:51.850Z [INFO] TestCatalogServiceNodes.server.raft: entering leader state: leader="Node at 127.0.0.1:29657 [Leader]" writer.go:29: 2021-01-29T19:33:51.850Z [INFO] TestCatalogServiceNodes.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:51.851Z [INFO] TestCatalogServiceNodes.server: New leader elected: payload=Node-2361578a-a4fd-34ed-bb80-212962471d33 writer.go:29: 2021-01-29T19:33:51.853Z [DEBUG] TestCatalogServiceNodes.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29657 writer.go:29: 2021-01-29T19:33:51.869Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:51.882Z [INFO] TestCatalogServiceNodes.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:51.883Z [INFO] TestCatalogServiceNodes.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:51.884Z [DEBUG] TestCatalogServiceNodes.server: Skipping self join check for node since the cluster is too small: node=Node-2361578a-a4fd-34ed-bb80-212962471d33 writer.go:29: 2021-01-29T19:33:51.885Z [INFO] TestCatalogServiceNodes.server: member joined, marking health alive: member=Node-2361578a-a4fd-34ed-bb80-212962471d33 === CONT TestCatalogServiceNodes_NodeMetaFilter writer.go:29: 2021-01-29T19:33:51.959Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:51.960Z [INFO] TestCatalogServiceNodes_NodeMetaFilter: Endpoints down --- PASS: TestCatalogServiceNodes_NodeMetaFilter (1.04s) === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:51.967Z [WARN] TestCatalogRegister_checkRegistration: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:51.968Z [DEBUG] TestCatalogRegister_checkRegistration.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:51.969Z [DEBUG] TestCatalogRegister_checkRegistration.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:51.984Z [INFO] TestCatalogRegister_checkRegistration.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cfca29bb-b468-71de-5829-76cfafa072b7 Address:127.0.0.1:29663}]" writer.go:29: 2021-01-29T19:33:51.986Z [INFO] TestCatalogRegister_checkRegistration.server.serf.wan: serf: EventMemberJoin: Node-cfca29bb-b468-71de-5829-76cfafa072b7.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:51.986Z [INFO] TestCatalogRegister_checkRegistration.server.raft: entering follower state: follower="Node at 127.0.0.1:29663 [Follower]" leader= writer.go:29: 2021-01-29T19:33:51.988Z [INFO] TestCatalogRegister_checkRegistration.server.serf.lan: serf: EventMemberJoin: Node-cfca29bb-b468-71de-5829-76cfafa072b7 127.0.0.1 writer.go:29: 2021-01-29T19:33:51.994Z [INFO] TestCatalogRegister_checkRegistration.server: Adding LAN server: server="Node-cfca29bb-b468-71de-5829-76cfafa072b7 (Addr: tcp/127.0.0.1:29663) (DC: dc1)" writer.go:29: 2021-01-29T19:33:51.994Z [INFO] TestCatalogRegister_checkRegistration.server: Handled event for server in area: event=member-join server=Node-cfca29bb-b468-71de-5829-76cfafa072b7.dc1 area=wan writer.go:29: 2021-01-29T19:33:51.996Z [INFO] TestCatalogRegister_checkRegistration: Started DNS server: address=127.0.0.1:29658 network=udp === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:52.004Z [INFO] TestCatalogServiceNodes_Filter: Waiting for endpoints to shut down === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.000Z [INFO] TestCatalogRegister_checkRegistration: Started DNS server: address=127.0.0.1:29658 network=tcp === CONT TestCatalogServiceNodes_Filter writer.go:29: 2021-01-29T19:33:52.005Z [INFO] TestCatalogServiceNodes_Filter: Endpoints down --- PASS: TestCatalogServiceNodes_Filter (1.11s) === CONT TestCatalogServices_NodeMetaFilter === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.011Z [INFO] TestCatalogRegister_checkRegistration: Started HTTP server: address=127.0.0.1:29659 network=tcp writer.go:29: 2021-01-29T19:33:52.013Z [INFO] TestCatalogRegister_checkRegistration: started state syncer === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.038Z [WARN] TestCatalogServices_NodeMetaFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:52.039Z [DEBUG] TestCatalogServices_NodeMetaFilter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:52.041Z [DEBUG] TestCatalogServices_NodeMetaFilter.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.044Z [WARN] TestCatalogRegister_checkRegistration.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.045Z [INFO] TestCatalogServices_NodeMetaFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f748e157-daae-ff36-fc33-ad1ebf53fe09 Address:127.0.0.1:29669}]" === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.045Z [INFO] TestCatalogRegister_checkRegistration.server.raft: entering candidate state: node="Node at 127.0.0.1:29663 [Candidate]" term=2 === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.047Z [INFO] TestCatalogServices_NodeMetaFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:29669 [Follower]" leader= writer.go:29: 2021-01-29T19:33:52.047Z [INFO] TestCatalogServices_NodeMetaFilter.server.serf.wan: serf: EventMemberJoin: Node-f748e157-daae-ff36-fc33-ad1ebf53fe09.dc1 127.0.0.1 === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.050Z [DEBUG] TestCatalogRegister_checkRegistration.server.raft: votes: needed=1 === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.051Z [INFO] TestCatalogServices_NodeMetaFilter.server.serf.lan: serf: EventMemberJoin: Node-f748e157-daae-ff36-fc33-ad1ebf53fe09 127.0.0.1 === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.051Z [DEBUG] TestCatalogRegister_checkRegistration.server.raft: vote granted: from=cfca29bb-b468-71de-5829-76cfafa072b7 term=2 tally=1 writer.go:29: 2021-01-29T19:33:52.052Z [INFO] TestCatalogRegister_checkRegistration.server.raft: election won: tally=1 === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.053Z [INFO] TestCatalogServices_NodeMetaFilter.server: Adding LAN server: server="Node-f748e157-daae-ff36-fc33-ad1ebf53fe09 (Addr: tcp/127.0.0.1:29669) (DC: dc1)" writer.go:29: 2021-01-29T19:33:52.054Z [INFO] TestCatalogServices_NodeMetaFilter.server: Handled event for server in area: event=member-join server=Node-f748e157-daae-ff36-fc33-ad1ebf53fe09.dc1 area=wan === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.054Z [INFO] TestCatalogRegister_checkRegistration.server.raft: entering leader state: leader="Node at 127.0.0.1:29663 [Leader]" writer.go:29: 2021-01-29T19:33:52.055Z [INFO] TestCatalogRegister_checkRegistration.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:52.056Z [INFO] TestCatalogRegister_checkRegistration.server: New leader elected: payload=Node-cfca29bb-b468-71de-5829-76cfafa072b7 writer.go:29: 2021-01-29T19:33:52.056Z [DEBUG] TestCatalogRegister_checkRegistration.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29663 === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:52.062Z [INFO] TestConnectCAConfig/updating_config_with_same_state: Waiting for endpoints to shut down === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.063Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestConnectCAConfig/updating_config_with_same_state writer.go:29: 2021-01-29T19:33:52.063Z [INFO] TestConnectCAConfig/updating_config_with_same_state: Endpoints down --- PASS: TestConnectCAConfig (6.05s) --- PASS: TestConnectCAConfig/basic (0.89s) --- PASS: TestConnectCAConfig/basic_with_IntermediateCertTTL (0.88s) --- PASS: TestConnectCAConfig/force_without_cross_sign_CamelCase (1.02s) --- PASS: TestConnectCAConfig/force_without_cross_sign_snake_case (1.05s) --- PASS: TestConnectCAConfig/setting_state_fails (1.02s) --- PASS: TestConnectCAConfig/updating_config_with_same_state (1.18s) === CONT TestCatalogServices === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.054Z [INFO] TestCatalogServices_NodeMetaFilter: Started DNS server: address=127.0.0.1:29664 network=udp === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.071Z [WARN] TestCatalogServices: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:52.072Z [DEBUG] TestCatalogServices.tlsutil: Update: version=1 === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.067Z [INFO] TestCatalogRegister_checkRegistration.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.075Z [DEBUG] TestCatalogServices.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.071Z [INFO] TestCatalogServices_NodeMetaFilter: Started DNS server: address=127.0.0.1:29664 network=tcp writer.go:29: 2021-01-29T19:33:52.085Z [INFO] TestCatalogServices_NodeMetaFilter: Started HTTP server: address=127.0.0.1:29665 network=tcp === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.077Z [INFO] TestCatalogRegister_checkRegistration.leader: started routine: routine="CA root pruning" === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.086Z [INFO] TestCatalogServices_NodeMetaFilter: started state syncer === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.090Z [INFO] TestCatalogServices.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a78c30de-3af4-5e13-5713-21522946fcbd Address:127.0.0.1:29675}]" writer.go:29: 2021-01-29T19:33:52.092Z [INFO] TestCatalogServices.server.raft: entering follower state: follower="Node at 127.0.0.1:29675 [Follower]" leader= writer.go:29: 2021-01-29T19:33:52.092Z [INFO] TestCatalogServices.server.serf.wan: serf: EventMemberJoin: Node-a78c30de-3af4-5e13-5713-21522946fcbd.dc1 127.0.0.1 === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.088Z [DEBUG] TestCatalogRegister_checkRegistration.server: Skipping self join check for node since the cluster is too small: node=Node-cfca29bb-b468-71de-5829-76cfafa072b7 === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.099Z [WARN] TestCatalogServices_NodeMetaFilter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:52.101Z [INFO] TestCatalogServices_NodeMetaFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:29669 [Candidate]" term=2 === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.097Z [INFO] TestCatalogRegister_checkRegistration.server: member joined, marking health alive: member=Node-cfca29bb-b468-71de-5829-76cfafa072b7 === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.103Z [DEBUG] TestCatalogServices_NodeMetaFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:52.104Z [DEBUG] TestCatalogServices_NodeMetaFilter.server.raft: vote granted: from=f748e157-daae-ff36-fc33-ad1ebf53fe09 term=2 tally=1 writer.go:29: 2021-01-29T19:33:52.105Z [INFO] TestCatalogServices_NodeMetaFilter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:52.106Z [INFO] TestCatalogServices_NodeMetaFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:29669 [Leader]" === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.096Z [INFO] TestCatalogServices.server.serf.lan: serf: EventMemberJoin: Node-a78c30de-3af4-5e13-5713-21522946fcbd 127.0.0.1 writer.go:29: 2021-01-29T19:33:52.110Z [INFO] TestCatalogServices.server: Adding LAN server: server="Node-a78c30de-3af4-5e13-5713-21522946fcbd (Addr: tcp/127.0.0.1:29675) (DC: dc1)" writer.go:29: 2021-01-29T19:33:52.110Z [INFO] TestCatalogServices.server: Handled event for server in area: event=member-join server=Node-a78c30de-3af4-5e13-5713-21522946fcbd.dc1 area=wan === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.106Z [INFO] TestCatalogServices_NodeMetaFilter.server: cluster leadership acquired === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.114Z [INFO] TestCatalogServices: Started DNS server: address=127.0.0.1:29670 network=tcp writer.go:29: 2021-01-29T19:33:52.115Z [INFO] TestCatalogServices: Started DNS server: address=127.0.0.1:29670 network=udp === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.107Z [INFO] TestCatalogServices_NodeMetaFilter.server: New leader elected: payload=Node-f748e157-daae-ff36-fc33-ad1ebf53fe09 === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.117Z [INFO] TestCatalogServices: Started HTTP server: address=127.0.0.1:29671 network=tcp writer.go:29: 2021-01-29T19:33:52.119Z [INFO] TestCatalogServices: started state syncer === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.108Z [DEBUG] TestCatalogServices_NodeMetaFilter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29669 === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.133Z [DEBUG] TestCatalogRegister_checkRegistration: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:52.135Z [INFO] TestCatalogRegister_checkRegistration: Synced node info writer.go:29: 2021-01-29T19:33:52.136Z [DEBUG] TestCatalogRegister_checkRegistration: Node info in sync === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.138Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:52.144Z [INFO] TestCatalogServices_NodeMetaFilter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:52.146Z [INFO] TestCatalogServices_NodeMetaFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.147Z [DEBUG] TestCatalogServices_NodeMetaFilter.server: Skipping self join check for node since the cluster is too small: node=Node-f748e157-daae-ff36-fc33-ad1ebf53fe09 writer.go:29: 2021-01-29T19:33:52.149Z [INFO] TestCatalogServices_NodeMetaFilter.server: member joined, marking health alive: member=Node-f748e157-daae-ff36-fc33-ad1ebf53fe09 === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.160Z [WARN] TestCatalogServices.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:52.162Z [INFO] TestCatalogServices.server.raft: entering candidate state: node="Node at 127.0.0.1:29675 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:52.164Z [DEBUG] TestCatalogServices.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:52.165Z [DEBUG] TestCatalogServices.server.raft: vote granted: from=a78c30de-3af4-5e13-5713-21522946fcbd term=2 tally=1 writer.go:29: 2021-01-29T19:33:52.166Z [INFO] TestCatalogServices.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:52.166Z [INFO] TestCatalogServices.server.raft: entering leader state: leader="Node at 127.0.0.1:29675 [Leader]" === CONT TestCatalogServiceNodes writer.go:29: 2021-01-29T19:33:52.170Z [DEBUG] TestCatalogServiceNodes: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.169Z [INFO] TestCatalogServices.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:52.169Z [INFO] TestCatalogServices.server: New leader elected: payload=Node-a78c30de-3af4-5e13-5713-21522946fcbd === CONT TestCatalogServiceNodes writer.go:29: 2021-01-29T19:33:52.172Z [INFO] TestCatalogServiceNodes: Synced node info writer.go:29: 2021-01-29T19:33:52.173Z [DEBUG] TestCatalogServiceNodes: Node info in sync === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.181Z [DEBUG] TestCatalogServices.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29675 writer.go:29: 2021-01-29T19:33:52.197Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:52.218Z [INFO] TestCatalogServices.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.224Z [INFO] TestCatalogRegister_checkRegistration: Requesting shutdown === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.219Z [INFO] TestCatalogServices.leader: started routine: routine="CA root pruning" === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.226Z [INFO] TestCatalogRegister_checkRegistration.server: shutting down server writer.go:29: 2021-01-29T19:33:52.227Z [DEBUG] TestCatalogRegister_checkRegistration.leader: stopping routine: routine="CA root pruning" === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.227Z [DEBUG] TestCatalogServices.server: Skipping self join check for node since the cluster is too small: node=Node-a78c30de-3af4-5e13-5713-21522946fcbd === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.227Z [WARN] TestCatalogRegister_checkRegistration.server.serf.lan: serf: Shutdown without a Leave === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.228Z [INFO] TestCatalogServices.server: member joined, marking health alive: member=Node-a78c30de-3af4-5e13-5713-21522946fcbd === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.229Z [DEBUG] TestCatalogRegister_checkRegistration.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.230Z [WARN] TestCatalogRegister_checkRegistration.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:52.232Z [INFO] TestCatalogRegister_checkRegistration.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:52.232Z [INFO] TestCatalogRegister_checkRegistration: consul server down writer.go:29: 2021-01-29T19:33:52.235Z [INFO] TestCatalogRegister_checkRegistration: shutdown complete writer.go:29: 2021-01-29T19:33:52.235Z [INFO] TestCatalogRegister_checkRegistration: Stopping server: protocol=DNS address=127.0.0.1:29658 network=tcp writer.go:29: 2021-01-29T19:33:52.236Z [INFO] TestCatalogRegister_checkRegistration: Stopping server: protocol=DNS address=127.0.0.1:29658 network=udp writer.go:29: 2021-01-29T19:33:52.237Z [INFO] TestCatalogRegister_checkRegistration: Stopping server: protocol=HTTP address=127.0.0.1:29659 network=tcp === CONT TestCatalogServiceNodes writer.go:29: 2021-01-29T19:33:52.239Z [INFO] TestCatalogServiceNodes: Requesting shutdown writer.go:29: 2021-01-29T19:33:52.240Z [INFO] TestCatalogServiceNodes.server: shutting down server writer.go:29: 2021-01-29T19:33:52.241Z [DEBUG] TestCatalogServiceNodes.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.243Z [WARN] TestCatalogServiceNodes.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:52.243Z [DEBUG] TestCatalogServiceNodes.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.249Z [WARN] TestCatalogServiceNodes.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:52.252Z [INFO] TestCatalogServiceNodes.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:52.252Z [INFO] TestCatalogServiceNodes: consul server down writer.go:29: 2021-01-29T19:33:52.253Z [INFO] TestCatalogServiceNodes: shutdown complete writer.go:29: 2021-01-29T19:33:52.254Z [INFO] TestCatalogServiceNodes: Stopping server: protocol=DNS address=127.0.0.1:29652 network=tcp writer.go:29: 2021-01-29T19:33:52.255Z [INFO] TestCatalogServiceNodes: Stopping server: protocol=DNS address=127.0.0.1:29652 network=udp === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.251Z [INFO] TestCatalogServices: Requesting shutdown === CONT TestCatalogServiceNodes writer.go:29: 2021-01-29T19:33:52.256Z [INFO] TestCatalogServiceNodes: Stopping server: protocol=HTTP address=127.0.0.1:29653 network=tcp === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.257Z [INFO] TestCatalogServices.server: shutting down server writer.go:29: 2021-01-29T19:33:52.260Z [DEBUG] TestCatalogServices.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.261Z [ERROR] TestCatalogServices.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:52.263Z [DEBUG] TestCatalogServices.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.264Z [WARN] TestCatalogServices.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:52.285Z [WARN] TestCatalogServices.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:52.287Z [INFO] TestCatalogServices.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:52.289Z [INFO] TestCatalogServices: consul server down writer.go:29: 2021-01-29T19:33:52.289Z [INFO] TestCatalogServices: shutdown complete writer.go:29: 2021-01-29T19:33:52.290Z [INFO] TestCatalogServices: Stopping server: protocol=DNS address=127.0.0.1:29670 network=tcp writer.go:29: 2021-01-29T19:33:52.291Z [INFO] TestCatalogServices: Stopping server: protocol=DNS address=127.0.0.1:29670 network=udp writer.go:29: 2021-01-29T19:33:52.292Z [INFO] TestCatalogServices: Stopping server: protocol=HTTP address=127.0.0.1:29671 network=tcp === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:52.498Z [DEBUG] TestCatalogServices_NodeMetaFilter: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:52.505Z [INFO] TestCatalogServices_NodeMetaFilter: Synced node info writer.go:29: 2021-01-29T19:33:52.505Z [INFO] TestCatalogServices_NodeMetaFilter: Requesting shutdown writer.go:29: 2021-01-29T19:33:52.508Z [INFO] TestCatalogServices_NodeMetaFilter.server: shutting down server writer.go:29: 2021-01-29T19:33:52.509Z [DEBUG] TestCatalogServices_NodeMetaFilter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.510Z [WARN] TestCatalogServices_NodeMetaFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:52.511Z [DEBUG] TestCatalogServices_NodeMetaFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.512Z [WARN] TestCatalogServices_NodeMetaFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:52.535Z [INFO] TestCatalogServices_NodeMetaFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:52.536Z [INFO] TestCatalogServices_NodeMetaFilter: consul server down writer.go:29: 2021-01-29T19:33:52.539Z [INFO] TestCatalogServices_NodeMetaFilter: shutdown complete writer.go:29: 2021-01-29T19:33:52.540Z [INFO] TestCatalogServices_NodeMetaFilter: Stopping server: protocol=DNS address=127.0.0.1:29664 network=tcp writer.go:29: 2021-01-29T19:33:52.542Z [INFO] TestCatalogServices_NodeMetaFilter: Stopping server: protocol=DNS address=127.0.0.1:29664 network=udp writer.go:29: 2021-01-29T19:33:52.544Z [INFO] TestCatalogServices_NodeMetaFilter: Stopping server: protocol=HTTP address=127.0.0.1:29665 network=tcp === CONT TestCatalogRegister_checkRegistration writer.go:29: 2021-01-29T19:33:52.738Z [INFO] TestCatalogRegister_checkRegistration: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:52.740Z [INFO] TestCatalogRegister_checkRegistration: Endpoints down --- PASS: TestCatalogRegister_checkRegistration (0.78s) === CONT TestCatalogNodes_DistanceSort === CONT TestCatalogServiceNodes writer.go:29: 2021-01-29T19:33:52.757Z [INFO] TestCatalogServiceNodes: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:52.759Z [INFO] TestCatalogServiceNodes: Endpoints down --- PASS: TestCatalogServiceNodes (1.00s) === CONT TestCatalogNodes_Blocking === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.763Z [WARN] TestCatalogNodes_DistanceSort: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:52.772Z [DEBUG] TestCatalogNodes_DistanceSort.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:52.774Z [DEBUG] TestCatalogNodes_DistanceSort.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.782Z [WARN] TestCatalogNodes_Blocking: bootstrap = true: do not enable unless necessary === CONT TestCatalogServices writer.go:29: 2021-01-29T19:33:52.793Z [INFO] TestCatalogServices: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:52.795Z [INFO] TestCatalogServices: Endpoints down --- PASS: TestCatalogServices (0.73s) === CONT TestCatalogNodes_Filter === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.799Z [INFO] TestCatalogNodes_DistanceSort.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:14a5e0d6-cad8-edf5-910d-5076d535aa78 Address:127.0.0.1:29687}]" === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.792Z [DEBUG] TestCatalogNodes_Blocking.tlsutil: Update: version=1 === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.801Z [INFO] TestCatalogNodes_DistanceSort.server.raft: entering follower state: follower="Node at 127.0.0.1:29687 [Follower]" leader= writer.go:29: 2021-01-29T19:33:52.802Z [INFO] TestCatalogNodes_DistanceSort.server.serf.wan: serf: EventMemberJoin: Node-14a5e0d6-cad8-edf5-910d-5076d535aa78.dc1 127.0.0.1 === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.803Z [DEBUG] TestCatalogNodes_Blocking.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.806Z [INFO] TestCatalogNodes_DistanceSort.server.serf.lan: serf: EventMemberJoin: Node-14a5e0d6-cad8-edf5-910d-5076d535aa78 127.0.0.1 writer.go:29: 2021-01-29T19:33:52.808Z [INFO] TestCatalogNodes_DistanceSort.server: Handled event for server in area: event=member-join server=Node-14a5e0d6-cad8-edf5-910d-5076d535aa78.dc1 area=wan writer.go:29: 2021-01-29T19:33:52.809Z [INFO] TestCatalogNodes_DistanceSort: Started DNS server: address=127.0.0.1:29682 network=udp writer.go:29: 2021-01-29T19:33:52.810Z [INFO] TestCatalogNodes_DistanceSort: Started DNS server: address=127.0.0.1:29682 network=tcp writer.go:29: 2021-01-29T19:33:52.810Z [INFO] TestCatalogNodes_DistanceSort.server: Adding LAN server: server="Node-14a5e0d6-cad8-edf5-910d-5076d535aa78 (Addr: tcp/127.0.0.1:29687) (DC: dc1)" === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.810Z [WARN] TestCatalogNodes_Filter: bootstrap = true: do not enable unless necessary === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.813Z [INFO] TestCatalogNodes_DistanceSort: Started HTTP server: address=127.0.0.1:29683 network=tcp writer.go:29: 2021-01-29T19:33:52.814Z [INFO] TestCatalogNodes_DistanceSort: started state syncer === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.815Z [INFO] TestCatalogNodes_Blocking.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:62972d88-52fc-eb35-5a5f-45fa17e66fa7 Address:127.0.0.1:29681}]" writer.go:29: 2021-01-29T19:33:52.818Z [INFO] TestCatalogNodes_Blocking.server.raft: entering follower state: follower="Node at 127.0.0.1:29681 [Follower]" leader= === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.814Z [DEBUG] TestCatalogNodes_Filter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:52.822Z [DEBUG] TestCatalogNodes_Filter.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.818Z [INFO] TestCatalogNodes_Blocking.server.serf.wan: serf: EventMemberJoin: Node-62972d88-52fc-eb35-5a5f-45fa17e66fa7.dc1 127.0.0.1 === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.833Z [INFO] TestCatalogNodes_Filter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:64cb505b-8c12-7cb3-bbfb-b9472ccb44cd Address:127.0.0.1:29693}]" === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.835Z [INFO] TestCatalogNodes_Blocking.server.serf.lan: serf: EventMemberJoin: Node-62972d88-52fc-eb35-5a5f-45fa17e66fa7 127.0.0.1 writer.go:29: 2021-01-29T19:33:52.839Z [INFO] TestCatalogNodes_Blocking.server: Adding LAN server: server="Node-62972d88-52fc-eb35-5a5f-45fa17e66fa7 (Addr: tcp/127.0.0.1:29681) (DC: dc1)" writer.go:29: 2021-01-29T19:33:52.839Z [INFO] TestCatalogNodes_Blocking.server: Handled event for server in area: event=member-join server=Node-62972d88-52fc-eb35-5a5f-45fa17e66fa7.dc1 area=wan === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.837Z [INFO] TestCatalogNodes_Filter.server.raft: entering follower state: follower="Node at 127.0.0.1:29693 [Follower]" leader= === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.840Z [INFO] TestCatalogNodes_Blocking: Started DNS server: address=127.0.0.1:29676 network=udp === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.840Z [WARN] TestCatalogNodes_DistanceSort.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.842Z [INFO] TestCatalogNodes_Blocking: Started DNS server: address=127.0.0.1:29676 network=tcp === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.842Z [INFO] TestCatalogNodes_DistanceSort.server.raft: entering candidate state: node="Node at 127.0.0.1:29687 [Candidate]" term=2 === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.844Z [INFO] TestCatalogNodes_Blocking: Started HTTP server: address=127.0.0.1:29677 network=tcp writer.go:29: 2021-01-29T19:33:52.845Z [INFO] TestCatalogNodes_Blocking: started state syncer === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.846Z [DEBUG] TestCatalogNodes_DistanceSort.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:52.847Z [DEBUG] TestCatalogNodes_DistanceSort.server.raft: vote granted: from=14a5e0d6-cad8-edf5-910d-5076d535aa78 term=2 tally=1 writer.go:29: 2021-01-29T19:33:52.847Z [INFO] TestCatalogNodes_DistanceSort.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:52.848Z [INFO] TestCatalogNodes_DistanceSort.server.raft: entering leader state: leader="Node at 127.0.0.1:29687 [Leader]" writer.go:29: 2021-01-29T19:33:52.850Z [INFO] TestCatalogNodes_DistanceSort.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:52.851Z [INFO] TestCatalogNodes_DistanceSort.server: New leader elected: payload=Node-14a5e0d6-cad8-edf5-910d-5076d535aa78 writer.go:29: 2021-01-29T19:33:52.852Z [DEBUG] TestCatalogNodes_DistanceSort.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29687 === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.847Z [INFO] TestCatalogNodes_Filter.server.serf.wan: serf: EventMemberJoin: Node-64cb505b-8c12-7cb3-bbfb-b9472ccb44cd.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:52.856Z [INFO] TestCatalogNodes_Filter.server.serf.lan: serf: EventMemberJoin: Node-64cb505b-8c12-7cb3-bbfb-b9472ccb44cd 127.0.0.1 === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.858Z [WARN] TestCatalogNodes_Blocking.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:52.860Z [INFO] TestCatalogNodes_Blocking.server.raft: entering candidate state: node="Node at 127.0.0.1:29681 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:52.862Z [DEBUG] TestCatalogNodes_Blocking.server.raft: votes: needed=1 === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.863Z [INFO] TestCatalogNodes_Filter.server: Adding LAN server: server="Node-64cb505b-8c12-7cb3-bbfb-b9472ccb44cd (Addr: tcp/127.0.0.1:29693) (DC: dc1)" === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.864Z [DEBUG] TestCatalogNodes_Blocking.server.raft: vote granted: from=62972d88-52fc-eb35-5a5f-45fa17e66fa7 term=2 tally=1 === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.865Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.865Z [INFO] TestCatalogNodes_Blocking.server.raft: election won: tally=1 === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.864Z [INFO] TestCatalogNodes_Filter: Started DNS server: address=127.0.0.1:29688 network=udp === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.866Z [INFO] TestCatalogNodes_Blocking.server.raft: entering leader state: leader="Node at 127.0.0.1:29681 [Leader]" writer.go:29: 2021-01-29T19:33:52.868Z [INFO] TestCatalogNodes_Blocking.server: cluster leadership acquired === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.868Z [INFO] TestCatalogNodes_Filter: Started DNS server: address=127.0.0.1:29688 network=tcp === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.870Z [INFO] TestCatalogNodes_Blocking.server: New leader elected: payload=Node-62972d88-52fc-eb35-5a5f-45fa17e66fa7 === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.864Z [INFO] TestCatalogNodes_Filter.server: Handled event for server in area: event=member-join server=Node-64cb505b-8c12-7cb3-bbfb-b9472ccb44cd.dc1 area=wan === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.875Z [INFO] TestCatalogNodes_DistanceSort.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:52.876Z [INFO] TestCatalogNodes_DistanceSort.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.877Z [DEBUG] TestCatalogNodes_DistanceSort.server: Skipping self join check for node since the cluster is too small: node=Node-14a5e0d6-cad8-edf5-910d-5076d535aa78 === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.873Z [INFO] TestCatalogNodes_Filter: Started HTTP server: address=127.0.0.1:29689 network=tcp === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.876Z [DEBUG] TestCatalogNodes_Blocking.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29681 === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:52.879Z [INFO] TestCatalogNodes_DistanceSort.server: member joined, marking health alive: member=Node-14a5e0d6-cad8-edf5-910d-5076d535aa78 === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.880Z [INFO] TestCatalogNodes_Filter: started state syncer === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.885Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:52.891Z [INFO] TestCatalogNodes_Blocking.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:52.891Z [INFO] TestCatalogNodes_Blocking.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.893Z [DEBUG] TestCatalogNodes_Blocking.server: Skipping self join check for node since the cluster is too small: node=Node-62972d88-52fc-eb35-5a5f-45fa17e66fa7 writer.go:29: 2021-01-29T19:33:52.893Z [INFO] TestCatalogNodes_Blocking.server: member joined, marking health alive: member=Node-62972d88-52fc-eb35-5a5f-45fa17e66fa7 === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:52.898Z [WARN] TestCatalogNodes_Filter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:52.899Z [INFO] TestCatalogNodes_Filter.server.raft: entering candidate state: node="Node at 127.0.0.1:29693 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:52.902Z [DEBUG] TestCatalogNodes_Filter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:52.903Z [DEBUG] TestCatalogNodes_Filter.server.raft: vote granted: from=64cb505b-8c12-7cb3-bbfb-b9472ccb44cd term=2 tally=1 writer.go:29: 2021-01-29T19:33:52.904Z [INFO] TestCatalogNodes_Filter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:52.906Z [INFO] TestCatalogNodes_Filter.server.raft: entering leader state: leader="Node at 127.0.0.1:29693 [Leader]" writer.go:29: 2021-01-29T19:33:52.907Z [INFO] TestCatalogNodes_Filter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:52.908Z [INFO] TestCatalogNodes_Filter.server: New leader elected: payload=Node-64cb505b-8c12-7cb3-bbfb-b9472ccb44cd writer.go:29: 2021-01-29T19:33:52.909Z [DEBUG] TestCatalogNodes_Filter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29693 writer.go:29: 2021-01-29T19:33:52.914Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:52.920Z [INFO] TestCatalogNodes_Filter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:52.923Z [INFO] TestCatalogNodes_Filter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:52.926Z [DEBUG] TestCatalogNodes_Filter.server: Skipping self join check for node since the cluster is too small: node=Node-64cb505b-8c12-7cb3-bbfb-b9472ccb44cd writer.go:29: 2021-01-29T19:33:52.929Z [INFO] TestCatalogNodes_Filter.server: member joined, marking health alive: member=Node-64cb505b-8c12-7cb3-bbfb-b9472ccb44cd writer.go:29: 2021-01-29T19:33:52.930Z [INFO] TestCatalogNodes_Filter: Synced node info === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:52.941Z [DEBUG] TestCatalogNodes_Blocking: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:52.962Z [INFO] TestCatalogNodes_Blocking: Synced node info === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:53.011Z [DEBUG] TestCatalogNodes_DistanceSort: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:53.015Z [INFO] TestCatalogNodes_DistanceSort: Synced node info writer.go:29: 2021-01-29T19:33:53.018Z [DEBUG] TestCatalogNodes_DistanceSort: Node info in sync === CONT TestCatalogServices_NodeMetaFilter writer.go:29: 2021-01-29T19:33:53.045Z [INFO] TestCatalogServices_NodeMetaFilter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:53.047Z [INFO] TestCatalogServices_NodeMetaFilter: Endpoints down --- PASS: TestCatalogServices_NodeMetaFilter (1.04s) === CONT TestCatalogNodes_MetaFilter writer.go:29: 2021-01-29T19:33:53.061Z [WARN] TestCatalogNodes_MetaFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:53.064Z [DEBUG] TestCatalogNodes_MetaFilter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:53.066Z [DEBUG] TestCatalogNodes_MetaFilter.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:53.071Z [INFO] TestCatalogNodes_MetaFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:90450315-87ca-ab30-5c76-87f2f95bd2e6 Address:127.0.0.1:29699}]" writer.go:29: 2021-01-29T19:33:53.075Z [INFO] TestCatalogNodes_MetaFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:29699 [Follower]" leader= writer.go:29: 2021-01-29T19:33:53.075Z [INFO] TestCatalogNodes_MetaFilter.server.serf.wan: serf: EventMemberJoin: Node-90450315-87ca-ab30-5c76-87f2f95bd2e6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:53.092Z [INFO] TestCatalogNodes_MetaFilter.server.serf.lan: serf: EventMemberJoin: Node-90450315-87ca-ab30-5c76-87f2f95bd2e6 127.0.0.1 writer.go:29: 2021-01-29T19:33:53.108Z [INFO] TestCatalogNodes_MetaFilter.server: Handled event for server in area: event=member-join server=Node-90450315-87ca-ab30-5c76-87f2f95bd2e6.dc1 area=wan writer.go:29: 2021-01-29T19:33:53.109Z [INFO] TestCatalogNodes_MetaFilter.server: Adding LAN server: server="Node-90450315-87ca-ab30-5c76-87f2f95bd2e6 (Addr: tcp/127.0.0.1:29699) (DC: dc1)" writer.go:29: 2021-01-29T19:33:53.114Z [INFO] TestCatalogNodes_MetaFilter: Started DNS server: address=127.0.0.1:29694 network=udp writer.go:29: 2021-01-29T19:33:53.116Z [INFO] TestCatalogNodes_MetaFilter: Started DNS server: address=127.0.0.1:29694 network=tcp writer.go:29: 2021-01-29T19:33:53.118Z [INFO] TestCatalogNodes_MetaFilter: Started HTTP server: address=127.0.0.1:29695 network=tcp writer.go:29: 2021-01-29T19:33:53.120Z [INFO] TestCatalogNodes_MetaFilter: started state syncer writer.go:29: 2021-01-29T19:33:53.138Z [WARN] TestCatalogNodes_MetaFilter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:53.139Z [INFO] TestCatalogNodes_MetaFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:29699 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:53.142Z [DEBUG] TestCatalogNodes_MetaFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:53.143Z [DEBUG] TestCatalogNodes_MetaFilter.server.raft: vote granted: from=90450315-87ca-ab30-5c76-87f2f95bd2e6 term=2 tally=1 writer.go:29: 2021-01-29T19:33:53.144Z [INFO] TestCatalogNodes_MetaFilter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:53.145Z [INFO] TestCatalogNodes_MetaFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:29699 [Leader]" writer.go:29: 2021-01-29T19:33:53.146Z [INFO] TestCatalogNodes_MetaFilter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:53.146Z [INFO] TestCatalogNodes_MetaFilter.server: New leader elected: payload=Node-90450315-87ca-ab30-5c76-87f2f95bd2e6 writer.go:29: 2021-01-29T19:33:53.147Z [DEBUG] TestCatalogNodes_MetaFilter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29699 writer.go:29: 2021-01-29T19:33:53.158Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:53.165Z [INFO] TestCatalogNodes_MetaFilter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:53.166Z [INFO] TestCatalogNodes_MetaFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.167Z [DEBUG] TestCatalogNodes_MetaFilter.server: Skipping self join check for node since the cluster is too small: node=Node-90450315-87ca-ab30-5c76-87f2f95bd2e6 writer.go:29: 2021-01-29T19:33:53.168Z [INFO] TestCatalogNodes_MetaFilter.server: member joined, marking health alive: member=Node-90450315-87ca-ab30-5c76-87f2f95bd2e6 === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:53.186Z [INFO] TestCatalogNodes_Filter: Requesting shutdown writer.go:29: 2021-01-29T19:33:53.188Z [INFO] TestCatalogNodes_Filter.server: shutting down server writer.go:29: 2021-01-29T19:33:53.189Z [DEBUG] TestCatalogNodes_Filter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.190Z [WARN] TestCatalogNodes_Filter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:53.190Z [DEBUG] TestCatalogNodes_Filter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.193Z [WARN] TestCatalogNodes_Filter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:53.197Z [INFO] TestCatalogNodes_Filter.server.router.manager: shutting down === CONT TestCatalogNodes_MetaFilter writer.go:29: 2021-01-29T19:33:53.199Z [INFO] TestCatalogNodes_MetaFilter: Requesting shutdown === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:53.197Z [INFO] TestCatalogNodes_Filter: consul server down === CONT TestCatalogNodes_MetaFilter writer.go:29: 2021-01-29T19:33:53.200Z [INFO] TestCatalogNodes_MetaFilter.server: shutting down server === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:53.200Z [INFO] TestCatalogNodes_Filter: shutdown complete === CONT TestCatalogNodes_MetaFilter writer.go:29: 2021-01-29T19:33:53.201Z [DEBUG] TestCatalogNodes_MetaFilter.leader: stopping routine: routine="CA root pruning" === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:53.202Z [INFO] TestCatalogNodes_Filter: Stopping server: protocol=DNS address=127.0.0.1:29688 network=tcp === CONT TestCatalogNodes_MetaFilter writer.go:29: 2021-01-29T19:33:53.202Z [WARN] TestCatalogNodes_MetaFilter.server.serf.lan: serf: Shutdown without a Leave === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:53.203Z [INFO] TestCatalogNodes_Filter: Stopping server: protocol=DNS address=127.0.0.1:29688 network=udp writer.go:29: 2021-01-29T19:33:53.206Z [INFO] TestCatalogNodes_Filter: Stopping server: protocol=HTTP address=127.0.0.1:29689 network=tcp === CONT TestCatalogNodes_MetaFilter writer.go:29: 2021-01-29T19:33:53.202Z [ERROR] TestCatalogNodes_MetaFilter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:53.202Z [DEBUG] TestCatalogNodes_MetaFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.206Z [WARN] TestCatalogNodes_MetaFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:53.213Z [INFO] TestCatalogNodes_MetaFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:53.213Z [INFO] TestCatalogNodes_MetaFilter: consul server down writer.go:29: 2021-01-29T19:33:53.216Z [INFO] TestCatalogNodes_MetaFilter: shutdown complete writer.go:29: 2021-01-29T19:33:53.217Z [INFO] TestCatalogNodes_MetaFilter: Stopping server: protocol=DNS address=127.0.0.1:29694 network=tcp writer.go:29: 2021-01-29T19:33:53.219Z [INFO] TestCatalogNodes_MetaFilter: Stopping server: protocol=DNS address=127.0.0.1:29694 network=udp writer.go:29: 2021-01-29T19:33:53.220Z [INFO] TestCatalogNodes_MetaFilter: Stopping server: protocol=HTTP address=127.0.0.1:29695 network=tcp === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:53.363Z [INFO] TestCatalogNodes_Blocking: Requesting shutdown writer.go:29: 2021-01-29T19:33:53.364Z [INFO] TestCatalogNodes_Blocking.server: shutting down server writer.go:29: 2021-01-29T19:33:53.366Z [DEBUG] TestCatalogNodes_Blocking.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.368Z [WARN] TestCatalogNodes_Blocking.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:53.371Z [DEBUG] TestCatalogNodes_Blocking.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.374Z [WARN] TestCatalogNodes_Blocking.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:53.391Z [INFO] TestCatalogNodes_Blocking.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:53.392Z [INFO] TestCatalogNodes_Blocking: consul server down writer.go:29: 2021-01-29T19:33:53.393Z [INFO] TestCatalogNodes_Blocking: shutdown complete writer.go:29: 2021-01-29T19:33:53.394Z [INFO] TestCatalogNodes_Blocking: Stopping server: protocol=DNS address=127.0.0.1:29676 network=tcp writer.go:29: 2021-01-29T19:33:53.395Z [INFO] TestCatalogNodes_Blocking: Stopping server: protocol=DNS address=127.0.0.1:29676 network=udp writer.go:29: 2021-01-29T19:33:53.396Z [INFO] TestCatalogNodes_Blocking: Stopping server: protocol=HTTP address=127.0.0.1:29677 network=tcp === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:53.535Z [INFO] TestCatalogNodes_DistanceSort: Requesting shutdown writer.go:29: 2021-01-29T19:33:53.541Z [INFO] TestCatalogNodes_DistanceSort.server: shutting down server writer.go:29: 2021-01-29T19:33:53.547Z [DEBUG] TestCatalogNodes_DistanceSort.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.549Z [WARN] TestCatalogNodes_DistanceSort.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:53.551Z [DEBUG] TestCatalogNodes_DistanceSort.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.554Z [WARN] TestCatalogNodes_DistanceSort.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:53.559Z [INFO] TestCatalogNodes_DistanceSort.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:53.564Z [INFO] TestCatalogNodes_DistanceSort: consul server down writer.go:29: 2021-01-29T19:33:53.568Z [INFO] TestCatalogNodes_DistanceSort: shutdown complete writer.go:29: 2021-01-29T19:33:53.571Z [INFO] TestCatalogNodes_DistanceSort: Stopping server: protocol=DNS address=127.0.0.1:29682 network=tcp writer.go:29: 2021-01-29T19:33:53.575Z [INFO] TestCatalogNodes_DistanceSort: Stopping server: protocol=DNS address=127.0.0.1:29682 network=udp writer.go:29: 2021-01-29T19:33:53.579Z [INFO] TestCatalogNodes_DistanceSort: Stopping server: protocol=HTTP address=127.0.0.1:29683 network=tcp === CONT TestCatalogNodes_Filter writer.go:29: 2021-01-29T19:33:53.707Z [INFO] TestCatalogNodes_Filter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:53.709Z [INFO] TestCatalogNodes_Filter: Endpoints down --- PASS: TestCatalogNodes_Filter (0.91s) === CONT TestCatalogNodes === CONT TestCatalogNodes_MetaFilter writer.go:29: 2021-01-29T19:33:53.720Z [INFO] TestCatalogNodes_MetaFilter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:53.724Z [INFO] TestCatalogNodes_MetaFilter: Endpoints down --- PASS: TestCatalogNodes_MetaFilter (0.68s) === CONT TestCatalogDatacenters === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.734Z [WARN] TestCatalogNodes: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:53.735Z [DEBUG] TestCatalogNodes.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:53.737Z [DEBUG] TestCatalogNodes.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.748Z [WARN] TestCatalogDatacenters: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:53.749Z [DEBUG] TestCatalogDatacenters.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:53.752Z [DEBUG] TestCatalogDatacenters.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.754Z [INFO] TestCatalogNodes.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a3dbd9f3-b71a-e239-ef3a-4cf36ddda281 Address:127.0.0.1:29717}]" writer.go:29: 2021-01-29T19:33:53.756Z [INFO] TestCatalogNodes.server.serf.wan: serf: EventMemberJoin: Node-a3dbd9f3-b71a-e239-ef3a-4cf36ddda281.dc1 127.0.0.1 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.757Z [INFO] TestCatalogDatacenters.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0cc6a293-e21b-ccac-f1b0-f595df016d18 Address:127.0.0.1:29711}]" === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.756Z [INFO] TestCatalogNodes.server.raft: entering follower state: follower="Node at 127.0.0.1:29717 [Follower]" leader= === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.760Z [INFO] TestCatalogDatacenters.server.raft: entering follower state: follower="Node at 127.0.0.1:29711 [Follower]" leader= === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.758Z [INFO] TestCatalogNodes.server.serf.lan: serf: EventMemberJoin: Node-a3dbd9f3-b71a-e239-ef3a-4cf36ddda281 127.0.0.1 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.761Z [INFO] TestCatalogDatacenters.server.serf.wan: serf: EventMemberJoin: Node-0cc6a293-e21b-ccac-f1b0-f595df016d18.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:53.764Z [INFO] TestCatalogDatacenters.server.serf.lan: serf: EventMemberJoin: Node-0cc6a293-e21b-ccac-f1b0-f595df016d18 127.0.0.1 === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.764Z [INFO] TestCatalogNodes.server: Handled event for server in area: event=member-join server=Node-a3dbd9f3-b71a-e239-ef3a-4cf36ddda281.dc1 area=wan === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.767Z [INFO] TestCatalogDatacenters: Started DNS server: address=127.0.0.1:29706 network=udp === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.765Z [INFO] TestCatalogNodes: Started DNS server: address=127.0.0.1:29712 network=udp === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.768Z [INFO] TestCatalogDatacenters.server: Adding LAN server: server="Node-0cc6a293-e21b-ccac-f1b0-f595df016d18 (Addr: tcp/127.0.0.1:29711) (DC: dc1)" writer.go:29: 2021-01-29T19:33:53.768Z [INFO] TestCatalogDatacenters.server: Handled event for server in area: event=member-join server=Node-0cc6a293-e21b-ccac-f1b0-f595df016d18.dc1 area=wan === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.768Z [INFO] TestCatalogNodes: Started DNS server: address=127.0.0.1:29712 network=tcp writer.go:29: 2021-01-29T19:33:53.772Z [INFO] TestCatalogNodes: Started HTTP server: address=127.0.0.1:29713 network=tcp === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.770Z [INFO] TestCatalogDatacenters: Started DNS server: address=127.0.0.1:29706 network=tcp === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.774Z [INFO] TestCatalogNodes: started state syncer === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.778Z [INFO] TestCatalogDatacenters: Started HTTP server: address=127.0.0.1:29707 network=tcp === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.773Z [INFO] TestCatalogNodes.server: Adding LAN server: server="Node-a3dbd9f3-b71a-e239-ef3a-4cf36ddda281 (Addr: tcp/127.0.0.1:29717) (DC: dc1)" === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.780Z [INFO] TestCatalogDatacenters: started state syncer === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.797Z [WARN] TestCatalogNodes.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:53.799Z [INFO] TestCatalogNodes.server.raft: entering candidate state: node="Node at 127.0.0.1:29717 [Candidate]" term=2 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.801Z [WARN] TestCatalogDatacenters.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:53.803Z [INFO] TestCatalogDatacenters.server.raft: entering candidate state: node="Node at 127.0.0.1:29711 [Candidate]" term=2 === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.803Z [DEBUG] TestCatalogNodes.server.raft: votes: needed=1 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.810Z [DEBUG] TestCatalogDatacenters.server.raft: votes: needed=1 === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.807Z [DEBUG] TestCatalogNodes.server.raft: vote granted: from=a3dbd9f3-b71a-e239-ef3a-4cf36ddda281 term=2 tally=1 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.811Z [DEBUG] TestCatalogDatacenters.server.raft: vote granted: from=0cc6a293-e21b-ccac-f1b0-f595df016d18 term=2 tally=1 writer.go:29: 2021-01-29T19:33:53.813Z [INFO] TestCatalogDatacenters.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:53.814Z [INFO] TestCatalogDatacenters.server.raft: entering leader state: leader="Node at 127.0.0.1:29711 [Leader]" writer.go:29: 2021-01-29T19:33:53.816Z [INFO] TestCatalogDatacenters.server: cluster leadership acquired === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.812Z [INFO] TestCatalogNodes.server.raft: election won: tally=1 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.817Z [INFO] TestCatalogDatacenters.server: New leader elected: payload=Node-0cc6a293-e21b-ccac-f1b0-f595df016d18 writer.go:29: 2021-01-29T19:33:53.818Z [DEBUG] TestCatalogDatacenters.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29711 === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.818Z [INFO] TestCatalogNodes.server.raft: entering leader state: leader="Node at 127.0.0.1:29717 [Leader]" writer.go:29: 2021-01-29T19:33:53.823Z [INFO] TestCatalogNodes.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:53.823Z [INFO] TestCatalogNodes.server: New leader elected: payload=Node-a3dbd9f3-b71a-e239-ef3a-4cf36ddda281 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.833Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:53.839Z [INFO] TestCatalogDatacenters.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.832Z [DEBUG] TestCatalogNodes.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29717 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.840Z [INFO] TestCatalogDatacenters.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.841Z [DEBUG] TestCatalogDatacenters.server: Skipping self join check for node since the cluster is too small: node=Node-0cc6a293-e21b-ccac-f1b0-f595df016d18 writer.go:29: 2021-01-29T19:33:53.843Z [INFO] TestCatalogDatacenters.server: member joined, marking health alive: member=Node-0cc6a293-e21b-ccac-f1b0-f595df016d18 === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.846Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:53.859Z [INFO] TestCatalogNodes.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:53.863Z [INFO] TestCatalogNodes.leader: started routine: routine="CA root pruning" === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.875Z [INFO] TestCatalogDatacenters: Requesting shutdown writer.go:29: 2021-01-29T19:33:53.878Z [INFO] TestCatalogDatacenters.server: shutting down server writer.go:29: 2021-01-29T19:33:53.880Z [DEBUG] TestCatalogDatacenters.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.881Z [WARN] TestCatalogDatacenters.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:53.880Z [ERROR] TestCatalogDatacenters.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:53.881Z [DEBUG] TestCatalogDatacenters.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:53.885Z [WARN] TestCatalogDatacenters.server.serf.wan: serf: Shutdown without a Leave === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.868Z [DEBUG] TestCatalogNodes.server: Skipping self join check for node since the cluster is too small: node=Node-a3dbd9f3-b71a-e239-ef3a-4cf36ddda281 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.890Z [INFO] TestCatalogDatacenters.server.router.manager: shutting down === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.890Z [INFO] TestCatalogNodes.server: member joined, marking health alive: member=Node-a3dbd9f3-b71a-e239-ef3a-4cf36ddda281 === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.890Z [INFO] TestCatalogDatacenters: consul server down writer.go:29: 2021-01-29T19:33:53.893Z [INFO] TestCatalogDatacenters: shutdown complete === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:53.877Z [INFO] TestCatalogNodes: Synced node info === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.894Z [INFO] TestCatalogDatacenters: Stopping server: protocol=DNS address=127.0.0.1:29706 network=tcp === CONT TestCatalogNodes_Blocking writer.go:29: 2021-01-29T19:33:53.897Z [INFO] TestCatalogNodes_Blocking: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:53.899Z [INFO] TestCatalogNodes_Blocking: Endpoints down --- PASS: TestCatalogNodes_Blocking (1.14s) === CONT TestCatalogDeregister === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:53.896Z [INFO] TestCatalogDatacenters: Stopping server: protocol=DNS address=127.0.0.1:29706 network=udp writer.go:29: 2021-01-29T19:33:53.903Z [INFO] TestCatalogDatacenters: Stopping server: protocol=HTTP address=127.0.0.1:29707 network=tcp === CONT TestCatalogDeregister writer.go:29: 2021-01-29T19:33:53.919Z [WARN] TestCatalogDeregister: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:53.922Z [DEBUG] TestCatalogDeregister.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:53.927Z [DEBUG] TestCatalogDeregister.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:53.943Z [INFO] TestCatalogDeregister.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:13c8970d-8628-155b-f5f6-8ccba9231eeb Address:127.0.0.1:29705}]" writer.go:29: 2021-01-29T19:33:53.946Z [INFO] TestCatalogDeregister.server.raft: entering follower state: follower="Node at 127.0.0.1:29705 [Follower]" leader= writer.go:29: 2021-01-29T19:33:53.948Z [INFO] TestCatalogDeregister.server.serf.wan: serf: EventMemberJoin: Node-13c8970d-8628-155b-f5f6-8ccba9231eeb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:53.953Z [INFO] TestCatalogDeregister.server.serf.lan: serf: EventMemberJoin: Node-13c8970d-8628-155b-f5f6-8ccba9231eeb 127.0.0.1 writer.go:29: 2021-01-29T19:33:53.958Z [INFO] TestCatalogDeregister.server: Handled event for server in area: event=member-join server=Node-13c8970d-8628-155b-f5f6-8ccba9231eeb.dc1 area=wan writer.go:29: 2021-01-29T19:33:53.958Z [INFO] TestCatalogDeregister.server: Adding LAN server: server="Node-13c8970d-8628-155b-f5f6-8ccba9231eeb (Addr: tcp/127.0.0.1:29705) (DC: dc1)" writer.go:29: 2021-01-29T19:33:53.959Z [INFO] TestCatalogDeregister: Started DNS server: address=127.0.0.1:29700 network=udp writer.go:29: 2021-01-29T19:33:53.963Z [INFO] TestCatalogDeregister: Started DNS server: address=127.0.0.1:29700 network=tcp writer.go:29: 2021-01-29T19:33:53.965Z [INFO] TestCatalogDeregister: Started HTTP server: address=127.0.0.1:29701 network=tcp writer.go:29: 2021-01-29T19:33:53.966Z [INFO] TestCatalogDeregister: started state syncer writer.go:29: 2021-01-29T19:33:54.006Z [WARN] TestCatalogDeregister.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:54.008Z [INFO] TestCatalogDeregister.server.raft: entering candidate state: node="Node at 127.0.0.1:29705 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:54.011Z [DEBUG] TestCatalogDeregister.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:54.013Z [DEBUG] TestCatalogDeregister.server.raft: vote granted: from=13c8970d-8628-155b-f5f6-8ccba9231eeb term=2 tally=1 writer.go:29: 2021-01-29T19:33:54.015Z [INFO] TestCatalogDeregister.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:54.017Z [INFO] TestCatalogDeregister.server.raft: entering leader state: leader="Node at 127.0.0.1:29705 [Leader]" writer.go:29: 2021-01-29T19:33:54.019Z [INFO] TestCatalogDeregister.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:54.020Z [INFO] TestCatalogDeregister.server: New leader elected: payload=Node-13c8970d-8628-155b-f5f6-8ccba9231eeb writer.go:29: 2021-01-29T19:33:54.029Z [DEBUG] TestCatalogDeregister.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29705 writer.go:29: 2021-01-29T19:33:54.044Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:54.052Z [INFO] TestCatalogDeregister.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:54.054Z [INFO] TestCatalogDeregister.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.056Z [DEBUG] TestCatalogDeregister.server: Skipping self join check for node since the cluster is too small: node=Node-13c8970d-8628-155b-f5f6-8ccba9231eeb writer.go:29: 2021-01-29T19:33:54.058Z [INFO] TestCatalogDeregister.server: member joined, marking health alive: member=Node-13c8970d-8628-155b-f5f6-8ccba9231eeb === CONT TestCatalogNodes_DistanceSort writer.go:29: 2021-01-29T19:33:54.087Z [INFO] TestCatalogNodes_DistanceSort: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:54.089Z [INFO] TestCatalogNodes_DistanceSort: Endpoints down === CONT TestCatalogRegister_Service_InvalidAddress --- PASS: TestCatalogNodes_DistanceSort (1.35s) === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.102Z [WARN] TestCatalogRegister_Service_InvalidAddress: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:54.104Z [DEBUG] TestCatalogRegister_Service_InvalidAddress.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:54.107Z [DEBUG] TestCatalogRegister_Service_InvalidAddress.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:54.113Z [INFO] TestCatalogRegister_Service_InvalidAddress.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7866edde-2864-3162-6d25-698b1358403c Address:127.0.0.1:29723}]" writer.go:29: 2021-01-29T19:33:54.116Z [INFO] TestCatalogRegister_Service_InvalidAddress.server.raft: entering follower state: follower="Node at 127.0.0.1:29723 [Follower]" leader= writer.go:29: 2021-01-29T19:33:54.117Z [INFO] TestCatalogRegister_Service_InvalidAddress.server.serf.wan: serf: EventMemberJoin: Node-7866edde-2864-3162-6d25-698b1358403c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:54.121Z [INFO] TestCatalogRegister_Service_InvalidAddress.server.serf.lan: serf: EventMemberJoin: Node-7866edde-2864-3162-6d25-698b1358403c 127.0.0.1 writer.go:29: 2021-01-29T19:33:54.124Z [INFO] TestCatalogRegister_Service_InvalidAddress.server: Adding LAN server: server="Node-7866edde-2864-3162-6d25-698b1358403c (Addr: tcp/127.0.0.1:29723) (DC: dc1)" writer.go:29: 2021-01-29T19:33:54.125Z [INFO] TestCatalogRegister_Service_InvalidAddress: Started DNS server: address=127.0.0.1:29718 network=udp writer.go:29: 2021-01-29T19:33:54.125Z [INFO] TestCatalogRegister_Service_InvalidAddress.server: Handled event for server in area: event=member-join server=Node-7866edde-2864-3162-6d25-698b1358403c.dc1 area=wan writer.go:29: 2021-01-29T19:33:54.128Z [INFO] TestCatalogRegister_Service_InvalidAddress: Started DNS server: address=127.0.0.1:29718 network=tcp writer.go:29: 2021-01-29T19:33:54.132Z [INFO] TestCatalogRegister_Service_InvalidAddress: Started HTTP server: address=127.0.0.1:29719 network=tcp === CONT TestCatalogDeregister writer.go:29: 2021-01-29T19:33:54.132Z [INFO] TestCatalogDeregister: Requesting shutdown === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.133Z [INFO] TestCatalogRegister_Service_InvalidAddress: started state syncer === CONT TestCatalogDeregister writer.go:29: 2021-01-29T19:33:54.133Z [INFO] TestCatalogDeregister.server: shutting down server writer.go:29: 2021-01-29T19:33:54.135Z [DEBUG] TestCatalogDeregister.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.136Z [WARN] TestCatalogDeregister.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:54.139Z [DEBUG] TestCatalogDeregister.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.137Z [ERROR] TestCatalogDeregister.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:54.144Z [WARN] TestCatalogDeregister.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:54.148Z [INFO] TestCatalogDeregister.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:54.149Z [INFO] TestCatalogDeregister: consul server down writer.go:29: 2021-01-29T19:33:54.150Z [INFO] TestCatalogDeregister: shutdown complete writer.go:29: 2021-01-29T19:33:54.151Z [INFO] TestCatalogDeregister: Stopping server: protocol=DNS address=127.0.0.1:29700 network=tcp writer.go:29: 2021-01-29T19:33:54.152Z [INFO] TestCatalogDeregister: Stopping server: protocol=DNS address=127.0.0.1:29700 network=udp writer.go:29: 2021-01-29T19:33:54.153Z [INFO] TestCatalogDeregister: Stopping server: protocol=HTTP address=127.0.0.1:29701 network=tcp === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.155Z [WARN] TestCatalogRegister_Service_InvalidAddress.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:54.156Z [INFO] TestCatalogRegister_Service_InvalidAddress.server.raft: entering candidate state: node="Node at 127.0.0.1:29723 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:54.158Z [DEBUG] TestCatalogRegister_Service_InvalidAddress.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:54.158Z [DEBUG] TestCatalogRegister_Service_InvalidAddress.server.raft: vote granted: from=7866edde-2864-3162-6d25-698b1358403c term=2 tally=1 writer.go:29: 2021-01-29T19:33:54.159Z [INFO] TestCatalogRegister_Service_InvalidAddress.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:54.160Z [INFO] TestCatalogRegister_Service_InvalidAddress.server.raft: entering leader state: leader="Node at 127.0.0.1:29723 [Leader]" === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:54.159Z [INFO] TestCatalogNodes: Requesting shutdown === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.161Z [INFO] TestCatalogRegister_Service_InvalidAddress.server: cluster leadership acquired === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:54.162Z [INFO] TestCatalogNodes.server: shutting down server === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.164Z [INFO] TestCatalogRegister_Service_InvalidAddress.server: New leader elected: payload=Node-7866edde-2864-3162-6d25-698b1358403c === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:54.164Z [DEBUG] TestCatalogNodes.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.166Z [WARN] TestCatalogNodes.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:54.166Z [DEBUG] TestCatalogNodes.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.187Z [WARN] TestCatalogNodes.server.serf.wan: serf: Shutdown without a Leave === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.188Z [DEBUG] TestCatalogRegister_Service_InvalidAddress.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29723 === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:54.190Z [INFO] TestCatalogNodes.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:54.190Z [INFO] TestCatalogNodes: consul server down === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.193Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:54.194Z [INFO] TestCatalogNodes: shutdown complete writer.go:29: 2021-01-29T19:33:54.196Z [INFO] TestCatalogNodes: Stopping server: protocol=DNS address=127.0.0.1:29712 network=tcp writer.go:29: 2021-01-29T19:33:54.198Z [INFO] TestCatalogNodes: Stopping server: protocol=DNS address=127.0.0.1:29712 network=udp writer.go:29: 2021-01-29T19:33:54.200Z [INFO] TestCatalogNodes: Stopping server: protocol=HTTP address=127.0.0.1:29713 network=tcp === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.215Z [INFO] TestCatalogRegister_Service_InvalidAddress.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:54.216Z [INFO] TestCatalogRegister_Service_InvalidAddress.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.217Z [DEBUG] TestCatalogRegister_Service_InvalidAddress.server: Skipping self join check for node since the cluster is too small: node=Node-7866edde-2864-3162-6d25-698b1358403c writer.go:29: 2021-01-29T19:33:54.217Z [INFO] TestCatalogRegister_Service_InvalidAddress.server: member joined, marking health alive: member=Node-7866edde-2864-3162-6d25-698b1358403c === RUN TestCatalogRegister_Service_InvalidAddress/addr_0.0.0.0 === RUN TestCatalogRegister_Service_InvalidAddress/addr_:: === RUN TestCatalogRegister_Service_InvalidAddress/addr_[::] === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.299Z [INFO] TestCatalogRegister_Service_InvalidAddress: Requesting shutdown writer.go:29: 2021-01-29T19:33:54.301Z [INFO] TestCatalogRegister_Service_InvalidAddress.server: shutting down server writer.go:29: 2021-01-29T19:33:54.302Z [DEBUG] TestCatalogRegister_Service_InvalidAddress.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.302Z [WARN] TestCatalogRegister_Service_InvalidAddress.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:54.302Z [ERROR] TestCatalogRegister_Service_InvalidAddress.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:54.302Z [DEBUG] TestCatalogRegister_Service_InvalidAddress.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.304Z [WARN] TestCatalogRegister_Service_InvalidAddress.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:54.308Z [INFO] TestCatalogRegister_Service_InvalidAddress.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:54.309Z [INFO] TestCatalogRegister_Service_InvalidAddress: consul server down writer.go:29: 2021-01-29T19:33:54.311Z [INFO] TestCatalogRegister_Service_InvalidAddress: shutdown complete writer.go:29: 2021-01-29T19:33:54.312Z [INFO] TestCatalogRegister_Service_InvalidAddress: Stopping server: protocol=DNS address=127.0.0.1:29718 network=tcp writer.go:29: 2021-01-29T19:33:54.314Z [INFO] TestCatalogRegister_Service_InvalidAddress: Stopping server: protocol=DNS address=127.0.0.1:29718 network=udp writer.go:29: 2021-01-29T19:33:54.315Z [INFO] TestCatalogRegister_Service_InvalidAddress: Stopping server: protocol=HTTP address=127.0.0.1:29719 network=tcp === CONT TestCatalogDatacenters writer.go:29: 2021-01-29T19:33:54.405Z [INFO] TestCatalogDatacenters: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:54.406Z [INFO] TestCatalogDatacenters: Endpoints down --- PASS: TestCatalogDatacenters (0.68s) === CONT TestBlacklist === RUN TestBlacklist/nothing_blocked_root === RUN TestBlacklist/nothing_blocked_path === RUN TestBlacklist/exact_match_1 === RUN TestBlacklist/exact_match_2 === RUN TestBlacklist/subpath === RUN TestBlacklist/longer_prefix === RUN TestBlacklist/longer_subpath === RUN TestBlacklist/partial_prefix === RUN TestBlacklist/no_match --- PASS: TestBlacklist (0.02s) --- PASS: TestBlacklist/nothing_blocked_root (0.00s) --- PASS: TestBlacklist/nothing_blocked_path (0.00s) --- PASS: TestBlacklist/exact_match_1 (0.00s) --- PASS: TestBlacklist/exact_match_2 (0.00s) --- PASS: TestBlacklist/subpath (0.00s) --- PASS: TestBlacklist/longer_prefix (0.00s) --- PASS: TestBlacklist/longer_subpath (0.00s) --- PASS: TestBlacklist/partial_prefix (0.00s) --- PASS: TestBlacklist/no_match (0.00s) === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.453Z [WARN] TestAgent_RerouteNewHTTPChecks: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:54.456Z [DEBUG] TestAgent_RerouteNewHTTPChecks.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:54.461Z [DEBUG] TestAgent_RerouteNewHTTPChecks.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:54.468Z [INFO] TestAgent_RerouteNewHTTPChecks.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:de8f6658-ee79-cbf2-dc88-1da9179aea9b Address:127.0.0.1:29735}]" writer.go:29: 2021-01-29T19:33:54.471Z [INFO] TestAgent_RerouteNewHTTPChecks.server.raft: entering follower state: follower="Node at 127.0.0.1:29735 [Follower]" leader= writer.go:29: 2021-01-29T19:33:54.476Z [INFO] TestAgent_RerouteNewHTTPChecks.server.serf.wan: serf: EventMemberJoin: Node-de8f6658-ee79-cbf2-dc88-1da9179aea9b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:54.482Z [INFO] TestAgent_RerouteNewHTTPChecks.server.serf.lan: serf: EventMemberJoin: Node-de8f6658-ee79-cbf2-dc88-1da9179aea9b 127.0.0.1 writer.go:29: 2021-01-29T19:33:54.487Z [INFO] TestAgent_RerouteNewHTTPChecks.server: Adding LAN server: server="Node-de8f6658-ee79-cbf2-dc88-1da9179aea9b (Addr: tcp/127.0.0.1:29735) (DC: dc1)" writer.go:29: 2021-01-29T19:33:54.489Z [INFO] TestAgent_RerouteNewHTTPChecks.server: Handled event for server in area: event=member-join server=Node-de8f6658-ee79-cbf2-dc88-1da9179aea9b.dc1 area=wan writer.go:29: 2021-01-29T19:33:54.494Z [INFO] TestAgent_RerouteNewHTTPChecks: Started DNS server: address=127.0.0.1:29730 network=tcp writer.go:29: 2021-01-29T19:33:54.497Z [INFO] TestAgent_RerouteNewHTTPChecks: Started DNS server: address=127.0.0.1:29730 network=udp writer.go:29: 2021-01-29T19:33:54.501Z [INFO] TestAgent_RerouteNewHTTPChecks: Started HTTP server: address=127.0.0.1:29731 network=tcp writer.go:29: 2021-01-29T19:33:54.502Z [INFO] TestAgent_RerouteNewHTTPChecks: started state syncer writer.go:29: 2021-01-29T19:33:54.521Z [WARN] TestAgent_RerouteNewHTTPChecks.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:54.524Z [INFO] TestAgent_RerouteNewHTTPChecks.server.raft: entering candidate state: node="Node at 127.0.0.1:29735 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:54.527Z [DEBUG] TestAgent_RerouteNewHTTPChecks.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:54.529Z [DEBUG] TestAgent_RerouteNewHTTPChecks.server.raft: vote granted: from=de8f6658-ee79-cbf2-dc88-1da9179aea9b term=2 tally=1 writer.go:29: 2021-01-29T19:33:54.531Z [INFO] TestAgent_RerouteNewHTTPChecks.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:54.533Z [INFO] TestAgent_RerouteNewHTTPChecks.server.raft: entering leader state: leader="Node at 127.0.0.1:29735 [Leader]" writer.go:29: 2021-01-29T19:33:54.534Z [INFO] TestAgent_RerouteNewHTTPChecks.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:54.535Z [INFO] TestAgent_RerouteNewHTTPChecks.server: New leader elected: payload=Node-de8f6658-ee79-cbf2-dc88-1da9179aea9b writer.go:29: 2021-01-29T19:33:54.536Z [DEBUG] TestAgent_RerouteNewHTTPChecks.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29735 writer.go:29: 2021-01-29T19:33:54.556Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:54.560Z [INFO] TestAgent_RerouteNewHTTPChecks: Synced node info writer.go:29: 2021-01-29T19:33:54.561Z [INFO] TestAgent_RerouteNewHTTPChecks.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:54.562Z [INFO] TestAgent_RerouteNewHTTPChecks.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.563Z [DEBUG] TestAgent_RerouteNewHTTPChecks.server: Skipping self join check for node since the cluster is too small: node=Node-de8f6658-ee79-cbf2-dc88-1da9179aea9b writer.go:29: 2021-01-29T19:33:54.564Z [INFO] TestAgent_RerouteNewHTTPChecks.server: member joined, marking health alive: member=Node-de8f6658-ee79-cbf2-dc88-1da9179aea9b === CONT TestCatalogDeregister writer.go:29: 2021-01-29T19:33:54.655Z [INFO] TestCatalogDeregister: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:54.656Z [INFO] TestCatalogDeregister: Endpoints down --- PASS: TestCatalogDeregister (0.76s) === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:54.669Z [WARN] TestAgent_RerouteExistingHTTPChecks: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:54.671Z [DEBUG] TestAgent_RerouteExistingHTTPChecks.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:54.673Z [DEBUG] TestAgent_RerouteExistingHTTPChecks.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:54.678Z [INFO] TestAgent_RerouteExistingHTTPChecks.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b1d56f67-c252-194c-cc53-8f5d5ffb131c Address:127.0.0.1:29729}]" writer.go:29: 2021-01-29T19:33:54.681Z [INFO] TestAgent_RerouteExistingHTTPChecks.server.raft: entering follower state: follower="Node at 127.0.0.1:29729 [Follower]" leader= writer.go:29: 2021-01-29T19:33:54.684Z [INFO] TestAgent_RerouteExistingHTTPChecks.server.serf.wan: serf: EventMemberJoin: Node-b1d56f67-c252-194c-cc53-8f5d5ffb131c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:54.692Z [INFO] TestAgent_RerouteExistingHTTPChecks.server.serf.lan: serf: EventMemberJoin: Node-b1d56f67-c252-194c-cc53-8f5d5ffb131c 127.0.0.1 writer.go:29: 2021-01-29T19:33:54.699Z [INFO] TestAgent_RerouteExistingHTTPChecks.server: Adding LAN server: server="Node-b1d56f67-c252-194c-cc53-8f5d5ffb131c (Addr: tcp/127.0.0.1:29729) (DC: dc1)" writer.go:29: 2021-01-29T19:33:54.700Z [INFO] TestAgent_RerouteExistingHTTPChecks: Started DNS server: address=127.0.0.1:29724 network=udp === CONT TestCatalogNodes writer.go:29: 2021-01-29T19:33:54.702Z [INFO] TestCatalogNodes: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:54.703Z [INFO] TestCatalogNodes: Endpoints down --- PASS: TestCatalogNodes (0.99s) === CONT TestAgent_consulConfig_RaftTrailingLogs === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:54.702Z [INFO] TestAgent_RerouteExistingHTTPChecks: Started DNS server: address=127.0.0.1:29724 network=tcp writer.go:29: 2021-01-29T19:33:54.700Z [INFO] TestAgent_RerouteExistingHTTPChecks.server: Handled event for server in area: event=member-join server=Node-b1d56f67-c252-194c-cc53-8f5d5ffb131c.dc1 area=wan writer.go:29: 2021-01-29T19:33:54.705Z [INFO] TestAgent_RerouteExistingHTTPChecks: Started HTTP server: address=127.0.0.1:29725 network=tcp writer.go:29: 2021-01-29T19:33:54.707Z [INFO] TestAgent_RerouteExistingHTTPChecks: started state syncer === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:54.711Z [WARN] TestAgent_consulConfig_RaftTrailingLogs: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:54.712Z [DEBUG] TestAgent_consulConfig_RaftTrailingLogs.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:54.713Z [DEBUG] TestAgent_consulConfig_RaftTrailingLogs.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:54.719Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:959e9ac4-380d-16ff-db18-e6914248d0ec Address:127.0.0.1:29747}]" writer.go:29: 2021-01-29T19:33:54.721Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server.raft: entering follower state: follower="Node at 127.0.0.1:29747 [Follower]" leader= === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:54.725Z [WARN] TestAgent_RerouteExistingHTTPChecks.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:54.726Z [INFO] TestAgent_RerouteExistingHTTPChecks.server.raft: entering candidate state: node="Node at 127.0.0.1:29729 [Candidate]" term=2 === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:54.721Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server.serf.wan: serf: EventMemberJoin: Node-959e9ac4-380d-16ff-db18-e6914248d0ec.dc1 127.0.0.1 === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:54.728Z [DEBUG] TestAgent_RerouteExistingHTTPChecks.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:54.729Z [DEBUG] TestAgent_RerouteExistingHTTPChecks.server.raft: vote granted: from=b1d56f67-c252-194c-cc53-8f5d5ffb131c term=2 tally=1 writer.go:29: 2021-01-29T19:33:54.731Z [INFO] TestAgent_RerouteExistingHTTPChecks.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:54.733Z [INFO] TestAgent_RerouteExistingHTTPChecks.server.raft: entering leader state: leader="Node at 127.0.0.1:29729 [Leader]" === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:54.732Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server.serf.lan: serf: EventMemberJoin: Node-959e9ac4-380d-16ff-db18-e6914248d0ec 127.0.0.1 === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:54.734Z [INFO] TestAgent_RerouteExistingHTTPChecks.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:54.735Z [INFO] TestAgent_RerouteExistingHTTPChecks.server: New leader elected: payload=Node-b1d56f67-c252-194c-cc53-8f5d5ffb131c === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:54.737Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server: Adding LAN server: server="Node-959e9ac4-380d-16ff-db18-e6914248d0ec (Addr: tcp/127.0.0.1:29747) (DC: dc1)" writer.go:29: 2021-01-29T19:33:54.739Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server: Handled event for server in area: event=member-join server=Node-959e9ac4-380d-16ff-db18-e6914248d0ec.dc1 area=wan === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:54.739Z [DEBUG] TestAgent_RerouteExistingHTTPChecks.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29729 writer.go:29: 2021-01-29T19:33:54.752Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:54.758Z [INFO] TestAgent_RerouteExistingHTTPChecks.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:54.759Z [INFO] TestAgent_RerouteExistingHTTPChecks.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.760Z [DEBUG] TestAgent_RerouteExistingHTTPChecks.server: Skipping self join check for node since the cluster is too small: node=Node-b1d56f67-c252-194c-cc53-8f5d5ffb131c writer.go:29: 2021-01-29T19:33:54.761Z [INFO] TestAgent_RerouteExistingHTTPChecks.server: member joined, marking health alive: member=Node-b1d56f67-c252-194c-cc53-8f5d5ffb131c === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:54.759Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: Started DNS server: address=127.0.0.1:29742 network=tcp writer.go:29: 2021-01-29T19:33:54.765Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: Started DNS server: address=127.0.0.1:29742 network=udp writer.go:29: 2021-01-29T19:33:54.766Z [WARN] TestAgent_consulConfig_RaftTrailingLogs.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:54.771Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server.raft: entering candidate state: node="Node at 127.0.0.1:29747 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:54.768Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: Started HTTP server: address=127.0.0.1:29743 network=tcp writer.go:29: 2021-01-29T19:33:54.778Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: started state syncer writer.go:29: 2021-01-29T19:33:54.785Z [DEBUG] TestAgent_consulConfig_RaftTrailingLogs.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:54.786Z [DEBUG] TestAgent_consulConfig_RaftTrailingLogs.server.raft: vote granted: from=959e9ac4-380d-16ff-db18-e6914248d0ec term=2 tally=1 writer.go:29: 2021-01-29T19:33:54.788Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:54.789Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server.raft: entering leader state: leader="Node at 127.0.0.1:29747 [Leader]" writer.go:29: 2021-01-29T19:33:54.791Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:54.791Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server: New leader elected: payload=Node-959e9ac4-380d-16ff-db18-e6914248d0ec writer.go:29: 2021-01-29T19:33:54.798Z [DEBUG] TestAgent_consulConfig_RaftTrailingLogs.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29747 writer.go:29: 2021-01-29T19:33:54.810Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:54.820Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:54.823Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.leader: started routine: routine="CA root pruning" === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.817Z [INFO] TestCatalogRegister_Service_InvalidAddress: Waiting for endpoints to shut down === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:54.826Z [DEBUG] TestAgent_consulConfig_RaftTrailingLogs.server: Skipping self join check for node since the cluster is too small: node=Node-959e9ac4-380d-16ff-db18-e6914248d0ec === CONT TestCatalogRegister_Service_InvalidAddress writer.go:29: 2021-01-29T19:33:54.829Z [INFO] TestCatalogRegister_Service_InvalidAddress: Endpoints down --- PASS: TestCatalogRegister_Service_InvalidAddress (0.74s) --- PASS: TestCatalogRegister_Service_InvalidAddress/addr_0.0.0.0 (0.00s) --- PASS: TestCatalogRegister_Service_InvalidAddress/addr_:: (0.00s) --- PASS: TestCatalogRegister_Service_InvalidAddress/addr_[::] (0.00s) === CONT TestAgent_consulConfig_AutoEncryptAllowTLS === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:54.829Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server: member joined, marking health alive: member=Node-959e9ac4-380d-16ff-db18-e6914248d0ec === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.840Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:54.842Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Node info in sync writer.go:29: 2021-01-29T19:33:54.843Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Node info in sync === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.845Z [WARN] TestAgent_consulConfig_AutoEncryptAllowTLS: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:54.846Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:54.848Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:54.855Z [DEBUG] TestAgent_RerouteExistingHTTPChecks: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:54.868Z [INFO] TestAgent_RerouteExistingHTTPChecks: Synced node info === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.870Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:083dbf2f-d4ef-1343-aa7d-5fa8baeec4b2 Address:127.0.0.1:29759}]" writer.go:29: 2021-01-29T19:33:54.872Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server.raft: entering follower state: follower="Node at 127.0.0.1:29759 [Follower]" leader= writer.go:29: 2021-01-29T19:33:54.873Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server.serf.wan: serf: EventMemberJoin: Node-083dbf2f-d4ef-1343-aa7d-5fa8baeec4b2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:54.875Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS.tlsutil: UpdateAutoEncryptCA: version=2 === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:54.871Z [DEBUG] TestAgent_RerouteExistingHTTPChecks: Node info in sync === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.883Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server.serf.lan: serf: EventMemberJoin: Node-083dbf2f-d4ef-1343-aa7d-5fa8baeec4b2 127.0.0.1 writer.go:29: 2021-01-29T19:33:54.895Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server: Adding LAN server: server="Node-083dbf2f-d4ef-1343-aa7d-5fa8baeec4b2 (Addr: tcp/127.0.0.1:29759) (DC: dc1)" writer.go:29: 2021-01-29T19:33:54.897Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server: Handled event for server in area: event=member-join server=Node-083dbf2f-d4ef-1343-aa7d-5fa8baeec4b2.dc1 area=wan writer.go:29: 2021-01-29T19:33:54.903Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: Started DNS server: address=127.0.0.1:29754 network=tcp writer.go:29: 2021-01-29T19:33:54.905Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: Started DNS server: address=127.0.0.1:29754 network=udp writer.go:29: 2021-01-29T19:33:54.915Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: Started HTTP server: address=127.0.0.1:29755 network=tcp === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.920Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Node info in sync writer.go:29: 2021-01-29T19:33:54.920Z [WARN] TestAgent_RerouteNewHTTPChecks: check has interval below minimum: check=http minimum_interval=1s writer.go:29: 2021-01-29T19:33:54.924Z [DEBUG] TestAgent_RerouteNewHTTPChecks.tlsutil: OutgoingTLSConfigForCheck: version=1 writer.go:29: 2021-01-29T19:33:54.924Z [INFO] TestAgent_RerouteNewHTTPChecks: Synced service: service=web === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.919Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: started state syncer === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.926Z [INFO] TestAgent_RerouteNewHTTPChecks: Synced service: service=web-proxy writer.go:29: 2021-01-29T19:33:54.927Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Node info in sync writer.go:29: 2021-01-29T19:33:54.928Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Service in sync: service=web writer.go:29: 2021-01-29T19:33:54.929Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Service in sync: service=web-proxy === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.930Z [WARN] TestAgent_consulConfig_AutoEncryptAllowTLS.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:54.932Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server.raft: entering candidate state: node="Node at 127.0.0.1:29759 [Candidate]" term=2 === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:54.935Z [DEBUG] TestAgent_consulConfig_RaftTrailingLogs: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.930Z [WARN] TestAgent_RerouteNewHTTPChecks: check has interval below minimum: check=grpc minimum_interval=1s writer.go:29: 2021-01-29T19:33:54.930Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Node info in sync writer.go:29: 2021-01-29T19:33:54.942Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Service in sync: service=web writer.go:29: 2021-01-29T19:33:54.944Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Service in sync: service=web-proxy writer.go:29: 2021-01-29T19:33:54.946Z [INFO] TestAgent_RerouteNewHTTPChecks: Synced check: check=http writer.go:29: 2021-01-29T19:33:54.948Z [INFO] TestAgent_RerouteNewHTTPChecks: Requesting shutdown writer.go:29: 2021-01-29T19:33:54.948Z [INFO] TestAgent_RerouteNewHTTPChecks.server: shutting down server === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:54.948Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: Synced node info === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.949Z [DEBUG] TestAgent_RerouteNewHTTPChecks.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.947Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS.server.raft: votes: needed=1 === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.950Z [WARN] TestAgent_RerouteNewHTTPChecks.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.950Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS.server.raft: vote granted: from=083dbf2f-d4ef-1343-aa7d-5fa8baeec4b2 term=2 tally=1 writer.go:29: 2021-01-29T19:33:54.952Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:54.953Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server.raft: entering leader state: leader="Node at 127.0.0.1:29759 [Leader]" === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.951Z [DEBUG] TestAgent_RerouteNewHTTPChecks.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.954Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server: cluster leadership acquired === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.951Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Node info in sync === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.955Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server: New leader elected: payload=Node-083dbf2f-d4ef-1343-aa7d-5fa8baeec4b2 === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.956Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Service in sync: service=web writer.go:29: 2021-01-29T19:33:54.952Z [WARN] TestAgent_RerouteNewHTTPChecks.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.957Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29759 === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.957Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Service in sync: service=web-proxy === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.966Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.968Z [DEBUG] TestAgent_RerouteNewHTTPChecks: Check in sync: check=http === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:54.975Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:54.977Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:54.975Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS.tlsutil: UpdateAutoEncryptCA: version=3 writer.go:29: 2021-01-29T19:33:54.978Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS.server: Skipping self join check for node since the cluster is too small: node=Node-083dbf2f-d4ef-1343-aa7d-5fa8baeec4b2 writer.go:29: 2021-01-29T19:33:54.979Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server: member joined, marking health alive: member=Node-083dbf2f-d4ef-1343-aa7d-5fa8baeec4b2 === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:54.963Z [INFO] TestAgent_RerouteNewHTTPChecks.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:54.963Z [INFO] TestAgent_RerouteNewHTTPChecks: consul server down writer.go:29: 2021-01-29T19:33:54.972Z [WARN] TestAgent_RerouteNewHTTPChecks: Syncing check failed.: check=grpc error="No cluster leader" writer.go:29: 2021-01-29T19:33:54.986Z [ERROR] TestAgent_RerouteNewHTTPChecks.anti_entropy: failed to sync changes: error="No cluster leader" writer.go:29: 2021-01-29T19:33:54.983Z [INFO] TestAgent_RerouteNewHTTPChecks: shutdown complete writer.go:29: 2021-01-29T19:33:54.990Z [INFO] TestAgent_RerouteNewHTTPChecks: Stopping server: protocol=DNS address=127.0.0.1:29730 network=tcp writer.go:29: 2021-01-29T19:33:54.992Z [INFO] TestAgent_RerouteNewHTTPChecks: Stopping server: protocol=DNS address=127.0.0.1:29730 network=udp writer.go:29: 2021-01-29T19:33:54.993Z [INFO] TestAgent_RerouteNewHTTPChecks: Stopping server: protocol=HTTP address=127.0.0.1:29731 network=tcp === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:55.045Z [WARN] TestAgent_RerouteExistingHTTPChecks: check has interval below minimum: check=http minimum_interval=1s writer.go:29: 2021-01-29T19:33:55.046Z [DEBUG] TestAgent_RerouteExistingHTTPChecks.tlsutil: OutgoingTLSConfigForCheck: version=1 writer.go:29: 2021-01-29T19:33:55.048Z [WARN] TestAgent_RerouteExistingHTTPChecks: check has interval below minimum: check=grpc minimum_interval=1s writer.go:29: 2021-01-29T19:33:55.058Z [INFO] TestAgent_RerouteExistingHTTPChecks: Requesting shutdown writer.go:29: 2021-01-29T19:33:55.059Z [INFO] TestAgent_RerouteExistingHTTPChecks.server: shutting down server writer.go:29: 2021-01-29T19:33:55.059Z [DEBUG] TestAgent_RerouteExistingHTTPChecks.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:55.061Z [WARN] TestAgent_RerouteExistingHTTPChecks.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:55.065Z [DEBUG] TestAgent_RerouteExistingHTTPChecks.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:55.068Z [WARN] TestAgent_RerouteExistingHTTPChecks.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:55.068Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: Requesting shutdown writer.go:29: 2021-01-29T19:33:55.070Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server: shutting down server === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:55.073Z [INFO] TestAgent_RerouteExistingHTTPChecks.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:55.074Z [INFO] TestAgent_RerouteExistingHTTPChecks: consul server down === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:55.072Z [DEBUG] TestAgent_consulConfig_RaftTrailingLogs.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:55.076Z [INFO] TestAgent_RerouteExistingHTTPChecks: shutdown complete writer.go:29: 2021-01-29T19:33:55.077Z [INFO] TestAgent_RerouteExistingHTTPChecks: Stopping server: protocol=DNS address=127.0.0.1:29724 network=tcp writer.go:29: 2021-01-29T19:33:55.078Z [INFO] TestAgent_RerouteExistingHTTPChecks: Stopping server: protocol=DNS address=127.0.0.1:29724 network=udp writer.go:29: 2021-01-29T19:33:55.080Z [INFO] TestAgent_RerouteExistingHTTPChecks: Stopping server: protocol=HTTP address=127.0.0.1:29725 network=tcp === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:55.077Z [WARN] TestAgent_consulConfig_RaftTrailingLogs.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:55.082Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:55.083Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: Synced node info writer.go:29: 2021-01-29T19:33:55.085Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS: Node info in sync === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:55.077Z [DEBUG] TestAgent_consulConfig_RaftTrailingLogs.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:55.083Z [WARN] TestAgent_consulConfig_RaftTrailingLogs.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:55.089Z [INFO] TestAgent_consulConfig_RaftTrailingLogs.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:55.091Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: consul server down writer.go:29: 2021-01-29T19:33:55.094Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: shutdown complete writer.go:29: 2021-01-29T19:33:55.097Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: Stopping server: protocol=DNS address=127.0.0.1:29742 network=tcp writer.go:29: 2021-01-29T19:33:55.099Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: Stopping server: protocol=DNS address=127.0.0.1:29742 network=udp writer.go:29: 2021-01-29T19:33:55.101Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: Stopping server: protocol=HTTP address=127.0.0.1:29743 network=tcp === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:55.160Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: Requesting shutdown writer.go:29: 2021-01-29T19:33:55.171Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server: shutting down server writer.go:29: 2021-01-29T19:33:55.173Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:55.174Z [WARN] TestAgent_consulConfig_AutoEncryptAllowTLS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:55.177Z [DEBUG] TestAgent_consulConfig_AutoEncryptAllowTLS.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:55.183Z [WARN] TestAgent_consulConfig_AutoEncryptAllowTLS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:55.188Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:55.189Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: consul server down writer.go:29: 2021-01-29T19:33:55.192Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: shutdown complete writer.go:29: 2021-01-29T19:33:55.193Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: Stopping server: protocol=DNS address=127.0.0.1:29754 network=tcp writer.go:29: 2021-01-29T19:33:55.195Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: Stopping server: protocol=DNS address=127.0.0.1:29754 network=udp writer.go:29: 2021-01-29T19:33:55.199Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: Stopping server: protocol=HTTP address=127.0.0.1:29755 network=tcp === CONT TestAgent_RerouteNewHTTPChecks writer.go:29: 2021-01-29T19:33:55.495Z [INFO] TestAgent_RerouteNewHTTPChecks: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:55.497Z [INFO] TestAgent_RerouteNewHTTPChecks: Endpoints down --- PASS: TestAgent_RerouteNewHTTPChecks (1.07s) === CONT TestAgent_ReloadConfigTLSConfigFailure === CONT TestAgent_RerouteExistingHTTPChecks writer.go:29: 2021-01-29T19:33:55.582Z [INFO] TestAgent_RerouteExistingHTTPChecks: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:55.584Z [INFO] TestAgent_RerouteExistingHTTPChecks: Endpoints down --- PASS: TestAgent_RerouteExistingHTTPChecks (0.93s) === CONT TestAgent_ReloadConfigIncomingRPCConfig === CONT TestAgent_consulConfig_RaftTrailingLogs writer.go:29: 2021-01-29T19:33:55.606Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:55.608Z [INFO] TestAgent_consulConfig_RaftTrailingLogs: Endpoints down --- PASS: TestAgent_consulConfig_RaftTrailingLogs (0.91s) === CONT TestAgent_ReloadConfigOutgoingRPCConfig === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.622Z [WARN] TestAgent_ReloadConfigTLSConfigFailure: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:55.628Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:55.631Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_consulConfig_AutoEncryptAllowTLS writer.go:29: 2021-01-29T19:33:55.702Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:55.706Z [INFO] TestAgent_consulConfig_AutoEncryptAllowTLS: Endpoints down === CONT TestAgent_loadTokens --- PASS: TestAgent_consulConfig_AutoEncryptAllowTLS (0.88s) === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.750Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9af6b50a-1453-f014-44b6-a053beff6f9d Address:127.0.0.1:29753}]" writer.go:29: 2021-01-29T19:33:55.781Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server.serf.wan: serf: EventMemberJoin: Node-9af6b50a-1453-f014-44b6-a053beff6f9d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:55.822Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server.raft: entering follower state: follower="Node at 127.0.0.1:29753 [Follower]" leader= writer.go:29: 2021-01-29T19:33:55.882Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server.serf.lan: serf: EventMemberJoin: Node-9af6b50a-1453-f014-44b6-a053beff6f9d 127.0.0.1 writer.go:29: 2021-01-29T19:33:55.892Z [WARN] TestAgent_ReloadConfigTLSConfigFailure.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:55.894Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server.raft: entering candidate state: node="Node at 127.0.0.1:29753 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:55.898Z [WARN] TestAgent_ReloadConfigTLSConfigFailure.server.raft: unable to get address for sever, using fallback address: id=9af6b50a-1453-f014-44b6-a053beff6f9d fallback=127.0.0.1:29753 error="Could not find address for server id 9af6b50a-1453-f014-44b6-a053beff6f9d" === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.900Z [WARN] TestAgent_ReloadConfigOutgoingRPCConfig: bootstrap = true: do not enable unless necessary === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.901Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.server.raft: votes: needed=1 === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.902Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.tlsutil: Update: version=1 === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.902Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.server.raft: vote granted: from=9af6b50a-1453-f014-44b6-a053beff6f9d term=2 tally=1 writer.go:29: 2021-01-29T19:33:55.904Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:55.905Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server.raft: entering leader state: leader="Node at 127.0.0.1:29753 [Leader]" writer.go:29: 2021-01-29T19:33:55.907Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:55.909Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: Started DNS server: address=127.0.0.1:29748 network=udp writer.go:29: 2021-01-29T19:33:55.910Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: Started DNS server: address=127.0.0.1:29748 network=tcp === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.911Z [WARN] TestAgent_ReloadConfigIncomingRPCConfig: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:55.913Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.tlsutil: Update: version=1 === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.913Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: Started HTTP server: address=127.0.0.1:29749 network=tcp === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.914Z [WARN] TestAgent_loadTokens: bootstrap = true: do not enable unless necessary === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.914Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.909Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server: Handled event for server in area: event=member-join server=Node-9af6b50a-1453-f014-44b6-a053beff6f9d.dc1 area=wan === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.915Z [DEBUG] TestAgent_loadTokens.tlsutil: Update: version=1 === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.909Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server: Adding LAN server: server="Node-9af6b50a-1453-f014-44b6-a053beff6f9d (Addr: tcp/127.0.0.1:29753) (DC: dc1)" === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.917Z [DEBUG] TestAgent_loadTokens.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.917Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server: New leader elected: payload=Node-9af6b50a-1453-f014-44b6-a053beff6f9d === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.904Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.918Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: started state syncer writer.go:29: 2021-01-29T19:33:55.922Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.tlsutil: IncomingRPCConfig: version=1 === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.924Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:078716fd-95c1-351a-eea0-0f43dcb702a7 Address:127.0.0.1:29765}]" === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.928Z [INFO] TestAgent_loadTokens.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:396ffabf-ba05-16de-38c2-d85e2bd33d29 Address:127.0.0.1:29771}]" writer.go:29: 2021-01-29T19:33:55.930Z [INFO] TestAgent_loadTokens.server.serf.wan: serf: EventMemberJoin: Node-396ffabf-ba05-16de-38c2-d85e2bd33d29.dc1 127.0.0.1 === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.922Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29753 writer.go:29: 2021-01-29T19:33:55.924Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: Synced node info === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.934Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server.serf.wan: serf: EventMemberJoin: Node-078716fd-95c1-351a-eea0-0f43dcb702a7.dc1 127.0.0.1 === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.934Z [INFO] TestAgent_loadTokens.server.raft: entering follower state: follower="Node at 127.0.0.1:29771 [Follower]" leader= === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.939Z [WARN] TestAgent_ReloadConfigTLSConfigFailure: bootstrap = true: do not enable unless necessary === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.938Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server.raft: entering follower state: follower="Node at 127.0.0.1:29765 [Follower]" leader= === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.935Z [INFO] TestAgent_loadTokens.server.serf.lan: serf: EventMemberJoin: Node-396ffabf-ba05-16de-38c2-d85e2bd33d29 127.0.0.1 writer.go:29: 2021-01-29T19:33:55.943Z [INFO] TestAgent_loadTokens.server: Adding LAN server: server="Node-396ffabf-ba05-16de-38c2-d85e2bd33d29 (Addr: tcp/127.0.0.1:29771) (DC: dc1)" === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.940Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.tlsutil: Update: version=1 === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.943Z [INFO] TestAgent_loadTokens.server: Handled event for server in area: event=member-join server=Node-396ffabf-ba05-16de-38c2-d85e2bd33d29.dc1 area=wan === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.946Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.tlsutil: IncomingRPCConfig: version=1 === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.946Z [INFO] TestAgent_loadTokens: Started DNS server: address=127.0.0.1:29766 network=udp === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.940Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server.serf.lan: serf: EventMemberJoin: Node-078716fd-95c1-351a-eea0-0f43dcb702a7 127.0.0.1 === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.941Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.951Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server: Adding LAN server: server="Node-078716fd-95c1-351a-eea0-0f43dcb702a7 (Addr: tcp/127.0.0.1:29765) (DC: dc1)" writer.go:29: 2021-01-29T19:33:55.951Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server: Handled event for server in area: event=member-join server=Node-078716fd-95c1-351a-eea0-0f43dcb702a7.dc1 area=wan === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.953Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: Requesting shutdown writer.go:29: 2021-01-29T19:33:55.954Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server: shutting down server writer.go:29: 2021-01-29T19:33:55.958Z [WARN] TestAgent_ReloadConfigTLSConfigFailure.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.955Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:db9243d3-b71c-0f31-56a0-4c2453a79dac Address:127.0.0.1:29741}]" === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.948Z [INFO] TestAgent_loadTokens: Started DNS server: address=127.0.0.1:29766 network=tcp === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.960Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: Started DNS server: address=127.0.0.1:29760 network=udp === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.965Z [INFO] TestAgent_loadTokens: Started HTTP server: address=127.0.0.1:29767 network=tcp writer.go:29: 2021-01-29T19:33:55.966Z [INFO] TestAgent_loadTokens: started state syncer === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.965Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server.raft: entering follower state: follower="Node at 127.0.0.1:29741 [Follower]" leader= === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.964Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: Started DNS server: address=127.0.0.1:29760 network=tcp === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.968Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:55.969Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.leader: started routine: routine="CA root pruning" === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.967Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server.serf.wan: serf: EventMemberJoin: Node-db9243d3-b71c-0f31-56a0-4c2453a79dac.dc1 127.0.0.1 === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.969Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.server: Skipping self join check for node since the cluster is too small: node=Node-9af6b50a-1453-f014-44b6-a053beff6f9d writer.go:29: 2021-01-29T19:33:55.970Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server: member joined, marking health alive: member=Node-9af6b50a-1453-f014-44b6-a053beff6f9d writer.go:29: 2021-01-29T19:33:55.970Z [WARN] TestAgent_ReloadConfigTLSConfigFailure.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.972Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server.serf.lan: serf: EventMemberJoin: Node-db9243d3-b71c-0f31-56a0-4c2453a79dac 127.0.0.1 === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.974Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:55.975Z [DEBUG] TestAgent_ReloadConfigTLSConfigFailure.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.970Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: Started HTTP server: address=127.0.0.1:29761 network=tcp === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.975Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server: Adding LAN server: server="Node-db9243d3-b71c-0f31-56a0-4c2453a79dac (Addr: tcp/127.0.0.1:29741) (DC: dc1)" writer.go:29: 2021-01-29T19:33:55.977Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: Started DNS server: address=127.0.0.1:29736 network=udp writer.go:29: 2021-01-29T19:33:55.979Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: Started DNS server: address=127.0.0.1:29736 network=tcp === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.976Z [INFO] TestAgent_ReloadConfigTLSConfigFailure.server.router.manager: shutting down === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.978Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: started state syncer === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.977Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: consul server down === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.977Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server: Handled event for server in area: event=member-join server=Node-db9243d3-b71c-0f31-56a0-4c2453a79dac.dc1 area=wan === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.982Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: shutdown complete === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.981Z [WARN] TestAgent_ReloadConfigIncomingRPCConfig.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.984Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: Stopping server: protocol=DNS address=127.0.0.1:29748 network=tcp writer.go:29: 2021-01-29T19:33:55.985Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: Stopping server: protocol=DNS address=127.0.0.1:29748 network=udp === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.981Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: Started HTTP server: address=127.0.0.1:29737 network=tcp === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:55.987Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: Stopping server: protocol=HTTP address=127.0.0.1:29749 network=tcp === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.985Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server.raft: entering candidate state: node="Node at 127.0.0.1:29765 [Candidate]" term=2 === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:55.987Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: started state syncer === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.990Z [WARN] TestAgent_loadTokens.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:55.993Z [INFO] TestAgent_loadTokens.server.raft: entering candidate state: node="Node at 127.0.0.1:29771 [Candidate]" term=2 === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.995Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:55.997Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.server.raft: vote granted: from=078716fd-95c1-351a-eea0-0f43dcb702a7 term=2 tally=1 === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.996Z [DEBUG] TestAgent_loadTokens.server.raft: votes: needed=1 === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:55.998Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server.raft: election won: tally=1 === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:55.994Z [DEBUG] TestAgent_loadTokens.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:33:55.998Z [DEBUG] TestAgent_loadTokens.server.raft: vote granted: from=396ffabf-ba05-16de-38c2-d85e2bd33d29 term=2 tally=1 === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:56.008Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server.raft: entering leader state: leader="Node at 127.0.0.1:29765 [Leader]" writer.go:29: 2021-01-29T19:33:56.010Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:56.011Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server: New leader elected: payload=Node-078716fd-95c1-351a-eea0-0f43dcb702a7 === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.010Z [INFO] TestAgent_loadTokens.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:56.012Z [INFO] TestAgent_loadTokens.server.raft: entering leader state: leader="Node at 127.0.0.1:29771 [Leader]" === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:56.017Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29765 === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.018Z [INFO] TestAgent_loadTokens.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:56.019Z [INFO] TestAgent_loadTokens.server: New leader elected: payload=Node-396ffabf-ba05-16de-38c2-d85e2bd33d29 === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:56.021Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.022Z [INFO] TestAgent_loadTokens.server: initializing acls === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:56.026Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:56.027Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.028Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.server: Skipping self join check for node since the cluster is too small: node=Node-078716fd-95c1-351a-eea0-0f43dcb702a7 === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:56.034Z [WARN] TestAgent_ReloadConfigOutgoingRPCConfig.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:56.035Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server.raft: entering candidate state: node="Node at 127.0.0.1:29741 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:56.037Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:56.037Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.server.raft: vote granted: from=db9243d3-b71c-0f31-56a0-4c2453a79dac term=2 tally=1 writer.go:29: 2021-01-29T19:33:56.038Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:56.039Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server.raft: entering leader state: leader="Node at 127.0.0.1:29741 [Leader]" writer.go:29: 2021-01-29T19:33:56.040Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:56.041Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server: New leader elected: payload=Node-db9243d3-b71c-0f31-56a0-4c2453a79dac === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.053Z [INFO] TestAgent_loadTokens.server: Created ACL 'global-management' policy === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:56.029Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server: member joined, marking health alive: member=Node-078716fd-95c1-351a-eea0-0f43dcb702a7 === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:56.055Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29741 === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.056Z [INFO] TestAgent_loadTokens.server: Created ACL anonymous token from configuration === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:56.058Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.058Z [INFO] TestAgent_loadTokens.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:56.062Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:56.064Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.065Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.server: Skipping self join check for node since the cluster is too small: node=Node-db9243d3-b71c-0f31-56a0-4c2453a79dac writer.go:29: 2021-01-29T19:33:56.067Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server: member joined, marking health alive: member=Node-db9243d3-b71c-0f31-56a0-4c2453a79dac === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.061Z [INFO] TestAgent_loadTokens.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:56.070Z [INFO] TestAgent_loadTokens.server.serf.lan: serf: EventMemberUpdate: Node-396ffabf-ba05-16de-38c2-d85e2bd33d29 writer.go:29: 2021-01-29T19:33:56.073Z [INFO] TestAgent_loadTokens.server: Updating LAN server: server="Node-396ffabf-ba05-16de-38c2-d85e2bd33d29 (Addr: tcp/127.0.0.1:29771) (DC: dc1)" writer.go:29: 2021-01-29T19:33:56.074Z [INFO] TestAgent_loadTokens.server.serf.wan: serf: EventMemberUpdate: Node-396ffabf-ba05-16de-38c2-d85e2bd33d29.dc1 writer.go:29: 2021-01-29T19:33:56.076Z [INFO] TestAgent_loadTokens.server: Handled event for server in area: event=member-update server=Node-396ffabf-ba05-16de-38c2-d85e2bd33d29.dc1 area=wan === RUN TestAgent_loadTokens/original-configuration === RUN TestAgent_loadTokens/updated-configuration === RUN TestAgent_loadTokens/persisted-tokens === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.100Z [WARN] TestAgent_loadTokens: "default" token present in both the configuration and persisted token store, using the persisted token writer.go:29: 2021-01-29T19:33:56.102Z [WARN] TestAgent_loadTokens: "agent" token present in both the configuration and persisted token store, using the persisted token writer.go:29: 2021-01-29T19:33:56.105Z [WARN] TestAgent_loadTokens: "agent_master" token present in both the configuration and persisted token store, using the persisted token writer.go:29: 2021-01-29T19:33:56.102Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:56.106Z [WARN] TestAgent_loadTokens: "replication" token present in both the configuration and persisted token store, using the persisted token === RUN TestAgent_loadTokens/persisted-tokens-override === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.112Z [INFO] TestAgent_loadTokens.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:56.114Z [INFO] TestAgent_loadTokens.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.114Z [WARN] TestAgent_loadTokens: "default" token present in both the configuration and persisted token store, using the persisted token writer.go:29: 2021-01-29T19:33:56.117Z [WARN] TestAgent_loadTokens: "agent" token present in both the configuration and persisted token store, using the persisted token writer.go:29: 2021-01-29T19:33:56.119Z [WARN] TestAgent_loadTokens: "agent_master" token present in both the configuration and persisted token store, using the persisted token writer.go:29: 2021-01-29T19:33:56.120Z [WARN] TestAgent_loadTokens: "replication" token present in both the configuration and persisted token store, using the persisted token writer.go:29: 2021-01-29T19:33:56.119Z [DEBUG] TestAgent_loadTokens.server: Skipping self join check for node since the cluster is too small: node=Node-396ffabf-ba05-16de-38c2-d85e2bd33d29 writer.go:29: 2021-01-29T19:33:56.123Z [INFO] TestAgent_loadTokens.server: member joined, marking health alive: member=Node-396ffabf-ba05-16de-38c2-d85e2bd33d29 writer.go:29: 2021-01-29T19:33:56.126Z [DEBUG] TestAgent_loadTokens.server: Skipping self join check for node since the cluster is too small: node=Node-396ffabf-ba05-16de-38c2-d85e2bd33d29 === RUN TestAgent_loadTokens/partial-persisted === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.133Z [WARN] TestAgent_loadTokens: "agent" token present in both the configuration and persisted token store, using the persisted token writer.go:29: 2021-01-29T19:33:56.135Z [WARN] TestAgent_loadTokens: "agent_master" token present in both the configuration and persisted token store, using the persisted token === RUN TestAgent_loadTokens/persistence-error-not-json === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.150Z [WARN] TestAgent_loadTokens: unable to load persisted tokens: error="failed to decode tokens file "/tmp/TestAgent_loadTokens-agent207392417/acl-tokens.json": invalid character '\x01' looking for beginning of value" === RUN TestAgent_loadTokens/persistence-error-wrong-top-level === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.157Z [WARN] TestAgent_loadTokens: unable to load persisted tokens: error="failed to decode tokens file "/tmp/TestAgent_loadTokens-agent207392417/acl-tokens.json": json: cannot unmarshal array into Go value of type agent.persistedTokens" writer.go:29: 2021-01-29T19:33:56.163Z [INFO] TestAgent_loadTokens: Requesting shutdown writer.go:29: 2021-01-29T19:33:56.165Z [INFO] TestAgent_loadTokens.server: shutting down server writer.go:29: 2021-01-29T19:33:56.169Z [DEBUG] TestAgent_loadTokens.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:56.170Z [DEBUG] TestAgent_loadTokens.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:56.169Z [ERROR] TestAgent_loadTokens.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:56.170Z [DEBUG] TestAgent_loadTokens.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:33:56.172Z [DEBUG] TestAgent_loadTokens.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.172Z [DEBUG] TestAgent_loadTokens.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:33:56.183Z [WARN] TestAgent_loadTokens.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:56.183Z [DEBUG] TestAgent_loadTokens.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.189Z [WARN] TestAgent_loadTokens.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:56.195Z [INFO] TestAgent_loadTokens.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:56.195Z [INFO] TestAgent_loadTokens: consul server down writer.go:29: 2021-01-29T19:33:56.196Z [INFO] TestAgent_loadTokens: shutdown complete writer.go:29: 2021-01-29T19:33:56.197Z [INFO] TestAgent_loadTokens: Stopping server: protocol=DNS address=127.0.0.1:29766 network=tcp writer.go:29: 2021-01-29T19:33:56.198Z [INFO] TestAgent_loadTokens: Stopping server: protocol=DNS address=127.0.0.1:29766 network=udp writer.go:29: 2021-01-29T19:33:56.199Z [INFO] TestAgent_loadTokens: Stopping server: protocol=HTTP address=127.0.0.1:29767 network=tcp === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:56.200Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:56.202Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: Synced node info === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:56.216Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.tlsutil: IncomingRPCConfig: version=1 writer.go:29: 2021-01-29T19:33:56.218Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.tlsutil: IncomingRPCConfig: version=1 writer.go:29: 2021-01-29T19:33:56.236Z [WARN] TestAgent_ReloadConfigIncomingRPCConfig: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:56.238Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.tlsutil: Update: version=2 writer.go:29: 2021-01-29T19:33:56.239Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.tlsutil: IncomingRPCConfig: version=2 writer.go:29: 2021-01-29T19:33:56.241Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: Requesting shutdown writer.go:29: 2021-01-29T19:33:56.243Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server: shutting down server writer.go:29: 2021-01-29T19:33:56.244Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.244Z [WARN] TestAgent_ReloadConfigIncomingRPCConfig.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:56.244Z [ERROR] TestAgent_ReloadConfigIncomingRPCConfig.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:56.244Z [DEBUG] TestAgent_ReloadConfigIncomingRPCConfig.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.246Z [WARN] TestAgent_ReloadConfigIncomingRPCConfig.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:56.253Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:56.253Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: consul server down writer.go:29: 2021-01-29T19:33:56.255Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: shutdown complete writer.go:29: 2021-01-29T19:33:56.255Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: Stopping server: protocol=DNS address=127.0.0.1:29760 network=tcp writer.go:29: 2021-01-29T19:33:56.256Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: Stopping server: protocol=DNS address=127.0.0.1:29760 network=udp writer.go:29: 2021-01-29T19:33:56.259Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: Stopping server: protocol=HTTP address=127.0.0.1:29761 network=tcp === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:56.408Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.tlsutil: OutgoingRPCConfig: version=1 writer.go:29: 2021-01-29T19:33:56.426Z [WARN] TestAgent_ReloadConfigOutgoingRPCConfig: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:56.429Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.tlsutil: Update: version=2 writer.go:29: 2021-01-29T19:33:56.432Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.tlsutil: OutgoingRPCConfig: version=2 writer.go:29: 2021-01-29T19:33:56.437Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: Requesting shutdown writer.go:29: 2021-01-29T19:33:56.439Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server: shutting down server writer.go:29: 2021-01-29T19:33:56.440Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.442Z [WARN] TestAgent_ReloadConfigOutgoingRPCConfig.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:56.443Z [DEBUG] TestAgent_ReloadConfigOutgoingRPCConfig.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.446Z [WARN] TestAgent_ReloadConfigOutgoingRPCConfig.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:56.450Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:56.451Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: consul server down writer.go:29: 2021-01-29T19:33:56.454Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: shutdown complete writer.go:29: 2021-01-29T19:33:56.455Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: Stopping server: protocol=DNS address=127.0.0.1:29736 network=tcp writer.go:29: 2021-01-29T19:33:56.457Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: Stopping server: protocol=DNS address=127.0.0.1:29736 network=udp writer.go:29: 2021-01-29T19:33:56.459Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: Stopping server: protocol=HTTP address=127.0.0.1:29737 network=tcp === CONT TestAgent_ReloadConfigTLSConfigFailure writer.go:29: 2021-01-29T19:33:56.488Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:56.490Z [INFO] TestAgent_ReloadConfigTLSConfigFailure: Endpoints down --- PASS: TestAgent_ReloadConfigTLSConfigFailure (0.99s) === CONT TestAgent_reloadWatchesHTTPS 2021-01-29T19:33:56.497Z [WARN] TestAgent_reloadWatchesHTTPS: bootstrap = true: do not enable unless necessary 2021-01-29T19:33:56.498Z [DEBUG] TestAgent_reloadWatchesHTTPS.tlsutil: Update: version=1 2021-01-29T19:33:56.498Z [DEBUG] TestAgent_reloadWatchesHTTPS.tlsutil: OutgoingRPCWrapper: version=1 2021-01-29T19:33:56.502Z [INFO] TestAgent_reloadWatchesHTTPS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7cad8e73-c7be-4e36-1bf2-00a9813e242e Address:127.0.0.1:29777}]" 2021-01-29T19:33:56.502Z [INFO] TestAgent_reloadWatchesHTTPS.server.raft: entering follower state: follower="Node at 127.0.0.1:29777 [Follower]" leader= 2021-01-29T19:33:56.504Z [INFO] TestAgent_reloadWatchesHTTPS.server.serf.wan: serf: EventMemberJoin: Node-7cad8e73-c7be-4e36-1bf2-00a9813e242e.dc1 127.0.0.1 2021-01-29T19:33:56.506Z [INFO] TestAgent_reloadWatchesHTTPS.server.serf.lan: serf: EventMemberJoin: Node-7cad8e73-c7be-4e36-1bf2-00a9813e242e 127.0.0.1 2021-01-29T19:33:56.507Z [INFO] TestAgent_reloadWatchesHTTPS.server: Adding LAN server: server="Node-7cad8e73-c7be-4e36-1bf2-00a9813e242e (Addr: tcp/127.0.0.1:29777) (DC: dc1)" 2021-01-29T19:33:56.508Z [INFO] TestAgent_reloadWatchesHTTPS.server: Handled event for server in area: event=member-join server=Node-7cad8e73-c7be-4e36-1bf2-00a9813e242e.dc1 area=wan 2021-01-29T19:33:56.510Z [INFO] TestAgent_reloadWatchesHTTPS: Started DNS server: address=127.0.0.1:29772 network=tcp 2021-01-29T19:33:56.511Z [INFO] TestAgent_reloadWatchesHTTPS: Started DNS server: address=127.0.0.1:29772 network=udp 2021-01-29T19:33:56.511Z [DEBUG] TestAgent_reloadWatchesHTTPS.tlsutil: IncomingHTTPSConfig: version=1 2021-01-29T19:33:56.512Z [INFO] TestAgent_reloadWatchesHTTPS: Started HTTPS server: address=127.0.0.1:29774 network=tcp 2021-01-29T19:33:56.512Z [INFO] TestAgent_reloadWatchesHTTPS: started state syncer 2021-01-29T19:33:56.544Z [WARN] TestAgent_reloadWatchesHTTPS.server.raft: heartbeat timeout reached, starting election: last-leader= 2021-01-29T19:33:56.544Z [INFO] TestAgent_reloadWatchesHTTPS.server.raft: entering candidate state: node="Node at 127.0.0.1:29777 [Candidate]" term=2 2021-01-29T19:33:56.545Z [DEBUG] TestAgent_reloadWatchesHTTPS.server.raft: votes: needed=1 2021-01-29T19:33:56.545Z [DEBUG] TestAgent_reloadWatchesHTTPS.server.raft: vote granted: from=7cad8e73-c7be-4e36-1bf2-00a9813e242e term=2 tally=1 2021-01-29T19:33:56.545Z [INFO] TestAgent_reloadWatchesHTTPS.server.raft: election won: tally=1 2021-01-29T19:33:56.545Z [INFO] TestAgent_reloadWatchesHTTPS.server.raft: entering leader state: leader="Node at 127.0.0.1:29777 [Leader]" 2021-01-29T19:33:56.545Z [INFO] TestAgent_reloadWatchesHTTPS.server: cluster leadership acquired 2021-01-29T19:33:56.546Z [INFO] TestAgent_reloadWatchesHTTPS.server: New leader elected: payload=Node-7cad8e73-c7be-4e36-1bf2-00a9813e242e 2021-01-29T19:33:56.547Z [DEBUG] TestAgent_reloadWatchesHTTPS.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29777 2021-01-29T19:33:56.551Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true 2021-01-29T19:33:56.557Z [INFO] TestAgent_reloadWatchesHTTPS.server.connect: initialized primary datacenter CA with provider: provider=consul 2021-01-29T19:33:56.557Z [INFO] TestAgent_reloadWatchesHTTPS.leader: started routine: routine="CA root pruning" 2021-01-29T19:33:56.557Z [DEBUG] TestAgent_reloadWatchesHTTPS.server: Skipping self join check for node since the cluster is too small: node=Node-7cad8e73-c7be-4e36-1bf2-00a9813e242e 2021-01-29T19:33:56.557Z [INFO] TestAgent_reloadWatchesHTTPS.server: member joined, marking health alive: member=Node-7cad8e73-c7be-4e36-1bf2-00a9813e242e 2021-01-29T19:33:56.612Z [INFO] TestAgent_reloadWatchesHTTPS: Requesting shutdown 2021-01-29T19:33:56.612Z [INFO] TestAgent_reloadWatchesHTTPS.server: shutting down server 2021-01-29T19:33:56.612Z [DEBUG] TestAgent_reloadWatchesHTTPS.leader: stopping routine: routine="CA root pruning" 2021-01-29T19:33:56.612Z [WARN] TestAgent_reloadWatchesHTTPS.server.serf.lan: serf: Shutdown without a Leave 2021-01-29T19:33:56.617Z [ERROR] TestAgent_reloadWatchesHTTPS.anti_entropy: failed to sync remote state: error="No cluster leader" 2021-01-29T19:33:56.617Z [DEBUG] TestAgent_reloadWatchesHTTPS.leader: stopped routine: routine="CA root pruning" 2021-01-29T19:33:56.617Z [DEBUG] TestAgent_reloadWatchesHTTPS.tlsutil: IncomingHTTPSConfig: version=1 2021/01/29 19:33:56 http: TLS handshake error from 127.0.0.1:50816: tls: no certificates configured 2021-01-29T19:33:56.619Z [ERROR] watch.watch: Watch errored: type=key error="Get "https://127.0.0.1:29774/v1/kv/asdf": remote error: tls: unrecognized name" retry=5s 2021-01-29T19:33:56.631Z [WARN] TestAgent_reloadWatchesHTTPS.server.serf.wan: serf: Shutdown without a Leave 2021-01-29T19:33:56.632Z [INFO] TestAgent_reloadWatchesHTTPS.server.router.manager: shutting down 2021-01-29T19:33:56.633Z [INFO] TestAgent_reloadWatchesHTTPS: consul server down 2021-01-29T19:33:56.633Z [INFO] TestAgent_reloadWatchesHTTPS: shutdown complete 2021-01-29T19:33:56.633Z [INFO] TestAgent_reloadWatchesHTTPS: Stopping server: protocol=DNS address=127.0.0.1:29772 network=tcp 2021-01-29T19:33:56.633Z [INFO] TestAgent_reloadWatchesHTTPS: Stopping server: protocol=DNS address=127.0.0.1:29772 network=udp 2021-01-29T19:33:56.633Z [INFO] TestAgent_reloadWatchesHTTPS: Stopping server: protocol=HTTPS address=127.0.0.1:29774 network=tcp === CONT TestAgent_loadTokens writer.go:29: 2021-01-29T19:33:56.700Z [INFO] TestAgent_loadTokens: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:56.701Z [INFO] TestAgent_loadTokens: Endpoints down --- PASS: TestAgent_loadTokens (0.99s) --- PASS: TestAgent_loadTokens/original-configuration (0.00s) --- PASS: TestAgent_loadTokens/updated-configuration (0.00s) --- PASS: TestAgent_loadTokens/persisted-tokens (0.02s) --- PASS: TestAgent_loadTokens/persisted-tokens-override (0.02s) --- PASS: TestAgent_loadTokens/partial-persisted (0.01s) --- PASS: TestAgent_loadTokens/persistence-error-not-json (0.01s) --- PASS: TestAgent_loadTokens/persistence-error-wrong-top-level (0.01s) === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.714Z [WARN] TestAgent_reloadWatches: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:56.716Z [DEBUG] TestAgent_reloadWatches.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:56.717Z [DEBUG] TestAgent_reloadWatches.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:56.732Z [INFO] TestAgent_reloadWatches.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:77313d16-54f9-8b3f-0a4d-763fb0deeca1 Address:127.0.0.1:29783}]" writer.go:29: 2021-01-29T19:33:56.739Z [INFO] TestAgent_reloadWatches.server.raft: entering follower state: follower="Node at 127.0.0.1:29783 [Follower]" leader= writer.go:29: 2021-01-29T19:33:56.741Z [INFO] TestAgent_reloadWatches.server.serf.wan: serf: EventMemberJoin: Node-77313d16-54f9-8b3f-0a4d-763fb0deeca1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:56.750Z [INFO] TestAgent_reloadWatches.server.serf.lan: serf: EventMemberJoin: Node-77313d16-54f9-8b3f-0a4d-763fb0deeca1 127.0.0.1 === CONT TestAgent_ReloadConfigIncomingRPCConfig writer.go:29: 2021-01-29T19:33:56.760Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:56.761Z [INFO] TestAgent_ReloadConfigIncomingRPCConfig: Endpoints down --- PASS: TestAgent_ReloadConfigIncomingRPCConfig (1.18s) === CONT TestAgent_GetCoordinate === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.764Z [INFO] TestAgent_reloadWatches.server: Adding LAN server: server="Node-77313d16-54f9-8b3f-0a4d-763fb0deeca1 (Addr: tcp/127.0.0.1:29783) (DC: dc1)" writer.go:29: 2021-01-29T19:33:56.764Z [INFO] TestAgent_reloadWatches: Started DNS server: address=127.0.0.1:29778 network=udp writer.go:29: 2021-01-29T19:33:56.765Z [INFO] TestAgent_reloadWatches.server: Handled event for server in area: event=member-join server=Node-77313d16-54f9-8b3f-0a4d-763fb0deeca1.dc1 area=wan writer.go:29: 2021-01-29T19:33:56.767Z [INFO] TestAgent_reloadWatches: Started DNS server: address=127.0.0.1:29778 network=tcp writer.go:29: 2021-01-29T19:33:56.769Z [INFO] TestAgent_reloadWatches: Started HTTP server: address=127.0.0.1:29779 network=tcp writer.go:29: 2021-01-29T19:33:56.769Z [INFO] TestAgent_reloadWatches: started state syncer === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.777Z [WARN] TestAgent_GetCoordinate: bootstrap = true: do not enable unless necessary === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.779Z [WARN] TestAgent_reloadWatches.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.778Z [DEBUG] TestAgent_GetCoordinate.tlsutil: Update: version=1 === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.780Z [INFO] TestAgent_reloadWatches.server.raft: entering candidate state: node="Node at 127.0.0.1:29783 [Candidate]" term=2 === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.782Z [DEBUG] TestAgent_GetCoordinate.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.783Z [DEBUG] TestAgent_reloadWatches.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:56.784Z [DEBUG] TestAgent_reloadWatches.server.raft: vote granted: from=77313d16-54f9-8b3f-0a4d-763fb0deeca1 term=2 tally=1 writer.go:29: 2021-01-29T19:33:56.785Z [INFO] TestAgent_reloadWatches.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:56.786Z [INFO] TestAgent_reloadWatches.server.raft: entering leader state: leader="Node at 127.0.0.1:29783 [Leader]" writer.go:29: 2021-01-29T19:33:56.797Z [INFO] TestAgent_reloadWatches.server: cluster leadership acquired === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.798Z [INFO] TestAgent_GetCoordinate.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:514d526f-d5a4-8fc3-4f13-8ed1ab0496eb Address:127.0.0.1:29789}]" === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.797Z [INFO] TestAgent_reloadWatches.server: New leader elected: payload=Node-77313d16-54f9-8b3f-0a4d-763fb0deeca1 === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.800Z [INFO] TestAgent_GetCoordinate.server.raft: entering follower state: follower="Node at 127.0.0.1:29789 [Follower]" leader= writer.go:29: 2021-01-29T19:33:56.802Z [INFO] TestAgent_GetCoordinate.server.serf.wan: serf: EventMemberJoin: Node-514d526f-d5a4-8fc3-4f13-8ed1ab0496eb.dc1 127.0.0.1 === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.798Z [DEBUG] TestAgent_reloadWatches.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29783 === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.805Z [INFO] TestAgent_GetCoordinate.server.serf.lan: serf: EventMemberJoin: Node-514d526f-d5a4-8fc3-4f13-8ed1ab0496eb 127.0.0.1 writer.go:29: 2021-01-29T19:33:56.807Z [INFO] TestAgent_GetCoordinate.server: Handled event for server in area: event=member-join server=Node-514d526f-d5a4-8fc3-4f13-8ed1ab0496eb.dc1 area=wan writer.go:29: 2021-01-29T19:33:56.808Z [INFO] TestAgent_GetCoordinate.server: Adding LAN server: server="Node-514d526f-d5a4-8fc3-4f13-8ed1ab0496eb (Addr: tcp/127.0.0.1:29789) (DC: dc1)" === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.810Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.809Z [INFO] TestAgent_GetCoordinate: Started DNS server: address=127.0.0.1:29784 network=udp writer.go:29: 2021-01-29T19:33:56.811Z [INFO] TestAgent_GetCoordinate: Started DNS server: address=127.0.0.1:29784 network=tcp writer.go:29: 2021-01-29T19:33:56.814Z [INFO] TestAgent_GetCoordinate: Started HTTP server: address=127.0.0.1:29785 network=tcp writer.go:29: 2021-01-29T19:33:56.815Z [INFO] TestAgent_GetCoordinate: started state syncer === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.817Z [INFO] TestAgent_reloadWatches.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:56.818Z [INFO] TestAgent_reloadWatches.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.819Z [DEBUG] TestAgent_reloadWatches.server: Skipping self join check for node since the cluster is too small: node=Node-77313d16-54f9-8b3f-0a4d-763fb0deeca1 writer.go:29: 2021-01-29T19:33:56.820Z [INFO] TestAgent_reloadWatches.server: member joined, marking health alive: member=Node-77313d16-54f9-8b3f-0a4d-763fb0deeca1 writer.go:29: 2021-01-29T19:33:56.859Z [INFO] TestAgent_reloadWatches: Requesting shutdown writer.go:29: 2021-01-29T19:33:56.861Z [INFO] TestAgent_reloadWatches.server: shutting down server writer.go:29: 2021-01-29T19:33:56.862Z [DEBUG] TestAgent_reloadWatches.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.863Z [WARN] TestAgent_reloadWatches.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:56.862Z [ERROR] TestAgent_reloadWatches.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:56.863Z [DEBUG] TestAgent_reloadWatches.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.865Z [WARN] TestAgent_reloadWatches.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:56.869Z [INFO] TestAgent_reloadWatches.server.router.manager: shutting down === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.870Z [WARN] TestAgent_GetCoordinate.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.869Z [INFO] TestAgent_reloadWatches: consul server down === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.871Z [INFO] TestAgent_GetCoordinate.server.raft: entering candidate state: node="Node at 127.0.0.1:29789 [Candidate]" term=2 === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.872Z [INFO] TestAgent_reloadWatches: shutdown complete === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.873Z [DEBUG] TestAgent_GetCoordinate.server.raft: votes: needed=1 === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.873Z [INFO] TestAgent_reloadWatches: Stopping server: protocol=DNS address=127.0.0.1:29778 network=tcp === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.874Z [DEBUG] TestAgent_GetCoordinate.server.raft: vote granted: from=514d526f-d5a4-8fc3-4f13-8ed1ab0496eb term=2 tally=1 writer.go:29: 2021-01-29T19:33:56.874Z [INFO] TestAgent_GetCoordinate.server.raft: election won: tally=1 === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.874Z [INFO] TestAgent_reloadWatches: Stopping server: protocol=DNS address=127.0.0.1:29778 network=udp === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.875Z [INFO] TestAgent_GetCoordinate.server.raft: entering leader state: leader="Node at 127.0.0.1:29789 [Leader]" writer.go:29: 2021-01-29T19:33:56.876Z [INFO] TestAgent_GetCoordinate.server: cluster leadership acquired === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:56.876Z [INFO] TestAgent_reloadWatches: Stopping server: protocol=HTTP address=127.0.0.1:29779 network=tcp === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:56.877Z [INFO] TestAgent_GetCoordinate.server: New leader elected: payload=Node-514d526f-d5a4-8fc3-4f13-8ed1ab0496eb writer.go:29: 2021-01-29T19:33:56.879Z [DEBUG] TestAgent_GetCoordinate.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29789 writer.go:29: 2021-01-29T19:33:56.884Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:56.890Z [INFO] TestAgent_GetCoordinate.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:56.891Z [INFO] TestAgent_GetCoordinate.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.892Z [DEBUG] TestAgent_GetCoordinate.server: Skipping self join check for node since the cluster is too small: node=Node-514d526f-d5a4-8fc3-4f13-8ed1ab0496eb writer.go:29: 2021-01-29T19:33:56.893Z [INFO] TestAgent_GetCoordinate.server: member joined, marking health alive: member=Node-514d526f-d5a4-8fc3-4f13-8ed1ab0496eb writer.go:29: 2021-01-29T19:33:56.920Z [INFO] TestAgent_GetCoordinate: Requesting shutdown writer.go:29: 2021-01-29T19:33:56.922Z [INFO] TestAgent_GetCoordinate.server: shutting down server writer.go:29: 2021-01-29T19:33:56.922Z [DEBUG] TestAgent_GetCoordinate.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.923Z [WARN] TestAgent_GetCoordinate.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:56.924Z [ERROR] TestAgent_GetCoordinate.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:56.925Z [DEBUG] TestAgent_GetCoordinate.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.927Z [WARN] TestAgent_GetCoordinate.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:56.929Z [INFO] TestAgent_GetCoordinate.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:56.930Z [INFO] TestAgent_GetCoordinate: consul server down writer.go:29: 2021-01-29T19:33:56.931Z [INFO] TestAgent_GetCoordinate: shutdown complete writer.go:29: 2021-01-29T19:33:56.932Z [INFO] TestAgent_GetCoordinate: Stopping server: protocol=DNS address=127.0.0.1:29784 network=tcp writer.go:29: 2021-01-29T19:33:56.933Z [INFO] TestAgent_GetCoordinate: Stopping server: protocol=DNS address=127.0.0.1:29784 network=udp writer.go:29: 2021-01-29T19:33:56.934Z [INFO] TestAgent_GetCoordinate: Stopping server: protocol=HTTP address=127.0.0.1:29785 network=tcp === CONT TestAgent_ReloadConfigOutgoingRPCConfig writer.go:29: 2021-01-29T19:33:56.961Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:56.963Z [INFO] TestAgent_ReloadConfigOutgoingRPCConfig: Endpoints down --- PASS: TestAgent_ReloadConfigOutgoingRPCConfig (1.35s) === CONT TestAgent_purgeCheckState writer.go:29: 2021-01-29T19:33:56.972Z [WARN] TestAgent_purgeCheckState: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:56.973Z [DEBUG] TestAgent_purgeCheckState.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:56.974Z [DEBUG] TestAgent_purgeCheckState.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:56.984Z [INFO] TestAgent_purgeCheckState.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0d512f03-838e-671e-b4ae-7fdbac9c28da Address:127.0.0.1:29801}]" writer.go:29: 2021-01-29T19:33:56.986Z [INFO] TestAgent_purgeCheckState.server.raft: entering follower state: follower="Node at 127.0.0.1:29801 [Follower]" leader= writer.go:29: 2021-01-29T19:33:56.986Z [INFO] TestAgent_purgeCheckState.server.serf.wan: serf: EventMemberJoin: Node-0d512f03-838e-671e-b4ae-7fdbac9c28da.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:56.992Z [INFO] TestAgent_purgeCheckState.server.serf.lan: serf: EventMemberJoin: Node-0d512f03-838e-671e-b4ae-7fdbac9c28da 127.0.0.1 writer.go:29: 2021-01-29T19:33:56.996Z [INFO] TestAgent_purgeCheckState.server: Adding LAN server: server="Node-0d512f03-838e-671e-b4ae-7fdbac9c28da (Addr: tcp/127.0.0.1:29801) (DC: dc1)" writer.go:29: 2021-01-29T19:33:56.997Z [INFO] TestAgent_purgeCheckState.server: Handled event for server in area: event=member-join server=Node-0d512f03-838e-671e-b4ae-7fdbac9c28da.dc1 area=wan writer.go:29: 2021-01-29T19:33:57.000Z [INFO] TestAgent_purgeCheckState: Started DNS server: address=127.0.0.1:29796 network=tcp writer.go:29: 2021-01-29T19:33:57.002Z [INFO] TestAgent_purgeCheckState: Started DNS server: address=127.0.0.1:29796 network=udp writer.go:29: 2021-01-29T19:33:57.004Z [INFO] TestAgent_purgeCheckState: Started HTTP server: address=127.0.0.1:29797 network=tcp writer.go:29: 2021-01-29T19:33:57.005Z [INFO] TestAgent_purgeCheckState: started state syncer writer.go:29: 2021-01-29T19:33:57.032Z [WARN] TestAgent_purgeCheckState.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:57.033Z [INFO] TestAgent_purgeCheckState.server.raft: entering candidate state: node="Node at 127.0.0.1:29801 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:57.036Z [DEBUG] TestAgent_purgeCheckState.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:57.037Z [DEBUG] TestAgent_purgeCheckState.server.raft: vote granted: from=0d512f03-838e-671e-b4ae-7fdbac9c28da term=2 tally=1 writer.go:29: 2021-01-29T19:33:57.037Z [INFO] TestAgent_purgeCheckState.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:57.038Z [INFO] TestAgent_purgeCheckState.server.raft: entering leader state: leader="Node at 127.0.0.1:29801 [Leader]" writer.go:29: 2021-01-29T19:33:57.040Z [INFO] TestAgent_purgeCheckState.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:57.041Z [INFO] TestAgent_purgeCheckState.server: New leader elected: payload=Node-0d512f03-838e-671e-b4ae-7fdbac9c28da writer.go:29: 2021-01-29T19:33:57.052Z [DEBUG] TestAgent_purgeCheckState.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29801 writer.go:29: 2021-01-29T19:33:57.071Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:57.084Z [INFO] TestAgent_purgeCheckState.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:57.085Z [INFO] TestAgent_purgeCheckState.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.086Z [DEBUG] TestAgent_purgeCheckState.server: Skipping self join check for node since the cluster is too small: node=Node-0d512f03-838e-671e-b4ae-7fdbac9c28da writer.go:29: 2021-01-29T19:33:57.087Z [INFO] TestAgent_purgeCheckState.server: member joined, marking health alive: member=Node-0d512f03-838e-671e-b4ae-7fdbac9c28da 2021-01-29T19:33:57.134Z [INFO] TestAgent_reloadWatchesHTTPS: Waiting for endpoints to shut down 2021-01-29T19:33:57.134Z [INFO] TestAgent_reloadWatchesHTTPS: Endpoints down --- PASS: TestAgent_reloadWatchesHTTPS (0.64s) === CONT TestAgent_loadCheckState === CONT TestAgent_purgeCheckState writer.go:29: 2021-01-29T19:33:57.200Z [INFO] TestAgent_purgeCheckState: Requesting shutdown writer.go:29: 2021-01-29T19:33:57.203Z [INFO] TestAgent_purgeCheckState.server: shutting down server writer.go:29: 2021-01-29T19:33:57.205Z [DEBUG] TestAgent_purgeCheckState.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.207Z [WARN] TestAgent_purgeCheckState.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_loadCheckState writer.go:29: 2021-01-29T19:33:57.198Z [WARN] TestAgent_loadCheckState: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:57.212Z [DEBUG] TestAgent_loadCheckState.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:57.218Z [DEBUG] TestAgent_loadCheckState.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_purgeCheckState writer.go:29: 2021-01-29T19:33:57.211Z [ERROR] TestAgent_purgeCheckState.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:57.211Z [DEBUG] TestAgent_purgeCheckState.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.221Z [WARN] TestAgent_purgeCheckState.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.242Z [INFO] TestAgent_purgeCheckState.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:57.244Z [INFO] TestAgent_purgeCheckState: consul server down writer.go:29: 2021-01-29T19:33:57.245Z [INFO] TestAgent_purgeCheckState: shutdown complete writer.go:29: 2021-01-29T19:33:57.247Z [INFO] TestAgent_purgeCheckState: Stopping server: protocol=DNS address=127.0.0.1:29796 network=tcp writer.go:29: 2021-01-29T19:33:57.248Z [INFO] TestAgent_purgeCheckState: Stopping server: protocol=DNS address=127.0.0.1:29796 network=udp writer.go:29: 2021-01-29T19:33:57.250Z [INFO] TestAgent_purgeCheckState: Stopping server: protocol=HTTP address=127.0.0.1:29797 network=tcp === CONT TestAgent_loadCheckState writer.go:29: 2021-01-29T19:33:57.293Z [INFO] TestAgent_loadCheckState.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bd1ca531-0f6c-08fe-8ecb-930cd98bac52 Address:127.0.0.1:29795}]" writer.go:29: 2021-01-29T19:33:57.296Z [INFO] TestAgent_loadCheckState.server.serf.wan: serf: EventMemberJoin: Node-bd1ca531-0f6c-08fe-8ecb-930cd98bac52.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.299Z [INFO] TestAgent_loadCheckState.server.raft: entering follower state: follower="Node at 127.0.0.1:29795 [Follower]" leader= writer.go:29: 2021-01-29T19:33:57.300Z [INFO] TestAgent_loadCheckState.server.serf.lan: serf: EventMemberJoin: Node-bd1ca531-0f6c-08fe-8ecb-930cd98bac52 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.315Z [INFO] TestAgent_loadCheckState.server: Handled event for server in area: event=member-join server=Node-bd1ca531-0f6c-08fe-8ecb-930cd98bac52.dc1 area=wan writer.go:29: 2021-01-29T19:33:57.317Z [INFO] TestAgent_loadCheckState.server: Adding LAN server: server="Node-bd1ca531-0f6c-08fe-8ecb-930cd98bac52 (Addr: tcp/127.0.0.1:29795) (DC: dc1)" writer.go:29: 2021-01-29T19:33:57.320Z [INFO] TestAgent_loadCheckState: Started DNS server: address=127.0.0.1:29790 network=udp writer.go:29: 2021-01-29T19:33:57.323Z [INFO] TestAgent_loadCheckState: Started DNS server: address=127.0.0.1:29790 network=tcp writer.go:29: 2021-01-29T19:33:57.331Z [INFO] TestAgent_loadCheckState: Started HTTP server: address=127.0.0.1:29791 network=tcp writer.go:29: 2021-01-29T19:33:57.335Z [INFO] TestAgent_loadCheckState: started state syncer writer.go:29: 2021-01-29T19:33:57.361Z [WARN] TestAgent_loadCheckState.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:57.365Z [INFO] TestAgent_loadCheckState.server.raft: entering candidate state: node="Node at 127.0.0.1:29795 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:57.380Z [DEBUG] TestAgent_loadCheckState.server.raft: votes: needed=1 === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:57.377Z [INFO] TestAgent_reloadWatches: Waiting for endpoints to shut down === CONT TestAgent_loadCheckState writer.go:29: 2021-01-29T19:33:57.380Z [DEBUG] TestAgent_loadCheckState.server.raft: vote granted: from=bd1ca531-0f6c-08fe-8ecb-930cd98bac52 term=2 tally=1 writer.go:29: 2021-01-29T19:33:57.381Z [INFO] TestAgent_loadCheckState.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:57.382Z [INFO] TestAgent_loadCheckState.server.raft: entering leader state: leader="Node at 127.0.0.1:29795 [Leader]" === CONT TestAgent_reloadWatches writer.go:29: 2021-01-29T19:33:57.381Z [INFO] TestAgent_reloadWatches: Endpoints down --- PASS: TestAgent_reloadWatches (0.68s) === CONT TestAgent_persistCheckState === CONT TestAgent_loadCheckState writer.go:29: 2021-01-29T19:33:57.383Z [INFO] TestAgent_loadCheckState.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:57.386Z [INFO] TestAgent_loadCheckState.server: New leader elected: payload=Node-bd1ca531-0f6c-08fe-8ecb-930cd98bac52 writer.go:29: 2021-01-29T19:33:57.390Z [DEBUG] TestAgent_loadCheckState.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29795 writer.go:29: 2021-01-29T19:33:57.404Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:57.423Z [INFO] TestAgent_loadCheckState.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:57.426Z [INFO] TestAgent_loadCheckState.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.429Z [DEBUG] TestAgent_loadCheckState.server: Skipping self join check for node since the cluster is too small: node=Node-bd1ca531-0f6c-08fe-8ecb-930cd98bac52 writer.go:29: 2021-01-29T19:33:57.433Z [INFO] TestAgent_loadCheckState.server: member joined, marking health alive: member=Node-bd1ca531-0f6c-08fe-8ecb-930cd98bac52 === CONT TestAgent_persistCheckState writer.go:29: 2021-01-29T19:33:57.434Z [WARN] TestAgent_persistCheckState: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:57.436Z [DEBUG] TestAgent_persistCheckState.tlsutil: Update: version=1 === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:57.435Z [INFO] TestAgent_GetCoordinate: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:57.439Z [INFO] TestAgent_GetCoordinate: Endpoints down === CONT TestAgent_persistCheckState writer.go:29: 2021-01-29T19:33:57.439Z [DEBUG] TestAgent_persistCheckState.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:57.452Z [INFO] TestAgent_persistCheckState.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:85bbc49f-45ce-f93b-6330-d98f019b0069 Address:127.0.0.1:29813}]" writer.go:29: 2021-01-29T19:33:57.456Z [INFO] TestAgent_persistCheckState.server.serf.wan: serf: EventMemberJoin: Node-85bbc49f-45ce-f93b-6330-d98f019b0069.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.461Z [INFO] TestAgent_persistCheckState.server.serf.lan: serf: EventMemberJoin: Node-85bbc49f-45ce-f93b-6330-d98f019b0069 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.472Z [INFO] TestAgent_persistCheckState: Started DNS server: address=127.0.0.1:29808 network=udp === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:57.475Z [WARN] TestAgent_GetCoordinate: bootstrap = true: do not enable unless necessary === CONT TestAgent_persistCheckState writer.go:29: 2021-01-29T19:33:57.464Z [INFO] TestAgent_persistCheckState.server.raft: entering follower state: follower="Node at 127.0.0.1:29813 [Follower]" leader= writer.go:29: 2021-01-29T19:33:57.477Z [INFO] TestAgent_persistCheckState.server: Handled event for server in area: event=member-join server=Node-85bbc49f-45ce-f93b-6330-d98f019b0069.dc1 area=wan writer.go:29: 2021-01-29T19:33:57.496Z [INFO] TestAgent_persistCheckState: Started DNS server: address=127.0.0.1:29808 network=tcp === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:57.505Z [DEBUG] TestAgent_GetCoordinate.tlsutil: Update: version=1 === CONT TestAgent_persistCheckState writer.go:29: 2021-01-29T19:33:57.496Z [INFO] TestAgent_persistCheckState.server: Adding LAN server: server="Node-85bbc49f-45ce-f93b-6330-d98f019b0069 (Addr: tcp/127.0.0.1:29813) (DC: dc1)" writer.go:29: 2021-01-29T19:33:57.511Z [INFO] TestAgent_persistCheckState: Started HTTP server: address=127.0.0.1:29809 network=tcp writer.go:29: 2021-01-29T19:33:57.513Z [INFO] TestAgent_persistCheckState: started state syncer === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:57.510Z [DEBUG] TestAgent_GetCoordinate.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:57.532Z [INFO] TestAgent_GetCoordinate.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:92ad6b6e-3c13-7c17-01a7-94fd6580b221 Address:127.0.0.1:29807}]" writer.go:29: 2021-01-29T19:33:57.535Z [INFO] TestAgent_GetCoordinate.server.raft: entering follower state: follower="Node at 127.0.0.1:29807 [Follower]" leader= writer.go:29: 2021-01-29T19:33:57.536Z [INFO] TestAgent_GetCoordinate.server.serf.wan: serf: EventMemberJoin: Node-92ad6b6e-3c13-7c17-01a7-94fd6580b221.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.541Z [INFO] TestAgent_GetCoordinate.server.serf.lan: serf: EventMemberJoin: Node-92ad6b6e-3c13-7c17-01a7-94fd6580b221 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.546Z [INFO] TestAgent_GetCoordinate.server: Handled event for server in area: event=member-join server=Node-92ad6b6e-3c13-7c17-01a7-94fd6580b221.dc1 area=wan writer.go:29: 2021-01-29T19:33:57.548Z [INFO] TestAgent_GetCoordinate.server: Adding LAN server: server="Node-92ad6b6e-3c13-7c17-01a7-94fd6580b221 (Addr: tcp/127.0.0.1:29807) (DC: dc1)" writer.go:29: 2021-01-29T19:33:57.550Z [INFO] TestAgent_GetCoordinate: Started DNS server: address=127.0.0.1:29802 network=tcp === CONT TestAgent_persistCheckState writer.go:29: 2021-01-29T19:33:57.552Z [WARN] TestAgent_persistCheckState.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:57.554Z [INFO] TestAgent_persistCheckState.server.raft: entering candidate state: node="Node at 127.0.0.1:29813 [Candidate]" term=2 === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:57.554Z [INFO] TestAgent_GetCoordinate: Started DNS server: address=127.0.0.1:29802 network=udp writer.go:29: 2021-01-29T19:33:57.557Z [INFO] TestAgent_GetCoordinate: Started HTTP server: address=127.0.0.1:29803 network=tcp writer.go:29: 2021-01-29T19:33:57.558Z [INFO] TestAgent_GetCoordinate: started state syncer === CONT TestAgent_persistCheckState writer.go:29: 2021-01-29T19:33:57.561Z [DEBUG] TestAgent_persistCheckState.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:57.563Z [DEBUG] TestAgent_persistCheckState.server.raft: vote granted: from=85bbc49f-45ce-f93b-6330-d98f019b0069 term=2 tally=1 writer.go:29: 2021-01-29T19:33:57.564Z [INFO] TestAgent_persistCheckState.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:57.566Z [INFO] TestAgent_persistCheckState.server.raft: entering leader state: leader="Node at 127.0.0.1:29813 [Leader]" writer.go:29: 2021-01-29T19:33:57.568Z [INFO] TestAgent_persistCheckState.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:57.568Z [INFO] TestAgent_persistCheckState.server: New leader elected: payload=Node-85bbc49f-45ce-f93b-6330-d98f019b0069 writer.go:29: 2021-01-29T19:33:57.569Z [INFO] TestAgent_persistCheckState: Requesting shutdown writer.go:29: 2021-01-29T19:33:57.570Z [DEBUG] TestAgent_persistCheckState.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29813 writer.go:29: 2021-01-29T19:33:57.573Z [INFO] TestAgent_persistCheckState.server: shutting down server writer.go:29: 2021-01-29T19:33:57.577Z [WARN] TestAgent_persistCheckState.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.577Z [ERROR] TestAgent_persistCheckState.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:57.578Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:57.580Z [WARN] TestAgent_persistCheckState.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.586Z [INFO] TestAgent_persistCheckState.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:57.586Z [ERROR] TestAgent_persistCheckState.server.connect: Raft apply failed: error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:57.589Z [ERROR] TestAgent_persistCheckState.server: failed to establish leadership: error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:57.587Z [INFO] TestAgent_persistCheckState: consul server down writer.go:29: 2021-01-29T19:33:57.591Z [INFO] TestAgent_persistCheckState: shutdown complete writer.go:29: 2021-01-29T19:33:57.593Z [INFO] TestAgent_persistCheckState: Stopping server: protocol=DNS address=127.0.0.1:29808 network=tcp writer.go:29: 2021-01-29T19:33:57.591Z [ERROR] TestAgent_persistCheckState.server: failed to transfer leadership attempt, will retry: attempt=0 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:57.595Z [INFO] TestAgent_persistCheckState: Stopping server: protocol=DNS address=127.0.0.1:29808 network=udp writer.go:29: 2021-01-29T19:33:57.598Z [INFO] TestAgent_persistCheckState: Stopping server: protocol=HTTP address=127.0.0.1:29809 network=tcp === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:57.607Z [WARN] TestAgent_GetCoordinate.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:57.608Z [INFO] TestAgent_GetCoordinate.server.raft: entering candidate state: node="Node at 127.0.0.1:29807 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:57.609Z [DEBUG] TestAgent_GetCoordinate.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:57.610Z [DEBUG] TestAgent_GetCoordinate.server.raft: vote granted: from=92ad6b6e-3c13-7c17-01a7-94fd6580b221 term=2 tally=1 writer.go:29: 2021-01-29T19:33:57.611Z [INFO] TestAgent_GetCoordinate.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:57.612Z [INFO] TestAgent_GetCoordinate.server.raft: entering leader state: leader="Node at 127.0.0.1:29807 [Leader]" writer.go:29: 2021-01-29T19:33:57.613Z [INFO] TestAgent_GetCoordinate.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:57.614Z [INFO] TestAgent_GetCoordinate.server: New leader elected: payload=Node-92ad6b6e-3c13-7c17-01a7-94fd6580b221 writer.go:29: 2021-01-29T19:33:57.616Z [INFO] TestAgent_GetCoordinate: Requesting shutdown writer.go:29: 2021-01-29T19:33:57.617Z [INFO] TestAgent_GetCoordinate.server: shutting down server writer.go:29: 2021-01-29T19:33:57.619Z [WARN] TestAgent_GetCoordinate.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.617Z [DEBUG] TestAgent_GetCoordinate.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29807 writer.go:29: 2021-01-29T19:33:57.619Z [ERROR] TestAgent_GetCoordinate.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:57.622Z [WARN] TestAgent_GetCoordinate.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.626Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:57.632Z [INFO] TestAgent_GetCoordinate.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:57.633Z [INFO] TestAgent_GetCoordinate: consul server down writer.go:29: 2021-01-29T19:33:57.635Z [INFO] TestAgent_GetCoordinate: shutdown complete writer.go:29: 2021-01-29T19:33:57.633Z [ERROR] TestAgent_GetCoordinate.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: leadership lost while committing log" writer.go:29: 2021-01-29T19:33:57.637Z [INFO] TestAgent_GetCoordinate: Stopping server: protocol=DNS address=127.0.0.1:29802 network=tcp writer.go:29: 2021-01-29T19:33:57.639Z [INFO] TestAgent_GetCoordinate: Stopping server: protocol=DNS address=127.0.0.1:29802 network=udp writer.go:29: 2021-01-29T19:33:57.638Z [ERROR] TestAgent_GetCoordinate.server: failed to transfer leadership attempt, will retry: attempt=0 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:57.640Z [ERROR] TestAgent_GetCoordinate.server: failed to transfer leadership attempt, will retry: attempt=1 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:57.641Z [ERROR] TestAgent_GetCoordinate.server: failed to transfer leadership attempt, will retry: attempt=2 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:57.640Z [INFO] TestAgent_GetCoordinate: Stopping server: protocol=HTTP address=127.0.0.1:29803 network=tcp writer.go:29: 2021-01-29T19:33:57.644Z [ERROR] TestAgent_GetCoordinate.server: failed to transfer leadership: error="failed to transfer leadership in 3 attempts" === CONT TestAgent_loadCheckState writer.go:29: 2021-01-29T19:33:57.675Z [DEBUG] TestAgent_loadCheckState: check state expired, not restoring: check=check1 writer.go:29: 2021-01-29T19:33:57.677Z [INFO] TestAgent_loadCheckState: Requesting shutdown writer.go:29: 2021-01-29T19:33:57.679Z [INFO] TestAgent_loadCheckState.server: shutting down server writer.go:29: 2021-01-29T19:33:57.680Z [DEBUG] TestAgent_loadCheckState.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.680Z [ERROR] TestAgent_loadCheckState.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:57.683Z [DEBUG] TestAgent_loadCheckState.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.684Z [WARN] TestAgent_loadCheckState.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.697Z [WARN] TestAgent_loadCheckState.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.700Z [INFO] TestAgent_loadCheckState.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:57.701Z [INFO] TestAgent_loadCheckState: consul server down writer.go:29: 2021-01-29T19:33:57.702Z [INFO] TestAgent_loadCheckState: shutdown complete writer.go:29: 2021-01-29T19:33:57.704Z [INFO] TestAgent_loadCheckState: Stopping server: protocol=DNS address=127.0.0.1:29790 network=tcp writer.go:29: 2021-01-29T19:33:57.705Z [INFO] TestAgent_loadCheckState: Stopping server: protocol=DNS address=127.0.0.1:29790 network=udp writer.go:29: 2021-01-29T19:33:57.707Z [INFO] TestAgent_loadCheckState: Stopping server: protocol=HTTP address=127.0.0.1:29791 network=tcp === CONT TestAgent_purgeCheckState writer.go:29: 2021-01-29T19:33:57.782Z [INFO] TestAgent_purgeCheckState: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:57.784Z [INFO] TestAgent_purgeCheckState: Endpoints down --- PASS: TestAgent_purgeCheckState (0.82s) === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:57.798Z [WARN] TestAgent_loadChecks_checkFails: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:57.800Z [DEBUG] TestAgent_loadChecks_checkFails.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:57.803Z [DEBUG] TestAgent_loadChecks_checkFails.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:57.809Z [INFO] TestAgent_loadChecks_checkFails.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2bc2ae95-6d3b-183a-a101-6c5c51883995 Address:127.0.0.1:29819}]" writer.go:29: 2021-01-29T19:33:57.812Z [INFO] TestAgent_loadChecks_checkFails.server.raft: entering follower state: follower="Node at 127.0.0.1:29819 [Follower]" leader= writer.go:29: 2021-01-29T19:33:57.812Z [INFO] TestAgent_loadChecks_checkFails.server.serf.wan: serf: EventMemberJoin: Node-2bc2ae95-6d3b-183a-a101-6c5c51883995.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.819Z [INFO] TestAgent_loadChecks_checkFails.server.serf.lan: serf: EventMemberJoin: Node-2bc2ae95-6d3b-183a-a101-6c5c51883995 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.824Z [INFO] TestAgent_loadChecks_checkFails.server: Adding LAN server: server="Node-2bc2ae95-6d3b-183a-a101-6c5c51883995 (Addr: tcp/127.0.0.1:29819) (DC: dc1)" writer.go:29: 2021-01-29T19:33:57.826Z [INFO] TestAgent_loadChecks_checkFails.server: Handled event for server in area: event=member-join server=Node-2bc2ae95-6d3b-183a-a101-6c5c51883995.dc1 area=wan writer.go:29: 2021-01-29T19:33:57.833Z [INFO] TestAgent_loadChecks_checkFails: Started DNS server: address=127.0.0.1:29814 network=tcp writer.go:29: 2021-01-29T19:33:57.834Z [INFO] TestAgent_loadChecks_checkFails: Started DNS server: address=127.0.0.1:29814 network=udp writer.go:29: 2021-01-29T19:33:57.838Z [INFO] TestAgent_loadChecks_checkFails: Started HTTP server: address=127.0.0.1:29815 network=tcp writer.go:29: 2021-01-29T19:33:57.840Z [INFO] TestAgent_loadChecks_checkFails: started state syncer writer.go:29: 2021-01-29T19:33:57.874Z [WARN] TestAgent_loadChecks_checkFails.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:57.877Z [INFO] TestAgent_loadChecks_checkFails.server.raft: entering candidate state: node="Node at 127.0.0.1:29819 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:57.880Z [DEBUG] TestAgent_loadChecks_checkFails.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:57.882Z [DEBUG] TestAgent_loadChecks_checkFails.server.raft: vote granted: from=2bc2ae95-6d3b-183a-a101-6c5c51883995 term=2 tally=1 writer.go:29: 2021-01-29T19:33:57.883Z [INFO] TestAgent_loadChecks_checkFails.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:57.885Z [INFO] TestAgent_loadChecks_checkFails.server.raft: entering leader state: leader="Node at 127.0.0.1:29819 [Leader]" writer.go:29: 2021-01-29T19:33:57.887Z [INFO] TestAgent_loadChecks_checkFails.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:57.887Z [INFO] TestAgent_loadChecks_checkFails.server: New leader elected: payload=Node-2bc2ae95-6d3b-183a-a101-6c5c51883995 writer.go:29: 2021-01-29T19:33:57.889Z [DEBUG] TestAgent_loadChecks_checkFails.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29819 writer.go:29: 2021-01-29T19:33:57.894Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:57.899Z [INFO] TestAgent_loadChecks_checkFails.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:57.900Z [INFO] TestAgent_loadChecks_checkFails.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.901Z [DEBUG] TestAgent_loadChecks_checkFails.server: Skipping self join check for node since the cluster is too small: node=Node-2bc2ae95-6d3b-183a-a101-6c5c51883995 writer.go:29: 2021-01-29T19:33:57.902Z [INFO] TestAgent_loadChecks_checkFails.server: member joined, marking health alive: member=Node-2bc2ae95-6d3b-183a-a101-6c5c51883995 === CONT TestAgent_persistCheckState writer.go:29: 2021-01-29T19:33:58.100Z [INFO] TestAgent_persistCheckState: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:58.101Z [INFO] TestAgent_persistCheckState: Endpoints down --- PASS: TestAgent_persistCheckState (0.72s) === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.119Z [WARN] TestAgent_checkStateSnapshot: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:58.124Z [DEBUG] TestAgent_checkStateSnapshot.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:58.126Z [DEBUG] TestAgent_checkStateSnapshot.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:58.132Z [INFO] TestAgent_checkStateSnapshot.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b6ce503d-50f2-83bd-b870-ceb42c7e3ebb Address:127.0.0.1:29825}]" writer.go:29: 2021-01-29T19:33:58.135Z [INFO] TestAgent_checkStateSnapshot.server.raft: entering follower state: follower="Node at 127.0.0.1:29825 [Follower]" leader= writer.go:29: 2021-01-29T19:33:58.139Z [INFO] TestAgent_checkStateSnapshot.server.serf.wan: serf: EventMemberJoin: Node-b6ce503d-50f2-83bd-b870-ceb42c7e3ebb.dc1 127.0.0.1 === CONT TestAgent_GetCoordinate writer.go:29: 2021-01-29T19:33:58.144Z [INFO] TestAgent_GetCoordinate: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:58.145Z [INFO] TestAgent_GetCoordinate: Endpoints down --- PASS: TestAgent_GetCoordinate (1.38s) === CONT TestAgent_NodeMaintenanceMode === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.146Z [INFO] TestAgent_checkStateSnapshot.server.serf.lan: serf: EventMemberJoin: Node-b6ce503d-50f2-83bd-b870-ceb42c7e3ebb 127.0.0.1 writer.go:29: 2021-01-29T19:33:58.149Z [INFO] TestAgent_checkStateSnapshot.server: Adding LAN server: server="Node-b6ce503d-50f2-83bd-b870-ceb42c7e3ebb (Addr: tcp/127.0.0.1:29825) (DC: dc1)" writer.go:29: 2021-01-29T19:33:58.151Z [INFO] TestAgent_checkStateSnapshot.server: Handled event for server in area: event=member-join server=Node-b6ce503d-50f2-83bd-b870-ceb42c7e3ebb.dc1 area=wan writer.go:29: 2021-01-29T19:33:58.160Z [INFO] TestAgent_checkStateSnapshot: Started DNS server: address=127.0.0.1:29820 network=tcp writer.go:29: 2021-01-29T19:33:58.162Z [INFO] TestAgent_checkStateSnapshot: Started DNS server: address=127.0.0.1:29820 network=udp writer.go:29: 2021-01-29T19:33:58.166Z [INFO] TestAgent_checkStateSnapshot: Started HTTP server: address=127.0.0.1:29821 network=tcp === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.165Z [WARN] TestAgent_NodeMaintenanceMode: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:58.168Z [DEBUG] TestAgent_NodeMaintenanceMode.tlsutil: Update: version=1 === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.171Z [DEBUG] TestAgent_loadChecks_checkFails: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.171Z [DEBUG] TestAgent_NodeMaintenanceMode.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.168Z [INFO] TestAgent_checkStateSnapshot: started state syncer === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.174Z [INFO] TestAgent_loadChecks_checkFails: Synced node info writer.go:29: 2021-01-29T19:33:58.175Z [DEBUG] TestAgent_loadChecks_checkFails: Node info in sync === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.176Z [INFO] TestAgent_NodeMaintenanceMode.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9de446bc-7794-1514-0fbf-f0b8907896bb Address:127.0.0.1:29837}]" writer.go:29: 2021-01-29T19:33:58.179Z [INFO] TestAgent_NodeMaintenanceMode.server.raft: entering follower state: follower="Node at 127.0.0.1:29837 [Follower]" leader= writer.go:29: 2021-01-29T19:33:58.182Z [INFO] TestAgent_NodeMaintenanceMode.server.serf.wan: serf: EventMemberJoin: Node-9de446bc-7794-1514-0fbf-f0b8907896bb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:58.186Z [INFO] TestAgent_NodeMaintenanceMode.server.serf.lan: serf: EventMemberJoin: Node-9de446bc-7794-1514-0fbf-f0b8907896bb 127.0.0.1 writer.go:29: 2021-01-29T19:33:58.191Z [INFO] TestAgent_NodeMaintenanceMode.server: Handled event for server in area: event=member-join server=Node-9de446bc-7794-1514-0fbf-f0b8907896bb.dc1 area=wan writer.go:29: 2021-01-29T19:33:58.191Z [INFO] TestAgent_NodeMaintenanceMode.server: Adding LAN server: server="Node-9de446bc-7794-1514-0fbf-f0b8907896bb (Addr: tcp/127.0.0.1:29837) (DC: dc1)" writer.go:29: 2021-01-29T19:33:58.197Z [INFO] TestAgent_NodeMaintenanceMode: Started DNS server: address=127.0.0.1:29832 network=tcp === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.200Z [WARN] TestAgent_checkStateSnapshot.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.201Z [INFO] TestAgent_NodeMaintenanceMode: Started DNS server: address=127.0.0.1:29832 network=udp === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.203Z [INFO] TestAgent_checkStateSnapshot.server.raft: entering candidate state: node="Node at 127.0.0.1:29825 [Candidate]" term=2 === CONT TestAgent_loadCheckState writer.go:29: 2021-01-29T19:33:58.208Z [INFO] TestAgent_loadCheckState: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:58.210Z [INFO] TestAgent_loadCheckState: Endpoints down --- PASS: TestAgent_loadCheckState (1.08s) === CONT TestAgent_AddCheck_restoresSnapshot === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.212Z [INFO] TestAgent_NodeMaintenanceMode: Started HTTP server: address=127.0.0.1:29833 network=tcp writer.go:29: 2021-01-29T19:33:58.214Z [INFO] TestAgent_NodeMaintenanceMode: started state syncer === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.217Z [DEBUG] TestAgent_checkStateSnapshot.server.raft: votes: needed=1 === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.219Z [WARN] TestAgent_AddCheck_restoresSnapshot: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:58.220Z [DEBUG] TestAgent_AddCheck_restoresSnapshot.tlsutil: Update: version=1 === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.219Z [DEBUG] TestAgent_checkStateSnapshot.server.raft: vote granted: from=b6ce503d-50f2-83bd-b870-ceb42c7e3ebb term=2 tally=1 writer.go:29: 2021-01-29T19:33:58.221Z [INFO] TestAgent_checkStateSnapshot.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:58.222Z [INFO] TestAgent_checkStateSnapshot.server.raft: entering leader state: leader="Node at 127.0.0.1:29825 [Leader]" writer.go:29: 2021-01-29T19:33:58.224Z [INFO] TestAgent_checkStateSnapshot.server: cluster leadership acquired === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.221Z [DEBUG] TestAgent_AddCheck_restoresSnapshot.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:58.231Z [INFO] TestAgent_AddCheck_restoresSnapshot.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:feef7e7e-e81e-e14f-2ad8-7c021200c4fd Address:127.0.0.1:29843}]" writer.go:29: 2021-01-29T19:33:58.233Z [INFO] TestAgent_AddCheck_restoresSnapshot.server.raft: entering follower state: follower="Node at 127.0.0.1:29843 [Follower]" leader= === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.225Z [INFO] TestAgent_checkStateSnapshot.server: New leader elected: payload=Node-b6ce503d-50f2-83bd-b870-ceb42c7e3ebb writer.go:29: 2021-01-29T19:33:58.231Z [DEBUG] TestAgent_checkStateSnapshot.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29825 === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.250Z [WARN] TestAgent_NodeMaintenanceMode.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:58.253Z [INFO] TestAgent_NodeMaintenanceMode.server.raft: entering candidate state: node="Node at 127.0.0.1:29837 [Candidate]" term=2 === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.255Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.251Z [INFO] TestAgent_AddCheck_restoresSnapshot.server.serf.wan: serf: EventMemberJoin: Node-feef7e7e-e81e-e14f-2ad8-7c021200c4fd.dc1 127.0.0.1 === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.255Z [DEBUG] TestAgent_NodeMaintenanceMode.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:58.257Z [DEBUG] TestAgent_NodeMaintenanceMode.server.raft: vote granted: from=9de446bc-7794-1514-0fbf-f0b8907896bb term=2 tally=1 writer.go:29: 2021-01-29T19:33:58.258Z [INFO] TestAgent_NodeMaintenanceMode.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:58.260Z [INFO] TestAgent_NodeMaintenanceMode.server.raft: entering leader state: leader="Node at 127.0.0.1:29837 [Leader]" writer.go:29: 2021-01-29T19:33:58.261Z [INFO] TestAgent_NodeMaintenanceMode.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:58.262Z [INFO] TestAgent_NodeMaintenanceMode.server: New leader elected: payload=Node-9de446bc-7794-1514-0fbf-f0b8907896bb === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.260Z [INFO] TestAgent_AddCheck_restoresSnapshot.server.serf.lan: serf: EventMemberJoin: Node-feef7e7e-e81e-e14f-2ad8-7c021200c4fd 127.0.0.1 === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.263Z [DEBUG] TestAgent_NodeMaintenanceMode.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29837 === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.266Z [INFO] TestAgent_AddCheck_restoresSnapshot.server: Adding LAN server: server="Node-feef7e7e-e81e-e14f-2ad8-7c021200c4fd (Addr: tcp/127.0.0.1:29843) (DC: dc1)" === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.268Z [WARN] TestAgent_loadChecks_checkFails: Failed to restore check: check=service:redis error="ServiceID "nope" does not exist" === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.267Z [INFO] TestAgent_AddCheck_restoresSnapshot.server: Handled event for server in area: event=member-join server=Node-feef7e7e-e81e-e14f-2ad8-7c021200c4fd.dc1 area=wan === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.267Z [INFO] TestAgent_checkStateSnapshot.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.269Z [DEBUG] TestAgent_loadChecks_checkFails: restored health check from file: check=service:redis file=/tmp/TestAgent_loadChecks_checkFails-agent348971113/checks/60a2ef12de014a05ecdc850d9aab46da writer.go:29: 2021-01-29T19:33:58.270Z [INFO] TestAgent_loadChecks_checkFails: Requesting shutdown === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.269Z [INFO] TestAgent_checkStateSnapshot.leader: started routine: routine="CA root pruning" === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.270Z [INFO] TestAgent_loadChecks_checkFails.server: shutting down server === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.269Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.271Z [DEBUG] TestAgent_loadChecks_checkFails.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:58.272Z [WARN] TestAgent_loadChecks_checkFails.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.273Z [INFO] TestAgent_AddCheck_restoresSnapshot: Started DNS server: address=127.0.0.1:29838 network=udp writer.go:29: 2021-01-29T19:33:58.273Z [WARN] TestAgent_AddCheck_restoresSnapshot.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:58.274Z [INFO] TestAgent_AddCheck_restoresSnapshot.server.raft: entering candidate state: node="Node at 127.0.0.1:29843 [Candidate]" term=2 === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.273Z [DEBUG] TestAgent_loadChecks_checkFails.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.276Z [INFO] TestAgent_AddCheck_restoresSnapshot: Started DNS server: address=127.0.0.1:29838 network=tcp === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.271Z [DEBUG] TestAgent_checkStateSnapshot.server: Skipping self join check for node since the cluster is too small: node=Node-b6ce503d-50f2-83bd-b870-ceb42c7e3ebb === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.275Z [WARN] TestAgent_loadChecks_checkFails.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.278Z [INFO] TestAgent_checkStateSnapshot.server: member joined, marking health alive: member=Node-b6ce503d-50f2-83bd-b870-ceb42c7e3ebb === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.279Z [INFO] TestAgent_AddCheck_restoresSnapshot: Started HTTP server: address=127.0.0.1:29839 network=tcp === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.279Z [INFO] TestAgent_NodeMaintenanceMode.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:58.292Z [INFO] TestAgent_NodeMaintenanceMode.leader: started routine: routine="CA root pruning" === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.292Z [INFO] TestAgent_AddCheck_restoresSnapshot: started state syncer === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.292Z [DEBUG] TestAgent_NodeMaintenanceMode.server: Skipping self join check for node since the cluster is too small: node=Node-9de446bc-7794-1514-0fbf-f0b8907896bb === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.280Z [INFO] TestAgent_loadChecks_checkFails.server.router.manager: shutting down === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.293Z [INFO] TestAgent_NodeMaintenanceMode.server: member joined, marking health alive: member=Node-9de446bc-7794-1514-0fbf-f0b8907896bb === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.280Z [INFO] TestAgent_loadChecks_checkFails: consul server down writer.go:29: 2021-01-29T19:33:58.294Z [INFO] TestAgent_loadChecks_checkFails: shutdown complete writer.go:29: 2021-01-29T19:33:58.295Z [INFO] TestAgent_loadChecks_checkFails: Stopping server: protocol=DNS address=127.0.0.1:29814 network=tcp === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.279Z [DEBUG] TestAgent_AddCheck_restoresSnapshot.server.raft: votes: needed=1 === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.296Z [INFO] TestAgent_loadChecks_checkFails: Stopping server: protocol=DNS address=127.0.0.1:29814 network=udp writer.go:29: 2021-01-29T19:33:58.297Z [INFO] TestAgent_loadChecks_checkFails: Stopping server: protocol=HTTP address=127.0.0.1:29815 network=tcp === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.296Z [DEBUG] TestAgent_AddCheck_restoresSnapshot.server.raft: vote granted: from=feef7e7e-e81e-e14f-2ad8-7c021200c4fd term=2 tally=1 writer.go:29: 2021-01-29T19:33:58.299Z [INFO] TestAgent_AddCheck_restoresSnapshot.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:58.301Z [INFO] TestAgent_AddCheck_restoresSnapshot.server.raft: entering leader state: leader="Node at 127.0.0.1:29843 [Leader]" writer.go:29: 2021-01-29T19:33:58.305Z [INFO] TestAgent_AddCheck_restoresSnapshot.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:58.307Z [INFO] TestAgent_AddCheck_restoresSnapshot.server: New leader elected: payload=Node-feef7e7e-e81e-e14f-2ad8-7c021200c4fd writer.go:29: 2021-01-29T19:33:58.310Z [DEBUG] TestAgent_AddCheck_restoresSnapshot.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29843 writer.go:29: 2021-01-29T19:33:58.316Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:58.327Z [INFO] TestAgent_AddCheck_restoresSnapshot.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:58.329Z [INFO] TestAgent_AddCheck_restoresSnapshot.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:58.330Z [DEBUG] TestAgent_AddCheck_restoresSnapshot.server: Skipping self join check for node since the cluster is too small: node=Node-feef7e7e-e81e-e14f-2ad8-7c021200c4fd writer.go:29: 2021-01-29T19:33:58.330Z [INFO] TestAgent_AddCheck_restoresSnapshot.server: member joined, marking health alive: member=Node-feef7e7e-e81e-e14f-2ad8-7c021200c4fd === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.333Z [INFO] TestAgent_NodeMaintenanceMode: Node entered maintenance mode writer.go:29: 2021-01-29T19:33:58.335Z [DEBUG] TestAgent_NodeMaintenanceMode: removed check: check=_node_maintenance writer.go:29: 2021-01-29T19:33:58.336Z [INFO] TestAgent_NodeMaintenanceMode: Node left maintenance mode writer.go:29: 2021-01-29T19:33:58.338Z [INFO] TestAgent_NodeMaintenanceMode: Node entered maintenance mode writer.go:29: 2021-01-29T19:33:58.341Z [INFO] TestAgent_NodeMaintenanceMode: Requesting shutdown writer.go:29: 2021-01-29T19:33:58.344Z [INFO] TestAgent_NodeMaintenanceMode.server: shutting down server writer.go:29: 2021-01-29T19:33:58.348Z [DEBUG] TestAgent_NodeMaintenanceMode.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:58.351Z [WARN] TestAgent_NodeMaintenanceMode.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:58.348Z [ERROR] TestAgent_NodeMaintenanceMode.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:58.354Z [DEBUG] TestAgent_NodeMaintenanceMode.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:58.363Z [WARN] TestAgent_NodeMaintenanceMode.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:58.368Z [INFO] TestAgent_NodeMaintenanceMode.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:58.369Z [INFO] TestAgent_NodeMaintenanceMode: consul server down writer.go:29: 2021-01-29T19:33:58.372Z [INFO] TestAgent_NodeMaintenanceMode: shutdown complete writer.go:29: 2021-01-29T19:33:58.374Z [INFO] TestAgent_NodeMaintenanceMode: Stopping server: protocol=DNS address=127.0.0.1:29832 network=tcp writer.go:29: 2021-01-29T19:33:58.375Z [INFO] TestAgent_NodeMaintenanceMode: Stopping server: protocol=DNS address=127.0.0.1:29832 network=udp writer.go:29: 2021-01-29T19:33:58.377Z [INFO] TestAgent_NodeMaintenanceMode: Stopping server: protocol=HTTP address=127.0.0.1:29833 network=tcp === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.471Z [DEBUG] TestAgent_checkStateSnapshot: removed check: check=service:redis writer.go:29: 2021-01-29T19:33:58.472Z [DEBUG] TestAgent_checkStateSnapshot: restored health check from file: check=service:redis file=/tmp/TestAgent_checkStateSnapshot-agent598790324/checks/60a2ef12de014a05ecdc850d9aab46da writer.go:29: 2021-01-29T19:33:58.475Z [INFO] TestAgent_checkStateSnapshot: Requesting shutdown writer.go:29: 2021-01-29T19:33:58.477Z [INFO] TestAgent_checkStateSnapshot.server: shutting down server writer.go:29: 2021-01-29T19:33:58.478Z [DEBUG] TestAgent_checkStateSnapshot.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:58.479Z [WARN] TestAgent_checkStateSnapshot.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:58.478Z [ERROR] TestAgent_checkStateSnapshot.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:58.479Z [DEBUG] TestAgent_checkStateSnapshot.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.484Z [INFO] TestAgent_AddCheck_restoresSnapshot: Requesting shutdown writer.go:29: 2021-01-29T19:33:58.486Z [INFO] TestAgent_AddCheck_restoresSnapshot.server: shutting down server writer.go:29: 2021-01-29T19:33:58.487Z [DEBUG] TestAgent_AddCheck_restoresSnapshot.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:58.488Z [WARN] TestAgent_AddCheck_restoresSnapshot.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:58.487Z [ERROR] TestAgent_AddCheck_restoresSnapshot.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:58.488Z [DEBUG] TestAgent_AddCheck_restoresSnapshot.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.483Z [WARN] TestAgent_checkStateSnapshot.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.490Z [WARN] TestAgent_AddCheck_restoresSnapshot.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:58.494Z [INFO] TestAgent_AddCheck_restoresSnapshot.server.router.manager: shutting down === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.493Z [INFO] TestAgent_checkStateSnapshot.server.router.manager: shutting down === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.495Z [INFO] TestAgent_AddCheck_restoresSnapshot: consul server down === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.493Z [INFO] TestAgent_checkStateSnapshot: consul server down === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.495Z [INFO] TestAgent_AddCheck_restoresSnapshot: shutdown complete writer.go:29: 2021-01-29T19:33:58.496Z [INFO] TestAgent_AddCheck_restoresSnapshot: Stopping server: protocol=DNS address=127.0.0.1:29838 network=tcp === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.496Z [INFO] TestAgent_checkStateSnapshot: shutdown complete === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.497Z [INFO] TestAgent_AddCheck_restoresSnapshot: Stopping server: protocol=DNS address=127.0.0.1:29838 network=udp writer.go:29: 2021-01-29T19:33:58.498Z [INFO] TestAgent_AddCheck_restoresSnapshot: Stopping server: protocol=HTTP address=127.0.0.1:29839 network=tcp === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:58.498Z [INFO] TestAgent_checkStateSnapshot: Stopping server: protocol=DNS address=127.0.0.1:29820 network=tcp writer.go:29: 2021-01-29T19:33:58.499Z [INFO] TestAgent_checkStateSnapshot: Stopping server: protocol=DNS address=127.0.0.1:29820 network=udp writer.go:29: 2021-01-29T19:33:58.501Z [INFO] TestAgent_checkStateSnapshot: Stopping server: protocol=HTTP address=127.0.0.1:29821 network=tcp === CONT TestAgent_loadChecks_checkFails writer.go:29: 2021-01-29T19:33:58.799Z [INFO] TestAgent_loadChecks_checkFails: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:58.801Z [INFO] TestAgent_loadChecks_checkFails: Endpoints down --- PASS: TestAgent_loadChecks_checkFails (1.02s) === CONT TestTxnEndpoint_KV_Actions === RUN TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:58.826Z [WARN] TestTxnEndpoint_KV_Actions/#00: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:58.828Z [DEBUG] TestTxnEndpoint_KV_Actions/#00.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:58.830Z [DEBUG] TestTxnEndpoint_KV_Actions/#00.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:58.835Z [INFO] TestTxnEndpoint_KV_Actions/#00.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5e26b193-2806-3fe2-1f95-effebf52ccb0 Address:127.0.0.1:29831}]" writer.go:29: 2021-01-29T19:33:58.838Z [INFO] TestTxnEndpoint_KV_Actions/#00.server.raft: entering follower state: follower="Node at 127.0.0.1:29831 [Follower]" leader= writer.go:29: 2021-01-29T19:33:58.842Z [INFO] TestTxnEndpoint_KV_Actions/#00.server.serf.wan: serf: EventMemberJoin: Node-5e26b193-2806-3fe2-1f95-effebf52ccb0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:58.847Z [INFO] TestTxnEndpoint_KV_Actions/#00.server.serf.lan: serf: EventMemberJoin: Node-5e26b193-2806-3fe2-1f95-effebf52ccb0 127.0.0.1 writer.go:29: 2021-01-29T19:33:58.849Z [INFO] TestTxnEndpoint_KV_Actions/#00.server: Handled event for server in area: event=member-join server=Node-5e26b193-2806-3fe2-1f95-effebf52ccb0.dc1 area=wan writer.go:29: 2021-01-29T19:33:58.850Z [INFO] TestTxnEndpoint_KV_Actions/#00.server: Adding LAN server: server="Node-5e26b193-2806-3fe2-1f95-effebf52ccb0 (Addr: tcp/127.0.0.1:29831) (DC: dc1)" writer.go:29: 2021-01-29T19:33:58.851Z [INFO] TestTxnEndpoint_KV_Actions/#00: Started DNS server: address=127.0.0.1:29826 network=udp writer.go:29: 2021-01-29T19:33:58.853Z [INFO] TestTxnEndpoint_KV_Actions/#00: Started DNS server: address=127.0.0.1:29826 network=tcp writer.go:29: 2021-01-29T19:33:58.856Z [INFO] TestTxnEndpoint_KV_Actions/#00: Started HTTP server: address=127.0.0.1:29827 network=tcp writer.go:29: 2021-01-29T19:33:58.857Z [INFO] TestTxnEndpoint_KV_Actions/#00: started state syncer === CONT TestAgent_NodeMaintenanceMode writer.go:29: 2021-01-29T19:33:58.879Z [INFO] TestAgent_NodeMaintenanceMode: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:58.880Z [INFO] TestAgent_NodeMaintenanceMode: Endpoints down --- PASS: TestAgent_NodeMaintenanceMode (0.74s) === CONT TestUiNodes_Filter === CONT TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:58.886Z [WARN] TestTxnEndpoint_KV_Actions/#00.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:58.888Z [INFO] TestTxnEndpoint_KV_Actions/#00.server.raft: entering candidate state: node="Node at 127.0.0.1:29831 [Candidate]" term=2 === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:58.903Z [WARN] TestUiNodes_Filter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:58.904Z [DEBUG] TestUiNodes_Filter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:58.906Z [DEBUG] TestUiNodes_Filter.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:58.916Z [DEBUG] TestTxnEndpoint_KV_Actions/#00.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:58.917Z [DEBUG] TestTxnEndpoint_KV_Actions/#00.server.raft: vote granted: from=5e26b193-2806-3fe2-1f95-effebf52ccb0 term=2 tally=1 === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:58.919Z [INFO] TestUiNodes_Filter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7a769d4a-bfc7-2cea-2e46-99778152194b Address:127.0.0.1:29855}]" writer.go:29: 2021-01-29T19:33:58.921Z [INFO] TestUiNodes_Filter.server.raft: entering follower state: follower="Node at 127.0.0.1:29855 [Follower]" leader= === CONT TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:58.920Z [INFO] TestTxnEndpoint_KV_Actions/#00.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:58.926Z [INFO] TestTxnEndpoint_KV_Actions/#00.server.raft: entering leader state: leader="Node at 127.0.0.1:29831 [Leader]" === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:58.926Z [INFO] TestUiNodes_Filter.server.serf.wan: serf: EventMemberJoin: Node-7a769d4a-bfc7-2cea-2e46-99778152194b.dc1 127.0.0.1 === CONT TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:58.929Z [INFO] TestTxnEndpoint_KV_Actions/#00.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:58.930Z [INFO] TestTxnEndpoint_KV_Actions/#00.server: New leader elected: payload=Node-5e26b193-2806-3fe2-1f95-effebf52ccb0 === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:58.933Z [INFO] TestUiNodes_Filter.server.serf.lan: serf: EventMemberJoin: Node-7a769d4a-bfc7-2cea-2e46-99778152194b 127.0.0.1 === CONT TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:58.933Z [DEBUG] TestTxnEndpoint_KV_Actions/#00.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29831 === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:58.936Z [INFO] TestUiNodes_Filter: Started DNS server: address=127.0.0.1:29850 network=udp writer.go:29: 2021-01-29T19:33:58.938Z [INFO] TestUiNodes_Filter.server: Adding LAN server: server="Node-7a769d4a-bfc7-2cea-2e46-99778152194b (Addr: tcp/127.0.0.1:29855) (DC: dc1)" === CONT TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:58.938Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:58.938Z [INFO] TestUiNodes_Filter: Started DNS server: address=127.0.0.1:29850 network=tcp writer.go:29: 2021-01-29T19:33:58.938Z [INFO] TestUiNodes_Filter.server: Handled event for server in area: event=member-join server=Node-7a769d4a-bfc7-2cea-2e46-99778152194b.dc1 area=wan writer.go:29: 2021-01-29T19:33:58.942Z [INFO] TestUiNodes_Filter: Started HTTP server: address=127.0.0.1:29851 network=tcp writer.go:29: 2021-01-29T19:33:58.946Z [INFO] TestUiNodes_Filter: started state syncer === CONT TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:58.943Z [INFO] TestTxnEndpoint_KV_Actions/#00.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:58.954Z [INFO] TestTxnEndpoint_KV_Actions/#00.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:58.958Z [DEBUG] TestTxnEndpoint_KV_Actions/#00.server: Skipping self join check for node since the cluster is too small: node=Node-5e26b193-2806-3fe2-1f95-effebf52ccb0 === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:58.962Z [WARN] TestUiNodes_Filter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:58.963Z [INFO] TestUiNodes_Filter.server.raft: entering candidate state: node="Node at 127.0.0.1:29855 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:58.968Z [DEBUG] TestUiNodes_Filter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:58.970Z [DEBUG] TestUiNodes_Filter.server.raft: vote granted: from=7a769d4a-bfc7-2cea-2e46-99778152194b term=2 tally=1 === CONT TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:58.963Z [INFO] TestTxnEndpoint_KV_Actions/#00.server: member joined, marking health alive: member=Node-5e26b193-2806-3fe2-1f95-effebf52ccb0 writer.go:29: 2021-01-29T19:33:58.967Z [INFO] TestTxnEndpoint_KV_Actions/#00: Synced node info writer.go:29: 2021-01-29T19:33:58.975Z [DEBUG] TestTxnEndpoint_KV_Actions/#00: Node info in sync === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:58.972Z [INFO] TestUiNodes_Filter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:58.980Z [INFO] TestUiNodes_Filter.server.raft: entering leader state: leader="Node at 127.0.0.1:29855 [Leader]" writer.go:29: 2021-01-29T19:33:58.984Z [INFO] TestUiNodes_Filter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:58.985Z [INFO] TestUiNodes_Filter.server: New leader elected: payload=Node-7a769d4a-bfc7-2cea-2e46-99778152194b writer.go:29: 2021-01-29T19:33:58.988Z [DEBUG] TestUiNodes_Filter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29855 writer.go:29: 2021-01-29T19:33:58.999Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_checkStateSnapshot writer.go:29: 2021-01-29T19:33:59.002Z [INFO] TestAgent_checkStateSnapshot: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:59.003Z [INFO] TestAgent_checkStateSnapshot: Endpoints down --- PASS: TestAgent_checkStateSnapshot (0.90s) === CONT TestUiNodes === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:59.010Z [INFO] TestUiNodes_Filter.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:58.999Z [INFO] TestAgent_AddCheck_restoresSnapshot: Waiting for endpoints to shut down === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:59.015Z [INFO] TestUiNodes_Filter.leader: started routine: routine="CA root pruning" === CONT TestAgent_AddCheck_restoresSnapshot writer.go:29: 2021-01-29T19:33:59.016Z [INFO] TestAgent_AddCheck_restoresSnapshot: Endpoints down --- PASS: TestAgent_AddCheck_restoresSnapshot (0.81s) === CONT TestUiIndex === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:59.020Z [DEBUG] TestUiNodes_Filter.server: Skipping self join check for node since the cluster is too small: node=Node-7a769d4a-bfc7-2cea-2e46-99778152194b === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.020Z [WARN] TestUiNodes: bootstrap = true: do not enable unless necessary === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:59.025Z [INFO] TestUiNodes_Filter.server: member joined, marking health alive: member=Node-7a769d4a-bfc7-2cea-2e46-99778152194b === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.028Z [DEBUG] TestUiNodes.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:59.033Z [DEBUG] TestUiNodes.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.049Z [WARN] TestUiIndex: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:59.050Z [DEBUG] TestUiIndex.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:59.053Z [DEBUG] TestUiIndex.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:59.058Z [INFO] TestUiIndex.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b58db6c4-bc40-611e-bf6f-2e16edd2e87f Address:127.0.0.1:29849}]" writer.go:29: 2021-01-29T19:33:59.061Z [INFO] TestUiIndex.server.raft: entering follower state: follower="Node at 127.0.0.1:29849 [Follower]" leader= === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.058Z [INFO] TestUiNodes.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:17af74e8-1b49-24c8-31ed-e6976539085d Address:127.0.0.1:29861}]" === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.063Z [INFO] TestUiIndex.server.serf.wan: serf: EventMemberJoin: Node-b58db6c4-bc40-611e-bf6f-2e16edd2e87f.dc1 127.0.0.1 === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.064Z [INFO] TestUiNodes.server.raft: entering follower state: follower="Node at 127.0.0.1:29861 [Follower]" leader= === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.068Z [INFO] TestUiIndex.server.serf.lan: serf: EventMemberJoin: Node-b58db6c4-bc40-611e-bf6f-2e16edd2e87f 127.0.0.1 writer.go:29: 2021-01-29T19:33:59.072Z [INFO] TestUiIndex.server: Adding LAN server: server="Node-b58db6c4-bc40-611e-bf6f-2e16edd2e87f (Addr: tcp/127.0.0.1:29849) (DC: dc1)" writer.go:29: 2021-01-29T19:33:59.073Z [INFO] TestUiIndex.server: Handled event for server in area: event=member-join server=Node-b58db6c4-bc40-611e-bf6f-2e16edd2e87f.dc1 area=wan writer.go:29: 2021-01-29T19:33:59.074Z [INFO] TestUiIndex: Started DNS server: address=127.0.0.1:29844 network=udp writer.go:29: 2021-01-29T19:33:59.075Z [INFO] TestUiIndex: Started DNS server: address=127.0.0.1:29844 network=tcp writer.go:29: 2021-01-29T19:33:59.077Z [INFO] TestUiIndex: Started HTTP server: address=127.0.0.1:29845 network=tcp writer.go:29: 2021-01-29T19:33:59.078Z [INFO] TestUiIndex: started state syncer === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.071Z [INFO] TestUiNodes.server.serf.wan: serf: EventMemberJoin: Node-17af74e8-1b49-24c8-31ed-e6976539085d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:59.082Z [INFO] TestUiNodes.server.serf.lan: serf: EventMemberJoin: Node-17af74e8-1b49-24c8-31ed-e6976539085d 127.0.0.1 writer.go:29: 2021-01-29T19:33:59.086Z [INFO] TestUiNodes.server: Adding LAN server: server="Node-17af74e8-1b49-24c8-31ed-e6976539085d (Addr: tcp/127.0.0.1:29861) (DC: dc1)" writer.go:29: 2021-01-29T19:33:59.087Z [INFO] TestUiNodes.server: Handled event for server in area: event=member-join server=Node-17af74e8-1b49-24c8-31ed-e6976539085d.dc1 area=wan writer.go:29: 2021-01-29T19:33:59.089Z [INFO] TestUiNodes: Started DNS server: address=127.0.0.1:29856 network=tcp writer.go:29: 2021-01-29T19:33:59.090Z [INFO] TestUiNodes: Started DNS server: address=127.0.0.1:29856 network=udp writer.go:29: 2021-01-29T19:33:59.103Z [INFO] TestUiNodes: Started HTTP server: address=127.0.0.1:29857 network=tcp writer.go:29: 2021-01-29T19:33:59.105Z [INFO] TestUiNodes: started state syncer === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.111Z [WARN] TestUiIndex.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:59.113Z [INFO] TestUiIndex.server.raft: entering candidate state: node="Node at 127.0.0.1:29849 [Candidate]" term=2 === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:59.117Z [INFO] TestUiNodes_Filter: Requesting shutdown writer.go:29: 2021-01-29T19:33:59.119Z [INFO] TestUiNodes_Filter.server: shutting down server writer.go:29: 2021-01-29T19:33:59.121Z [DEBUG] TestUiNodes_Filter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.123Z [WARN] TestUiNodes_Filter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:59.123Z [DEBUG] TestUiNodes_Filter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.121Z [ERROR] TestUiNodes_Filter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:59.131Z [WARN] TestUiNodes_Filter.server.serf.wan: serf: Shutdown without a Leave === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.131Z [DEBUG] TestUiIndex.server.raft: votes: needed=1 === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:59.133Z [INFO] TestUiNodes_Filter.server.router.manager: shutting down === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.133Z [DEBUG] TestUiIndex.server.raft: vote granted: from=b58db6c4-bc40-611e-bf6f-2e16edd2e87f term=2 tally=1 === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:59.134Z [INFO] TestUiNodes_Filter: consul server down === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.134Z [INFO] TestUiIndex.server.raft: election won: tally=1 === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:59.135Z [INFO] TestUiNodes_Filter: shutdown complete writer.go:29: 2021-01-29T19:33:59.136Z [INFO] TestUiNodes_Filter: Stopping server: protocol=DNS address=127.0.0.1:29850 network=tcp === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.136Z [INFO] TestUiIndex.server.raft: entering leader state: leader="Node at 127.0.0.1:29849 [Leader]" writer.go:29: 2021-01-29T19:33:59.138Z [INFO] TestUiIndex.server: cluster leadership acquired === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.138Z [WARN] TestUiNodes.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.139Z [INFO] TestUiIndex.server: New leader elected: payload=Node-b58db6c4-bc40-611e-bf6f-2e16edd2e87f === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:59.137Z [INFO] TestUiNodes_Filter: Stopping server: protocol=DNS address=127.0.0.1:29850 network=udp writer.go:29: 2021-01-29T19:33:59.140Z [INFO] TestUiNodes_Filter: Stopping server: protocol=HTTP address=127.0.0.1:29851 network=tcp === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.139Z [INFO] TestUiNodes.server.raft: entering candidate state: node="Node at 127.0.0.1:29861 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:59.143Z [DEBUG] TestUiNodes.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:59.145Z [DEBUG] TestUiNodes.server.raft: vote granted: from=17af74e8-1b49-24c8-31ed-e6976539085d term=2 tally=1 writer.go:29: 2021-01-29T19:33:59.147Z [INFO] TestUiNodes.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:59.151Z [INFO] TestUiNodes.server.raft: entering leader state: leader="Node at 127.0.0.1:29861 [Leader]" === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.148Z [DEBUG] TestUiIndex.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29849 === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.154Z [INFO] TestUiNodes.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:59.155Z [INFO] TestUiNodes.server: New leader elected: payload=Node-17af74e8-1b49-24c8-31ed-e6976539085d === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.159Z [INFO] TestUiIndex: Synced node info writer.go:29: 2021-01-29T19:33:59.179Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.178Z [DEBUG] TestUiNodes.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29861 writer.go:29: 2021-01-29T19:33:59.185Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.185Z [INFO] TestUiIndex.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:59.187Z [INFO] TestUiIndex.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.189Z [DEBUG] TestUiIndex.server: Skipping self join check for node since the cluster is too small: node=Node-b58db6c4-bc40-611e-bf6f-2e16edd2e87f writer.go:29: 2021-01-29T19:33:59.192Z [INFO] TestUiIndex.server: member joined, marking health alive: member=Node-b58db6c4-bc40-611e-bf6f-2e16edd2e87f === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.206Z [INFO] TestUiNodes.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:59.208Z [INFO] TestUiNodes.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.209Z [DEBUG] TestUiNodes.server: Skipping self join check for node since the cluster is too small: node=Node-17af74e8-1b49-24c8-31ed-e6976539085d writer.go:29: 2021-01-29T19:33:59.211Z [INFO] TestUiNodes.server: member joined, marking health alive: member=Node-17af74e8-1b49-24c8-31ed-e6976539085d === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.264Z [INFO] TestUiIndex: Requesting shutdown writer.go:29: 2021-01-29T19:33:59.264Z [INFO] TestUiIndex.server: shutting down server writer.go:29: 2021-01-29T19:33:59.266Z [DEBUG] TestUiIndex.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.267Z [DEBUG] TestUiIndex.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.271Z [WARN] TestUiIndex.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:59.275Z [WARN] TestUiIndex.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:59.278Z [INFO] TestUiIndex.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:59.278Z [INFO] TestUiIndex: consul server down writer.go:29: 2021-01-29T19:33:59.282Z [INFO] TestUiIndex: shutdown complete writer.go:29: 2021-01-29T19:33:59.284Z [INFO] TestUiIndex: Stopping server: protocol=DNS address=127.0.0.1:29844 network=tcp writer.go:29: 2021-01-29T19:33:59.286Z [INFO] TestUiIndex: Stopping server: protocol=DNS address=127.0.0.1:29844 network=udp writer.go:29: 2021-01-29T19:33:59.287Z [INFO] TestUiIndex: Stopping server: protocol=HTTP address=127.0.0.1:29845 network=tcp === CONT TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:59.314Z [INFO] TestTxnEndpoint_KV_Actions/#00: Requesting shutdown writer.go:29: 2021-01-29T19:33:59.315Z [INFO] TestTxnEndpoint_KV_Actions/#00.server: shutting down server writer.go:29: 2021-01-29T19:33:59.316Z [DEBUG] TestTxnEndpoint_KV_Actions/#00.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.317Z [WARN] TestTxnEndpoint_KV_Actions/#00.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:59.317Z [DEBUG] TestTxnEndpoint_KV_Actions/#00.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.320Z [WARN] TestTxnEndpoint_KV_Actions/#00.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:59.322Z [INFO] TestTxnEndpoint_KV_Actions/#00.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:59.323Z [INFO] TestTxnEndpoint_KV_Actions/#00: consul server down writer.go:29: 2021-01-29T19:33:59.326Z [INFO] TestTxnEndpoint_KV_Actions/#00: shutdown complete writer.go:29: 2021-01-29T19:33:59.327Z [INFO] TestTxnEndpoint_KV_Actions/#00: Stopping server: protocol=DNS address=127.0.0.1:29826 network=tcp writer.go:29: 2021-01-29T19:33:59.328Z [INFO] TestTxnEndpoint_KV_Actions/#00: Stopping server: protocol=DNS address=127.0.0.1:29826 network=udp writer.go:29: 2021-01-29T19:33:59.329Z [INFO] TestTxnEndpoint_KV_Actions/#00: Stopping server: protocol=HTTP address=127.0.0.1:29827 network=tcp === CONT TestUiNodes writer.go:29: 2021-01-29T19:33:59.394Z [DEBUG] TestUiNodes: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:59.397Z [INFO] TestUiNodes: Synced node info writer.go:29: 2021-01-29T19:33:59.564Z [INFO] TestUiNodes: Requesting shutdown writer.go:29: 2021-01-29T19:33:59.566Z [INFO] TestUiNodes.server: shutting down server writer.go:29: 2021-01-29T19:33:59.567Z [DEBUG] TestUiNodes.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.569Z [DEBUG] TestUiNodes.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.569Z [WARN] TestUiNodes.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:59.573Z [WARN] TestUiNodes.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:59.577Z [INFO] TestUiNodes.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:59.579Z [INFO] TestUiNodes: consul server down writer.go:29: 2021-01-29T19:33:59.586Z [INFO] TestUiNodes: shutdown complete writer.go:29: 2021-01-29T19:33:59.590Z [INFO] TestUiNodes: Stopping server: protocol=DNS address=127.0.0.1:29856 network=tcp writer.go:29: 2021-01-29T19:33:59.595Z [INFO] TestUiNodes: Stopping server: protocol=DNS address=127.0.0.1:29856 network=udp writer.go:29: 2021-01-29T19:33:59.596Z [INFO] TestUiNodes: Stopping server: protocol=HTTP address=127.0.0.1:29857 network=tcp === CONT TestUiNodes_Filter writer.go:29: 2021-01-29T19:33:59.641Z [INFO] TestUiNodes_Filter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:59.642Z [INFO] TestUiNodes_Filter: Endpoints down --- PASS: TestUiNodes_Filter (0.76s) === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:33:59.682Z [WARN] TestTxnEndpoint_UpdateCheck: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:59.685Z [DEBUG] TestTxnEndpoint_UpdateCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:59.687Z [DEBUG] TestTxnEndpoint_UpdateCheck.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.790Z [INFO] TestUiIndex: Waiting for endpoints to shut down === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:33:59.795Z [INFO] TestTxnEndpoint_UpdateCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:21558cfd-1a69-94ab-b405-8fe415b1ded9 Address:127.0.0.1:29879}]" === CONT TestUiIndex writer.go:29: 2021-01-29T19:33:59.795Z [INFO] TestUiIndex: Endpoints down --- PASS: TestUiIndex (0.78s) === CONT TestTxnEndpoint_Bad_Size_Item === RUN TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:33:59.800Z [INFO] TestTxnEndpoint_UpdateCheck.server.serf.wan: serf: EventMemberJoin: Node-21558cfd-1a69-94ab-b405-8fe415b1ded9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:59.805Z [INFO] TestTxnEndpoint_UpdateCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:29879 [Follower]" leader= === CONT TestTxnEndpoint_KV_Actions/#00 writer.go:29: 2021-01-29T19:33:59.830Z [INFO] TestTxnEndpoint_KV_Actions/#00: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:59.832Z [INFO] TestTxnEndpoint_KV_Actions/#00: Endpoints down === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:33:59.838Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:59.841Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:59.845Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.tlsutil: OutgoingRPCWrapper: version=1 === RUN TestTxnEndpoint_KV_Actions/#01 === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:33:59.863Z [WARN] TestTxnEndpoint_UpdateCheck.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:59.866Z [INFO] TestTxnEndpoint_UpdateCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:29879 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:59.869Z [WARN] TestTxnEndpoint_UpdateCheck.server.raft: unable to get address for sever, using fallback address: id=21558cfd-1a69-94ab-b405-8fe415b1ded9 fallback=127.0.0.1:29879 error="Could not find address for server id 21558cfd-1a69-94ab-b405-8fe415b1ded9" writer.go:29: 2021-01-29T19:33:59.871Z [DEBUG] TestTxnEndpoint_UpdateCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:59.872Z [DEBUG] TestTxnEndpoint_UpdateCheck.server.raft: vote granted: from=21558cfd-1a69-94ab-b405-8fe415b1ded9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:59.874Z [INFO] TestTxnEndpoint_UpdateCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:59.875Z [INFO] TestTxnEndpoint_UpdateCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:29879 [Leader]" === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:33:59.877Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5b60b00c-a4d9-e820-a103-7b4718e10307 Address:127.0.0.1:29885}]" writer.go:29: 2021-01-29T19:33:59.881Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.serf.wan: serf: EventMemberJoin: Node-5b60b00c-a4d9-e820-a103-7b4718e10307.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:59.896Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.serf.lan: serf: EventMemberJoin: Node-5b60b00c-a4d9-e820-a103-7b4718e10307 127.0.0.1 writer.go:29: 2021-01-29T19:33:59.892Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.raft: entering follower state: follower="Node at 127.0.0.1:29885 [Follower]" leader= writer.go:29: 2021-01-29T19:33:59.922Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server: Handled event for server in area: event=member-join server=Node-5b60b00c-a4d9-e820-a103-7b4718e10307.dc1 area=wan writer.go:29: 2021-01-29T19:33:59.924Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server: Adding LAN server: server="Node-5b60b00c-a4d9-e820-a103-7b4718e10307 (Addr: tcp/127.0.0.1:29885) (DC: dc1)" writer.go:29: 2021-01-29T19:33:59.926Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Started DNS server: address=127.0.0.1:29880 network=tcp === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:33:59.940Z [WARN] TestTxnEndpoint_KV_Actions/#01: bootstrap = true: do not enable unless necessary === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:33:59.942Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Started DNS server: address=127.0.0.1:29880 network=udp writer.go:29: 2021-01-29T19:33:59.952Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:59.953Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.raft: entering candidate state: node="Node at 127.0.0.1:29885 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:59.956Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Started HTTP server: address=127.0.0.1:29881 network=tcp writer.go:29: 2021-01-29T19:33:59.958Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: started state syncer === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:33:59.941Z [DEBUG] TestTxnEndpoint_KV_Actions/#01.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:59.975Z [DEBUG] TestTxnEndpoint_KV_Actions/#01.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:33:59.985Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:59.987Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.raft: vote granted: from=5b60b00c-a4d9-e820-a103-7b4718e10307 term=2 tally=1 writer.go:29: 2021-01-29T19:33:59.988Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:59.989Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.raft: entering leader state: leader="Node at 127.0.0.1:29885 [Leader]" writer.go:29: 2021-01-29T19:34:00.004Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:00.008Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server: New leader elected: payload=Node-5b60b00c-a4d9-e820-a103-7b4718e10307 writer.go:29: 2021-01-29T19:34:00.017Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29885 === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.058Z [INFO] TestTxnEndpoint_KV_Actions/#01.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:78854108-297c-4f75-2543-cf02611796fd Address:127.0.0.1:29873}]" writer.go:29: 2021-01-29T19:34:00.077Z [INFO] TestTxnEndpoint_KV_Actions/#01.server.raft: entering follower state: follower="Node at 127.0.0.1:29873 [Follower]" leader= writer.go:29: 2021-01-29T19:34:00.086Z [INFO] TestTxnEndpoint_KV_Actions/#01.server.serf.wan: serf: EventMemberJoin: Node-78854108-297c-4f75-2543-cf02611796fd.dc1 127.0.0.1 === CONT TestUiNodes writer.go:29: 2021-01-29T19:34:00.104Z [INFO] TestUiNodes: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:00.106Z [INFO] TestUiNodes: Endpoints down --- PASS: TestUiNodes (1.11s) === CONT TestTxnEndpoint_Bad_Size_Ops === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:34:00.110Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.101Z [INFO] TestTxnEndpoint_KV_Actions/#01.server.serf.lan: serf: EventMemberJoin: Node-78854108-297c-4f75-2543-cf02611796fd 127.0.0.1 writer.go:29: 2021-01-29T19:34:00.131Z [INFO] TestTxnEndpoint_KV_Actions/#01: Started DNS server: address=127.0.0.1:29868 network=udp writer.go:29: 2021-01-29T19:34:00.163Z [INFO] TestTxnEndpoint_KV_Actions/#01.server: Handled event for server in area: event=member-join server=Node-78854108-297c-4f75-2543-cf02611796fd.dc1 area=wan writer.go:29: 2021-01-29T19:34:00.182Z [INFO] TestTxnEndpoint_KV_Actions/#01.server: Adding LAN server: server="Node-78854108-297c-4f75-2543-cf02611796fd (Addr: tcp/127.0.0.1:29873) (DC: dc1)" writer.go:29: 2021-01-29T19:34:00.182Z [WARN] TestTxnEndpoint_KV_Actions/#01.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:00.195Z [INFO] TestTxnEndpoint_KV_Actions/#01.server.raft: entering candidate state: node="Node at 127.0.0.1:29873 [Candidate]" term=2 === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:34:00.220Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Synced node info writer.go:29: 2021-01-29T19:34:00.221Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Node info in sync === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.219Z [INFO] TestTxnEndpoint_KV_Actions/#01: Started DNS server: address=127.0.0.1:29868 network=tcp writer.go:29: 2021-01-29T19:34:00.238Z [INFO] TestTxnEndpoint_KV_Actions/#01: Started HTTP server: address=127.0.0.1:29869 network=tcp writer.go:29: 2021-01-29T19:34:00.239Z [INFO] TestTxnEndpoint_KV_Actions/#01: started state syncer writer.go:29: 2021-01-29T19:34:00.244Z [DEBUG] TestTxnEndpoint_KV_Actions/#01.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:00.256Z [DEBUG] TestTxnEndpoint_KV_Actions/#01.server.raft: vote granted: from=78854108-297c-4f75-2543-cf02611796fd term=2 tally=1 === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.258Z [INFO] TestTxnEndpoint_UpdateCheck.server.serf.lan: serf: EventMemberJoin: Node-21558cfd-1a69-94ab-b405-8fe415b1ded9 127.0.0.1 === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.259Z [INFO] TestTxnEndpoint_KV_Actions/#01.server.raft: election won: tally=1 === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.264Z [INFO] TestTxnEndpoint_UpdateCheck: Started DNS server: address=127.0.0.1:29874 network=udp === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.264Z [INFO] TestTxnEndpoint_KV_Actions/#01.server.raft: entering leader state: leader="Node at 127.0.0.1:29873 [Leader]" writer.go:29: 2021-01-29T19:34:00.267Z [INFO] TestTxnEndpoint_KV_Actions/#01.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:00.278Z [INFO] TestTxnEndpoint_KV_Actions/#01.server: New leader elected: payload=Node-78854108-297c-4f75-2543-cf02611796fd === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:34:00.276Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.282Z [DEBUG] TestTxnEndpoint_KV_Actions/#01.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29873 === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:34:00.283Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.leader: started routine: routine="CA root pruning" === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.276Z [INFO] TestTxnEndpoint_UpdateCheck.server: Adding LAN server: server="Node-21558cfd-1a69-94ab-b405-8fe415b1ded9 (Addr: tcp/127.0.0.1:29879) (DC: dc1)" === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:34:00.286Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server: Skipping self join check for node since the cluster is too small: node=Node-5b60b00c-a4d9-e820-a103-7b4718e10307 writer.go:29: 2021-01-29T19:34:00.292Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server: member joined, marking health alive: member=Node-5b60b00c-a4d9-e820-a103-7b4718e10307 === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.278Z [INFO] TestTxnEndpoint_UpdateCheck: Started DNS server: address=127.0.0.1:29874 network=tcp === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.296Z [WARN] TestTxnEndpoint_Bad_Size_Ops: bootstrap = true: do not enable unless necessary === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.297Z [INFO] TestTxnEndpoint_UpdateCheck.server: cluster leadership acquired === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.296Z [DEBUG] TestTxnEndpoint_Bad_Size_Ops.tlsutil: Update: version=1 === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.298Z [INFO] TestTxnEndpoint_UpdateCheck.server: New leader elected: payload=Node-21558cfd-1a69-94ab-b405-8fe415b1ded9 writer.go:29: 2021-01-29T19:34:00.279Z [INFO] TestTxnEndpoint_UpdateCheck.server: Handled event for server in area: event=member-join server=Node-21558cfd-1a69-94ab-b405-8fe415b1ded9.dc1 area=wan === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.299Z [DEBUG] TestTxnEndpoint_Bad_Size_Ops.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.301Z [INFO] TestTxnEndpoint_UpdateCheck: Started HTTP server: address=127.0.0.1:29875 network=tcp writer.go:29: 2021-01-29T19:34:00.302Z [INFO] TestTxnEndpoint_UpdateCheck: started state syncer === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.306Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.303Z [DEBUG] TestTxnEndpoint_UpdateCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29879 === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.308Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3829b1d6-ddd7-cb01-cc8e-24f22f6d9a3c Address:127.0.0.1:29867}]" writer.go:29: 2021-01-29T19:34:00.310Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server.serf.wan: serf: EventMemberJoin: Node-3829b1d6-ddd7-cb01-cc8e-24f22f6d9a3c.dc1 127.0.0.1 === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.311Z [INFO] TestTxnEndpoint_UpdateCheck: Synced node info === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.311Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server.raft: entering follower state: follower="Node at 127.0.0.1:29867 [Follower]" leader= === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.314Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.316Z [INFO] TestTxnEndpoint_KV_Actions/#01.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.313Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server.serf.lan: serf: EventMemberJoin: Node-3829b1d6-ddd7-cb01-cc8e-24f22f6d9a3c 127.0.0.1 === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.318Z [INFO] TestTxnEndpoint_KV_Actions/#01.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:00.319Z [DEBUG] TestTxnEndpoint_KV_Actions/#01.server: Skipping self join check for node since the cluster is too small: node=Node-78854108-297c-4f75-2543-cf02611796fd === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.322Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server: Adding LAN server: server="Node-3829b1d6-ddd7-cb01-cc8e-24f22f6d9a3c (Addr: tcp/127.0.0.1:29867) (DC: dc1)" writer.go:29: 2021-01-29T19:34:00.324Z [INFO] TestTxnEndpoint_Bad_Size_Ops: Started DNS server: address=127.0.0.1:29862 network=udp writer.go:29: 2021-01-29T19:34:00.322Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server: Handled event for server in area: event=member-join server=Node-3829b1d6-ddd7-cb01-cc8e-24f22f6d9a3c.dc1 area=wan writer.go:29: 2021-01-29T19:34:00.326Z [INFO] TestTxnEndpoint_Bad_Size_Ops: Started DNS server: address=127.0.0.1:29862 network=tcp writer.go:29: 2021-01-29T19:34:00.330Z [INFO] TestTxnEndpoint_Bad_Size_Ops: Started HTTP server: address=127.0.0.1:29863 network=tcp === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.321Z [INFO] TestTxnEndpoint_KV_Actions/#01.server: member joined, marking health alive: member=Node-78854108-297c-4f75-2543-cf02611796fd === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.333Z [INFO] TestTxnEndpoint_Bad_Size_Ops: started state syncer === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.338Z [INFO] TestTxnEndpoint_UpdateCheck.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:00.340Z [INFO] TestTxnEndpoint_UpdateCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:00.341Z [DEBUG] TestTxnEndpoint_UpdateCheck.server: Skipping self join check for node since the cluster is too small: node=Node-21558cfd-1a69-94ab-b405-8fe415b1ded9 writer.go:29: 2021-01-29T19:34:00.343Z [INFO] TestTxnEndpoint_UpdateCheck.server: member joined, marking health alive: member=Node-21558cfd-1a69-94ab-b405-8fe415b1ded9 writer.go:29: 2021-01-29T19:34:00.346Z [DEBUG] TestTxnEndpoint_UpdateCheck.server: Skipping self join check for node since the cluster is too small: node=Node-21558cfd-1a69-94ab-b405-8fe415b1ded9 === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.354Z [WARN] TestTxnEndpoint_Bad_Size_Ops.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:00.356Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server.raft: entering candidate state: node="Node at 127.0.0.1:29867 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:00.358Z [DEBUG] TestTxnEndpoint_Bad_Size_Ops.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:00.360Z [DEBUG] TestTxnEndpoint_Bad_Size_Ops.server.raft: vote granted: from=3829b1d6-ddd7-cb01-cc8e-24f22f6d9a3c term=2 tally=1 writer.go:29: 2021-01-29T19:34:00.361Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:00.363Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server.raft: entering leader state: leader="Node at 127.0.0.1:29867 [Leader]" writer.go:29: 2021-01-29T19:34:00.364Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:00.365Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server: New leader elected: payload=Node-3829b1d6-ddd7-cb01-cc8e-24f22f6d9a3c === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.380Z [INFO] TestTxnEndpoint_UpdateCheck: Requesting shutdown === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.371Z [DEBUG] TestTxnEndpoint_Bad_Size_Ops.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29867 === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.382Z [INFO] TestTxnEndpoint_UpdateCheck.server: shutting down server writer.go:29: 2021-01-29T19:34:00.385Z [DEBUG] TestTxnEndpoint_UpdateCheck.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:00.386Z [WARN] TestTxnEndpoint_UpdateCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:00.389Z [DEBUG] TestTxnEndpoint_UpdateCheck.leader: stopped routine: routine="CA root pruning" === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.393Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.392Z [WARN] TestTxnEndpoint_UpdateCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:00.396Z [INFO] TestTxnEndpoint_UpdateCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:00.396Z [INFO] TestTxnEndpoint_UpdateCheck: consul server down writer.go:29: 2021-01-29T19:34:00.399Z [INFO] TestTxnEndpoint_UpdateCheck: shutdown complete writer.go:29: 2021-01-29T19:34:00.401Z [INFO] TestTxnEndpoint_UpdateCheck: Stopping server: protocol=DNS address=127.0.0.1:29874 network=tcp writer.go:29: 2021-01-29T19:34:00.403Z [INFO] TestTxnEndpoint_UpdateCheck: Stopping server: protocol=DNS address=127.0.0.1:29874 network=udp writer.go:29: 2021-01-29T19:34:00.404Z [INFO] TestTxnEndpoint_UpdateCheck: Stopping server: protocol=HTTP address=127.0.0.1:29875 network=tcp === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.406Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:00.407Z [INFO] TestTxnEndpoint_Bad_Size_Ops.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:00.409Z [DEBUG] TestTxnEndpoint_Bad_Size_Ops.server: Skipping self join check for node since the cluster is too small: node=Node-3829b1d6-ddd7-cb01-cc8e-24f22f6d9a3c writer.go:29: 2021-01-29T19:34:00.413Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server: member joined, marking health alive: member=Node-3829b1d6-ddd7-cb01-cc8e-24f22f6d9a3c writer.go:29: 2021-01-29T19:34:00.436Z [INFO] TestTxnEndpoint_Bad_Size_Ops: Synced node info === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.497Z [DEBUG] TestTxnEndpoint_KV_Actions/#01: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:00.500Z [INFO] TestTxnEndpoint_KV_Actions/#01: Synced node info === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:34:00.536Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Requesting shutdown writer.go:29: 2021-01-29T19:34:00.541Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server: shutting down server writer.go:29: 2021-01-29T19:34:00.543Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:00.545Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:00.545Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:00.552Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:00.561Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:00.563Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: consul server down writer.go:29: 2021-01-29T19:34:00.565Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: shutdown complete writer.go:29: 2021-01-29T19:34:00.566Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Stopping server: protocol=DNS address=127.0.0.1:29880 network=tcp writer.go:29: 2021-01-29T19:34:00.569Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Stopping server: protocol=DNS address=127.0.0.1:29880 network=udp writer.go:29: 2021-01-29T19:34:00.571Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Stopping server: protocol=HTTP address=127.0.0.1:29881 network=tcp === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.581Z [INFO] TestTxnEndpoint_Bad_Size_Ops: Requesting shutdown writer.go:29: 2021-01-29T19:34:00.582Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server: shutting down server writer.go:29: 2021-01-29T19:34:00.583Z [DEBUG] TestTxnEndpoint_Bad_Size_Ops.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:00.585Z [WARN] TestTxnEndpoint_Bad_Size_Ops.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:00.586Z [DEBUG] TestTxnEndpoint_Bad_Size_Ops.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:00.587Z [WARN] TestTxnEndpoint_Bad_Size_Ops.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:00.591Z [INFO] TestTxnEndpoint_Bad_Size_Ops.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:00.591Z [INFO] TestTxnEndpoint_Bad_Size_Ops: consul server down writer.go:29: 2021-01-29T19:34:00.602Z [INFO] TestTxnEndpoint_Bad_Size_Ops: shutdown complete writer.go:29: 2021-01-29T19:34:00.603Z [INFO] TestTxnEndpoint_Bad_Size_Ops: Stopping server: protocol=DNS address=127.0.0.1:29862 network=tcp === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.608Z [INFO] TestTxnEndpoint_KV_Actions/#01: Requesting shutdown writer.go:29: 2021-01-29T19:34:00.609Z [INFO] TestTxnEndpoint_KV_Actions/#01.server: shutting down server writer.go:29: 2021-01-29T19:34:00.611Z [DEBUG] TestTxnEndpoint_KV_Actions/#01.leader: stopping routine: routine="CA root pruning" === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:00.605Z [INFO] TestTxnEndpoint_Bad_Size_Ops: Stopping server: protocol=DNS address=127.0.0.1:29862 network=udp writer.go:29: 2021-01-29T19:34:00.616Z [INFO] TestTxnEndpoint_Bad_Size_Ops: Stopping server: protocol=HTTP address=127.0.0.1:29863 network=tcp === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:00.612Z [WARN] TestTxnEndpoint_KV_Actions/#01.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:00.613Z [DEBUG] TestTxnEndpoint_KV_Actions/#01.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:00.626Z [WARN] TestTxnEndpoint_KV_Actions/#01.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:00.629Z [INFO] TestTxnEndpoint_KV_Actions/#01.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:00.629Z [INFO] TestTxnEndpoint_KV_Actions/#01: consul server down writer.go:29: 2021-01-29T19:34:00.632Z [INFO] TestTxnEndpoint_KV_Actions/#01: shutdown complete writer.go:29: 2021-01-29T19:34:00.633Z [INFO] TestTxnEndpoint_KV_Actions/#01: Stopping server: protocol=DNS address=127.0.0.1:29868 network=tcp writer.go:29: 2021-01-29T19:34:00.635Z [INFO] TestTxnEndpoint_KV_Actions/#01: Stopping server: protocol=DNS address=127.0.0.1:29868 network=udp writer.go:29: 2021-01-29T19:34:00.637Z [INFO] TestTxnEndpoint_KV_Actions/#01: Stopping server: protocol=HTTP address=127.0.0.1:29869 network=tcp === CONT TestTxnEndpoint_UpdateCheck writer.go:29: 2021-01-29T19:34:00.906Z [INFO] TestTxnEndpoint_UpdateCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:00.908Z [INFO] TestTxnEndpoint_UpdateCheck: Endpoints down --- PASS: TestTxnEndpoint_UpdateCheck (1.27s) === CONT TestTxnEndpoint_Bad_Size_Net === RUN TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits writer.go:29: 2021-01-29T19:34:00.938Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:00.940Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:00.942Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:00.955Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7d48236c-4330-38f5-022b-51f1bed113b3 Address:127.0.0.1:29891}]" writer.go:29: 2021-01-29T19:34:00.958Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.raft: entering follower state: follower="Node at 127.0.0.1:29891 [Follower]" leader= writer.go:29: 2021-01-29T19:34:00.962Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.serf.wan: serf: EventMemberJoin: Node-7d48236c-4330-38f5-022b-51f1bed113b3.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:00.966Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.serf.lan: serf: EventMemberJoin: Node-7d48236c-4330-38f5-022b-51f1bed113b3 127.0.0.1 writer.go:29: 2021-01-29T19:34:00.970Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Started DNS server: address=127.0.0.1:29886 network=udp writer.go:29: 2021-01-29T19:34:00.972Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server: Handled event for server in area: event=member-join server=Node-7d48236c-4330-38f5-022b-51f1bed113b3.dc1 area=wan writer.go:29: 2021-01-29T19:34:00.975Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server: Adding LAN server: server="Node-7d48236c-4330-38f5-022b-51f1bed113b3 (Addr: tcp/127.0.0.1:29891) (DC: dc1)" writer.go:29: 2021-01-29T19:34:00.979Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Started DNS server: address=127.0.0.1:29886 network=tcp writer.go:29: 2021-01-29T19:34:00.999Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Started HTTP server: address=127.0.0.1:29887 network=tcp writer.go:29: 2021-01-29T19:34:01.006Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:01.008Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.raft: entering candidate state: node="Node at 127.0.0.1:29891 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:01.010Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: started state syncer writer.go:29: 2021-01-29T19:34:01.011Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:01.013Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.raft: vote granted: from=7d48236c-4330-38f5-022b-51f1bed113b3 term=2 tally=1 writer.go:29: 2021-01-29T19:34:01.014Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:01.016Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.raft: entering leader state: leader="Node at 127.0.0.1:29891 [Leader]" writer.go:29: 2021-01-29T19:34:01.017Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:01.018Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server: New leader elected: payload=Node-7d48236c-4330-38f5-022b-51f1bed113b3 writer.go:29: 2021-01-29T19:34:01.029Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Synced node info writer.go:29: 2021-01-29T19:34:01.030Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29891 writer.go:29: 2021-01-29T19:34:01.036Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:01.045Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:01.046Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:01.048Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server: Skipping self join check for node since the cluster is too small: node=Node-7d48236c-4330-38f5-022b-51f1bed113b3 writer.go:29: 2021-01-29T19:34:01.049Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server: member joined, marking health alive: member=Node-7d48236c-4330-38f5-022b-51f1bed113b3 === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits writer.go:29: 2021-01-29T19:34:01.073Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:01.075Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits: Endpoints down === RUN TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.109Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:01.110Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:01.113Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTxnEndpoint_Bad_Size_Ops writer.go:29: 2021-01-29T19:34:01.117Z [INFO] TestTxnEndpoint_Bad_Size_Ops: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:01.119Z [INFO] TestTxnEndpoint_Bad_Size_Ops: Endpoints down --- PASS: TestTxnEndpoint_Bad_Size_Ops (1.01s) === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.132Z [WARN] TestTxnEndpoint_Bad_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:01.134Z [DEBUG] TestTxnEndpoint_Bad_JSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:01.137Z [DEBUG] TestTxnEndpoint_Bad_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTxnEndpoint_KV_Actions/#01 writer.go:29: 2021-01-29T19:34:01.139Z [INFO] TestTxnEndpoint_KV_Actions/#01: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:01.140Z [INFO] TestTxnEndpoint_KV_Actions/#01: Endpoints down --- PASS: TestTxnEndpoint_KV_Actions (2.34s) --- PASS: TestTxnEndpoint_KV_Actions/#00 (1.03s) --- PASS: TestTxnEndpoint_KV_Actions/#01 (1.28s) === CONT TestStatusPeersSecondary === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.182Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:12b5c126-44c8-25a8-43f2-1f4351af6626 Address:127.0.0.1:29903}]" writer.go:29: 2021-01-29T19:34:01.187Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.serf.wan: serf: EventMemberJoin: Node-12b5c126-44c8-25a8-43f2-1f4351af6626.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:01.236Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.raft: entering follower state: follower="Node at 127.0.0.1:29903 [Follower]" leader= writer.go:29: 2021-01-29T19:34:01.252Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.serf.lan: serf: EventMemberJoin: Node-12b5c126-44c8-25a8-43f2-1f4351af6626 127.0.0.1 writer.go:29: 2021-01-29T19:34:01.256Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Started DNS server: address=127.0.0.1:29898 network=udp writer.go:29: 2021-01-29T19:34:01.261Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server: Adding LAN server: server="Node-12b5c126-44c8-25a8-43f2-1f4351af6626 (Addr: tcp/127.0.0.1:29903) (DC: dc1)" writer.go:29: 2021-01-29T19:34:01.262Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server: Handled event for server in area: event=member-join server=Node-12b5c126-44c8-25a8-43f2-1f4351af6626.dc1 area=wan writer.go:29: 2021-01-29T19:34:01.267Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Started DNS server: address=127.0.0.1:29898 network=tcp === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.268Z [WARN] TestStatusPeersSecondary: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:01.271Z [DEBUG] TestStatusPeersSecondary.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:01.281Z [DEBUG] TestStatusPeersSecondary.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.302Z [INFO] TestTxnEndpoint_Bad_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f22463c7-d19b-0819-c968-8636f71f007e Address:127.0.0.1:29909}]" === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.271Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Started HTTP server: address=127.0.0.1:29899 network=tcp === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.309Z [INFO] TestTxnEndpoint_Bad_JSON.server.serf.wan: serf: EventMemberJoin: Node-f22463c7-d19b-0819-c968-8636f71f007e.dc1 127.0.0.1 === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.312Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: started state syncer writer.go:29: 2021-01-29T19:34:01.295Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:01.317Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.raft: entering candidate state: node="Node at 127.0.0.1:29903 [Candidate]" term=2 === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.313Z [INFO] TestTxnEndpoint_Bad_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:29909 [Follower]" leader= writer.go:29: 2021-01-29T19:34:01.320Z [INFO] TestTxnEndpoint_Bad_JSON.server.serf.lan: serf: EventMemberJoin: Node-f22463c7-d19b-0819-c968-8636f71f007e 127.0.0.1 === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.384Z [INFO] TestStatusPeersSecondary.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:64ce3813-190e-7002-ce83-30dd33301eb8 Address:127.0.0.1:29921}]" writer.go:29: 2021-01-29T19:34:01.423Z [INFO] TestStatusPeersSecondary.server.raft: entering follower state: follower="Node at 127.0.0.1:29921 [Follower]" leader= === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.404Z [WARN] TestTxnEndpoint_Bad_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:01.428Z [INFO] TestTxnEndpoint_Bad_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:29909 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:01.419Z [INFO] TestTxnEndpoint_Bad_JSON: Started DNS server: address=127.0.0.1:29904 network=tcp === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits writer.go:29: 2021-01-29T19:34:01.431Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.433Z [INFO] TestTxnEndpoint_Bad_JSON: Started DNS server: address=127.0.0.1:29904 network=udp writer.go:29: 2021-01-29T19:34:01.440Z [INFO] TestTxnEndpoint_Bad_JSON: Started HTTP server: address=127.0.0.1:29905 network=tcp writer.go:29: 2021-01-29T19:34:01.459Z [INFO] TestTxnEndpoint_Bad_JSON: started state syncer === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits writer.go:29: 2021-01-29T19:34:01.433Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Node info in sync writer.go:29: 2021-01-29T19:34:01.503Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Node info in sync === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.442Z [INFO] TestTxnEndpoint_Bad_JSON.server: Handled event for server in area: event=member-join server=Node-f22463c7-d19b-0819-c968-8636f71f007e.dc1 area=wan writer.go:29: 2021-01-29T19:34:01.459Z [INFO] TestTxnEndpoint_Bad_JSON.server: Adding LAN server: server="Node-f22463c7-d19b-0819-c968-8636f71f007e (Addr: tcp/127.0.0.1:29909) (DC: dc1)" === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.492Z [WARN] TestStatusPeersSecondary.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:01.513Z [INFO] TestStatusPeersSecondary.server.raft: entering candidate state: node="Node at 127.0.0.1:29921 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:01.525Z [WARN] TestStatusPeersSecondary.server.raft: unable to get address for sever, using fallback address: id=64ce3813-190e-7002-ce83-30dd33301eb8 fallback=127.0.0.1:29921 error="Could not find address for server id 64ce3813-190e-7002-ce83-30dd33301eb8" === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.531Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:01.542Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.raft: vote granted: from=12b5c126-44c8-25a8-43f2-1f4351af6626 term=2 tally=1 writer.go:29: 2021-01-29T19:34:01.543Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:01.544Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.raft: entering leader state: leader="Node at 127.0.0.1:29903 [Leader]" writer.go:29: 2021-01-29T19:34:01.546Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:01.552Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server: New leader elected: payload=Node-12b5c126-44c8-25a8-43f2-1f4351af6626 === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.544Z [INFO] TestStatusPeersSecondary.server.serf.wan: serf: EventMemberJoin: Node-64ce3813-190e-7002-ce83-30dd33301eb8.primary 127.0.0.1 === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.552Z [DEBUG] TestTxnEndpoint_Bad_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:01.578Z [DEBUG] TestTxnEndpoint_Bad_JSON.server.raft: vote granted: from=f22463c7-d19b-0819-c968-8636f71f007e term=2 tally=1 writer.go:29: 2021-01-29T19:34:01.584Z [INFO] TestTxnEndpoint_Bad_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:01.586Z [INFO] TestTxnEndpoint_Bad_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:29909 [Leader]" writer.go:29: 2021-01-29T19:34:01.594Z [INFO] TestTxnEndpoint_Bad_JSON.server: cluster leadership acquired === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.646Z [DEBUG] TestStatusPeersSecondary.server.raft: votes: needed=1 === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.649Z [INFO] TestTxnEndpoint_Bad_JSON.server: New leader elected: payload=Node-f22463c7-d19b-0819-c968-8636f71f007e === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.640Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29903 === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.648Z [DEBUG] TestStatusPeersSecondary.server.raft: vote granted: from=64ce3813-190e-7002-ce83-30dd33301eb8 term=2 tally=1 === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.671Z [INFO] TestTxnEndpoint_Bad_JSON: Requesting shutdown === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.678Z [INFO] TestStatusPeersSecondary.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:01.680Z [INFO] TestStatusPeersSecondary.server.raft: entering leader state: leader="Node at 127.0.0.1:29921 [Leader]" writer.go:29: 2021-01-29T19:34:01.665Z [INFO] TestStatusPeersSecondary.server.serf.lan: serf: EventMemberJoin: Node-64ce3813-190e-7002-ce83-30dd33301eb8 127.0.0.1 writer.go:29: 2021-01-29T19:34:01.691Z [INFO] TestStatusPeersSecondary: Started DNS server: address=127.0.0.1:29916 network=udp === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.679Z [INFO] TestTxnEndpoint_Bad_JSON.server: shutting down server writer.go:29: 2021-01-29T19:34:01.696Z [WARN] TestTxnEndpoint_Bad_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:01.686Z [DEBUG] TestTxnEndpoint_Bad_JSON.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29909 writer.go:29: 2021-01-29T19:34:01.717Z [ERROR] TestTxnEndpoint_Bad_JSON.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:01.723Z [WARN] TestTxnEndpoint_Bad_JSON.server.serf.wan: serf: Shutdown without a Leave === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.723Z [INFO] TestStatusPeersSecondary.server: Adding LAN server: server="Node-64ce3813-190e-7002-ce83-30dd33301eb8 (Addr: tcp/127.0.0.1:29921) (DC: primary)" writer.go:29: 2021-01-29T19:34:01.724Z [INFO] TestStatusPeersSecondary.server: Handled event for server in area: event=member-join server=Node-64ce3813-190e-7002-ce83-30dd33301eb8.primary area=wan === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.726Z [INFO] TestTxnEndpoint_Bad_JSON.server.router.manager: shutting down === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.727Z [INFO] TestStatusPeersSecondary.server: cluster leadership acquired === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.728Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true 2021-01-29T19:34:01.730Z [ERROR] watch.watch: Watch errored: type=key error="Get "https://127.0.0.1:29774/v1/kv/asdf": dial tcp 127.0.0.1:29774: connect: connection refused" retry=20s === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.730Z [INFO] TestTxnEndpoint_Bad_JSON: consul server down writer.go:29: 2021-01-29T19:34:01.731Z [INFO] TestTxnEndpoint_Bad_JSON: shutdown complete writer.go:29: 2021-01-29T19:34:01.732Z [INFO] TestTxnEndpoint_Bad_JSON: Stopping server: protocol=DNS address=127.0.0.1:29904 network=tcp writer.go:29: 2021-01-29T19:34:01.733Z [INFO] TestTxnEndpoint_Bad_JSON: Stopping server: protocol=DNS address=127.0.0.1:29904 network=udp writer.go:29: 2021-01-29T19:34:01.733Z [ERROR] TestTxnEndpoint_Bad_JSON.server.autopilot: failed to initialize config: error="leadership lost while committing log" === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.730Z [DEBUG] TestStatusPeersSecondary.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29921 === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.734Z [INFO] TestTxnEndpoint_Bad_JSON: Stopping server: protocol=HTTP address=127.0.0.1:29905 network=tcp === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.740Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:01.744Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:01.745Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server: Skipping self join check for node since the cluster is too small: node=Node-12b5c126-44c8-25a8-43f2-1f4351af6626 writer.go:29: 2021-01-29T19:34:01.746Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server: member joined, marking health alive: member=Node-12b5c126-44c8-25a8-43f2-1f4351af6626 === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:01.737Z [ERROR] TestTxnEndpoint_Bad_JSON.server: failed to establish leadership: error="raft is already shutdown" === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.730Z [INFO] TestStatusPeersSecondary.server: New leader elected: payload=Node-64ce3813-190e-7002-ce83-30dd33301eb8 writer.go:29: 2021-01-29T19:34:01.733Z [INFO] TestStatusPeersSecondary: Started DNS server: address=127.0.0.1:29916 network=tcp writer.go:29: 2021-01-29T19:34:01.759Z [INFO] TestStatusPeersSecondary: Started HTTP server: address=127.0.0.1:29917 network=tcp writer.go:29: 2021-01-29T19:34:01.763Z [INFO] TestStatusPeersSecondary: started state syncer writer.go:29: 2021-01-29T19:34:01.769Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:01.771Z [INFO] TestStatusPeersSecondary: Synced node info === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.778Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:01.782Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Synced node info writer.go:29: 2021-01-29T19:34:01.784Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Node info in sync === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.786Z [INFO] TestStatusPeersSecondary.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:01.787Z [INFO] TestStatusPeersSecondary.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:01.787Z [WARN] TestStatusPeersSecondary: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:01.789Z [DEBUG] TestStatusPeersSecondary.server: Skipping self join check for node since the cluster is too small: node=Node-64ce3813-190e-7002-ce83-30dd33301eb8 writer.go:29: 2021-01-29T19:34:01.792Z [INFO] TestStatusPeersSecondary.server: member joined, marking health alive: member=Node-64ce3813-190e-7002-ce83-30dd33301eb8 writer.go:29: 2021-01-29T19:34:01.791Z [DEBUG] TestStatusPeersSecondary.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:01.794Z [DEBUG] TestStatusPeersSecondary.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:01.798Z [DEBUG] TestStatusPeersSecondary.server: Skipping self join check for node since the cluster is too small: node=Node-64ce3813-190e-7002-ce83-30dd33301eb8 writer.go:29: 2021-01-29T19:34:01.801Z [INFO] TestStatusPeersSecondary.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f97fbf67-9ba1-b878-bcdb-da73d8603f9a Address:127.0.0.1:29915}]" writer.go:29: 2021-01-29T19:34:01.804Z [INFO] TestStatusPeersSecondary.server.raft: entering follower state: follower="Node at 127.0.0.1:29915 [Follower]" leader= writer.go:29: 2021-01-29T19:34:01.805Z [INFO] TestStatusPeersSecondary.server.serf.wan: serf: EventMemberJoin: Node-f97fbf67-9ba1-b878-bcdb-da73d8603f9a.secondary 127.0.0.1 writer.go:29: 2021-01-29T19:34:01.811Z [INFO] TestStatusPeersSecondary.server.serf.lan: serf: EventMemberJoin: Node-f97fbf67-9ba1-b878-bcdb-da73d8603f9a 127.0.0.1 writer.go:29: 2021-01-29T19:34:01.838Z [INFO] TestStatusPeersSecondary.server: Adding LAN server: server="Node-f97fbf67-9ba1-b878-bcdb-da73d8603f9a (Addr: tcp/127.0.0.1:29915) (DC: secondary)" writer.go:29: 2021-01-29T19:34:01.839Z [INFO] TestStatusPeersSecondary.server: Handled event for server in area: event=member-join server=Node-f97fbf67-9ba1-b878-bcdb-da73d8603f9a.secondary area=wan writer.go:29: 2021-01-29T19:34:01.844Z [INFO] TestStatusPeersSecondary: Started DNS server: address=127.0.0.1:29910 network=tcp writer.go:29: 2021-01-29T19:34:01.846Z [INFO] TestStatusPeersSecondary: Started DNS server: address=127.0.0.1:29910 network=udp writer.go:29: 2021-01-29T19:34:01.850Z [INFO] TestStatusPeersSecondary: Started HTTP server: address=127.0.0.1:29911 network=tcp === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.839Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Requesting shutdown === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.851Z [INFO] TestStatusPeersSecondary: started state syncer === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.852Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server: shutting down server writer.go:29: 2021-01-29T19:34:01.854Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.leader: stopping routine: routine="CA root pruning" === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.855Z [WARN] TestStatusPeersSecondary.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:01.857Z [INFO] TestStatusPeersSecondary.server.raft: entering candidate state: node="Node at 127.0.0.1:29915 [Candidate]" term=2 === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.856Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:01.858Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:01.882Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.serf.wan: serf: Shutdown without a Leave === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.887Z [DEBUG] TestStatusPeersSecondary.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:01.888Z [DEBUG] TestStatusPeersSecondary.server.raft: vote granted: from=f97fbf67-9ba1-b878-bcdb-da73d8603f9a term=2 tally=1 writer.go:29: 2021-01-29T19:34:01.890Z [INFO] TestStatusPeersSecondary.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:01.891Z [INFO] TestStatusPeersSecondary.server.raft: entering leader state: leader="Node at 127.0.0.1:29915 [Leader]" writer.go:29: 2021-01-29T19:34:01.893Z [INFO] TestStatusPeersSecondary.server: cluster leadership acquired === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.893Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len.server.router.manager: shutting down === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.895Z [INFO] TestStatusPeersSecondary.server: New leader elected: payload=Node-f97fbf67-9ba1-b878-bcdb-da73d8603f9a === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.895Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: consul server down === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.899Z [DEBUG] TestStatusPeersSecondary.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29915 === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:01.897Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: shutdown complete writer.go:29: 2021-01-29T19:34:01.907Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Stopping server: protocol=DNS address=127.0.0.1:29898 network=tcp writer.go:29: 2021-01-29T19:34:01.908Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Stopping server: protocol=DNS address=127.0.0.1:29898 network=udp writer.go:29: 2021-01-29T19:34:01.909Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Stopping server: protocol=HTTP address=127.0.0.1:29899 network=tcp === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.918Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits writer.go:29: 2021-01-29T19:34:01.957Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Requesting shutdown writer.go:29: 2021-01-29T19:34:01.958Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server: shutting down server writer.go:29: 2021-01-29T19:34:01.959Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:01.960Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:01.961Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.leader: stopped routine: routine="CA root pruning" === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.967Z [INFO] TestStatusPeersSecondary.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits writer.go:29: 2021-01-29T19:34:01.965Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:01.973Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits.server.router.manager: shutting down === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.971Z [INFO] TestStatusPeersSecondary.leader: started routine: routine="CA root pruning" === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits writer.go:29: 2021-01-29T19:34:01.975Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: consul server down === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:01.976Z [DEBUG] TestStatusPeersSecondary.server: Skipping self join check for node since the cluster is too small: node=Node-f97fbf67-9ba1-b878-bcdb-da73d8603f9a writer.go:29: 2021-01-29T19:34:01.976Z [INFO] TestStatusPeersSecondary.server: member joined, marking health alive: member=Node-f97fbf67-9ba1-b878-bcdb-da73d8603f9a === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits writer.go:29: 2021-01-29T19:34:01.976Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: shutdown complete writer.go:29: 2021-01-29T19:34:01.978Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Stopping server: protocol=DNS address=127.0.0.1:29886 network=tcp writer.go:29: 2021-01-29T19:34:01.979Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Stopping server: protocol=DNS address=127.0.0.1:29886 network=udp writer.go:29: 2021-01-29T19:34:01.981Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Stopping server: protocol=HTTP address=127.0.0.1:29887 network=tcp === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:02.192Z [INFO] TestStatusPeersSecondary: (WAN) joining: wan_addresses=[127.0.0.1:29920] writer.go:29: 2021-01-29T19:34:02.195Z [DEBUG] TestStatusPeersSecondary.server.memberlist.wan: memberlist: Initiating push/pull sync with: 127.0.0.1:29920 writer.go:29: 2021-01-29T19:34:02.196Z [DEBUG] TestStatusPeersSecondary.server.memberlist.wan: memberlist: Stream connection from=127.0.0.1:58986 writer.go:29: 2021-01-29T19:34:02.199Z [INFO] TestStatusPeersSecondary.server.serf.wan: serf: EventMemberJoin: Node-f97fbf67-9ba1-b878-bcdb-da73d8603f9a.secondary 127.0.0.1 writer.go:29: 2021-01-29T19:34:02.203Z [INFO] TestStatusPeersSecondary.server: Handled event for server in area: event=member-join server=Node-f97fbf67-9ba1-b878-bcdb-da73d8603f9a.secondary area=wan writer.go:29: 2021-01-29T19:34:02.203Z [INFO] TestStatusPeersSecondary.server.serf.wan: serf: EventMemberJoin: Node-64ce3813-190e-7002-ce83-30dd33301eb8.primary 127.0.0.1 writer.go:29: 2021-01-29T19:34:02.205Z [INFO] TestStatusPeersSecondary: (WAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:34:02.205Z [INFO] TestStatusPeersSecondary.server: Handled event for server in area: event=member-join server=Node-64ce3813-190e-7002-ce83-30dd33301eb8.primary area=wan writer.go:29: 2021-01-29T19:34:02.208Z [DEBUG] TestStatusPeersSecondary.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:02.218Z [DEBUG] TestStatusPeersSecondary.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:02.231Z [INFO] TestStatusPeersSecondary: Requesting shutdown writer.go:29: 2021-01-29T19:34:02.233Z [INFO] TestStatusPeersSecondary.server: shutting down server writer.go:29: 2021-01-29T19:34:02.236Z [DEBUG] TestStatusPeersSecondary.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:02.236Z [ERROR] TestStatusPeersSecondary.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:02.238Z [DEBUG] TestStatusPeersSecondary.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:02.238Z [WARN] TestStatusPeersSecondary.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:02.242Z [WARN] TestStatusPeersSecondary.server.serf.wan: serf: Shutdown without a Leave === CONT TestTxnEndpoint_Bad_JSON writer.go:29: 2021-01-29T19:34:02.245Z [INFO] TestTxnEndpoint_Bad_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:02.246Z [INFO] TestTxnEndpoint_Bad_JSON: Endpoints down --- PASS: TestTxnEndpoint_Bad_JSON (1.13s) === CONT TestAgent_FailCheck === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:02.246Z [INFO] TestStatusPeersSecondary.server.router.manager: shutting down === CONT TestAgent_FailCheck writer.go:29: 2021-01-29T19:34:02.254Z [WARN] TestAgent_FailCheck: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:02.255Z [DEBUG] TestAgent_FailCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:02.256Z [DEBUG] TestAgent_FailCheck.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:02.247Z [INFO] TestStatusPeersSecondary.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:02.247Z [INFO] TestStatusPeersSecondary: consul server down writer.go:29: 2021-01-29T19:34:02.260Z [INFO] TestStatusPeersSecondary: shutdown complete writer.go:29: 2021-01-29T19:34:02.262Z [INFO] TestStatusPeersSecondary: Stopping server: protocol=DNS address=127.0.0.1:29910 network=tcp writer.go:29: 2021-01-29T19:34:02.265Z [INFO] TestStatusPeersSecondary: Stopping server: protocol=DNS address=127.0.0.1:29910 network=udp writer.go:29: 2021-01-29T19:34:02.268Z [INFO] TestStatusPeersSecondary: Stopping server: protocol=HTTP address=127.0.0.1:29911 network=tcp === CONT TestAgent_FailCheck writer.go:29: 2021-01-29T19:34:02.274Z [INFO] TestAgent_FailCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6df46a0f-abad-a8a6-8f2c-392d694279a3 Address:127.0.0.1:29897}]" writer.go:29: 2021-01-29T19:34:02.276Z [INFO] TestAgent_FailCheck.server.serf.wan: serf: EventMemberJoin: Node-6df46a0f-abad-a8a6-8f2c-392d694279a3.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:02.277Z [INFO] TestAgent_FailCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:29897 [Follower]" leader= writer.go:29: 2021-01-29T19:34:02.279Z [INFO] TestAgent_FailCheck.server.serf.lan: serf: EventMemberJoin: Node-6df46a0f-abad-a8a6-8f2c-392d694279a3 127.0.0.1 writer.go:29: 2021-01-29T19:34:02.288Z [INFO] TestAgent_FailCheck.server: Handled event for server in area: event=member-join server=Node-6df46a0f-abad-a8a6-8f2c-392d694279a3.dc1 area=wan writer.go:29: 2021-01-29T19:34:02.289Z [INFO] TestAgent_FailCheck: Started DNS server: address=127.0.0.1:29892 network=udp writer.go:29: 2021-01-29T19:34:02.291Z [INFO] TestAgent_FailCheck: Started DNS server: address=127.0.0.1:29892 network=tcp writer.go:29: 2021-01-29T19:34:02.290Z [INFO] TestAgent_FailCheck.server: Adding LAN server: server="Node-6df46a0f-abad-a8a6-8f2c-392d694279a3 (Addr: tcp/127.0.0.1:29897) (DC: dc1)" writer.go:29: 2021-01-29T19:34:02.293Z [INFO] TestAgent_FailCheck: Started HTTP server: address=127.0.0.1:29893 network=tcp writer.go:29: 2021-01-29T19:34:02.297Z [INFO] TestAgent_FailCheck: started state syncer writer.go:29: 2021-01-29T19:34:02.325Z [WARN] TestAgent_FailCheck.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:02.328Z [INFO] TestAgent_FailCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:29897 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:02.331Z [DEBUG] TestAgent_FailCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:02.334Z [DEBUG] TestAgent_FailCheck.server.raft: vote granted: from=6df46a0f-abad-a8a6-8f2c-392d694279a3 term=2 tally=1 writer.go:29: 2021-01-29T19:34:02.336Z [INFO] TestAgent_FailCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:02.338Z [INFO] TestAgent_FailCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:29897 [Leader]" writer.go:29: 2021-01-29T19:34:02.340Z [INFO] TestAgent_FailCheck.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:02.341Z [INFO] TestAgent_FailCheck.server: New leader elected: payload=Node-6df46a0f-abad-a8a6-8f2c-392d694279a3 writer.go:29: 2021-01-29T19:34:02.344Z [DEBUG] TestAgent_FailCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29897 writer.go:29: 2021-01-29T19:34:02.352Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:02.367Z [INFO] TestAgent_FailCheck.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:02.368Z [INFO] TestAgent_FailCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:02.370Z [DEBUG] TestAgent_FailCheck.server: Skipping self join check for node since the cluster is too small: node=Node-6df46a0f-abad-a8a6-8f2c-392d694279a3 writer.go:29: 2021-01-29T19:34:02.372Z [INFO] TestAgent_FailCheck.server: member joined, marking health alive: member=Node-6df46a0f-abad-a8a6-8f2c-392d694279a3 === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:02.410Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:02.413Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len: Endpoints down === RUN TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size === CONT TestAgent_FailCheck writer.go:29: 2021-01-29T19:34:02.417Z [DEBUG] TestAgent_FailCheck: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:02.421Z [INFO] TestAgent_FailCheck: Synced node info === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.423Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:02.424Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:02.425Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:02.449Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d0ee53df-d37d-05a2-3aa4-d59a6ca00e0c Address:127.0.0.1:29927}]" writer.go:29: 2021-01-29T19:34:02.452Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.serf.wan: serf: EventMemberJoin: Node-d0ee53df-d37d-05a2-3aa4-d59a6ca00e0c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:02.453Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.raft: entering follower state: follower="Node at 127.0.0.1:29927 [Follower]" leader= writer.go:29: 2021-01-29T19:34:02.457Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.serf.lan: serf: EventMemberJoin: Node-d0ee53df-d37d-05a2-3aa4-d59a6ca00e0c 127.0.0.1 writer.go:29: 2021-01-29T19:34:02.469Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server: Adding LAN server: server="Node-d0ee53df-d37d-05a2-3aa4-d59a6ca00e0c (Addr: tcp/127.0.0.1:29927) (DC: dc1)" writer.go:29: 2021-01-29T19:34:02.469Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server: Handled event for server in area: event=member-join server=Node-d0ee53df-d37d-05a2-3aa4-d59a6ca00e0c.dc1 area=wan writer.go:29: 2021-01-29T19:34:02.473Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: Started DNS server: address=127.0.0.1:29922 network=tcp === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits writer.go:29: 2021-01-29T19:34:02.496Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:02.497Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits: Endpoints down === RUN TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.476Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: Started DNS server: address=127.0.0.1:29922 network=udp writer.go:29: 2021-01-29T19:34:02.500Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: Started HTTP server: address=127.0.0.1:29923 network=tcp writer.go:29: 2021-01-29T19:34:02.502Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: started state syncer writer.go:29: 2021-01-29T19:34:02.506Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:02.507Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: bootstrap = true: do not enable unless necessary === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.507Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.raft: entering candidate state: node="Node at 127.0.0.1:29927 [Candidate]" term=2 === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:02.509Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.tlsutil: Update: version=1 === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.510Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:02.513Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.raft: vote granted: from=d0ee53df-d37d-05a2-3aa4-d59a6ca00e0c term=2 tally=1 writer.go:29: 2021-01-29T19:34:02.515Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:02.528Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.raft: entering leader state: leader="Node at 127.0.0.1:29927 [Leader]" writer.go:29: 2021-01-29T19:34:02.530Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server: cluster leadership acquired === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:02.511Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.530Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server: New leader elected: payload=Node-d0ee53df-d37d-05a2-3aa4-d59a6ca00e0c === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:02.535Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8f44c7b9-a6bd-4a3c-c205-702566c46810 Address:127.0.0.1:29933}]" writer.go:29: 2021-01-29T19:34:02.536Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.raft: entering follower state: follower="Node at 127.0.0.1:29933 [Follower]" leader= === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.535Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29927 === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:02.536Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.serf.wan: serf: EventMemberJoin: Node-8f44c7b9-a6bd-4a3c-c205-702566c46810.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:02.540Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.serf.lan: serf: EventMemberJoin: Node-8f44c7b9-a6bd-4a3c-c205-702566c46810 127.0.0.1 writer.go:29: 2021-01-29T19:34:02.542Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server: Handled event for server in area: event=member-join server=Node-8f44c7b9-a6bd-4a3c-c205-702566c46810.dc1 area=wan writer.go:29: 2021-01-29T19:34:02.542Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server: Adding LAN server: server="Node-8f44c7b9-a6bd-4a3c-c205-702566c46810 (Addr: tcp/127.0.0.1:29933) (DC: dc1)" writer.go:29: 2021-01-29T19:34:02.543Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Started DNS server: address=127.0.0.1:29928 network=udp writer.go:29: 2021-01-29T19:34:02.545Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Started DNS server: address=127.0.0.1:29928 network=tcp writer.go:29: 2021-01-29T19:34:02.547Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Started HTTP server: address=127.0.0.1:29929 network=tcp writer.go:29: 2021-01-29T19:34:02.548Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: started state syncer === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.545Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:02.558Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:02.560Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:02.562Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server: Skipping self join check for node since the cluster is too small: node=Node-d0ee53df-d37d-05a2-3aa4-d59a6ca00e0c writer.go:29: 2021-01-29T19:34:02.564Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server: member joined, marking health alive: member=Node-d0ee53df-d37d-05a2-3aa4-d59a6ca00e0c === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:02.598Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:02.599Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.raft: entering candidate state: node="Node at 127.0.0.1:29933 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:02.601Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:02.602Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.raft: vote granted: from=8f44c7b9-a6bd-4a3c-c205-702566c46810 term=2 tally=1 writer.go:29: 2021-01-29T19:34:02.602Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:02.603Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.raft: entering leader state: leader="Node at 127.0.0.1:29933 [Leader]" writer.go:29: 2021-01-29T19:34:02.604Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:02.605Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server: New leader elected: payload=Node-8f44c7b9-a6bd-4a3c-c205-702566c46810 writer.go:29: 2021-01-29T19:34:02.606Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29933 writer.go:29: 2021-01-29T19:34:02.616Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:02.623Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:02.625Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:02.628Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server: Skipping self join check for node since the cluster is too small: node=Node-8f44c7b9-a6bd-4a3c-c205-702566c46810 writer.go:29: 2021-01-29T19:34:02.630Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server: member joined, marking health alive: member=Node-8f44c7b9-a6bd-4a3c-c205-702566c46810 === CONT TestAgent_FailCheck writer.go:29: 2021-01-29T19:34:02.683Z [DEBUG] TestAgent_FailCheck: Check status updated: check=test status=critical writer.go:29: 2021-01-29T19:34:02.687Z [DEBUG] TestAgent_FailCheck: Node info in sync writer.go:29: 2021-01-29T19:34:02.690Z [INFO] TestAgent_FailCheck: Synced check: check=test writer.go:29: 2021-01-29T19:34:02.692Z [INFO] TestAgent_FailCheck: Requesting shutdown writer.go:29: 2021-01-29T19:34:02.693Z [DEBUG] TestAgent_FailCheck: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:02.695Z [DEBUG] TestAgent_FailCheck: Node info in sync writer.go:29: 2021-01-29T19:34:02.693Z [INFO] TestAgent_FailCheck.server: shutting down server writer.go:29: 2021-01-29T19:34:02.697Z [DEBUG] TestAgent_FailCheck.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:02.696Z [DEBUG] TestAgent_FailCheck: Check in sync: check=test writer.go:29: 2021-01-29T19:34:02.701Z [DEBUG] TestAgent_FailCheck: Node info in sync writer.go:29: 2021-01-29T19:34:02.702Z [DEBUG] TestAgent_FailCheck: Check in sync: check=test writer.go:29: 2021-01-29T19:34:02.699Z [WARN] TestAgent_FailCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:02.699Z [DEBUG] TestAgent_FailCheck.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:02.735Z [WARN] TestAgent_FailCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:02.744Z [INFO] TestAgent_FailCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:02.749Z [INFO] TestAgent_FailCheck: consul server down writer.go:29: 2021-01-29T19:34:02.754Z [INFO] TestAgent_FailCheck: shutdown complete writer.go:29: 2021-01-29T19:34:02.755Z [INFO] TestAgent_FailCheck: Stopping server: protocol=DNS address=127.0.0.1:29892 network=tcp writer.go:29: 2021-01-29T19:34:02.756Z [INFO] TestAgent_FailCheck: Stopping server: protocol=DNS address=127.0.0.1:29892 network=udp writer.go:29: 2021-01-29T19:34:02.757Z [INFO] TestAgent_FailCheck: Stopping server: protocol=HTTP address=127.0.0.1:29893 network=tcp === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.755Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: Requesting shutdown writer.go:29: 2021-01-29T19:34:02.765Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server: shutting down server writer.go:29: 2021-01-29T19:34:02.769Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.leader: stopping routine: routine="CA root pruning" === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:02.773Z [INFO] TestStatusPeersSecondary: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:02.782Z [INFO] TestStatusPeersSecondary: Endpoints down writer.go:29: 2021-01-29T19:34:02.783Z [INFO] TestStatusPeersSecondary: Requesting shutdown === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.778Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.serf.lan: serf: Shutdown without a Leave === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:02.784Z [INFO] TestStatusPeersSecondary.server: shutting down server writer.go:29: 2021-01-29T19:34:02.785Z [DEBUG] TestStatusPeersSecondary.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:02.785Z [WARN] TestStatusPeersSecondary.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:02.787Z [DEBUG] TestStatusPeersSecondary.leader: stopped routine: routine="CA root pruning" === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.773Z [ERROR] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:02.786Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:02.790Z [WARN] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.serf.wan: serf: Shutdown without a Leave === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:02.813Z [WARN] TestStatusPeersSecondary.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:02.818Z [DEBUG] TestStatusPeersSecondary: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.822Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:02.823Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: consul server down === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:02.822Z [DEBUG] TestStatusPeersSecondary: Node info in sync === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:02.824Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: shutdown complete writer.go:29: 2021-01-29T19:34:02.825Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: Stopping server: protocol=DNS address=127.0.0.1:29922 network=tcp writer.go:29: 2021-01-29T19:34:02.826Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: Stopping server: protocol=DNS address=127.0.0.1:29922 network=udp writer.go:29: 2021-01-29T19:34:02.827Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: Stopping server: protocol=HTTP address=127.0.0.1:29923 network=tcp === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:02.824Z [DEBUG] TestStatusPeersSecondary: Node info in sync writer.go:29: 2021-01-29T19:34:02.833Z [INFO] TestStatusPeersSecondary.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:02.840Z [INFO] TestStatusPeersSecondary.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:02.846Z [INFO] TestStatusPeersSecondary: consul server down writer.go:29: 2021-01-29T19:34:02.852Z [INFO] TestStatusPeersSecondary: shutdown complete writer.go:29: 2021-01-29T19:34:02.853Z [INFO] TestStatusPeersSecondary: Stopping server: protocol=DNS address=127.0.0.1:29916 network=tcp writer.go:29: 2021-01-29T19:34:02.854Z [INFO] TestStatusPeersSecondary: Stopping server: protocol=DNS address=127.0.0.1:29916 network=udp writer.go:29: 2021-01-29T19:34:02.855Z [INFO] TestStatusPeersSecondary: Stopping server: protocol=HTTP address=127.0.0.1:29917 network=tcp === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:02.966Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:02.986Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Synced node info === CONT TestAgent_FailCheck writer.go:29: 2021-01-29T19:34:03.263Z [INFO] TestAgent_FailCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:03.267Z [INFO] TestAgent_FailCheck: Endpoints down --- PASS: TestAgent_FailCheck (1.02s) === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.317Z [WARN] TestAgent_unloadChecks: bootstrap = true: do not enable unless necessary === CONT TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:03.328Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:03.333Z [INFO] TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size: Endpoints down === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.332Z [DEBUG] TestAgent_unloadChecks.tlsutil: Update: version=1 === RUN TestTxnEndpoint_Bad_Size_Item/allowed === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.335Z [DEBUG] TestAgent_unloadChecks.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:03.347Z [INFO] TestAgent_unloadChecks.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e268807f-a1e3-5d28-c2a8-10fac00af245 Address:127.0.0.1:29945}]" writer.go:29: 2021-01-29T19:34:03.353Z [INFO] TestAgent_unloadChecks.server.raft: entering follower state: follower="Node at 127.0.0.1:29945 [Follower]" leader= === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:03.356Z [INFO] TestStatusPeersSecondary: Waiting for endpoints to shut down === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.356Z [WARN] TestTxnEndpoint_Bad_Size_Item/allowed: bootstrap = true: do not enable unless necessary === CONT TestStatusPeersSecondary writer.go:29: 2021-01-29T19:34:03.358Z [INFO] TestStatusPeersSecondary: Endpoints down --- PASS: TestStatusPeersSecondary (2.22s) === CONT TestAgent_loadChecks_token === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.358Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/allowed.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:03.360Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/allowed.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.357Z [INFO] TestAgent_unloadChecks.server.serf.wan: serf: EventMemberJoin: Node-e268807f-a1e3-5d28-c2a8-10fac00af245.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:03.372Z [INFO] TestAgent_unloadChecks.server.serf.lan: serf: EventMemberJoin: Node-e268807f-a1e3-5d28-c2a8-10fac00af245 127.0.0.1 writer.go:29: 2021-01-29T19:34:03.375Z [INFO] TestAgent_unloadChecks: Started DNS server: address=127.0.0.1:29940 network=udp writer.go:29: 2021-01-29T19:34:03.376Z [INFO] TestAgent_unloadChecks.server: Adding LAN server: server="Node-e268807f-a1e3-5d28-c2a8-10fac00af245 (Addr: tcp/127.0.0.1:29945) (DC: dc1)" writer.go:29: 2021-01-29T19:34:03.377Z [INFO] TestAgent_unloadChecks.server: Handled event for server in area: event=member-join server=Node-e268807f-a1e3-5d28-c2a8-10fac00af245.dc1 area=wan === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.378Z [WARN] TestAgent_loadChecks_token: bootstrap = true: do not enable unless necessary === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.379Z [INFO] TestAgent_unloadChecks: Started DNS server: address=127.0.0.1:29940 network=tcp === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.380Z [DEBUG] TestAgent_loadChecks_token.tlsutil: Update: version=1 === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.383Z [INFO] TestAgent_unloadChecks: Started HTTP server: address=127.0.0.1:29941 network=tcp === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.383Z [DEBUG] TestAgent_loadChecks_token.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.384Z [INFO] TestAgent_unloadChecks: started state syncer === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.387Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:481e0a69-5535-26b4-92e3-816abf8237ec Address:127.0.0.1:29951}]" writer.go:29: 2021-01-29T19:34:03.390Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server.raft: entering follower state: follower="Node at 127.0.0.1:29951 [Follower]" leader= writer.go:29: 2021-01-29T19:34:03.390Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server.serf.wan: serf: EventMemberJoin: Node-481e0a69-5535-26b4-92e3-816abf8237ec.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:03.394Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server.serf.lan: serf: EventMemberJoin: Node-481e0a69-5535-26b4-92e3-816abf8237ec 127.0.0.1 === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.396Z [INFO] TestAgent_loadChecks_token.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a7365729-45d2-aca4-f735-0ef54a1ebcbb Address:127.0.0.1:29969}]" === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.398Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: Started DNS server: address=127.0.0.1:29946 network=udp === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.399Z [INFO] TestAgent_loadChecks_token.server.serf.wan: serf: EventMemberJoin: Node-a7365729-45d2-aca4-f735-0ef54a1ebcbb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:03.402Z [INFO] TestAgent_loadChecks_token.server.raft: entering follower state: follower="Node at 127.0.0.1:29969 [Follower]" leader= writer.go:29: 2021-01-29T19:34:03.404Z [INFO] TestAgent_loadChecks_token.server.serf.lan: serf: EventMemberJoin: Node-a7365729-45d2-aca4-f735-0ef54a1ebcbb 127.0.0.1 writer.go:29: 2021-01-29T19:34:03.407Z [INFO] TestAgent_loadChecks_token: Started DNS server: address=127.0.0.1:29964 network=udp writer.go:29: 2021-01-29T19:34:03.408Z [INFO] TestAgent_loadChecks_token.server: Handled event for server in area: event=member-join server=Node-a7365729-45d2-aca4-f735-0ef54a1ebcbb.dc1 area=wan writer.go:29: 2021-01-29T19:34:03.409Z [INFO] TestAgent_loadChecks_token.server: Adding LAN server: server="Node-a7365729-45d2-aca4-f735-0ef54a1ebcbb (Addr: tcp/127.0.0.1:29969) (DC: dc1)" writer.go:29: 2021-01-29T19:34:03.411Z [INFO] TestAgent_loadChecks_token: Started DNS server: address=127.0.0.1:29964 network=tcp writer.go:29: 2021-01-29T19:34:03.414Z [INFO] TestAgent_loadChecks_token: Started HTTP server: address=127.0.0.1:29965 network=tcp writer.go:29: 2021-01-29T19:34:03.414Z [INFO] TestAgent_loadChecks_token: started state syncer === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.400Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server: Adding LAN server: server="Node-481e0a69-5535-26b4-92e3-816abf8237ec (Addr: tcp/127.0.0.1:29951) (DC: dc1)" writer.go:29: 2021-01-29T19:34:03.411Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server: Handled event for server in area: event=member-join server=Node-481e0a69-5535-26b4-92e3-816abf8237ec.dc1 area=wan writer.go:29: 2021-01-29T19:34:03.416Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: Started DNS server: address=127.0.0.1:29946 network=tcp writer.go:29: 2021-01-29T19:34:03.419Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: Started HTTP server: address=127.0.0.1:29947 network=tcp === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.419Z [WARN] TestAgent_unloadChecks.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.420Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: started state syncer === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.421Z [INFO] TestAgent_unloadChecks.server.raft: entering candidate state: node="Node at 127.0.0.1:29945 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:03.423Z [DEBUG] TestAgent_unloadChecks.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:03.425Z [DEBUG] TestAgent_unloadChecks.server.raft: vote granted: from=e268807f-a1e3-5d28-c2a8-10fac00af245 term=2 tally=1 writer.go:29: 2021-01-29T19:34:03.427Z [INFO] TestAgent_unloadChecks.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:03.428Z [INFO] TestAgent_unloadChecks.server.raft: entering leader state: leader="Node at 127.0.0.1:29945 [Leader]" writer.go:29: 2021-01-29T19:34:03.430Z [INFO] TestAgent_unloadChecks.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:03.431Z [INFO] TestAgent_unloadChecks.server: New leader elected: payload=Node-e268807f-a1e3-5d28-c2a8-10fac00af245 === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.433Z [WARN] TestTxnEndpoint_Bad_Size_Item/allowed.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:03.434Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server.raft: entering candidate state: node="Node at 127.0.0.1:29951 [Candidate]" term=2 === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.431Z [DEBUG] TestAgent_unloadChecks.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29945 === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.435Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/allowed.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:03.436Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/allowed.server.raft: vote granted: from=481e0a69-5535-26b4-92e3-816abf8237ec term=2 tally=1 writer.go:29: 2021-01-29T19:34:03.437Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:03.438Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server.raft: entering leader state: leader="Node at 127.0.0.1:29951 [Leader]" === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.438Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.438Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:03.439Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server: New leader elected: payload=Node-481e0a69-5535-26b4-92e3-816abf8237ec writer.go:29: 2021-01-29T19:34:03.440Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/allowed.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29951 writer.go:29: 2021-01-29T19:34:03.461Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.460Z [INFO] TestAgent_unloadChecks.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:03.463Z [INFO] TestAgent_unloadChecks.leader: started routine: routine="CA root pruning" === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:03.462Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Requesting shutdown === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.465Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.464Z [DEBUG] TestAgent_unloadChecks.server: Skipping self join check for node since the cluster is too small: node=Node-e268807f-a1e3-5d28-c2a8-10fac00af245 === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:03.465Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server: shutting down server === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.466Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:03.467Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/allowed.server: Skipping self join check for node since the cluster is too small: node=Node-481e0a69-5535-26b4-92e3-816abf8237ec === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.466Z [INFO] TestAgent_unloadChecks.server: member joined, marking health alive: member=Node-e268807f-a1e3-5d28-c2a8-10fac00af245 === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:03.466Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.leader: stopping routine: routine="CA root pruning" === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.467Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server: member joined, marking health alive: member=Node-481e0a69-5535-26b4-92e3-816abf8237ec === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.472Z [WARN] TestAgent_loadChecks_token.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:03.473Z [INFO] TestAgent_loadChecks_token.server.raft: entering candidate state: node="Node at 127.0.0.1:29969 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:03.475Z [DEBUG] TestAgent_loadChecks_token.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:03.476Z [DEBUG] TestAgent_loadChecks_token.server.raft: vote granted: from=a7365729-45d2-aca4-f735-0ef54a1ebcbb term=2 tally=1 === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:03.468Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.478Z [INFO] TestAgent_loadChecks_token.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:03.479Z [INFO] TestAgent_loadChecks_token.server.raft: entering leader state: leader="Node at 127.0.0.1:29969 [Leader]" === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:03.468Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:03.474Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:03.481Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Node info in sync === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.480Z [INFO] TestAgent_loadChecks_token.server: cluster leadership acquired === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:03.480Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:03.483Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Node info in sync === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.480Z [INFO] TestAgent_loadChecks_token.server: New leader elected: payload=Node-a7365729-45d2-aca4-f735-0ef54a1ebcbb === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:03.487Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:03.488Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: consul server down writer.go:29: 2021-01-29T19:34:03.489Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: shutdown complete === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.482Z [DEBUG] TestAgent_loadChecks_token.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29969 === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:03.489Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Stopping server: protocol=DNS address=127.0.0.1:29928 network=tcp writer.go:29: 2021-01-29T19:34:03.490Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Stopping server: protocol=DNS address=127.0.0.1:29928 network=udp writer.go:29: 2021-01-29T19:34:03.491Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Stopping server: protocol=HTTP address=127.0.0.1:29929 network=tcp === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.493Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.495Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/allowed: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:03.498Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: Synced node info writer.go:29: 2021-01-29T19:34:03.499Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/allowed: Node info in sync === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.511Z [INFO] TestAgent_loadChecks_token.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:03.514Z [INFO] TestAgent_loadChecks_token.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:03.516Z [DEBUG] TestAgent_loadChecks_token.server: Skipping self join check for node since the cluster is too small: node=Node-a7365729-45d2-aca4-f735-0ef54a1ebcbb writer.go:29: 2021-01-29T19:34:03.523Z [INFO] TestAgent_loadChecks_token.server: member joined, marking health alive: member=Node-a7365729-45d2-aca4-f735-0ef54a1ebcbb writer.go:29: 2021-01-29T19:34:03.627Z [DEBUG] TestAgent_loadChecks_token: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.642Z [DEBUG] TestAgent_unloadChecks: removed check: check=service:redis writer.go:29: 2021-01-29T19:34:03.644Z [INFO] TestAgent_unloadChecks: Requesting shutdown writer.go:29: 2021-01-29T19:34:03.644Z [INFO] TestAgent_unloadChecks.server: shutting down server writer.go:29: 2021-01-29T19:34:03.645Z [DEBUG] TestAgent_unloadChecks.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:03.646Z [WARN] TestAgent_unloadChecks.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:03.649Z [DEBUG] TestAgent_unloadChecks: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:03.652Z [DEBUG] TestAgent_unloadChecks.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:03.653Z [WARN] TestAgent_unloadChecks.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.677Z [INFO] TestAgent_loadChecks_token: Synced node info === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.680Z [INFO] TestAgent_unloadChecks.server.router.manager: shutting down === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.694Z [INFO] TestAgent_loadChecks_token: Synced check: check=rabbitmq writer.go:29: 2021-01-29T19:34:03.698Z [DEBUG] TestAgent_loadChecks_token: Node info in sync writer.go:29: 2021-01-29T19:34:03.701Z [DEBUG] TestAgent_loadChecks_token: Check in sync: check=rabbitmq === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:03.738Z [WARN] TestAgent_unloadChecks: Syncing node info failed.: error="leadership lost while committing log" writer.go:29: 2021-01-29T19:34:03.740Z [ERROR] TestAgent_unloadChecks.anti_entropy: failed to sync remote state: error="leadership lost while committing log" writer.go:29: 2021-01-29T19:34:03.742Z [INFO] TestAgent_unloadChecks: consul server down writer.go:29: 2021-01-29T19:34:03.743Z [INFO] TestAgent_unloadChecks: shutdown complete writer.go:29: 2021-01-29T19:34:03.745Z [INFO] TestAgent_unloadChecks: Stopping server: protocol=DNS address=127.0.0.1:29940 network=tcp writer.go:29: 2021-01-29T19:34:03.747Z [INFO] TestAgent_unloadChecks: Stopping server: protocol=DNS address=127.0.0.1:29940 network=udp writer.go:29: 2021-01-29T19:34:03.748Z [INFO] TestAgent_unloadChecks: Stopping server: protocol=HTTP address=127.0.0.1:29941 network=tcp === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:03.803Z [INFO] TestAgent_loadChecks_token: Requesting shutdown writer.go:29: 2021-01-29T19:34:03.804Z [INFO] TestAgent_loadChecks_token.server: shutting down server writer.go:29: 2021-01-29T19:34:03.805Z [DEBUG] TestAgent_loadChecks_token.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:03.806Z [WARN] TestAgent_loadChecks_token.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:03.807Z [DEBUG] TestAgent_loadChecks_token.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:03.832Z [WARN] TestAgent_loadChecks_token.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:03.845Z [INFO] TestAgent_loadChecks_token.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:03.847Z [INFO] TestAgent_loadChecks_token: consul server down writer.go:29: 2021-01-29T19:34:03.848Z [INFO] TestAgent_loadChecks_token: shutdown complete writer.go:29: 2021-01-29T19:34:03.849Z [INFO] TestAgent_loadChecks_token: Stopping server: protocol=DNS address=127.0.0.1:29964 network=tcp writer.go:29: 2021-01-29T19:34:03.850Z [INFO] TestAgent_loadChecks_token: Stopping server: protocol=DNS address=127.0.0.1:29964 network=udp writer.go:29: 2021-01-29T19:34:03.850Z [INFO] TestAgent_loadChecks_token: Stopping server: protocol=HTTP address=127.0.0.1:29965 network=tcp === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:03.890Z [WARN] TestTxnEndpoint_Bad_Size_Item/allowed.server.rpc: Attempting to apply large raft entry: size_in_bytes=1573029 writer.go:29: 2021-01-29T19:34:03.919Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: Requesting shutdown writer.go:29: 2021-01-29T19:34:03.921Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server: shutting down server writer.go:29: 2021-01-29T19:34:03.923Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/allowed.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:03.925Z [WARN] TestTxnEndpoint_Bad_Size_Item/allowed.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:03.925Z [DEBUG] TestTxnEndpoint_Bad_Size_Item/allowed.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:03.928Z [WARN] TestTxnEndpoint_Bad_Size_Item/allowed.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:03.930Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:03.930Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: consul server down writer.go:29: 2021-01-29T19:34:03.933Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: shutdown complete writer.go:29: 2021-01-29T19:34:03.934Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: Stopping server: protocol=DNS address=127.0.0.1:29946 network=tcp writer.go:29: 2021-01-29T19:34:03.936Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: Stopping server: protocol=DNS address=127.0.0.1:29946 network=udp writer.go:29: 2021-01-29T19:34:03.937Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: Stopping server: protocol=HTTP address=127.0.0.1:29947 network=tcp === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len writer.go:29: 2021-01-29T19:34:03.992Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:03.994Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len: Endpoints down === RUN TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:04.004Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:04.004Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:04.006Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:04.024Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cb6af5d5-80cf-adb9-773b-feabc3d68ac0 Address:127.0.0.1:29963}]" writer.go:29: 2021-01-29T19:34:04.026Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.raft: entering follower state: follower="Node at 127.0.0.1:29963 [Follower]" leader= writer.go:29: 2021-01-29T19:34:04.026Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.serf.wan: serf: EventMemberJoin: Node-cb6af5d5-80cf-adb9-773b-feabc3d68ac0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:04.034Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.serf.lan: serf: EventMemberJoin: Node-cb6af5d5-80cf-adb9-773b-feabc3d68ac0 127.0.0.1 writer.go:29: 2021-01-29T19:34:04.042Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server: Handled event for server in area: event=member-join server=Node-cb6af5d5-80cf-adb9-773b-feabc3d68ac0.dc1 area=wan writer.go:29: 2021-01-29T19:34:04.042Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server: Adding LAN server: server="Node-cb6af5d5-80cf-adb9-773b-feabc3d68ac0 (Addr: tcp/127.0.0.1:29963) (DC: dc1)" writer.go:29: 2021-01-29T19:34:04.049Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Started DNS server: address=127.0.0.1:29958 network=tcp writer.go:29: 2021-01-29T19:34:04.050Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Started DNS server: address=127.0.0.1:29958 network=udp writer.go:29: 2021-01-29T19:34:04.052Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Started HTTP server: address=127.0.0.1:29959 network=tcp writer.go:29: 2021-01-29T19:34:04.053Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: started state syncer writer.go:29: 2021-01-29T19:34:04.094Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:04.096Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.raft: entering candidate state: node="Node at 127.0.0.1:29963 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:04.098Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:04.099Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.raft: vote granted: from=cb6af5d5-80cf-adb9-773b-feabc3d68ac0 term=2 tally=1 writer.go:29: 2021-01-29T19:34:04.100Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:04.102Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.raft: entering leader state: leader="Node at 127.0.0.1:29963 [Leader]" writer.go:29: 2021-01-29T19:34:04.104Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:04.104Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server: New leader elected: payload=Node-cb6af5d5-80cf-adb9-773b-feabc3d68ac0 writer.go:29: 2021-01-29T19:34:04.109Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29963 writer.go:29: 2021-01-29T19:34:04.124Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:04.130Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:04.132Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:04.133Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server: Skipping self join check for node since the cluster is too small: node=Node-cb6af5d5-80cf-adb9-773b-feabc3d68ac0 writer.go:29: 2021-01-29T19:34:04.135Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server: member joined, marking health alive: member=Node-cb6af5d5-80cf-adb9-773b-feabc3d68ac0 === CONT TestAgent_unloadChecks writer.go:29: 2021-01-29T19:34:04.251Z [INFO] TestAgent_unloadChecks: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:04.253Z [INFO] TestAgent_unloadChecks: Endpoints down --- PASS: TestAgent_unloadChecks (0.99s) === CONT TestAgent_PurgeCheckOnDuplicate writer.go:29: 2021-01-29T19:34:04.283Z [WARN] TestAgent_PurgeCheckOnDuplicate: Node name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: node_name="Node f0617d1c-d8ac-9863-4a0d-297ab09c4e8e" writer.go:29: 2021-01-29T19:34:04.284Z [DEBUG] TestAgent_PurgeCheckOnDuplicate.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:04.303Z [INFO] TestAgent_PurgeCheckOnDuplicate.client.serf.lan: serf: EventMemberJoin: Node f0617d1c-d8ac-9863-4a0d-297ab09c4e8e 127.0.0.1 writer.go:29: 2021-01-29T19:34:04.307Z [INFO] TestAgent_PurgeCheckOnDuplicate: Started DNS server: address=127.0.0.1:29952 network=udp writer.go:29: 2021-01-29T19:34:04.309Z [INFO] TestAgent_PurgeCheckOnDuplicate: Started DNS server: address=127.0.0.1:29952 network=tcp writer.go:29: 2021-01-29T19:34:04.311Z [INFO] TestAgent_PurgeCheckOnDuplicate: Started HTTP server: address=127.0.0.1:29953 network=tcp writer.go:29: 2021-01-29T19:34:04.313Z [INFO] TestAgent_PurgeCheckOnDuplicate: started state syncer writer.go:29: 2021-01-29T19:34:04.314Z [WARN] TestAgent_PurgeCheckOnDuplicate.client.manager: No servers available writer.go:29: 2021-01-29T19:34:04.316Z [ERROR] TestAgent_PurgeCheckOnDuplicate.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:34:04.317Z [INFO] TestAgent_PurgeCheckOnDuplicate: Requesting shutdown writer.go:29: 2021-01-29T19:34:04.319Z [INFO] TestAgent_PurgeCheckOnDuplicate.client: shutting down client writer.go:29: 2021-01-29T19:34:04.320Z [WARN] TestAgent_PurgeCheckOnDuplicate.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:04.320Z [INFO] TestAgent_PurgeCheckOnDuplicate.client.manager: shutting down writer.go:29: 2021-01-29T19:34:04.324Z [INFO] TestAgent_PurgeCheckOnDuplicate: consul client down === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:04.325Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_PurgeCheckOnDuplicate writer.go:29: 2021-01-29T19:34:04.326Z [INFO] TestAgent_PurgeCheckOnDuplicate: shutdown complete === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:04.327Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Synced node info === CONT TestAgent_PurgeCheckOnDuplicate writer.go:29: 2021-01-29T19:34:04.328Z [INFO] TestAgent_PurgeCheckOnDuplicate: Stopping server: protocol=DNS address=127.0.0.1:29952 network=tcp writer.go:29: 2021-01-29T19:34:04.332Z [INFO] TestAgent_PurgeCheckOnDuplicate: Stopping server: protocol=DNS address=127.0.0.1:29952 network=udp writer.go:29: 2021-01-29T19:34:04.333Z [INFO] TestAgent_PurgeCheckOnDuplicate: Stopping server: protocol=HTTP address=127.0.0.1:29953 network=tcp === CONT TestAgent_loadChecks_token writer.go:29: 2021-01-29T19:34:04.351Z [INFO] TestAgent_loadChecks_token: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:04.353Z [INFO] TestAgent_loadChecks_token: Endpoints down --- PASS: TestAgent_loadChecks_token (1.00s) === CONT TestAgent_PersistCheck writer.go:29: 2021-01-29T19:34:04.369Z [DEBUG] TestAgent_PersistCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:04.373Z [INFO] TestAgent_PersistCheck.client.serf.lan: serf: EventMemberJoin: Node-497e0796-00a6-f583-7e42-d5081946ef1e 127.0.0.1 writer.go:29: 2021-01-29T19:34:04.378Z [INFO] TestAgent_PersistCheck: Started DNS server: address=127.0.0.1:29934 network=tcp writer.go:29: 2021-01-29T19:34:04.380Z [INFO] TestAgent_PersistCheck: Started DNS server: address=127.0.0.1:29934 network=udp writer.go:29: 2021-01-29T19:34:04.382Z [INFO] TestAgent_PersistCheck: Started HTTP server: address=127.0.0.1:29935 network=tcp writer.go:29: 2021-01-29T19:34:04.383Z [INFO] TestAgent_PersistCheck: started state syncer writer.go:29: 2021-01-29T19:34:04.384Z [WARN] TestAgent_PersistCheck.client.manager: No servers available writer.go:29: 2021-01-29T19:34:04.387Z [ERROR] TestAgent_PersistCheck.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:34:04.415Z [INFO] TestAgent_PersistCheck: Requesting shutdown writer.go:29: 2021-01-29T19:34:04.416Z [INFO] TestAgent_PersistCheck.client: shutting down client writer.go:29: 2021-01-29T19:34:04.418Z [WARN] TestAgent_PersistCheck.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:04.420Z [INFO] TestAgent_PersistCheck.client.manager: shutting down writer.go:29: 2021-01-29T19:34:04.433Z [INFO] TestAgent_PersistCheck: consul client down writer.go:29: 2021-01-29T19:34:04.434Z [INFO] TestAgent_PersistCheck: shutdown complete writer.go:29: 2021-01-29T19:34:04.435Z [INFO] TestAgent_PersistCheck: Stopping server: protocol=DNS address=127.0.0.1:29934 network=tcp writer.go:29: 2021-01-29T19:34:04.436Z [INFO] TestAgent_PersistCheck: Stopping server: protocol=DNS address=127.0.0.1:29934 network=udp writer.go:29: 2021-01-29T19:34:04.437Z [INFO] TestAgent_PersistCheck: Stopping server: protocol=HTTP address=127.0.0.1:29935 network=tcp === CONT TestTxnEndpoint_Bad_Size_Item/allowed writer.go:29: 2021-01-29T19:34:04.442Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:04.444Z [INFO] TestTxnEndpoint_Bad_Size_Item/allowed: Endpoints down --- PASS: TestTxnEndpoint_Bad_Size_Item (4.65s) --- PASS: TestTxnEndpoint_Bad_Size_Item/exceeds_default_limits (1.28s) --- PASS: TestTxnEndpoint_Bad_Size_Item/exceeds_configured_max_txn_len (1.34s) --- PASS: TestTxnEndpoint_Bad_Size_Item/exceeds_default_max_kv_value_size (0.92s) --- PASS: TestTxnEndpoint_Bad_Size_Item/allowed (1.11s) === CONT TestAgent_updateTTLCheck writer.go:29: 2021-01-29T19:34:04.459Z [WARN] TestAgent_updateTTLCheck: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:04.460Z [DEBUG] TestAgent_updateTTLCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:04.462Z [DEBUG] TestAgent_updateTTLCheck.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:04.498Z [INFO] TestAgent_updateTTLCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a65980a5-9266-de85-ebd6-6a23802157ab Address:127.0.0.1:29975}]" writer.go:29: 2021-01-29T19:34:04.501Z [INFO] TestAgent_updateTTLCheck.server.serf.wan: serf: EventMemberJoin: Node-a65980a5-9266-de85-ebd6-6a23802157ab.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:04.504Z [INFO] TestAgent_updateTTLCheck.server.serf.lan: serf: EventMemberJoin: Node-a65980a5-9266-de85-ebd6-6a23802157ab 127.0.0.1 writer.go:29: 2021-01-29T19:34:04.508Z [INFO] TestAgent_updateTTLCheck: Started DNS server: address=127.0.0.1:29970 network=udp writer.go:29: 2021-01-29T19:34:04.509Z [INFO] TestAgent_updateTTLCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:29975 [Follower]" leader= writer.go:29: 2021-01-29T19:34:04.512Z [INFO] TestAgent_updateTTLCheck.server: Adding LAN server: server="Node-a65980a5-9266-de85-ebd6-6a23802157ab (Addr: tcp/127.0.0.1:29975) (DC: dc1)" writer.go:29: 2021-01-29T19:34:04.515Z [INFO] TestAgent_updateTTLCheck: Started DNS server: address=127.0.0.1:29970 network=tcp writer.go:29: 2021-01-29T19:34:04.512Z [INFO] TestAgent_updateTTLCheck.server: Handled event for server in area: event=member-join server=Node-a65980a5-9266-de85-ebd6-6a23802157ab.dc1 area=wan writer.go:29: 2021-01-29T19:34:04.518Z [INFO] TestAgent_updateTTLCheck: Started HTTP server: address=127.0.0.1:29971 network=tcp writer.go:29: 2021-01-29T19:34:04.521Z [INFO] TestAgent_updateTTLCheck: started state syncer writer.go:29: 2021-01-29T19:34:04.582Z [WARN] TestAgent_updateTTLCheck.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:04.583Z [INFO] TestAgent_updateTTLCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:29975 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:04.624Z [DEBUG] TestAgent_updateTTLCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:04.624Z [DEBUG] TestAgent_updateTTLCheck.server.raft: vote granted: from=a65980a5-9266-de85-ebd6-6a23802157ab term=2 tally=1 writer.go:29: 2021-01-29T19:34:04.626Z [INFO] TestAgent_updateTTLCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:04.628Z [INFO] TestAgent_updateTTLCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:29975 [Leader]" writer.go:29: 2021-01-29T19:34:04.629Z [INFO] TestAgent_updateTTLCheck.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:04.630Z [INFO] TestAgent_updateTTLCheck.server: New leader elected: payload=Node-a65980a5-9266-de85-ebd6-6a23802157ab writer.go:29: 2021-01-29T19:34:04.655Z [DEBUG] TestAgent_updateTTLCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29975 writer.go:29: 2021-01-29T19:34:04.714Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:04.719Z [INFO] TestAgent_updateTTLCheck.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:04.720Z [INFO] TestAgent_updateTTLCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:04.721Z [DEBUG] TestAgent_updateTTLCheck.server: Skipping self join check for node since the cluster is too small: node=Node-a65980a5-9266-de85-ebd6-6a23802157ab writer.go:29: 2021-01-29T19:34:04.722Z [INFO] TestAgent_updateTTLCheck.server: member joined, marking health alive: member=Node-a65980a5-9266-de85-ebd6-6a23802157ab === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:04.811Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Requesting shutdown writer.go:29: 2021-01-29T19:34:04.813Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server: shutting down server writer.go:29: 2021-01-29T19:34:04.814Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:04.816Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:04.816Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:04.819Z [WARN] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:04.822Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:04.823Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: consul server down writer.go:29: 2021-01-29T19:34:04.824Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: shutdown complete writer.go:29: 2021-01-29T19:34:04.825Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Stopping server: protocol=DNS address=127.0.0.1:29958 network=tcp writer.go:29: 2021-01-29T19:34:04.826Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Stopping server: protocol=DNS address=127.0.0.1:29958 network=udp writer.go:29: 2021-01-29T19:34:04.827Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Stopping server: protocol=HTTP address=127.0.0.1:29959 network=tcp === CONT TestAgent_PurgeCheckOnDuplicate writer.go:29: 2021-01-29T19:34:04.836Z [INFO] TestAgent_PurgeCheckOnDuplicate: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:04.838Z [INFO] TestAgent_PurgeCheckOnDuplicate: Endpoints down writer.go:29: 2021-01-29T19:34:04.864Z [WARN] TestAgent_PurgeCheckOnDuplicate-a2: Node name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: node_name="Node f0617d1c-d8ac-9863-4a0d-297ab09c4e8e" writer.go:29: 2021-01-29T19:34:04.870Z [DEBUG] TestAgent_PurgeCheckOnDuplicate-a2.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:04.877Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2.client.serf.lan: serf: EventMemberJoin: Node f0617d1c-d8ac-9863-4a0d-297ab09c4e8e 127.0.0.1 writer.go:29: 2021-01-29T19:34:04.881Z [WARN] TestAgent_PurgeCheckOnDuplicate-a2.client.serf.lan: serf: Failed to re-join any previously known node writer.go:29: 2021-01-29T19:34:04.882Z [DEBUG] TestAgent_PurgeCheckOnDuplicate-a2: check exists, not restoring from file: check=mem file=/tmp/consul-test/TestAgent_PurgeCheckOnDuplicate-agent841283116/checks/afc4fc7e48a0710a1dc94ef3e8bc5764 === CONT TestAgent_updateTTLCheck writer.go:29: 2021-01-29T19:34:04.886Z [DEBUG] TestAgent_updateTTLCheck: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:04.888Z [INFO] TestAgent_updateTTLCheck: Synced node info === CONT TestAgent_PurgeCheckOnDuplicate writer.go:29: 2021-01-29T19:34:04.887Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: Started DNS server: address=127.0.0.1:29982 network=tcp writer.go:29: 2021-01-29T19:34:04.889Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: Started DNS server: address=127.0.0.1:29982 network=udp writer.go:29: 2021-01-29T19:34:04.892Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: Started HTTP server: address=127.0.0.1:29983 network=tcp writer.go:29: 2021-01-29T19:34:04.893Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: started state syncer writer.go:29: 2021-01-29T19:34:04.893Z [WARN] TestAgent_PurgeCheckOnDuplicate-a2.client.manager: No servers available writer.go:29: 2021-01-29T19:34:04.895Z [ERROR] TestAgent_PurgeCheckOnDuplicate-a2.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:34:04.896Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: Requesting shutdown writer.go:29: 2021-01-29T19:34:04.897Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2.client: shutting down client writer.go:29: 2021-01-29T19:34:04.899Z [WARN] TestAgent_PurgeCheckOnDuplicate-a2.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:04.899Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2.client.manager: shutting down writer.go:29: 2021-01-29T19:34:04.904Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: consul client down writer.go:29: 2021-01-29T19:34:04.905Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: shutdown complete writer.go:29: 2021-01-29T19:34:04.906Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: Stopping server: protocol=DNS address=127.0.0.1:29982 network=tcp writer.go:29: 2021-01-29T19:34:04.907Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: Stopping server: protocol=DNS address=127.0.0.1:29982 network=udp writer.go:29: 2021-01-29T19:34:04.909Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: Stopping server: protocol=HTTP address=127.0.0.1:29983 network=tcp === CONT TestAgent_PersistCheck writer.go:29: 2021-01-29T19:34:04.937Z [INFO] TestAgent_PersistCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:04.939Z [INFO] TestAgent_PersistCheck: Endpoints down writer.go:29: 2021-01-29T19:34:04.962Z [DEBUG] TestAgent_PersistCheck-a2.tlsutil: Update: version=1 === CONT TestAgent_updateTTLCheck writer.go:29: 2021-01-29T19:34:04.974Z [DEBUG] TestAgent_updateTTLCheck: Check status updated: check=mem status=passing writer.go:29: 2021-01-29T19:34:04.976Z [DEBUG] TestAgent_updateTTLCheck: Check status updated: check=mem status=critical === CONT TestAgent_PersistCheck writer.go:29: 2021-01-29T19:34:04.971Z [INFO] TestAgent_PersistCheck-a2.client.serf.lan: serf: EventMemberJoin: Node-7d5c5882-3470-fe0c-0108-836ab538ac28 127.0.0.1 === CONT TestAgent_updateTTLCheck writer.go:29: 2021-01-29T19:34:04.977Z [INFO] TestAgent_updateTTLCheck: Requesting shutdown writer.go:29: 2021-01-29T19:34:04.977Z [INFO] TestAgent_updateTTLCheck.server: shutting down server writer.go:29: 2021-01-29T19:34:04.978Z [DEBUG] TestAgent_updateTTLCheck.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_PersistCheck writer.go:29: 2021-01-29T19:34:04.978Z [DEBUG] TestAgent_PersistCheck-a2: restored health check from file: check=mem file=/tmp/consul-test/TestAgent_PersistCheck-agent194909595/checks/afc4fc7e48a0710a1dc94ef3e8bc5764 writer.go:29: 2021-01-29T19:34:04.979Z [INFO] TestAgent_PersistCheck-a2.client.serf.lan: serf: Attempting re-join to previously known node: Node-497e0796-00a6-f583-7e42-d5081946ef1e: 127.0.0.1:29937 writer.go:29: 2021-01-29T19:34:04.982Z [INFO] TestAgent_PersistCheck-a2: Started DNS server: address=127.0.0.1:29976 network=tcp writer.go:29: 2021-01-29T19:34:04.984Z [INFO] TestAgent_PersistCheck-a2: Started DNS server: address=127.0.0.1:29976 network=udp writer.go:29: 2021-01-29T19:34:04.983Z [DEBUG] TestAgent_PersistCheck-a2.client.memberlist.lan: memberlist: Failed to join 127.0.0.1: dial tcp 127.0.0.1:29937: connect: connection refused writer.go:29: 2021-01-29T19:34:04.986Z [WARN] TestAgent_PersistCheck-a2.client.serf.lan: serf: Failed to re-join any previously known node === CONT TestAgent_updateTTLCheck writer.go:29: 2021-01-29T19:34:04.980Z [WARN] TestAgent_updateTTLCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:04.980Z [DEBUG] TestAgent_updateTTLCheck.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_PersistCheck writer.go:29: 2021-01-29T19:34:04.986Z [INFO] TestAgent_PersistCheck-a2: Started HTTP server: address=127.0.0.1:29977 network=tcp writer.go:29: 2021-01-29T19:34:04.990Z [INFO] TestAgent_PersistCheck-a2: started state syncer writer.go:29: 2021-01-29T19:34:04.991Z [WARN] TestAgent_PersistCheck-a2.client.manager: No servers available === CONT TestAgent_updateTTLCheck writer.go:29: 2021-01-29T19:34:04.997Z [WARN] TestAgent_updateTTLCheck.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_PersistCheck writer.go:29: 2021-01-29T19:34:04.996Z [ERROR] TestAgent_PersistCheck-a2.anti_entropy: failed to sync remote state: error="No known Consul servers" === CONT TestAgent_updateTTLCheck writer.go:29: 2021-01-29T19:34:05.000Z [INFO] TestAgent_updateTTLCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:05.001Z [INFO] TestAgent_updateTTLCheck: consul server down writer.go:29: 2021-01-29T19:34:05.004Z [INFO] TestAgent_updateTTLCheck: shutdown complete writer.go:29: 2021-01-29T19:34:05.006Z [INFO] TestAgent_updateTTLCheck: Stopping server: protocol=DNS address=127.0.0.1:29970 network=tcp writer.go:29: 2021-01-29T19:34:05.008Z [INFO] TestAgent_updateTTLCheck: Stopping server: protocol=DNS address=127.0.0.1:29970 network=udp writer.go:29: 2021-01-29T19:34:05.010Z [INFO] TestAgent_updateTTLCheck: Stopping server: protocol=HTTP address=127.0.0.1:29971 network=tcp === CONT TestAgent_PersistCheck writer.go:29: 2021-01-29T19:34:05.010Z [INFO] TestAgent_PersistCheck-a2: Requesting shutdown writer.go:29: 2021-01-29T19:34:05.022Z [INFO] TestAgent_PersistCheck-a2.client: shutting down client writer.go:29: 2021-01-29T19:34:05.024Z [WARN] TestAgent_PersistCheck-a2.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.025Z [INFO] TestAgent_PersistCheck-a2.client.manager: shutting down writer.go:29: 2021-01-29T19:34:05.027Z [INFO] TestAgent_PersistCheck-a2: consul client down writer.go:29: 2021-01-29T19:34:05.028Z [INFO] TestAgent_PersistCheck-a2: shutdown complete writer.go:29: 2021-01-29T19:34:05.030Z [INFO] TestAgent_PersistCheck-a2: Stopping server: protocol=DNS address=127.0.0.1:29976 network=tcp writer.go:29: 2021-01-29T19:34:05.031Z [INFO] TestAgent_PersistCheck-a2: Stopping server: protocol=DNS address=127.0.0.1:29976 network=udp writer.go:29: 2021-01-29T19:34:05.034Z [INFO] TestAgent_PersistCheck-a2: Stopping server: protocol=HTTP address=127.0.0.1:29977 network=tcp === CONT TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size writer.go:29: 2021-01-29T19:34:05.328Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:05.330Z [INFO] TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size: Endpoints down === RUN TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:05.343Z [WARN] TestTxnEndpoint_Bad_Size_Net/allowed: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:05.345Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:05.347Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:05.352Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c9117cb9-3fc0-bfdc-3b5a-e4f255a3a509 Address:127.0.0.1:29993}]" writer.go:29: 2021-01-29T19:34:05.355Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server.raft: entering follower state: follower="Node at 127.0.0.1:29993 [Follower]" leader= writer.go:29: 2021-01-29T19:34:05.356Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server.serf.wan: serf: EventMemberJoin: Node-c9117cb9-3fc0-bfdc-3b5a-e4f255a3a509.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:05.363Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server.serf.lan: serf: EventMemberJoin: Node-c9117cb9-3fc0-bfdc-3b5a-e4f255a3a509 127.0.0.1 writer.go:29: 2021-01-29T19:34:05.367Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server: Handled event for server in area: event=member-join server=Node-c9117cb9-3fc0-bfdc-3b5a-e4f255a3a509.dc1 area=wan writer.go:29: 2021-01-29T19:34:05.367Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server: Adding LAN server: server="Node-c9117cb9-3fc0-bfdc-3b5a-e4f255a3a509 (Addr: tcp/127.0.0.1:29993) (DC: dc1)" writer.go:29: 2021-01-29T19:34:05.369Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: Started DNS server: address=127.0.0.1:29988 network=udp writer.go:29: 2021-01-29T19:34:05.371Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: Started DNS server: address=127.0.0.1:29988 network=tcp writer.go:29: 2021-01-29T19:34:05.374Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: Started HTTP server: address=127.0.0.1:29989 network=tcp writer.go:29: 2021-01-29T19:34:05.377Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: started state syncer === CONT TestAgent_PurgeCheckOnDuplicate writer.go:29: 2021-01-29T19:34:05.415Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:05.417Z [INFO] TestAgent_PurgeCheckOnDuplicate-a2: Endpoints down === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:05.419Z [WARN] TestTxnEndpoint_Bad_Size_Net/allowed.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:05.421Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server.raft: entering candidate state: node="Node at 127.0.0.1:29993 [Candidate]" term=2 === CONT TestAgent_RemoveCheck --- PASS: TestAgent_PurgeCheckOnDuplicate (1.17s) === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.435Z [WARN] TestAgent_RemoveCheck: bootstrap = true: do not enable unless necessary === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:05.437Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:05.438Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed.server.raft: vote granted: from=c9117cb9-3fc0-bfdc-3b5a-e4f255a3a509 term=2 tally=1 === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.437Z [DEBUG] TestAgent_RemoveCheck.tlsutil: Update: version=1 === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:05.439Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:05.440Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server.raft: entering leader state: leader="Node at 127.0.0.1:29993 [Leader]" writer.go:29: 2021-01-29T19:34:05.441Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server: cluster leadership acquired === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.440Z [DEBUG] TestAgent_RemoveCheck.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:05.441Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server: New leader elected: payload=Node-c9117cb9-3fc0-bfdc-3b5a-e4f255a3a509 writer.go:29: 2021-01-29T19:34:05.445Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29993 === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.447Z [INFO] TestAgent_RemoveCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e1f994c0-3c64-305d-f421-808008d92b41 Address:127.0.0.1:30005}]" writer.go:29: 2021-01-29T19:34:05.451Z [INFO] TestAgent_RemoveCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:30005 [Follower]" leader= writer.go:29: 2021-01-29T19:34:05.454Z [INFO] TestAgent_RemoveCheck.server.serf.wan: serf: EventMemberJoin: Node-e1f994c0-3c64-305d-f421-808008d92b41.dc1 127.0.0.1 === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:05.457Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.461Z [INFO] TestAgent_RemoveCheck.server.serf.lan: serf: EventMemberJoin: Node-e1f994c0-3c64-305d-f421-808008d92b41 127.0.0.1 === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:05.463Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:05.464Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.464Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed.server: Skipping self join check for node since the cluster is too small: node=Node-c9117cb9-3fc0-bfdc-3b5a-e4f255a3a509 writer.go:29: 2021-01-29T19:34:05.465Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server: member joined, marking health alive: member=Node-c9117cb9-3fc0-bfdc-3b5a-e4f255a3a509 === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.465Z [INFO] TestAgent_RemoveCheck.server: Adding LAN server: server="Node-e1f994c0-3c64-305d-f421-808008d92b41 (Addr: tcp/127.0.0.1:30005) (DC: dc1)" writer.go:29: 2021-01-29T19:34:05.468Z [INFO] TestAgent_RemoveCheck.server: Handled event for server in area: event=member-join server=Node-e1f994c0-3c64-305d-f421-808008d92b41.dc1 area=wan writer.go:29: 2021-01-29T19:34:05.484Z [INFO] TestAgent_RemoveCheck: Started DNS server: address=127.0.0.1:30000 network=tcp writer.go:29: 2021-01-29T19:34:05.487Z [INFO] TestAgent_RemoveCheck: Started DNS server: address=127.0.0.1:30000 network=udp writer.go:29: 2021-01-29T19:34:05.489Z [INFO] TestAgent_RemoveCheck: Started HTTP server: address=127.0.0.1:30001 network=tcp writer.go:29: 2021-01-29T19:34:05.490Z [INFO] TestAgent_RemoveCheck: started state syncer === CONT TestAgent_updateTTLCheck writer.go:29: 2021-01-29T19:34:05.512Z [INFO] TestAgent_updateTTLCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:05.514Z [INFO] TestAgent_updateTTLCheck: Endpoints down --- PASS: TestAgent_updateTTLCheck (1.07s) === CONT TestAgent_AddCheck_Alias_userAndSetToken === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.519Z [WARN] TestAgent_RemoveCheck.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:05.521Z [INFO] TestAgent_RemoveCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:30005 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:05.523Z [DEBUG] TestAgent_RemoveCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:05.525Z [DEBUG] TestAgent_RemoveCheck.server.raft: vote granted: from=e1f994c0-3c64-305d-f421-808008d92b41 term=2 tally=1 writer.go:29: 2021-01-29T19:34:05.526Z [INFO] TestAgent_RemoveCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:05.528Z [INFO] TestAgent_RemoveCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:30005 [Leader]" writer.go:29: 2021-01-29T19:34:05.530Z [INFO] TestAgent_RemoveCheck.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:05.530Z [INFO] TestAgent_RemoveCheck.server: New leader elected: payload=Node-e1f994c0-3c64-305d-f421-808008d92b41 writer.go:29: 2021-01-29T19:34:05.531Z [DEBUG] TestAgent_RemoveCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30005 === CONT TestAgent_PersistCheck writer.go:29: 2021-01-29T19:34:05.536Z [INFO] TestAgent_PersistCheck-a2: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:05.537Z [INFO] TestAgent_PersistCheck-a2: Endpoints down --- PASS: TestAgent_PersistCheck (1.19s) === CONT TestAgent_AddCheck_Alias_userToken === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.543Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.558Z [WARN] TestAgent_AddCheck_Alias_userAndSetToken: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:05.562Z [DEBUG] TestAgent_AddCheck_Alias_userAndSetToken.tlsutil: Update: version=1 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.565Z [WARN] TestAgent_AddCheck_Alias_userToken: bootstrap = true: do not enable unless necessary === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.569Z [INFO] TestAgent_RemoveCheck.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.569Z [DEBUG] TestAgent_AddCheck_Alias_userAndSetToken.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.570Z [DEBUG] TestAgent_AddCheck_Alias_userToken.tlsutil: Update: version=1 === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.571Z [INFO] TestAgent_RemoveCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.574Z [DEBUG] TestAgent_RemoveCheck.server: Skipping self join check for node since the cluster is too small: node=Node-e1f994c0-3c64-305d-f421-808008d92b41 writer.go:29: 2021-01-29T19:34:05.577Z [INFO] TestAgent_RemoveCheck.server: member joined, marking health alive: member=Node-e1f994c0-3c64-305d-f421-808008d92b41 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.575Z [DEBUG] TestAgent_AddCheck_Alias_userToken.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.578Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0d0cfbc2-4aa0-af80-c27e-637fa52aae62 Address:127.0.0.1:29999}]" writer.go:29: 2021-01-29T19:34:05.582Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server.raft: entering follower state: follower="Node at 127.0.0.1:29999 [Follower]" leader= === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.588Z [DEBUG] TestAgent_RemoveCheck: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:05.592Z [INFO] TestAgent_RemoveCheck: Synced node info === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.592Z [INFO] TestAgent_AddCheck_Alias_userToken.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:46a1f6f9-4e26-0788-3ca9-206e0ce07fec Address:127.0.0.1:30011}]" === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.596Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server.serf.wan: serf: EventMemberJoin: Node-0d0cfbc2-4aa0-af80-c27e-637fa52aae62.dc1 127.0.0.1 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.600Z [INFO] TestAgent_AddCheck_Alias_userToken.server.raft: entering follower state: follower="Node at 127.0.0.1:30011 [Follower]" leader= writer.go:29: 2021-01-29T19:34:05.604Z [INFO] TestAgent_AddCheck_Alias_userToken.server.serf.wan: serf: EventMemberJoin: Node-46a1f6f9-4e26-0788-3ca9-206e0ce07fec.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:05.613Z [INFO] TestAgent_AddCheck_Alias_userToken.server.serf.lan: serf: EventMemberJoin: Node-46a1f6f9-4e26-0788-3ca9-206e0ce07fec 127.0.0.1 writer.go:29: 2021-01-29T19:34:05.619Z [INFO] TestAgent_AddCheck_Alias_userToken: Started DNS server: address=127.0.0.1:30006 network=udp writer.go:29: 2021-01-29T19:34:05.621Z [INFO] TestAgent_AddCheck_Alias_userToken.server: Adding LAN server: server="Node-46a1f6f9-4e26-0788-3ca9-206e0ce07fec (Addr: tcp/127.0.0.1:30011) (DC: dc1)" writer.go:29: 2021-01-29T19:34:05.622Z [INFO] TestAgent_AddCheck_Alias_userToken.server: Handled event for server in area: event=member-join server=Node-46a1f6f9-4e26-0788-3ca9-206e0ce07fec.dc1 area=wan writer.go:29: 2021-01-29T19:34:05.627Z [INFO] TestAgent_AddCheck_Alias_userToken: Started DNS server: address=127.0.0.1:30006 network=tcp === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.628Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server.serf.lan: serf: EventMemberJoin: Node-0d0cfbc2-4aa0-af80-c27e-637fa52aae62 127.0.0.1 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.636Z [INFO] TestAgent_AddCheck_Alias_userToken: Started HTTP server: address=127.0.0.1:30007 network=tcp writer.go:29: 2021-01-29T19:34:05.637Z [INFO] TestAgent_AddCheck_Alias_userToken: started state syncer === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.637Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server: Adding LAN server: server="Node-0d0cfbc2-4aa0-af80-c27e-637fa52aae62 (Addr: tcp/127.0.0.1:29999) (DC: dc1)" writer.go:29: 2021-01-29T19:34:05.642Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: Started DNS server: address=127.0.0.1:29994 network=udp writer.go:29: 2021-01-29T19:34:05.647Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: Started DNS server: address=127.0.0.1:29994 network=tcp writer.go:29: 2021-01-29T19:34:05.644Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server: Handled event for server in area: event=member-join server=Node-0d0cfbc2-4aa0-af80-c27e-637fa52aae62.dc1 area=wan === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.657Z [WARN] TestAgent_AddCheck_Alias_userToken.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.645Z [WARN] TestAgent_AddCheck_Alias_userAndSetToken.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.658Z [INFO] TestAgent_AddCheck_Alias_userToken.server.raft: entering candidate state: node="Node at 127.0.0.1:30011 [Candidate]" term=2 === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.653Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: Started HTTP server: address=127.0.0.1:29995 network=tcp writer.go:29: 2021-01-29T19:34:05.659Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server.raft: entering candidate state: node="Node at 127.0.0.1:29999 [Candidate]" term=2 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.663Z [DEBUG] TestAgent_AddCheck_Alias_userToken.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:05.663Z [DEBUG] TestAgent_AddCheck_Alias_userToken.server.raft: vote granted: from=46a1f6f9-4e26-0788-3ca9-206e0ce07fec term=2 tally=1 writer.go:29: 2021-01-29T19:34:05.664Z [INFO] TestAgent_AddCheck_Alias_userToken.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:05.665Z [INFO] TestAgent_AddCheck_Alias_userToken.server.raft: entering leader state: leader="Node at 127.0.0.1:30011 [Leader]" writer.go:29: 2021-01-29T19:34:05.666Z [INFO] TestAgent_AddCheck_Alias_userToken.server: cluster leadership acquired === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.663Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: started state syncer === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.666Z [INFO] TestAgent_AddCheck_Alias_userToken.server: New leader elected: payload=Node-46a1f6f9-4e26-0788-3ca9-206e0ce07fec writer.go:29: 2021-01-29T19:34:05.667Z [DEBUG] TestAgent_AddCheck_Alias_userToken.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30011 === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.666Z [DEBUG] TestAgent_AddCheck_Alias_userAndSetToken.server.raft: votes: needed=1 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.676Z [INFO] TestAgent_AddCheck_Alias_userToken: Requesting shutdown writer.go:29: 2021-01-29T19:34:05.678Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.678Z [DEBUG] TestAgent_AddCheck_Alias_userAndSetToken.server.raft: vote granted: from=0d0cfbc2-4aa0-af80-c27e-637fa52aae62 term=2 tally=1 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.679Z [INFO] TestAgent_AddCheck_Alias_userToken.server: shutting down server writer.go:29: 2021-01-29T19:34:05.682Z [WARN] TestAgent_AddCheck_Alias_userToken.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.682Z [ERROR] TestAgent_AddCheck_Alias_userToken.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:05.685Z [INFO] TestAgent_AddCheck_Alias_userToken.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.681Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server.raft: election won: tally=1 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.688Z [INFO] TestAgent_AddCheck_Alias_userToken.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.688Z [WARN] TestAgent_AddCheck_Alias_userToken.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.690Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server.raft: entering leader state: leader="Node at 127.0.0.1:29999 [Leader]" === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.690Z [DEBUG] TestAgent_AddCheck_Alias_userToken.server: Skipping self join check for node since the cluster is too small: node=Node-46a1f6f9-4e26-0788-3ca9-206e0ce07fec writer.go:29: 2021-01-29T19:34:05.693Z [INFO] TestAgent_AddCheck_Alias_userToken.server.router.manager: shutting down === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.695Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server: cluster leadership acquired === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.694Z [INFO] TestAgent_AddCheck_Alias_userToken: consul server down writer.go:29: 2021-01-29T19:34:05.697Z [INFO] TestAgent_AddCheck_Alias_userToken: shutdown complete === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.696Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server: New leader elected: payload=Node-0d0cfbc2-4aa0-af80-c27e-637fa52aae62 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.696Z [INFO] TestAgent_AddCheck_Alias_userToken.server: member joined, marking health alive: member=Node-46a1f6f9-4e26-0788-3ca9-206e0ce07fec writer.go:29: 2021-01-29T19:34:05.699Z [ERROR] TestAgent_AddCheck_Alias_userToken.server: failed to reconcile member: member="{Node-46a1f6f9-4e26-0788-3ca9-206e0ce07fec 127.0.0.1 30009 map[acls:0 bootstrap:1 build:1.7.4: dc:dc1 id:46a1f6f9-4e26-0788-3ca9-206e0ce07fec port:30011 raft_vsn:3 role:consul segment: vsn:2 vsn_max:3 vsn_min:2 wan_join_port:30010] alive 1 5 2 2 5 4}" error="raft is already shutdown" writer.go:29: 2021-01-29T19:34:05.700Z [DEBUG] TestAgent_AddCheck_Alias_userToken.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.698Z [DEBUG] TestAgent_AddCheck_Alias_userAndSetToken.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29999 === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.698Z [INFO] TestAgent_AddCheck_Alias_userToken: Stopping server: protocol=DNS address=127.0.0.1:30006 network=tcp writer.go:29: 2021-01-29T19:34:05.701Z [DEBUG] TestAgent_AddCheck_Alias_userToken.leader: stopped routine: routine="CA root pruning" === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:05.706Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:05.707Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: Synced node info writer.go:29: 2021-01-29T19:34:05.708Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed: Node info in sync === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:05.702Z [INFO] TestAgent_AddCheck_Alias_userToken: Stopping server: protocol=DNS address=127.0.0.1:30006 network=udp writer.go:29: 2021-01-29T19:34:05.711Z [INFO] TestAgent_AddCheck_Alias_userToken: Stopping server: protocol=HTTP address=127.0.0.1:30007 network=tcp === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.714Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:05.722Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:05.723Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.725Z [DEBUG] TestAgent_AddCheck_Alias_userAndSetToken.server: Skipping self join check for node since the cluster is too small: node=Node-0d0cfbc2-4aa0-af80-c27e-637fa52aae62 writer.go:29: 2021-01-29T19:34:05.727Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server: member joined, marking health alive: member=Node-0d0cfbc2-4aa0-af80-c27e-637fa52aae62 === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:05.755Z [DEBUG] TestAgent_RemoveCheck: removed check: check=mem writer.go:29: 2021-01-29T19:34:05.757Z [DEBUG] TestAgent_RemoveCheck: removed check: check=mem writer.go:29: 2021-01-29T19:34:05.760Z [INFO] TestAgent_RemoveCheck: Requesting shutdown writer.go:29: 2021-01-29T19:34:05.761Z [INFO] TestAgent_RemoveCheck.server: shutting down server writer.go:29: 2021-01-29T19:34:05.763Z [DEBUG] TestAgent_RemoveCheck.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.765Z [WARN] TestAgent_RemoveCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.765Z [DEBUG] TestAgent_RemoveCheck.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.772Z [WARN] TestAgent_RemoveCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.774Z [INFO] TestAgent_RemoveCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:05.774Z [INFO] TestAgent_RemoveCheck: consul server down writer.go:29: 2021-01-29T19:34:05.777Z [INFO] TestAgent_RemoveCheck: shutdown complete writer.go:29: 2021-01-29T19:34:05.778Z [INFO] TestAgent_RemoveCheck: Stopping server: protocol=DNS address=127.0.0.1:30000 network=tcp writer.go:29: 2021-01-29T19:34:05.779Z [INFO] TestAgent_RemoveCheck: Stopping server: protocol=DNS address=127.0.0.1:30000 network=udp writer.go:29: 2021-01-29T19:34:05.780Z [INFO] TestAgent_RemoveCheck: Stopping server: protocol=HTTP address=127.0.0.1:30001 network=tcp === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:05.850Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: Requesting shutdown writer.go:29: 2021-01-29T19:34:05.852Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server: shutting down server writer.go:29: 2021-01-29T19:34:05.853Z [DEBUG] TestAgent_AddCheck_Alias_userAndSetToken.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.855Z [WARN] TestAgent_AddCheck_Alias_userAndSetToken.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.856Z [ERROR] TestAgent_AddCheck_Alias_userAndSetToken.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:05.856Z [DEBUG] TestAgent_AddCheck_Alias_userAndSetToken.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.883Z [WARN] TestAgent_AddCheck_Alias_userAndSetToken.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.886Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:05.887Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: consul server down writer.go:29: 2021-01-29T19:34:05.888Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: shutdown complete writer.go:29: 2021-01-29T19:34:05.889Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: Stopping server: protocol=DNS address=127.0.0.1:29994 network=tcp writer.go:29: 2021-01-29T19:34:05.890Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: Stopping server: protocol=DNS address=127.0.0.1:29994 network=udp writer.go:29: 2021-01-29T19:34:05.891Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: Stopping server: protocol=HTTP address=127.0.0.1:29995 network=tcp === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:05.972Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:05.972Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed: Node info in sync === CONT TestAgent_AddCheck_Alias_userToken writer.go:29: 2021-01-29T19:34:06.212Z [INFO] TestAgent_AddCheck_Alias_userToken: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:06.214Z [INFO] TestAgent_AddCheck_Alias_userToken: Endpoints down --- PASS: TestAgent_AddCheck_Alias_userToken (0.68s) === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.243Z [WARN] TestAgent_AddCheck_Alias_setToken: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:06.249Z [DEBUG] TestAgent_AddCheck_Alias_setToken.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:06.256Z [DEBUG] TestAgent_AddCheck_Alias_setToken.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RemoveCheck writer.go:29: 2021-01-29T19:34:06.283Z [INFO] TestAgent_RemoveCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:06.289Z [INFO] TestAgent_RemoveCheck: Endpoints down --- PASS: TestAgent_RemoveCheck (0.87s) === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.346Z [WARN] TestAgent_AddCheck_Alias: bootstrap = true: do not enable unless necessary === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.347Z [INFO] TestAgent_AddCheck_Alias_setToken.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fef49bd4-1289-8002-fb22-8b7b3d2f05ad Address:127.0.0.1:30023}]" writer.go:29: 2021-01-29T19:34:06.358Z [INFO] TestAgent_AddCheck_Alias_setToken.server.serf.wan: serf: EventMemberJoin: Node-fef49bd4-1289-8002-fb22-8b7b3d2f05ad.dc1 127.0.0.1 === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.366Z [DEBUG] TestAgent_AddCheck_Alias.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:06.387Z [DEBUG] TestAgent_AddCheck_Alias.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.360Z [INFO] TestAgent_AddCheck_Alias_setToken.server.raft: entering follower state: follower="Node at 127.0.0.1:30023 [Follower]" leader= === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:06.402Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: Waiting for endpoints to shut down === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.379Z [INFO] TestAgent_AddCheck_Alias_setToken.server.serf.lan: serf: EventMemberJoin: Node-fef49bd4-1289-8002-fb22-8b7b3d2f05ad 127.0.0.1 writer.go:29: 2021-01-29T19:34:06.428Z [INFO] TestAgent_AddCheck_Alias_setToken: Started DNS server: address=127.0.0.1:30018 network=udp === CONT TestAgent_AddCheck_Alias_userAndSetToken writer.go:29: 2021-01-29T19:34:06.415Z [INFO] TestAgent_AddCheck_Alias_userAndSetToken: Endpoints down --- PASS: TestAgent_AddCheck_Alias_userAndSetToken (0.92s) === CONT TestAgent_AddCheck_GRPC === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:06.435Z [WARN] TestTxnEndpoint_Bad_Size_Net/allowed.server.rpc: Attempting to apply large raft entry: size_in_bytes=4719032 === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.437Z [INFO] TestAgent_AddCheck_Alias_setToken.server: Adding LAN server: server="Node-fef49bd4-1289-8002-fb22-8b7b3d2f05ad (Addr: tcp/127.0.0.1:30023) (DC: dc1)" writer.go:29: 2021-01-29T19:34:06.442Z [INFO] TestAgent_AddCheck_Alias_setToken.server: Handled event for server in area: event=member-join server=Node-fef49bd4-1289-8002-fb22-8b7b3d2f05ad.dc1 area=wan writer.go:29: 2021-01-29T19:34:06.461Z [WARN] TestAgent_AddCheck_Alias_setToken.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:06.483Z [INFO] TestAgent_AddCheck_Alias_setToken.server.raft: entering candidate state: node="Node at 127.0.0.1:30023 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:06.498Z [INFO] TestAgent_AddCheck_Alias_setToken: Started DNS server: address=127.0.0.1:30018 network=tcp === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.503Z [INFO] TestAgent_AddCheck_Alias.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:23f3876e-6570-bb01-f26e-d24e6376783f Address:127.0.0.1:30017}]" === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.529Z [INFO] TestAgent_AddCheck_Alias_setToken: Started HTTP server: address=127.0.0.1:30019 network=tcp writer.go:29: 2021-01-29T19:34:06.531Z [INFO] TestAgent_AddCheck_Alias_setToken: started state syncer === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.521Z [INFO] TestAgent_AddCheck_Alias.server.raft: entering follower state: follower="Node at 127.0.0.1:30017 [Follower]" leader= writer.go:29: 2021-01-29T19:34:06.537Z [INFO] TestAgent_AddCheck_Alias.server.serf.wan: serf: EventMemberJoin: Node-23f3876e-6570-bb01-f26e-d24e6376783f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:06.563Z [INFO] TestAgent_AddCheck_Alias.server.serf.lan: serf: EventMemberJoin: Node-23f3876e-6570-bb01-f26e-d24e6376783f 127.0.0.1 === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.567Z [DEBUG] TestAgent_AddCheck_Alias_setToken.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:06.568Z [DEBUG] TestAgent_AddCheck_Alias_setToken.server.raft: vote granted: from=fef49bd4-1289-8002-fb22-8b7b3d2f05ad term=2 tally=1 writer.go:29: 2021-01-29T19:34:06.569Z [INFO] TestAgent_AddCheck_Alias_setToken.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:06.571Z [INFO] TestAgent_AddCheck_Alias_setToken.server.raft: entering leader state: leader="Node at 127.0.0.1:30023 [Leader]" writer.go:29: 2021-01-29T19:34:06.574Z [INFO] TestAgent_AddCheck_Alias_setToken.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:06.576Z [INFO] TestAgent_AddCheck_Alias_setToken.server: New leader elected: payload=Node-fef49bd4-1289-8002-fb22-8b7b3d2f05ad === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.595Z [INFO] TestAgent_AddCheck_Alias: Started DNS server: address=127.0.0.1:30012 network=udp writer.go:29: 2021-01-29T19:34:06.617Z [INFO] TestAgent_AddCheck_Alias: Started DNS server: address=127.0.0.1:30012 network=tcp writer.go:29: 2021-01-29T19:34:06.622Z [WARN] TestAgent_AddCheck_Alias.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:06.623Z [INFO] TestAgent_AddCheck_Alias.server.raft: entering candidate state: node="Node at 127.0.0.1:30017 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:06.623Z [INFO] TestAgent_AddCheck_Alias: Started HTTP server: address=127.0.0.1:30013 network=tcp writer.go:29: 2021-01-29T19:34:06.625Z [INFO] TestAgent_AddCheck_Alias: started state syncer === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.639Z [DEBUG] TestAgent_AddCheck_Alias_setToken.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30023 === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.609Z [INFO] TestAgent_AddCheck_Alias.server: Handled event for server in area: event=member-join server=Node-23f3876e-6570-bb01-f26e-d24e6376783f.dc1 area=wan writer.go:29: 2021-01-29T19:34:06.599Z [INFO] TestAgent_AddCheck_Alias.server: Adding LAN server: server="Node-23f3876e-6570-bb01-f26e-d24e6376783f (Addr: tcp/127.0.0.1:30017) (DC: dc1)" writer.go:29: 2021-01-29T19:34:06.651Z [WARN] TestAgent_AddCheck_Alias.server.raft: unable to get address for sever, using fallback address: id=23f3876e-6570-bb01-f26e-d24e6376783f fallback=127.0.0.1:30017 error="Could not find address for server id 23f3876e-6570-bb01-f26e-d24e6376783f" === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.653Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.657Z [WARN] TestAgent_AddCheck_GRPC: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:06.658Z [DEBUG] TestAgent_AddCheck_GRPC.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:06.662Z [DEBUG] TestAgent_AddCheck_GRPC.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.676Z [DEBUG] TestAgent_AddCheck_Alias.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:06.677Z [DEBUG] TestAgent_AddCheck_Alias.server.raft: vote granted: from=23f3876e-6570-bb01-f26e-d24e6376783f term=2 tally=1 writer.go:29: 2021-01-29T19:34:06.679Z [INFO] TestAgent_AddCheck_Alias.server.raft: election won: tally=1 === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.682Z [INFO] TestAgent_AddCheck_GRPC.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9f846364-5887-c5e1-aaf2-3d0f19e731b2 Address:127.0.0.1:30029}]" === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.683Z [INFO] TestAgent_AddCheck_Alias.server.raft: entering leader state: leader="Node at 127.0.0.1:30017 [Leader]" writer.go:29: 2021-01-29T19:34:06.685Z [INFO] TestAgent_AddCheck_Alias.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:06.687Z [INFO] TestAgent_AddCheck_Alias.server: New leader elected: payload=Node-23f3876e-6570-bb01-f26e-d24e6376783f === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.689Z [INFO] TestAgent_AddCheck_GRPC.server.raft: entering follower state: follower="Node at 127.0.0.1:30029 [Follower]" leader= === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.694Z [INFO] TestAgent_AddCheck_Alias_setToken.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:06.697Z [INFO] TestAgent_AddCheck_Alias_setToken.leader: started routine: routine="CA root pruning" === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.696Z [DEBUG] TestAgent_AddCheck_Alias.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30017 === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.697Z [INFO] TestAgent_AddCheck_GRPC.server.serf.wan: serf: EventMemberJoin: Node-9f846364-5887-c5e1-aaf2-3d0f19e731b2.dc1 127.0.0.1 === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.702Z [DEBUG] TestAgent_AddCheck_Alias_setToken.server: Skipping self join check for node since the cluster is too small: node=Node-fef49bd4-1289-8002-fb22-8b7b3d2f05ad writer.go:29: 2021-01-29T19:34:06.705Z [INFO] TestAgent_AddCheck_Alias_setToken.server: member joined, marking health alive: member=Node-fef49bd4-1289-8002-fb22-8b7b3d2f05ad === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.717Z [INFO] TestAgent_AddCheck_GRPC.server.serf.lan: serf: EventMemberJoin: Node-9f846364-5887-c5e1-aaf2-3d0f19e731b2 127.0.0.1 writer.go:29: 2021-01-29T19:34:06.721Z [INFO] TestAgent_AddCheck_GRPC: Started DNS server: address=127.0.0.1:30024 network=udp writer.go:29: 2021-01-29T19:34:06.722Z [INFO] TestAgent_AddCheck_GRPC.server: Handled event for server in area: event=member-join server=Node-9f846364-5887-c5e1-aaf2-3d0f19e731b2.dc1 area=wan writer.go:29: 2021-01-29T19:34:06.723Z [INFO] TestAgent_AddCheck_GRPC: Started DNS server: address=127.0.0.1:30024 network=tcp === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.721Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.724Z [INFO] TestAgent_AddCheck_GRPC.server: Adding LAN server: server="Node-9f846364-5887-c5e1-aaf2-3d0f19e731b2 (Addr: tcp/127.0.0.1:30029) (DC: dc1)" writer.go:29: 2021-01-29T19:34:06.726Z [INFO] TestAgent_AddCheck_GRPC: Started HTTP server: address=127.0.0.1:30025 network=tcp writer.go:29: 2021-01-29T19:34:06.731Z [INFO] TestAgent_AddCheck_GRPC: started state syncer === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.737Z [INFO] TestAgent_AddCheck_Alias.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:06.739Z [INFO] TestAgent_AddCheck_Alias.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.742Z [DEBUG] TestAgent_AddCheck_Alias.server: Skipping self join check for node since the cluster is too small: node=Node-23f3876e-6570-bb01-f26e-d24e6376783f writer.go:29: 2021-01-29T19:34:06.745Z [INFO] TestAgent_AddCheck_Alias.server: member joined, marking health alive: member=Node-23f3876e-6570-bb01-f26e-d24e6376783f === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.751Z [WARN] TestAgent_AddCheck_GRPC.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.758Z [INFO] TestAgent_AddCheck_Alias_setToken: Requesting shutdown writer.go:29: 2021-01-29T19:34:06.760Z [INFO] TestAgent_AddCheck_Alias_setToken.server: shutting down server writer.go:29: 2021-01-29T19:34:06.763Z [DEBUG] TestAgent_AddCheck_Alias_setToken.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.757Z [INFO] TestAgent_AddCheck_GRPC.server.raft: entering candidate state: node="Node at 127.0.0.1:30029 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:06.773Z [DEBUG] TestAgent_AddCheck_GRPC.server.raft: votes: needed=1 === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.765Z [WARN] TestAgent_AddCheck_Alias_setToken.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:06.763Z [ERROR] TestAgent_AddCheck_Alias_setToken.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:06.765Z [DEBUG] TestAgent_AddCheck_Alias_setToken.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.787Z [WARN] TestAgent_AddCheck_Alias_setToken.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.778Z [DEBUG] TestAgent_AddCheck_GRPC.server.raft: vote granted: from=9f846364-5887-c5e1-aaf2-3d0f19e731b2 term=2 tally=1 === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.789Z [INFO] TestAgent_AddCheck_Alias_setToken.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:06.791Z [INFO] TestAgent_AddCheck_Alias_setToken: consul server down writer.go:29: 2021-01-29T19:34:06.793Z [INFO] TestAgent_AddCheck_Alias_setToken: shutdown complete === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.790Z [INFO] TestAgent_AddCheck_GRPC.server.raft: election won: tally=1 === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:06.795Z [INFO] TestAgent_AddCheck_Alias_setToken: Stopping server: protocol=DNS address=127.0.0.1:30018 network=tcp writer.go:29: 2021-01-29T19:34:06.800Z [INFO] TestAgent_AddCheck_Alias_setToken: Stopping server: protocol=DNS address=127.0.0.1:30018 network=udp writer.go:29: 2021-01-29T19:34:06.802Z [INFO] TestAgent_AddCheck_Alias_setToken: Stopping server: protocol=HTTP address=127.0.0.1:30019 network=tcp === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.797Z [INFO] TestAgent_AddCheck_GRPC.server.raft: entering leader state: leader="Node at 127.0.0.1:30029 [Leader]" writer.go:29: 2021-01-29T19:34:06.818Z [INFO] TestAgent_AddCheck_GRPC.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:06.820Z [INFO] TestAgent_AddCheck_GRPC.server: New leader elected: payload=Node-9f846364-5887-c5e1-aaf2-3d0f19e731b2 === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:06.837Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: Requesting shutdown writer.go:29: 2021-01-29T19:34:06.838Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server: shutting down server === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.837Z [DEBUG] TestAgent_AddCheck_GRPC.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30029 === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:06.840Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.841Z [WARN] TestTxnEndpoint_Bad_Size_Net/allowed.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.844Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:06.843Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.845Z [WARN] TestTxnEndpoint_Bad_Size_Net/allowed.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:06.850Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed.server.router.manager: shutting down === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.853Z [INFO] TestAgent_AddCheck_GRPC.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:06.854Z [INFO] TestAgent_AddCheck_GRPC.leader: started routine: routine="CA root pruning" === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:06.850Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: consul server down === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.855Z [DEBUG] TestAgent_AddCheck_GRPC.server: Skipping self join check for node since the cluster is too small: node=Node-9f846364-5887-c5e1-aaf2-3d0f19e731b2 === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.856Z [INFO] TestAgent_AddCheck_Alias: Requesting shutdown writer.go:29: 2021-01-29T19:34:06.857Z [INFO] TestAgent_AddCheck_Alias.server: shutting down server writer.go:29: 2021-01-29T19:34:06.858Z [DEBUG] TestAgent_AddCheck_Alias.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.859Z [WARN] TestAgent_AddCheck_Alias.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.857Z [INFO] TestAgent_AddCheck_GRPC.server: member joined, marking health alive: member=Node-9f846364-5887-c5e1-aaf2-3d0f19e731b2 === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.861Z [ERROR] TestAgent_AddCheck_Alias.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:06.862Z [DEBUG] TestAgent_AddCheck_Alias.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.862Z [WARN] TestAgent_AddCheck_Alias.server.serf.wan: serf: Shutdown without a Leave === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:06.856Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: shutdown complete === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.867Z [INFO] TestAgent_AddCheck_Alias.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:06.868Z [INFO] TestAgent_AddCheck_Alias: consul server down writer.go:29: 2021-01-29T19:34:06.870Z [INFO] TestAgent_AddCheck_Alias: shutdown complete === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:06.868Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: Stopping server: protocol=DNS address=127.0.0.1:29988 network=tcp === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.871Z [INFO] TestAgent_AddCheck_Alias: Stopping server: protocol=DNS address=127.0.0.1:30012 network=tcp === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:06.871Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: Stopping server: protocol=DNS address=127.0.0.1:29988 network=udp === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.872Z [INFO] TestAgent_AddCheck_Alias: Stopping server: protocol=DNS address=127.0.0.1:30012 network=udp === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:06.873Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: Stopping server: protocol=HTTP address=127.0.0.1:29989 network=tcp === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:06.875Z [INFO] TestAgent_AddCheck_Alias: Stopping server: protocol=HTTP address=127.0.0.1:30013 network=tcp === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:06.902Z [INFO] TestAgent_AddCheck_GRPC: Requesting shutdown writer.go:29: 2021-01-29T19:34:06.907Z [INFO] TestAgent_AddCheck_GRPC.server: shutting down server writer.go:29: 2021-01-29T19:34:06.908Z [DEBUG] TestAgent_AddCheck_GRPC.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.910Z [WARN] TestAgent_AddCheck_GRPC.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:06.912Z [ERROR] TestAgent_AddCheck_GRPC.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:06.912Z [DEBUG] TestAgent_AddCheck_GRPC.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.927Z [WARN] TestAgent_AddCheck_GRPC.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:06.929Z [INFO] TestAgent_AddCheck_GRPC.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:06.930Z [INFO] TestAgent_AddCheck_GRPC: consul server down writer.go:29: 2021-01-29T19:34:06.931Z [INFO] TestAgent_AddCheck_GRPC: shutdown complete writer.go:29: 2021-01-29T19:34:06.932Z [INFO] TestAgent_AddCheck_GRPC: Stopping server: protocol=DNS address=127.0.0.1:30024 network=tcp writer.go:29: 2021-01-29T19:34:06.933Z [INFO] TestAgent_AddCheck_GRPC: Stopping server: protocol=DNS address=127.0.0.1:30024 network=udp writer.go:29: 2021-01-29T19:34:06.935Z [INFO] TestAgent_AddCheck_GRPC: Stopping server: protocol=HTTP address=127.0.0.1:30025 network=tcp === CONT TestAgent_AddCheck_Alias_setToken writer.go:29: 2021-01-29T19:34:07.306Z [INFO] TestAgent_AddCheck_Alias_setToken: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:07.308Z [INFO] TestAgent_AddCheck_Alias_setToken: Endpoints down --- PASS: TestAgent_AddCheck_Alias_setToken (1.09s) === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:07.324Z [WARN] TestAgent_AddCheck_ExecRemoteDisable: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:07.326Z [DEBUG] TestAgent_AddCheck_ExecRemoteDisable.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:07.328Z [DEBUG] TestAgent_AddCheck_ExecRemoteDisable.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:07.337Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4804a49d-f543-790e-8cfc-d2f68d151c3d Address:127.0.0.1:30035}]" writer.go:29: 2021-01-29T19:34:07.341Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server.raft: entering follower state: follower="Node at 127.0.0.1:30035 [Follower]" leader= writer.go:29: 2021-01-29T19:34:07.343Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server.serf.wan: serf: EventMemberJoin: Node-4804a49d-f543-790e-8cfc-d2f68d151c3d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:07.352Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server.serf.lan: serf: EventMemberJoin: Node-4804a49d-f543-790e-8cfc-d2f68d151c3d 127.0.0.1 writer.go:29: 2021-01-29T19:34:07.358Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server: Adding LAN server: server="Node-4804a49d-f543-790e-8cfc-d2f68d151c3d (Addr: tcp/127.0.0.1:30035) (DC: dc1)" writer.go:29: 2021-01-29T19:34:07.360Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server: Handled event for server in area: event=member-join server=Node-4804a49d-f543-790e-8cfc-d2f68d151c3d.dc1 area=wan writer.go:29: 2021-01-29T19:34:07.361Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: Started DNS server: address=127.0.0.1:30030 network=udp writer.go:29: 2021-01-29T19:34:07.365Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: Started DNS server: address=127.0.0.1:30030 network=tcp writer.go:29: 2021-01-29T19:34:07.367Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: Started HTTP server: address=127.0.0.1:30031 network=tcp writer.go:29: 2021-01-29T19:34:07.370Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: started state syncer === CONT TestTxnEndpoint_Bad_Size_Net/allowed writer.go:29: 2021-01-29T19:34:07.375Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:07.378Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed: Endpoints down === RUN TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible === CONT TestAgent_AddCheck_Alias writer.go:29: 2021-01-29T19:34:07.380Z [INFO] TestAgent_AddCheck_Alias: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:07.390Z [INFO] TestAgent_AddCheck_Alias: Endpoints down --- PASS: TestAgent_AddCheck_Alias (1.10s) === CONT TestAgent_AddCheck_ExecDisable === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:07.398Z [WARN] TestAgent_AddCheck_ExecRemoteDisable.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:07.399Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server.raft: entering candidate state: node="Node at 127.0.0.1:30035 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:07.401Z [DEBUG] TestAgent_AddCheck_ExecRemoteDisable.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:07.402Z [DEBUG] TestAgent_AddCheck_ExecRemoteDisable.server.raft: vote granted: from=4804a49d-f543-790e-8cfc-d2f68d151c3d term=2 tally=1 writer.go:29: 2021-01-29T19:34:07.403Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:07.403Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server.raft: entering leader state: leader="Node at 127.0.0.1:30035 [Leader]" === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.406Z [WARN] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: bootstrap = true: do not enable unless necessary === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:07.405Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server: cluster leadership acquired === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.407Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:07.409Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:07.405Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server: New leader elected: payload=Node-4804a49d-f543-790e-8cfc-d2f68d151c3d writer.go:29: 2021-01-29T19:34:07.406Z [DEBUG] TestAgent_AddCheck_ExecRemoteDisable.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30035 === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.418Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:98fef437-0054-2a4d-67f4-ed9eef966f48 Address:127.0.0.1:30053}]" writer.go:29: 2021-01-29T19:34:07.422Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.serf.wan: serf: EventMemberJoin: Node-98fef437-0054-2a4d-67f4-ed9eef966f48.dc1 127.0.0.1 === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.419Z [WARN] TestAgent_AddCheck_ExecDisable: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:07.426Z [DEBUG] TestAgent_AddCheck_ExecDisable.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:07.429Z [DEBUG] TestAgent_AddCheck_ExecDisable.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:07.434Z [INFO] TestAgent_AddCheck_ExecDisable.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:666accd0-7f6d-42f5-61c8-a837ac212abc Address:127.0.0.1:30047}]" writer.go:29: 2021-01-29T19:34:07.437Z [INFO] TestAgent_AddCheck_ExecDisable.server.raft: entering follower state: follower="Node at 127.0.0.1:30047 [Follower]" leader= === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.424Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.raft: entering follower state: follower="Node at 127.0.0.1:30053 [Follower]" leader= === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:07.440Z [INFO] TestAgent_AddCheck_GRPC: Waiting for endpoints to shut down === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.424Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.serf.lan: serf: EventMemberJoin: Node-98fef437-0054-2a4d-67f4-ed9eef966f48 127.0.0.1 === CONT TestAgent_AddCheck_GRPC writer.go:29: 2021-01-29T19:34:07.442Z [INFO] TestAgent_AddCheck_GRPC: Endpoints down --- PASS: TestAgent_AddCheck_GRPC (1.01s) === CONT TestAgent_AddCheck_RestoreState === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:07.442Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.444Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Started DNS server: address=127.0.0.1:30048 network=udp writer.go:29: 2021-01-29T19:34:07.446Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server: Handled event for server in area: event=member-join server=Node-98fef437-0054-2a4d-67f4-ed9eef966f48.dc1 area=wan writer.go:29: 2021-01-29T19:34:07.446Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server: Adding LAN server: server="Node-98fef437-0054-2a4d-67f4-ed9eef966f48 (Addr: tcp/127.0.0.1:30053) (DC: dc1)" writer.go:29: 2021-01-29T19:34:07.449Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Started DNS server: address=127.0.0.1:30048 network=tcp writer.go:29: 2021-01-29T19:34:07.454Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Started HTTP server: address=127.0.0.1:30049 network=tcp writer.go:29: 2021-01-29T19:34:07.455Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: started state syncer === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.458Z [INFO] TestAgent_AddCheck_ExecDisable.server.serf.wan: serf: EventMemberJoin: Node-666accd0-7f6d-42f5-61c8-a837ac212abc.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:07.461Z [INFO] TestAgent_AddCheck_ExecDisable.server.serf.lan: serf: EventMemberJoin: Node-666accd0-7f6d-42f5-61c8-a837ac212abc 127.0.0.1 === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:07.463Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:07.464Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:07.465Z [DEBUG] TestAgent_AddCheck_ExecRemoteDisable.server: Skipping self join check for node since the cluster is too small: node=Node-4804a49d-f543-790e-8cfc-d2f68d151c3d === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.465Z [INFO] TestAgent_AddCheck_ExecDisable.server: Adding LAN server: server="Node-666accd0-7f6d-42f5-61c8-a837ac212abc (Addr: tcp/127.0.0.1:30047) (DC: dc1)" === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:07.466Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server: member joined, marking health alive: member=Node-4804a49d-f543-790e-8cfc-d2f68d151c3d === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.467Z [INFO] TestAgent_AddCheck_ExecDisable.server: Handled event for server in area: event=member-join server=Node-666accd0-7f6d-42f5-61c8-a837ac212abc.dc1 area=wan writer.go:29: 2021-01-29T19:34:07.466Z [INFO] TestAgent_AddCheck_ExecDisable: Started DNS server: address=127.0.0.1:30042 network=udp writer.go:29: 2021-01-29T19:34:07.469Z [INFO] TestAgent_AddCheck_ExecDisable: Started DNS server: address=127.0.0.1:30042 network=tcp writer.go:29: 2021-01-29T19:34:07.481Z [WARN] TestAgent_AddCheck_ExecDisable.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:07.483Z [INFO] TestAgent_AddCheck_ExecDisable.server.raft: entering candidate state: node="Node at 127.0.0.1:30047 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:07.484Z [INFO] TestAgent_AddCheck_ExecDisable: Started HTTP server: address=127.0.0.1:30043 network=tcp writer.go:29: 2021-01-29T19:34:07.486Z [INFO] TestAgent_AddCheck_ExecDisable: started state syncer === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.487Z [WARN] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:07.489Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.raft: entering candidate state: node="Node at 127.0.0.1:30053 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:07.492Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:07.494Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.raft: vote granted: from=98fef437-0054-2a4d-67f4-ed9eef966f48 term=2 tally=1 writer.go:29: 2021-01-29T19:34:07.496Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.raft: election won: tally=1 === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.499Z [DEBUG] TestAgent_AddCheck_ExecDisable.server.raft: votes: needed=1 === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.504Z [WARN] TestAgent_AddCheck_RestoreState: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:07.505Z [DEBUG] TestAgent_AddCheck_RestoreState.tlsutil: Update: version=1 === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.503Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.raft: entering leader state: leader="Node at 127.0.0.1:30053 [Leader]" === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.505Z [DEBUG] TestAgent_AddCheck_ExecDisable.server.raft: vote granted: from=666accd0-7f6d-42f5-61c8-a837ac212abc term=2 tally=1 === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.507Z [DEBUG] TestAgent_AddCheck_RestoreState.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.507Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:07.508Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server: New leader elected: payload=Node-98fef437-0054-2a4d-67f4-ed9eef966f48 === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.509Z [INFO] TestAgent_AddCheck_ExecDisable.server.raft: election won: tally=1 === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.510Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30053 === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.511Z [INFO] TestAgent_AddCheck_ExecDisable.server.raft: entering leader state: leader="Node at 127.0.0.1:30047 [Leader]" writer.go:29: 2021-01-29T19:34:07.513Z [INFO] TestAgent_AddCheck_ExecDisable.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:07.514Z [INFO] TestAgent_AddCheck_ExecDisable.server: New leader elected: payload=Node-666accd0-7f6d-42f5-61c8-a837ac212abc === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.516Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.516Z [DEBUG] TestAgent_AddCheck_ExecDisable.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30047 === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.522Z [INFO] TestAgent_AddCheck_RestoreState.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:192881eb-4314-62b4-3861-c14fb65e53e4 Address:127.0.0.1:30041}]" === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.523Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.524Z [INFO] TestAgent_AddCheck_RestoreState.server.raft: entering follower state: follower="Node at 127.0.0.1:30041 [Follower]" leader= === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.535Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:07.537Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:07.538Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server: Skipping self join check for node since the cluster is too small: node=Node-98fef437-0054-2a4d-67f4-ed9eef966f48 writer.go:29: 2021-01-29T19:34:07.540Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server: member joined, marking health alive: member=Node-98fef437-0054-2a4d-67f4-ed9eef966f48 === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.532Z [INFO] TestAgent_AddCheck_RestoreState.server.serf.wan: serf: EventMemberJoin: Node-192881eb-4314-62b4-3861-c14fb65e53e4.dc1 127.0.0.1 === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.541Z [INFO] TestAgent_AddCheck_ExecDisable.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.551Z [INFO] TestAgent_AddCheck_RestoreState.server.serf.lan: serf: EventMemberJoin: Node-192881eb-4314-62b4-3861-c14fb65e53e4 127.0.0.1 writer.go:29: 2021-01-29T19:34:07.555Z [INFO] TestAgent_AddCheck_RestoreState.server: Adding LAN server: server="Node-192881eb-4314-62b4-3861-c14fb65e53e4 (Addr: tcp/127.0.0.1:30041) (DC: dc1)" writer.go:29: 2021-01-29T19:34:07.556Z [INFO] TestAgent_AddCheck_RestoreState.server: Handled event for server in area: event=member-join server=Node-192881eb-4314-62b4-3861-c14fb65e53e4.dc1 area=wan writer.go:29: 2021-01-29T19:34:07.559Z [INFO] TestAgent_AddCheck_RestoreState: Started DNS server: address=127.0.0.1:30036 network=udp === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.547Z [INFO] TestAgent_AddCheck_ExecDisable.leader: started routine: routine="CA root pruning" === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.560Z [INFO] TestAgent_AddCheck_RestoreState: Started DNS server: address=127.0.0.1:30036 network=tcp writer.go:29: 2021-01-29T19:34:07.563Z [INFO] TestAgent_AddCheck_RestoreState: Started HTTP server: address=127.0.0.1:30037 network=tcp writer.go:29: 2021-01-29T19:34:07.567Z [INFO] TestAgent_AddCheck_RestoreState: started state syncer === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.561Z [DEBUG] TestAgent_AddCheck_ExecDisable.server: Skipping self join check for node since the cluster is too small: node=Node-666accd0-7f6d-42f5-61c8-a837ac212abc === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.574Z [WARN] TestAgent_AddCheck_RestoreState.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.580Z [INFO] TestAgent_AddCheck_ExecDisable.server: member joined, marking health alive: member=Node-666accd0-7f6d-42f5-61c8-a837ac212abc === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.582Z [INFO] TestAgent_AddCheck_RestoreState.server.raft: entering candidate state: node="Node at 127.0.0.1:30041 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:07.587Z [DEBUG] TestAgent_AddCheck_RestoreState.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:07.591Z [DEBUG] TestAgent_AddCheck_RestoreState.server.raft: vote granted: from=192881eb-4314-62b4-3861-c14fb65e53e4 term=2 tally=1 writer.go:29: 2021-01-29T19:34:07.594Z [INFO] TestAgent_AddCheck_RestoreState.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:07.598Z [INFO] TestAgent_AddCheck_RestoreState.server.raft: entering leader state: leader="Node at 127.0.0.1:30041 [Leader]" writer.go:29: 2021-01-29T19:34:07.602Z [INFO] TestAgent_AddCheck_RestoreState.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:07.603Z [INFO] TestAgent_AddCheck_RestoreState.server: New leader elected: payload=Node-192881eb-4314-62b4-3861-c14fb65e53e4 writer.go:29: 2021-01-29T19:34:07.613Z [DEBUG] TestAgent_AddCheck_RestoreState.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30041 writer.go:29: 2021-01-29T19:34:07.630Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.637Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:07.638Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: Requesting shutdown writer.go:29: 2021-01-29T19:34:07.641Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server: shutting down server writer.go:29: 2021-01-29T19:34:07.643Z [DEBUG] TestAgent_AddCheck_ExecRemoteDisable.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:07.644Z [WARN] TestAgent_AddCheck_ExecRemoteDisable.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:07.644Z [ERROR] TestAgent_AddCheck_ExecRemoteDisable.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:07.640Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Synced node info === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:07.648Z [DEBUG] TestAgent_AddCheck_ExecRemoteDisable.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:07.666Z [WARN] TestAgent_AddCheck_ExecRemoteDisable.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.665Z [INFO] TestAgent_AddCheck_RestoreState.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:07.672Z [INFO] TestAgent_AddCheck_RestoreState.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:07.672Z [DEBUG] TestAgent_AddCheck_RestoreState.server: Skipping self join check for node since the cluster is too small: node=Node-192881eb-4314-62b4-3861-c14fb65e53e4 writer.go:29: 2021-01-29T19:34:07.673Z [INFO] TestAgent_AddCheck_RestoreState.server: member joined, marking health alive: member=Node-192881eb-4314-62b4-3861-c14fb65e53e4 === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:07.670Z [INFO] TestAgent_AddCheck_ExecRemoteDisable.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:07.670Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: consul server down writer.go:29: 2021-01-29T19:34:07.676Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: shutdown complete writer.go:29: 2021-01-29T19:34:07.680Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: Stopping server: protocol=DNS address=127.0.0.1:30030 network=tcp writer.go:29: 2021-01-29T19:34:07.682Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: Stopping server: protocol=DNS address=127.0.0.1:30030 network=udp writer.go:29: 2021-01-29T19:34:07.684Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: Stopping server: protocol=HTTP address=127.0.0.1:30031 network=tcp === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.772Z [INFO] TestAgent_AddCheck_RestoreState: Requesting shutdown writer.go:29: 2021-01-29T19:34:07.773Z [INFO] TestAgent_AddCheck_RestoreState.server: shutting down server writer.go:29: 2021-01-29T19:34:07.774Z [DEBUG] TestAgent_AddCheck_RestoreState.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:07.775Z [WARN] TestAgent_AddCheck_RestoreState.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:07.776Z [DEBUG] TestAgent_AddCheck_RestoreState.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:07.774Z [ERROR] TestAgent_AddCheck_RestoreState.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.783Z [DEBUG] TestAgent_AddCheck_ExecDisable: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.786Z [WARN] TestAgent_AddCheck_RestoreState.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.787Z [INFO] TestAgent_AddCheck_ExecDisable: Synced node info writer.go:29: 2021-01-29T19:34:07.788Z [DEBUG] TestAgent_AddCheck_ExecDisable: Node info in sync === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:07.789Z [INFO] TestAgent_AddCheck_RestoreState.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:07.791Z [INFO] TestAgent_AddCheck_RestoreState: consul server down writer.go:29: 2021-01-29T19:34:07.792Z [INFO] TestAgent_AddCheck_RestoreState: shutdown complete writer.go:29: 2021-01-29T19:34:07.793Z [INFO] TestAgent_AddCheck_RestoreState: Stopping server: protocol=DNS address=127.0.0.1:30036 network=tcp writer.go:29: 2021-01-29T19:34:07.794Z [INFO] TestAgent_AddCheck_RestoreState: Stopping server: protocol=DNS address=127.0.0.1:30036 network=udp writer.go:29: 2021-01-29T19:34:07.795Z [INFO] TestAgent_AddCheck_RestoreState: Stopping server: protocol=HTTP address=127.0.0.1:30037 network=tcp === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:07.855Z [INFO] TestAgent_AddCheck_ExecDisable: Requesting shutdown writer.go:29: 2021-01-29T19:34:07.858Z [INFO] TestAgent_AddCheck_ExecDisable.server: shutting down server writer.go:29: 2021-01-29T19:34:07.859Z [DEBUG] TestAgent_AddCheck_ExecDisable.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:07.860Z [WARN] TestAgent_AddCheck_ExecDisable.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:07.862Z [DEBUG] TestAgent_AddCheck_ExecDisable.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:07.867Z [WARN] TestAgent_AddCheck_ExecDisable.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:07.870Z [INFO] TestAgent_AddCheck_ExecDisable.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:07.871Z [INFO] TestAgent_AddCheck_ExecDisable: consul server down writer.go:29: 2021-01-29T19:34:07.874Z [INFO] TestAgent_AddCheck_ExecDisable: shutdown complete writer.go:29: 2021-01-29T19:34:07.875Z [INFO] TestAgent_AddCheck_ExecDisable: Stopping server: protocol=DNS address=127.0.0.1:30042 network=tcp writer.go:29: 2021-01-29T19:34:07.876Z [INFO] TestAgent_AddCheck_ExecDisable: Stopping server: protocol=DNS address=127.0.0.1:30042 network=udp writer.go:29: 2021-01-29T19:34:07.877Z [INFO] TestAgent_AddCheck_ExecDisable: Stopping server: protocol=HTTP address=127.0.0.1:30043 network=tcp === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:08.140Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:08.148Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Node info in sync writer.go:29: 2021-01-29T19:34:08.148Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Node info in sync === CONT TestAgent_AddCheck_ExecRemoteDisable writer.go:29: 2021-01-29T19:34:08.187Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:08.191Z [INFO] TestAgent_AddCheck_ExecRemoteDisable: Endpoints down --- PASS: TestAgent_AddCheck_ExecRemoteDisable (0.89s) === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.235Z [WARN] TestAgent_AddCheck_MissingService: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:08.237Z [DEBUG] TestAgent_AddCheck_MissingService.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:08.240Z [DEBUG] TestAgent_AddCheck_MissingService.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:08.252Z [INFO] TestAgent_AddCheck_MissingService.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8155644a-2871-7657-dffb-145acc54c6c0 Address:127.0.0.1:30077}]" writer.go:29: 2021-01-29T19:34:08.257Z [INFO] TestAgent_AddCheck_MissingService.server.serf.wan: serf: EventMemberJoin: Node-8155644a-2871-7657-dffb-145acc54c6c0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:08.265Z [INFO] TestAgent_AddCheck_MissingService.server.serf.lan: serf: EventMemberJoin: Node-8155644a-2871-7657-dffb-145acc54c6c0 127.0.0.1 writer.go:29: 2021-01-29T19:34:08.281Z [INFO] TestAgent_AddCheck_MissingService: Started DNS server: address=127.0.0.1:30072 network=udp writer.go:29: 2021-01-29T19:34:08.275Z [INFO] TestAgent_AddCheck_MissingService.server.raft: entering follower state: follower="Node at 127.0.0.1:30077 [Follower]" leader= writer.go:29: 2021-01-29T19:34:08.286Z [INFO] TestAgent_AddCheck_MissingService.server: Adding LAN server: server="Node-8155644a-2871-7657-dffb-145acc54c6c0 (Addr: tcp/127.0.0.1:30077) (DC: dc1)" writer.go:29: 2021-01-29T19:34:08.287Z [INFO] TestAgent_AddCheck_MissingService.server: Handled event for server in area: event=member-join server=Node-8155644a-2871-7657-dffb-145acc54c6c0.dc1 area=wan writer.go:29: 2021-01-29T19:34:08.297Z [INFO] TestAgent_AddCheck_MissingService: Started DNS server: address=127.0.0.1:30072 network=tcp writer.go:29: 2021-01-29T19:34:08.311Z [INFO] TestAgent_AddCheck_MissingService: Started HTTP server: address=127.0.0.1:30073 network=tcp writer.go:29: 2021-01-29T19:34:08.313Z [INFO] TestAgent_AddCheck_MissingService: started state syncer === CONT TestAgent_AddCheck_RestoreState writer.go:29: 2021-01-29T19:34:08.315Z [INFO] TestAgent_AddCheck_RestoreState: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:08.317Z [INFO] TestAgent_AddCheck_RestoreState: Endpoints down --- PASS: TestAgent_AddCheck_RestoreState (0.87s) === CONT TestAgent_AddCheck_MinInterval === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.349Z [WARN] TestAgent_AddCheck_MissingService.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:08.352Z [INFO] TestAgent_AddCheck_MissingService.server.raft: entering candidate state: node="Node at 127.0.0.1:30077 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:08.370Z [DEBUG] TestAgent_AddCheck_MissingService.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:08.372Z [DEBUG] TestAgent_AddCheck_MissingService.server.raft: vote granted: from=8155644a-2871-7657-dffb-145acc54c6c0 term=2 tally=1 writer.go:29: 2021-01-29T19:34:08.375Z [INFO] TestAgent_AddCheck_MissingService.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:08.377Z [INFO] TestAgent_AddCheck_MissingService.server.raft: entering leader state: leader="Node at 127.0.0.1:30077 [Leader]" writer.go:29: 2021-01-29T19:34:08.379Z [INFO] TestAgent_AddCheck_MissingService.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:08.382Z [INFO] TestAgent_AddCheck_MissingService.server: New leader elected: payload=Node-8155644a-2871-7657-dffb-145acc54c6c0 === CONT TestAgent_AddCheck_ExecDisable writer.go:29: 2021-01-29T19:34:08.384Z [INFO] TestAgent_AddCheck_ExecDisable: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:08.391Z [INFO] TestAgent_AddCheck_ExecDisable: Endpoints down --- PASS: TestAgent_AddCheck_ExecDisable (1.00s) === CONT TestAgent_AddCheck_StartPassing === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.402Z [DEBUG] TestAgent_AddCheck_MissingService.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30077 === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.427Z [WARN] TestAgent_AddCheck_StartPassing: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:08.428Z [DEBUG] TestAgent_AddCheck_StartPassing.tlsutil: Update: version=1 === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.428Z [WARN] TestAgent_AddCheck_MinInterval: bootstrap = true: do not enable unless necessary === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.430Z [DEBUG] TestAgent_AddCheck_StartPassing.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.433Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.430Z [DEBUG] TestAgent_AddCheck_MinInterval.tlsutil: Update: version=1 === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.438Z [INFO] TestAgent_AddCheck_StartPassing.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e43cf42c-f16b-ad4a-3720-6aad08f61a0e Address:127.0.0.1:30065}]" === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.442Z [DEBUG] TestAgent_AddCheck_MinInterval.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.447Z [INFO] TestAgent_AddCheck_MissingService.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:08.448Z [INFO] TestAgent_AddCheck_MissingService.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:08.450Z [DEBUG] TestAgent_AddCheck_MissingService.server: Skipping self join check for node since the cluster is too small: node=Node-8155644a-2871-7657-dffb-145acc54c6c0 writer.go:29: 2021-01-29T19:34:08.451Z [INFO] TestAgent_AddCheck_MissingService.server: member joined, marking health alive: member=Node-8155644a-2871-7657-dffb-145acc54c6c0 === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.456Z [INFO] TestAgent_AddCheck_MinInterval.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2bfaa898-9588-235c-dfd4-2b4037c7ea4c Address:127.0.0.1:30059}]" === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.444Z [INFO] TestAgent_AddCheck_StartPassing.server.raft: entering follower state: follower="Node at 127.0.0.1:30065 [Follower]" leader= === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.462Z [INFO] TestAgent_AddCheck_MinInterval.server.raft: entering follower state: follower="Node at 127.0.0.1:30059 [Follower]" leader= === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.460Z [INFO] TestAgent_AddCheck_StartPassing.server.serf.wan: serf: EventMemberJoin: Node-e43cf42c-f16b-ad4a-3720-6aad08f61a0e.dc1 127.0.0.1 === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.469Z [INFO] TestAgent_AddCheck_MinInterval.server.serf.wan: serf: EventMemberJoin: Node-2bfaa898-9588-235c-dfd4-2b4037c7ea4c.dc1 127.0.0.1 === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.476Z [INFO] TestAgent_AddCheck_StartPassing.server.serf.lan: serf: EventMemberJoin: Node-e43cf42c-f16b-ad4a-3720-6aad08f61a0e 127.0.0.1 === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.481Z [INFO] TestAgent_AddCheck_MinInterval.server.serf.lan: serf: EventMemberJoin: Node-2bfaa898-9588-235c-dfd4-2b4037c7ea4c 127.0.0.1 === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.483Z [INFO] TestAgent_AddCheck_StartPassing.server: Adding LAN server: server="Node-e43cf42c-f16b-ad4a-3720-6aad08f61a0e (Addr: tcp/127.0.0.1:30065) (DC: dc1)" writer.go:29: 2021-01-29T19:34:08.484Z [INFO] TestAgent_AddCheck_StartPassing.server: Handled event for server in area: event=member-join server=Node-e43cf42c-f16b-ad4a-3720-6aad08f61a0e.dc1 area=wan === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.485Z [INFO] TestAgent_AddCheck_MinInterval: Started DNS server: address=127.0.0.1:30054 network=udp === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.485Z [INFO] TestAgent_AddCheck_StartPassing: Started DNS server: address=127.0.0.1:30060 network=udp === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.487Z [INFO] TestAgent_AddCheck_MinInterval.server: Adding LAN server: server="Node-2bfaa898-9588-235c-dfd4-2b4037c7ea4c (Addr: tcp/127.0.0.1:30059) (DC: dc1)" writer.go:29: 2021-01-29T19:34:08.488Z [INFO] TestAgent_AddCheck_MinInterval.server: Handled event for server in area: event=member-join server=Node-2bfaa898-9588-235c-dfd4-2b4037c7ea4c.dc1 area=wan === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.488Z [INFO] TestAgent_AddCheck_StartPassing: Started DNS server: address=127.0.0.1:30060 network=tcp writer.go:29: 2021-01-29T19:34:08.492Z [INFO] TestAgent_AddCheck_StartPassing: Started HTTP server: address=127.0.0.1:30061 network=tcp writer.go:29: 2021-01-29T19:34:08.494Z [INFO] TestAgent_AddCheck_StartPassing: started state syncer === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.490Z [INFO] TestAgent_AddCheck_MinInterval: Started DNS server: address=127.0.0.1:30054 network=tcp writer.go:29: 2021-01-29T19:34:08.503Z [INFO] TestAgent_AddCheck_MinInterval: Started HTTP server: address=127.0.0.1:30055 network=tcp writer.go:29: 2021-01-29T19:34:08.504Z [INFO] TestAgent_AddCheck_MinInterval: started state syncer === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:08.531Z [WARN] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.rpc: Attempting to apply large raft entry: size_in_bytes=4719032 === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.533Z [WARN] TestAgent_AddCheck_StartPassing.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:08.534Z [INFO] TestAgent_AddCheck_StartPassing.server.raft: entering candidate state: node="Node at 127.0.0.1:30065 [Candidate]" term=2 === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.536Z [WARN] TestAgent_AddCheck_MinInterval.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.537Z [DEBUG] TestAgent_AddCheck_StartPassing.server.raft: votes: needed=1 === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.538Z [INFO] TestAgent_AddCheck_MinInterval.server.raft: entering candidate state: node="Node at 127.0.0.1:30059 [Candidate]" term=2 === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.538Z [DEBUG] TestAgent_AddCheck_StartPassing.server.raft: vote granted: from=e43cf42c-f16b-ad4a-3720-6aad08f61a0e term=2 tally=1 === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.540Z [DEBUG] TestAgent_AddCheck_MinInterval.server.raft: votes: needed=1 === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.540Z [INFO] TestAgent_AddCheck_StartPassing.server.raft: election won: tally=1 === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.541Z [DEBUG] TestAgent_AddCheck_MinInterval.server.raft: vote granted: from=2bfaa898-9588-235c-dfd4-2b4037c7ea4c term=2 tally=1 writer.go:29: 2021-01-29T19:34:08.543Z [INFO] TestAgent_AddCheck_MinInterval.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:08.545Z [INFO] TestAgent_AddCheck_MinInterval.server.raft: entering leader state: leader="Node at 127.0.0.1:30059 [Leader]" writer.go:29: 2021-01-29T19:34:08.546Z [INFO] TestAgent_AddCheck_MinInterval.server: cluster leadership acquired === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.547Z [INFO] TestAgent_AddCheck_MissingService: Requesting shutdown === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.542Z [INFO] TestAgent_AddCheck_StartPassing.server.raft: entering leader state: leader="Node at 127.0.0.1:30065 [Leader]" === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.549Z [INFO] TestAgent_AddCheck_MissingService.server: shutting down server === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.547Z [INFO] TestAgent_AddCheck_MinInterval.server: New leader elected: payload=Node-2bfaa898-9588-235c-dfd4-2b4037c7ea4c === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.552Z [DEBUG] TestAgent_AddCheck_MissingService.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:08.553Z [WARN] TestAgent_AddCheck_MissingService.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.553Z [INFO] TestAgent_AddCheck_StartPassing.server: cluster leadership acquired === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.556Z [ERROR] TestAgent_AddCheck_MissingService.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:08.558Z [DEBUG] TestAgent_AddCheck_MissingService.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.558Z [INFO] TestAgent_AddCheck_StartPassing.server: New leader elected: payload=Node-e43cf42c-f16b-ad4a-3720-6aad08f61a0e === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:08.560Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Requesting shutdown === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.561Z [WARN] TestAgent_AddCheck_MissingService.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:08.564Z [INFO] TestAgent_AddCheck_MissingService.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:08.566Z [INFO] TestAgent_AddCheck_MissingService: consul server down === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.562Z [DEBUG] TestAgent_AddCheck_StartPassing.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30065 === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.567Z [INFO] TestAgent_AddCheck_MissingService: shutdown complete writer.go:29: 2021-01-29T19:34:08.569Z [INFO] TestAgent_AddCheck_MissingService: Stopping server: protocol=DNS address=127.0.0.1:30072 network=tcp === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:08.561Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server: shutting down server === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.571Z [INFO] TestAgent_AddCheck_MissingService: Stopping server: protocol=DNS address=127.0.0.1:30072 network=udp === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.573Z [DEBUG] TestAgent_AddCheck_MinInterval.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30059 === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:08.573Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.579Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:08.576Z [WARN] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:08.576Z [INFO] TestAgent_AddCheck_MissingService: Stopping server: protocol=HTTP address=127.0.0.1:30073 network=tcp === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:08.582Z [DEBUG] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.584Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:08.585Z [WARN] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:08.593Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible.server.router.manager: shutting down === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.597Z [INFO] TestAgent_AddCheck_StartPassing.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:08.598Z [INFO] TestAgent_AddCheck_StartPassing.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:08.599Z [DEBUG] TestAgent_AddCheck_StartPassing.server: Skipping self join check for node since the cluster is too small: node=Node-e43cf42c-f16b-ad4a-3720-6aad08f61a0e writer.go:29: 2021-01-29T19:34:08.600Z [INFO] TestAgent_AddCheck_StartPassing.server: member joined, marking health alive: member=Node-e43cf42c-f16b-ad4a-3720-6aad08f61a0e === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:08.595Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: consul server down writer.go:29: 2021-01-29T19:34:08.602Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: shutdown complete writer.go:29: 2021-01-29T19:34:08.605Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Stopping server: protocol=DNS address=127.0.0.1:30048 network=tcp writer.go:29: 2021-01-29T19:34:08.610Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Stopping server: protocol=DNS address=127.0.0.1:30048 network=udp === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.612Z [INFO] TestAgent_AddCheck_MinInterval.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:08.616Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Stopping server: protocol=HTTP address=127.0.0.1:30049 network=tcp === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.617Z [INFO] TestAgent_AddCheck_MinInterval.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:08.623Z [DEBUG] TestAgent_AddCheck_MinInterval.server: Skipping self join check for node since the cluster is too small: node=Node-2bfaa898-9588-235c-dfd4-2b4037c7ea4c writer.go:29: 2021-01-29T19:34:08.628Z [INFO] TestAgent_AddCheck_MinInterval.server: member joined, marking health alive: member=Node-2bfaa898-9588-235c-dfd4-2b4037c7ea4c === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.648Z [INFO] TestAgent_AddCheck_StartPassing: Requesting shutdown writer.go:29: 2021-01-29T19:34:08.650Z [INFO] TestAgent_AddCheck_StartPassing.server: shutting down server writer.go:29: 2021-01-29T19:34:08.651Z [DEBUG] TestAgent_AddCheck_StartPassing.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:08.663Z [WARN] TestAgent_AddCheck_StartPassing.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:08.660Z [ERROR] TestAgent_AddCheck_StartPassing.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:08.666Z [DEBUG] TestAgent_AddCheck_StartPassing.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.669Z [DEBUG] TestAgent_AddCheck_MinInterval: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:08.683Z [INFO] TestAgent_AddCheck_MinInterval: Synced node info writer.go:29: 2021-01-29T19:34:08.685Z [DEBUG] TestAgent_AddCheck_MinInterval: Node info in sync === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:08.683Z [WARN] TestAgent_AddCheck_StartPassing.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:08.688Z [INFO] TestAgent_AddCheck_StartPassing.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:08.694Z [INFO] TestAgent_AddCheck_StartPassing: consul server down writer.go:29: 2021-01-29T19:34:08.695Z [INFO] TestAgent_AddCheck_StartPassing: shutdown complete writer.go:29: 2021-01-29T19:34:08.697Z [INFO] TestAgent_AddCheck_StartPassing: Stopping server: protocol=DNS address=127.0.0.1:30060 network=tcp writer.go:29: 2021-01-29T19:34:08.700Z [INFO] TestAgent_AddCheck_StartPassing: Stopping server: protocol=DNS address=127.0.0.1:30060 network=udp writer.go:29: 2021-01-29T19:34:08.701Z [INFO] TestAgent_AddCheck_StartPassing: Stopping server: protocol=HTTP address=127.0.0.1:30061 network=tcp === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:08.880Z [WARN] TestAgent_AddCheck_MinInterval: check has interval below minimum: check=mem minimum_interval=1s writer.go:29: 2021-01-29T19:34:08.881Z [INFO] TestAgent_AddCheck_MinInterval: Requesting shutdown writer.go:29: 2021-01-29T19:34:08.882Z [INFO] TestAgent_AddCheck_MinInterval.server: shutting down server writer.go:29: 2021-01-29T19:34:08.884Z [DEBUG] TestAgent_AddCheck_MinInterval.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:08.885Z [WARN] TestAgent_AddCheck_MinInterval.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:08.885Z [DEBUG] TestAgent_AddCheck_MinInterval.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:08.890Z [WARN] TestAgent_AddCheck_MinInterval.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:08.892Z [INFO] TestAgent_AddCheck_MinInterval.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:08.892Z [INFO] TestAgent_AddCheck_MinInterval: consul server down writer.go:29: 2021-01-29T19:34:08.896Z [INFO] TestAgent_AddCheck_MinInterval: shutdown complete writer.go:29: 2021-01-29T19:34:08.897Z [INFO] TestAgent_AddCheck_MinInterval: Stopping server: protocol=DNS address=127.0.0.1:30054 network=tcp writer.go:29: 2021-01-29T19:34:08.899Z [INFO] TestAgent_AddCheck_MinInterval: Stopping server: protocol=DNS address=127.0.0.1:30054 network=udp writer.go:29: 2021-01-29T19:34:08.901Z [INFO] TestAgent_AddCheck_MinInterval: Stopping server: protocol=HTTP address=127.0.0.1:30055 network=tcp === CONT TestAgent_AddCheck_MissingService writer.go:29: 2021-01-29T19:34:09.082Z [INFO] TestAgent_AddCheck_MissingService: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:09.084Z [INFO] TestAgent_AddCheck_MissingService: Endpoints down --- PASS: TestAgent_AddCheck_MissingService (0.89s) === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.098Z [WARN] TestAgent_AddCheck: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:09.100Z [DEBUG] TestAgent_AddCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:09.102Z [DEBUG] TestAgent_AddCheck.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:09.109Z [INFO] TestAgent_AddCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3a711362-3e73-b4c9-8b18-12b4c902266c Address:127.0.0.1:30071}]" writer.go:29: 2021-01-29T19:34:09.111Z [INFO] TestAgent_AddCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:30071 [Follower]" leader= writer.go:29: 2021-01-29T19:34:09.115Z [INFO] TestAgent_AddCheck.server.serf.wan: serf: EventMemberJoin: Node-3a711362-3e73-b4c9-8b18-12b4c902266c.dc1 127.0.0.1 === CONT TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible writer.go:29: 2021-01-29T19:34:09.122Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:09.124Z [INFO] TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible: Endpoints down --- PASS: TestTxnEndpoint_Bad_Size_Net (8.22s) --- PASS: TestTxnEndpoint_Bad_Size_Net/exceeds_default_limits (1.59s) --- PASS: TestTxnEndpoint_Bad_Size_Net/exceeds_configured_max_txn_len (1.50s) --- PASS: TestTxnEndpoint_Bad_Size_Net/exceeds_default_max_kv_value_size (1.34s) --- PASS: TestTxnEndpoint_Bad_Size_Net/allowed (2.05s) --- PASS: TestTxnEndpoint_Bad_Size_Net/allowed_kv_max_backward_compatible (1.74s) === CONT TestAgent_IndexChurn === RUN TestAgent_IndexChurn/no_tags === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.121Z [INFO] TestAgent_AddCheck.server.serf.lan: serf: EventMemberJoin: Node-3a711362-3e73-b4c9-8b18-12b4c902266c 127.0.0.1 writer.go:29: 2021-01-29T19:34:09.144Z [INFO] TestAgent_AddCheck.server: Handled event for server in area: event=member-join server=Node-3a711362-3e73-b4c9-8b18-12b4c902266c.dc1 area=wan === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.146Z [WARN] TestAgent_IndexChurn/no_tags: bootstrap = true: do not enable unless necessary === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.145Z [INFO] TestAgent_AddCheck.server: Adding LAN server: server="Node-3a711362-3e73-b4c9-8b18-12b4c902266c (Addr: tcp/127.0.0.1:30071) (DC: dc1)" === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.147Z [DEBUG] TestAgent_IndexChurn/no_tags.tlsutil: Update: version=1 === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.147Z [INFO] TestAgent_AddCheck: Started DNS server: address=127.0.0.1:30066 network=udp === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.149Z [DEBUG] TestAgent_IndexChurn/no_tags.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.149Z [INFO] TestAgent_AddCheck: Started DNS server: address=127.0.0.1:30066 network=tcp === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.154Z [INFO] TestAgent_IndexChurn/no_tags.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5c8f32f7-68d1-8b0c-2361-169acd066d19 Address:127.0.0.1:30083}]" writer.go:29: 2021-01-29T19:34:09.156Z [INFO] TestAgent_IndexChurn/no_tags.server.serf.wan: serf: EventMemberJoin: Node-5c8f32f7-68d1-8b0c-2361-169acd066d19.dc1 127.0.0.1 === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.152Z [INFO] TestAgent_AddCheck: Started HTTP server: address=127.0.0.1:30067 network=tcp === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.156Z [INFO] TestAgent_IndexChurn/no_tags.server.raft: entering follower state: follower="Node at 127.0.0.1:30083 [Follower]" leader= writer.go:29: 2021-01-29T19:34:09.158Z [INFO] TestAgent_IndexChurn/no_tags.server.serf.lan: serf: EventMemberJoin: Node-5c8f32f7-68d1-8b0c-2361-169acd066d19 127.0.0.1 writer.go:29: 2021-01-29T19:34:09.169Z [INFO] TestAgent_IndexChurn/no_tags.server: Adding LAN server: server="Node-5c8f32f7-68d1-8b0c-2361-169acd066d19 (Addr: tcp/127.0.0.1:30083) (DC: dc1)" writer.go:29: 2021-01-29T19:34:09.169Z [INFO] TestAgent_IndexChurn/no_tags.server: Handled event for server in area: event=member-join server=Node-5c8f32f7-68d1-8b0c-2361-169acd066d19.dc1 area=wan writer.go:29: 2021-01-29T19:34:09.171Z [INFO] TestAgent_IndexChurn/no_tags: Started DNS server: address=127.0.0.1:30078 network=tcp writer.go:29: 2021-01-29T19:34:09.172Z [INFO] TestAgent_IndexChurn/no_tags: Started DNS server: address=127.0.0.1:30078 network=udp === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.158Z [INFO] TestAgent_AddCheck: started state syncer === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.175Z [INFO] TestAgent_IndexChurn/no_tags: Started HTTP server: address=127.0.0.1:30079 network=tcp writer.go:29: 2021-01-29T19:34:09.176Z [INFO] TestAgent_IndexChurn/no_tags: started state syncer === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.182Z [WARN] TestAgent_AddCheck.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:09.183Z [INFO] TestAgent_AddCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:30071 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:09.185Z [DEBUG] TestAgent_AddCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:09.186Z [DEBUG] TestAgent_AddCheck.server.raft: vote granted: from=3a711362-3e73-b4c9-8b18-12b4c902266c term=2 tally=1 writer.go:29: 2021-01-29T19:34:09.188Z [INFO] TestAgent_AddCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:09.189Z [INFO] TestAgent_AddCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:30071 [Leader]" writer.go:29: 2021-01-29T19:34:09.191Z [INFO] TestAgent_AddCheck.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:09.194Z [INFO] TestAgent_AddCheck.server: New leader elected: payload=Node-3a711362-3e73-b4c9-8b18-12b4c902266c === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.211Z [WARN] TestAgent_IndexChurn/no_tags.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:09.213Z [INFO] TestAgent_IndexChurn/no_tags.server.raft: entering candidate state: node="Node at 127.0.0.1:30083 [Candidate]" term=2 === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.216Z [DEBUG] TestAgent_AddCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30071 === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:09.215Z [INFO] TestAgent_AddCheck_StartPassing: Waiting for endpoints to shut down === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.219Z [DEBUG] TestAgent_IndexChurn/no_tags.server.raft: votes: needed=1 === CONT TestAgent_AddCheck_StartPassing writer.go:29: 2021-01-29T19:34:09.218Z [INFO] TestAgent_AddCheck_StartPassing: Endpoints down --- PASS: TestAgent_AddCheck_StartPassing (0.83s) === CONT TestAgent_makeNodeID === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.220Z [DEBUG] TestAgent_IndexChurn/no_tags.server.raft: vote granted: from=5c8f32f7-68d1-8b0c-2361-169acd066d19 term=2 tally=1 writer.go:29: 2021-01-29T19:34:09.221Z [INFO] TestAgent_IndexChurn/no_tags.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:09.222Z [INFO] TestAgent_IndexChurn/no_tags.server.raft: entering leader state: leader="Node at 127.0.0.1:30083 [Leader]" === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.223Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.223Z [INFO] TestAgent_IndexChurn/no_tags.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:09.223Z [INFO] TestAgent_IndexChurn/no_tags.server: New leader elected: payload=Node-5c8f32f7-68d1-8b0c-2361-169acd066d19 === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.229Z [INFO] TestAgent_AddCheck.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:09.230Z [INFO] TestAgent_AddCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.230Z [DEBUG] TestAgent_AddCheck.server: Skipping self join check for node since the cluster is too small: node=Node-3a711362-3e73-b4c9-8b18-12b4c902266c writer.go:29: 2021-01-29T19:34:09.231Z [INFO] TestAgent_AddCheck.server: member joined, marking health alive: member=Node-3a711362-3e73-b4c9-8b18-12b4c902266c === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.224Z [DEBUG] TestAgent_IndexChurn/no_tags.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30083 writer.go:29: 2021-01-29T19:34:09.247Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:09.256Z [INFO] TestAgent_IndexChurn/no_tags.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:09.259Z [INFO] TestAgent_IndexChurn/no_tags.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.262Z [DEBUG] TestAgent_IndexChurn/no_tags.server: Skipping self join check for node since the cluster is too small: node=Node-5c8f32f7-68d1-8b0c-2361-169acd066d19 writer.go:29: 2021-01-29T19:34:09.264Z [INFO] TestAgent_IndexChurn/no_tags.server: member joined, marking health alive: member=Node-5c8f32f7-68d1-8b0c-2361-169acd066d19 === CONT TestAgent_makeNodeID writer.go:29: 2021-01-29T19:34:09.274Z [WARN] TestAgent_makeNodeID: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:09.277Z [DEBUG] TestAgent_makeNodeID: Using random ID as node ID: id=c146b7c3-2b57-64bf-2a29-64cecf4d3c8c writer.go:29: 2021-01-29T19:34:09.281Z [DEBUG] TestAgent_makeNodeID.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:09.284Z [DEBUG] TestAgent_makeNodeID.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:09.291Z [INFO] TestAgent_makeNodeID.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c146b7c3-2b57-64bf-2a29-64cecf4d3c8c Address:127.0.0.1:30089}]" writer.go:29: 2021-01-29T19:34:09.295Z [INFO] TestAgent_makeNodeID.server.raft: entering follower state: follower="Node at 127.0.0.1:30089 [Follower]" leader= writer.go:29: 2021-01-29T19:34:09.296Z [INFO] TestAgent_makeNodeID.server.serf.wan: serf: EventMemberJoin: Node-1e2aa212-9dff-2d12-8aa3-ddc52057f133.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:09.304Z [INFO] TestAgent_makeNodeID.server.serf.lan: serf: EventMemberJoin: Node-1e2aa212-9dff-2d12-8aa3-ddc52057f133 127.0.0.1 writer.go:29: 2021-01-29T19:34:09.308Z [INFO] TestAgent_makeNodeID.server: Adding LAN server: server="Node-1e2aa212-9dff-2d12-8aa3-ddc52057f133 (Addr: tcp/127.0.0.1:30089) (DC: dc1)" writer.go:29: 2021-01-29T19:34:09.311Z [INFO] TestAgent_makeNodeID.server: Handled event for server in area: event=member-join server=Node-1e2aa212-9dff-2d12-8aa3-ddc52057f133.dc1 area=wan writer.go:29: 2021-01-29T19:34:09.313Z [INFO] TestAgent_makeNodeID: Started DNS server: address=127.0.0.1:30084 network=tcp writer.go:29: 2021-01-29T19:34:09.314Z [INFO] TestAgent_makeNodeID: Started DNS server: address=127.0.0.1:30084 network=udp writer.go:29: 2021-01-29T19:34:09.316Z [INFO] TestAgent_makeNodeID: Started HTTP server: address=127.0.0.1:30085 network=tcp writer.go:29: 2021-01-29T19:34:09.318Z [INFO] TestAgent_makeNodeID: started state syncer writer.go:29: 2021-01-29T19:34:09.348Z [WARN] TestAgent_makeNodeID.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:09.350Z [INFO] TestAgent_makeNodeID.server.raft: entering candidate state: node="Node at 127.0.0.1:30089 [Candidate]" term=2 === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.352Z [INFO] TestAgent_AddCheck: Requesting shutdown === CONT TestAgent_makeNodeID writer.go:29: 2021-01-29T19:34:09.353Z [DEBUG] TestAgent_makeNodeID.server.raft: votes: needed=1 === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.354Z [INFO] TestAgent_AddCheck.server: shutting down server writer.go:29: 2021-01-29T19:34:09.356Z [DEBUG] TestAgent_AddCheck.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.356Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.358Z [WARN] TestAgent_AddCheck.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.359Z [INFO] TestAgent_IndexChurn/no_tags: Synced node info === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.357Z [ERROR] TestAgent_AddCheck.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.362Z [INFO] TestAgent_IndexChurn/no_tags: Synced service: service=redis writer.go:29: 2021-01-29T19:34:09.364Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.358Z [DEBUG] TestAgent_AddCheck.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.361Z [WARN] TestAgent_AddCheck.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_makeNodeID writer.go:29: 2021-01-29T19:34:09.355Z [DEBUG] TestAgent_makeNodeID.server.raft: vote granted: from=c146b7c3-2b57-64bf-2a29-64cecf4d3c8c term=2 tally=1 === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.367Z [INFO] TestAgent_IndexChurn/no_tags: Synced check: check=node-check === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.371Z [INFO] TestAgent_AddCheck.server.router.manager: shutting down === CONT TestAgent_makeNodeID writer.go:29: 2021-01-29T19:34:09.369Z [INFO] TestAgent_makeNodeID.server.raft: election won: tally=1 === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.373Z [INFO] TestAgent_AddCheck: consul server down === CONT TestAgent_makeNodeID writer.go:29: 2021-01-29T19:34:09.374Z [INFO] TestAgent_makeNodeID.server.raft: entering leader state: leader="Node at 127.0.0.1:30089 [Leader]" writer.go:29: 2021-01-29T19:34:09.377Z [INFO] TestAgent_makeNodeID.server: cluster leadership acquired === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.376Z [INFO] TestAgent_AddCheck: shutdown complete writer.go:29: 2021-01-29T19:34:09.379Z [INFO] TestAgent_AddCheck: Stopping server: protocol=DNS address=127.0.0.1:30066 network=tcp writer.go:29: 2021-01-29T19:34:09.380Z [INFO] TestAgent_AddCheck: Stopping server: protocol=DNS address=127.0.0.1:30066 network=udp writer.go:29: 2021-01-29T19:34:09.381Z [INFO] TestAgent_AddCheck: Stopping server: protocol=HTTP address=127.0.0.1:30067 network=tcp === CONT TestAgent_makeNodeID writer.go:29: 2021-01-29T19:34:09.381Z [INFO] TestAgent_makeNodeID.server: New leader elected: payload=Node-1e2aa212-9dff-2d12-8aa3-ddc52057f133 writer.go:29: 2021-01-29T19:34:09.382Z [DEBUG] TestAgent_makeNodeID.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30089 writer.go:29: 2021-01-29T19:34:09.394Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:09.402Z [INFO] TestAgent_makeNodeID.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_AddCheck_MinInterval writer.go:29: 2021-01-29T19:34:09.403Z [INFO] TestAgent_AddCheck_MinInterval: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:09.403Z [INFO] TestAgent_AddCheck_MinInterval: Endpoints down --- PASS: TestAgent_AddCheck_MinInterval (1.09s) === CONT TestAgent_setupNodeID === CONT TestAgent_makeNodeID writer.go:29: 2021-01-29T19:34:09.403Z [INFO] TestAgent_makeNodeID.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.405Z [DEBUG] TestAgent_makeNodeID.server: Skipping self join check for node since the cluster is too small: node=Node-1e2aa212-9dff-2d12-8aa3-ddc52057f133 writer.go:29: 2021-01-29T19:34:09.406Z [INFO] TestAgent_makeNodeID.server: member joined, marking health alive: member=Node-1e2aa212-9dff-2d12-8aa3-ddc52057f133 === CONT TestAgent_setupNodeID writer.go:29: 2021-01-29T19:34:09.414Z [WARN] TestAgent_setupNodeID: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:09.415Z [DEBUG] TestAgent_setupNodeID: Using random ID as node ID: id=018aff4f-d2c4-3833-8109-8cdaa4f9c231 writer.go:29: 2021-01-29T19:34:09.419Z [DEBUG] TestAgent_setupNodeID.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:09.423Z [DEBUG] TestAgent_setupNodeID.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:09.438Z [INFO] TestAgent_setupNodeID.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:018aff4f-d2c4-3833-8109-8cdaa4f9c231 Address:127.0.0.1:30095}]" writer.go:29: 2021-01-29T19:34:09.440Z [INFO] TestAgent_setupNodeID.server.raft: entering follower state: follower="Node at 127.0.0.1:30095 [Follower]" leader= writer.go:29: 2021-01-29T19:34:09.440Z [INFO] TestAgent_setupNodeID.server.serf.wan: serf: EventMemberJoin: Node-c7b7ab70-5583-be9f-5ef5-54f90ffd7759.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:09.450Z [INFO] TestAgent_setupNodeID.server.serf.lan: serf: EventMemberJoin: Node-c7b7ab70-5583-be9f-5ef5-54f90ffd7759 127.0.0.1 writer.go:29: 2021-01-29T19:34:09.456Z [INFO] TestAgent_setupNodeID.server: Adding LAN server: server="Node-c7b7ab70-5583-be9f-5ef5-54f90ffd7759 (Addr: tcp/127.0.0.1:30095) (DC: dc1)" writer.go:29: 2021-01-29T19:34:09.456Z [INFO] TestAgent_setupNodeID.server: Handled event for server in area: event=member-join server=Node-c7b7ab70-5583-be9f-5ef5-54f90ffd7759.dc1 area=wan writer.go:29: 2021-01-29T19:34:09.459Z [INFO] TestAgent_setupNodeID: Started DNS server: address=127.0.0.1:30090 network=udp writer.go:29: 2021-01-29T19:34:09.461Z [INFO] TestAgent_setupNodeID: Started DNS server: address=127.0.0.1:30090 network=tcp writer.go:29: 2021-01-29T19:34:09.464Z [INFO] TestAgent_setupNodeID: Started HTTP server: address=127.0.0.1:30091 network=tcp writer.go:29: 2021-01-29T19:34:09.465Z [INFO] TestAgent_setupNodeID: started state syncer writer.go:29: 2021-01-29T19:34:09.489Z [WARN] TestAgent_setupNodeID.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:09.492Z [INFO] TestAgent_setupNodeID.server.raft: entering candidate state: node="Node at 127.0.0.1:30095 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:09.495Z [DEBUG] TestAgent_setupNodeID.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:09.497Z [DEBUG] TestAgent_setupNodeID.server.raft: vote granted: from=018aff4f-d2c4-3833-8109-8cdaa4f9c231 term=2 tally=1 writer.go:29: 2021-01-29T19:34:09.499Z [INFO] TestAgent_setupNodeID.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:09.501Z [INFO] TestAgent_setupNodeID.server.raft: entering leader state: leader="Node at 127.0.0.1:30095 [Leader]" writer.go:29: 2021-01-29T19:34:09.503Z [INFO] TestAgent_setupNodeID.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:09.503Z [INFO] TestAgent_setupNodeID.server: New leader elected: payload=Node-c7b7ab70-5583-be9f-5ef5-54f90ffd7759 writer.go:29: 2021-01-29T19:34:09.504Z [DEBUG] TestAgent_setupNodeID.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30095 writer.go:29: 2021-01-29T19:34:09.513Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:09.520Z [INFO] TestAgent_setupNodeID: Requesting shutdown writer.go:29: 2021-01-29T19:34:09.522Z [INFO] TestAgent_setupNodeID.server: shutting down server writer.go:29: 2021-01-29T19:34:09.524Z [WARN] TestAgent_setupNodeID.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:09.524Z [ERROR] TestAgent_setupNodeID.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:09.526Z [WARN] TestAgent_setupNodeID.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:09.534Z [INFO] TestAgent_setupNodeID.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:09.534Z [ERROR] TestAgent_setupNodeID.server: failed to establish leadership: error="error generating CA root certificate: leadership lost while committing log" writer.go:29: 2021-01-29T19:34:09.534Z [INFO] TestAgent_setupNodeID: consul server down writer.go:29: 2021-01-29T19:34:09.548Z [INFO] TestAgent_setupNodeID: shutdown complete writer.go:29: 2021-01-29T19:34:09.552Z [INFO] TestAgent_setupNodeID: Stopping server: protocol=DNS address=127.0.0.1:30090 network=tcp writer.go:29: 2021-01-29T19:34:09.556Z [INFO] TestAgent_setupNodeID: Stopping server: protocol=DNS address=127.0.0.1:30090 network=udp writer.go:29: 2021-01-29T19:34:09.558Z [INFO] TestAgent_setupNodeID: Stopping server: protocol=HTTP address=127.0.0.1:30091 network=tcp === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.603Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.604Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync writer.go:29: 2021-01-29T19:34:09.605Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:09.606Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:09.608Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check === CONT TestAgent_makeNodeID writer.go:29: 2021-01-29T19:34:09.660Z [DEBUG] TestAgent_makeNodeID: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.663Z [INFO] TestAgent_makeNodeID: Synced node info writer.go:29: 2021-01-29T19:34:09.665Z [DEBUG] TestAgent_makeNodeID: Node info in sync writer.go:29: 2021-01-29T19:34:09.673Z [DEBUG] TestAgent_makeNodeID: Using random ID as node ID: id=eee920d1-7540-0a77-4279-e241cccae51b writer.go:29: 2021-01-29T19:34:09.675Z [DEBUG] TestAgent_makeNodeID: Using random ID as node ID: id=162de0e4-d47e-d02b-ccbc-fbafc22a6bbf writer.go:29: 2021-01-29T19:34:09.679Z [DEBUG] TestAgent_makeNodeID: Using unique ID from host as node ID: id=ac28a8ca-265b-bb99-fcdf-fd3c3241462a writer.go:29: 2021-01-29T19:34:09.682Z [DEBUG] TestAgent_makeNodeID: Using unique ID from host as node ID: id=ac28a8ca-265b-bb99-fcdf-fd3c3241462a writer.go:29: 2021-01-29T19:34:09.693Z [INFO] TestAgent_makeNodeID: Requesting shutdown writer.go:29: 2021-01-29T19:34:09.695Z [INFO] TestAgent_makeNodeID.server: shutting down server writer.go:29: 2021-01-29T19:34:09.697Z [DEBUG] TestAgent_makeNodeID.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.698Z [WARN] TestAgent_makeNodeID.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:09.698Z [DEBUG] TestAgent_makeNodeID.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.701Z [WARN] TestAgent_makeNodeID.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:09.703Z [INFO] TestAgent_makeNodeID.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:09.704Z [INFO] TestAgent_makeNodeID: consul server down writer.go:29: 2021-01-29T19:34:09.705Z [INFO] TestAgent_makeNodeID: shutdown complete writer.go:29: 2021-01-29T19:34:09.706Z [INFO] TestAgent_makeNodeID: Stopping server: protocol=DNS address=127.0.0.1:30084 network=tcp writer.go:29: 2021-01-29T19:34:09.707Z [INFO] TestAgent_makeNodeID: Stopping server: protocol=DNS address=127.0.0.1:30084 network=udp writer.go:29: 2021-01-29T19:34:09.708Z [INFO] TestAgent_makeNodeID: Stopping server: protocol=HTTP address=127.0.0.1:30085 network=tcp === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.871Z [DEBUG] TestAgent_IndexChurn/no_tags: Registered node: node="Node-level check" writer.go:29: 2021-01-29T19:34:09.873Z [DEBUG] TestAgent_IndexChurn/no_tags: Registered node: node="Serf Health Status" writer.go:29: 2021-01-29T19:34:09.874Z [DEBUG] TestAgent_IndexChurn/no_tags: Registered node: node="Service-level check" writer.go:29: 2021-01-29T19:34:09.876Z [INFO] TestAgent_IndexChurn/no_tags: Sync in progress: iteration=1 writer.go:29: 2021-01-29T19:34:09.877Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.879Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync writer.go:29: 2021-01-29T19:34:09.880Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:09.882Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check === CONT TestAgent_AddCheck writer.go:29: 2021-01-29T19:34:09.882Z [INFO] TestAgent_AddCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:09.883Z [INFO] TestAgent_AddCheck: Endpoints down --- PASS: TestAgent_AddCheck (0.80s) === CONT TestAgent_ReconnectConfigWanDisabled === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.883Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:09.884Z [INFO] TestAgent_IndexChurn/no_tags: Sync in progress: iteration=2 writer.go:29: 2021-01-29T19:34:09.886Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.888Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync writer.go:29: 2021-01-29T19:34:09.889Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:09.891Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:09.893Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:09.894Z [INFO] TestAgent_IndexChurn/no_tags: Sync in progress: iteration=3 writer.go:29: 2021-01-29T19:34:09.896Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.897Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync writer.go:29: 2021-01-29T19:34:09.899Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:09.900Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:09.902Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:09.903Z [INFO] TestAgent_IndexChurn/no_tags: Sync in progress: iteration=4 writer.go:29: 2021-01-29T19:34:09.905Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.906Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync writer.go:29: 2021-01-29T19:34:09.907Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:09.908Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:09.909Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:09.910Z [INFO] TestAgent_IndexChurn/no_tags: Sync in progress: iteration=5 writer.go:29: 2021-01-29T19:34:09.910Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.911Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync writer.go:29: 2021-01-29T19:34:09.912Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:09.913Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.912Z [WARN] TestAgent_ReconnectConfigWanDisabled: bootstrap = true: do not enable unless necessary === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.914Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:09.915Z [INFO] TestAgent_IndexChurn/no_tags: Sync in progress: iteration=6 === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.915Z [DEBUG] TestAgent_ReconnectConfigWanDisabled.tlsutil: Update: version=1 === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.916Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.917Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.917Z [DEBUG] TestAgent_ReconnectConfigWanDisabled.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.918Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:09.919Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:09.919Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:09.920Z [INFO] TestAgent_IndexChurn/no_tags: Sync in progress: iteration=7 writer.go:29: 2021-01-29T19:34:09.921Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.922Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.922Z [INFO] TestAgent_ReconnectConfigWanDisabled.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f287c82e-7a1b-fd21-c3dd-e5477439e2eb Address:127.0.0.1:30107}]" === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.923Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:09.924Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:09.925Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:09.926Z [INFO] TestAgent_IndexChurn/no_tags: Sync in progress: iteration=8 === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.925Z [INFO] TestAgent_ReconnectConfigWanDisabled.server.raft: entering follower state: follower="Node at 127.0.0.1:30107 [Follower]" leader= === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.927Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.929Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync writer.go:29: 2021-01-29T19:34:09.930Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.929Z [INFO] TestAgent_ReconnectConfigWanDisabled.server.serf.lan: serf: EventMemberJoin: Node-f287c82e-7a1b-fd21-c3dd-e5477439e2eb 127.0.0.1 === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.932Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:09.933Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.934Z [INFO] TestAgent_ReconnectConfigWanDisabled.server: Adding LAN server: server="Node-f287c82e-7a1b-fd21-c3dd-e5477439e2eb (Addr: tcp/127.0.0.1:30107) (DC: dc1)" === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.935Z [INFO] TestAgent_IndexChurn/no_tags: Sync in progress: iteration=9 writer.go:29: 2021-01-29T19:34:09.936Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.937Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync writer.go:29: 2021-01-29T19:34:09.938Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:09.939Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.938Z [INFO] TestAgent_ReconnectConfigWanDisabled: Started DNS server: address=127.0.0.1:30102 network=tcp === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.940Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:09.941Z [INFO] TestAgent_IndexChurn/no_tags: Sync in progress: iteration=10 === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.940Z [INFO] TestAgent_ReconnectConfigWanDisabled: Started DNS server: address=127.0.0.1:30102 network=udp === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.942Z [DEBUG] TestAgent_IndexChurn/no_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:09.942Z [DEBUG] TestAgent_IndexChurn/no_tags: Node info in sync writer.go:29: 2021-01-29T19:34:09.943Z [DEBUG] TestAgent_IndexChurn/no_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:09.944Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=redis-check === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.944Z [INFO] TestAgent_ReconnectConfigWanDisabled: Started HTTP server: address=127.0.0.1:30103 network=tcp === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.945Z [DEBUG] TestAgent_IndexChurn/no_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:09.946Z [INFO] TestAgent_IndexChurn/no_tags: Requesting shutdown === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.946Z [INFO] TestAgent_ReconnectConfigWanDisabled: started state syncer === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:09.947Z [INFO] TestAgent_IndexChurn/no_tags.server: shutting down server writer.go:29: 2021-01-29T19:34:09.948Z [DEBUG] TestAgent_IndexChurn/no_tags.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.949Z [WARN] TestAgent_IndexChurn/no_tags.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:09.949Z [DEBUG] TestAgent_IndexChurn/no_tags.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.951Z [WARN] TestAgent_IndexChurn/no_tags.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:09.956Z [INFO] TestAgent_IndexChurn/no_tags.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:09.956Z [INFO] TestAgent_IndexChurn/no_tags: consul server down writer.go:29: 2021-01-29T19:34:09.959Z [INFO] TestAgent_IndexChurn/no_tags: shutdown complete writer.go:29: 2021-01-29T19:34:09.960Z [INFO] TestAgent_IndexChurn/no_tags: Stopping server: protocol=DNS address=127.0.0.1:30078 network=tcp writer.go:29: 2021-01-29T19:34:09.961Z [INFO] TestAgent_IndexChurn/no_tags: Stopping server: protocol=DNS address=127.0.0.1:30078 network=udp writer.go:29: 2021-01-29T19:34:09.962Z [INFO] TestAgent_IndexChurn/no_tags: Stopping server: protocol=HTTP address=127.0.0.1:30079 network=tcp === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:09.977Z [WARN] TestAgent_ReconnectConfigWanDisabled.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:09.979Z [INFO] TestAgent_ReconnectConfigWanDisabled.server.raft: entering candidate state: node="Node at 127.0.0.1:30107 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:09.981Z [DEBUG] TestAgent_ReconnectConfigWanDisabled.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:09.983Z [DEBUG] TestAgent_ReconnectConfigWanDisabled.server.raft: vote granted: from=f287c82e-7a1b-fd21-c3dd-e5477439e2eb term=2 tally=1 writer.go:29: 2021-01-29T19:34:09.986Z [INFO] TestAgent_ReconnectConfigWanDisabled.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:09.988Z [INFO] TestAgent_ReconnectConfigWanDisabled.server.raft: entering leader state: leader="Node at 127.0.0.1:30107 [Leader]" writer.go:29: 2021-01-29T19:34:09.996Z [INFO] TestAgent_ReconnectConfigWanDisabled.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:09.998Z [INFO] TestAgent_ReconnectConfigWanDisabled.server: New leader elected: payload=Node-f287c82e-7a1b-fd21-c3dd-e5477439e2eb writer.go:29: 2021-01-29T19:34:09.999Z [DEBUG] TestAgent_ReconnectConfigWanDisabled.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30107 writer.go:29: 2021-01-29T19:34:10.012Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:10.019Z [INFO] TestAgent_ReconnectConfigWanDisabled.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:10.021Z [INFO] TestAgent_ReconnectConfigWanDisabled.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.023Z [DEBUG] TestAgent_ReconnectConfigWanDisabled.server: Skipping self join check for node since the cluster is too small: node=Node-f287c82e-7a1b-fd21-c3dd-e5477439e2eb writer.go:29: 2021-01-29T19:34:10.025Z [INFO] TestAgent_ReconnectConfigWanDisabled.server: member joined, marking health alive: member=Node-f287c82e-7a1b-fd21-c3dd-e5477439e2eb === CONT TestAgent_setupNodeID writer.go:29: 2021-01-29T19:34:10.059Z [INFO] TestAgent_setupNodeID: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:10.062Z [INFO] TestAgent_setupNodeID: Endpoints down --- PASS: TestAgent_setupNodeID (0.66s) === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.076Z [WARN] TestAgent_ReconnectConfigSettings: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:10.078Z [DEBUG] TestAgent_ReconnectConfigSettings.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:10.081Z [DEBUG] TestAgent_ReconnectConfigSettings.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:10.094Z [INFO] TestAgent_ReconnectConfigSettings.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7cde8c36-0e5c-9bdd-7b78-b19257bde8fb Address:127.0.0.1:30101}]" writer.go:29: 2021-01-29T19:34:10.097Z [INFO] TestAgent_ReconnectConfigSettings.server.raft: entering follower state: follower="Node at 127.0.0.1:30101 [Follower]" leader= writer.go:29: 2021-01-29T19:34:10.098Z [INFO] TestAgent_ReconnectConfigSettings.server.serf.wan: serf: EventMemberJoin: Node-7cde8c36-0e5c-9bdd-7b78-b19257bde8fb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.106Z [INFO] TestAgent_ReconnectConfigSettings.server.serf.lan: serf: EventMemberJoin: Node-7cde8c36-0e5c-9bdd-7b78-b19257bde8fb 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.112Z [INFO] TestAgent_ReconnectConfigSettings.server: Adding LAN server: server="Node-7cde8c36-0e5c-9bdd-7b78-b19257bde8fb (Addr: tcp/127.0.0.1:30101) (DC: dc1)" writer.go:29: 2021-01-29T19:34:10.114Z [INFO] TestAgent_ReconnectConfigSettings.server: Handled event for server in area: event=member-join server=Node-7cde8c36-0e5c-9bdd-7b78-b19257bde8fb.dc1 area=wan writer.go:29: 2021-01-29T19:34:10.116Z [INFO] TestAgent_ReconnectConfigSettings: Started DNS server: address=127.0.0.1:30096 network=tcp writer.go:29: 2021-01-29T19:34:10.118Z [INFO] TestAgent_ReconnectConfigSettings: Started DNS server: address=127.0.0.1:30096 network=udp writer.go:29: 2021-01-29T19:34:10.121Z [INFO] TestAgent_ReconnectConfigSettings: Started HTTP server: address=127.0.0.1:30097 network=tcp === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:10.123Z [DEBUG] TestAgent_ReconnectConfigWanDisabled: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.123Z [INFO] TestAgent_ReconnectConfigSettings: started state syncer === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:10.126Z [INFO] TestAgent_ReconnectConfigWanDisabled: Synced node info writer.go:29: 2021-01-29T19:34:10.128Z [DEBUG] TestAgent_ReconnectConfigWanDisabled: Node info in sync === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.165Z [WARN] TestAgent_ReconnectConfigSettings.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:10.167Z [INFO] TestAgent_ReconnectConfigSettings.server.raft: entering candidate state: node="Node at 127.0.0.1:30101 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:10.169Z [DEBUG] TestAgent_ReconnectConfigSettings.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:10.171Z [DEBUG] TestAgent_ReconnectConfigSettings.server.raft: vote granted: from=7cde8c36-0e5c-9bdd-7b78-b19257bde8fb term=2 tally=1 writer.go:29: 2021-01-29T19:34:10.173Z [INFO] TestAgent_ReconnectConfigSettings.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:10.177Z [INFO] TestAgent_ReconnectConfigSettings.server.raft: entering leader state: leader="Node at 127.0.0.1:30101 [Leader]" writer.go:29: 2021-01-29T19:34:10.179Z [INFO] TestAgent_ReconnectConfigSettings.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:10.182Z [INFO] TestAgent_ReconnectConfigSettings.server: New leader elected: payload=Node-7cde8c36-0e5c-9bdd-7b78-b19257bde8fb writer.go:29: 2021-01-29T19:34:10.198Z [DEBUG] TestAgent_ReconnectConfigSettings.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30101 === CONT TestAgent_makeNodeID writer.go:29: 2021-01-29T19:34:10.209Z [INFO] TestAgent_makeNodeID: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:10.213Z [INFO] TestAgent_makeNodeID: Endpoints down --- PASS: TestAgent_makeNodeID (1.00s) === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.234Z [WARN] TestAgent_TokenStore: bootstrap = true: do not enable unless necessary === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.235Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.236Z [DEBUG] TestAgent_TokenStore.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:10.239Z [DEBUG] TestAgent_TokenStore.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:10.249Z [INFO] TestAgent_TokenStore.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:66837bdf-3cec-10cc-5868-2f5f86e40622 Address:127.0.0.1:30113}]" writer.go:29: 2021-01-29T19:34:10.252Z [INFO] TestAgent_TokenStore.server.raft: entering follower state: follower="Node at 127.0.0.1:30113 [Follower]" leader= writer.go:29: 2021-01-29T19:34:10.253Z [INFO] TestAgent_TokenStore.server.serf.wan: serf: EventMemberJoin: Node-66837bdf-3cec-10cc-5868-2f5f86e40622.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.255Z [INFO] TestAgent_TokenStore.server.serf.lan: serf: EventMemberJoin: Node-66837bdf-3cec-10cc-5868-2f5f86e40622 127.0.0.1 === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.254Z [INFO] TestAgent_ReconnectConfigSettings.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.259Z [INFO] TestAgent_TokenStore.server: Adding LAN server: server="Node-66837bdf-3cec-10cc-5868-2f5f86e40622 (Addr: tcp/127.0.0.1:30113) (DC: dc1)" writer.go:29: 2021-01-29T19:34:10.259Z [INFO] TestAgent_TokenStore.server: Handled event for server in area: event=member-join server=Node-66837bdf-3cec-10cc-5868-2f5f86e40622.dc1 area=wan === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.258Z [INFO] TestAgent_ReconnectConfigSettings.leader: started routine: routine="CA root pruning" === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.261Z [INFO] TestAgent_TokenStore: Started DNS server: address=127.0.0.1:30108 network=tcp === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.261Z [DEBUG] TestAgent_ReconnectConfigSettings.server: Skipping self join check for node since the cluster is too small: node=Node-7cde8c36-0e5c-9bdd-7b78-b19257bde8fb === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.262Z [INFO] TestAgent_TokenStore: Started DNS server: address=127.0.0.1:30108 network=udp writer.go:29: 2021-01-29T19:34:10.266Z [INFO] TestAgent_TokenStore: Started HTTP server: address=127.0.0.1:30109 network=tcp writer.go:29: 2021-01-29T19:34:10.268Z [INFO] TestAgent_TokenStore: started state syncer === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:10.270Z [INFO] TestAgent_ReconnectConfigWanDisabled: Requesting shutdown writer.go:29: 2021-01-29T19:34:10.273Z [INFO] TestAgent_ReconnectConfigWanDisabled.server: shutting down server writer.go:29: 2021-01-29T19:34:10.276Z [DEBUG] TestAgent_ReconnectConfigWanDisabled.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.277Z [WARN] TestAgent_ReconnectConfigWanDisabled.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.277Z [DEBUG] TestAgent_ReconnectConfigWanDisabled.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.263Z [INFO] TestAgent_ReconnectConfigSettings.server: member joined, marking health alive: member=Node-7cde8c36-0e5c-9bdd-7b78-b19257bde8fb === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:10.281Z [INFO] TestAgent_ReconnectConfigWanDisabled: consul server down writer.go:29: 2021-01-29T19:34:10.283Z [INFO] TestAgent_ReconnectConfigWanDisabled: shutdown complete writer.go:29: 2021-01-29T19:34:10.284Z [INFO] TestAgent_ReconnectConfigWanDisabled: Stopping server: protocol=DNS address=127.0.0.1:30102 network=tcp writer.go:29: 2021-01-29T19:34:10.286Z [INFO] TestAgent_ReconnectConfigWanDisabled: Stopping server: protocol=DNS address=127.0.0.1:30102 network=udp writer.go:29: 2021-01-29T19:34:10.288Z [INFO] TestAgent_ReconnectConfigWanDisabled: Stopping server: protocol=HTTP address=127.0.0.1:30103 network=tcp === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.290Z [INFO] TestAgent_ReconnectConfigSettings: Requesting shutdown writer.go:29: 2021-01-29T19:34:10.293Z [INFO] TestAgent_ReconnectConfigSettings.server: shutting down server writer.go:29: 2021-01-29T19:34:10.295Z [DEBUG] TestAgent_ReconnectConfigSettings.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.297Z [WARN] TestAgent_ReconnectConfigSettings.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.297Z [DEBUG] TestAgent_ReconnectConfigSettings.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.295Z [ERROR] TestAgent_ReconnectConfigSettings.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:10.300Z [WARN] TestAgent_ReconnectConfigSettings.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.308Z [INFO] TestAgent_ReconnectConfigSettings.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:10.308Z [INFO] TestAgent_ReconnectConfigSettings: consul server down writer.go:29: 2021-01-29T19:34:10.312Z [INFO] TestAgent_ReconnectConfigSettings: shutdown complete writer.go:29: 2021-01-29T19:34:10.313Z [INFO] TestAgent_ReconnectConfigSettings: Stopping server: protocol=DNS address=127.0.0.1:30096 network=tcp writer.go:29: 2021-01-29T19:34:10.314Z [INFO] TestAgent_ReconnectConfigSettings: Stopping server: protocol=DNS address=127.0.0.1:30096 network=udp writer.go:29: 2021-01-29T19:34:10.315Z [INFO] TestAgent_ReconnectConfigSettings: Stopping server: protocol=HTTP address=127.0.0.1:30097 network=tcp === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.323Z [WARN] TestAgent_TokenStore.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:10.324Z [INFO] TestAgent_TokenStore.server.raft: entering candidate state: node="Node at 127.0.0.1:30113 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:10.326Z [DEBUG] TestAgent_TokenStore.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:10.327Z [DEBUG] TestAgent_TokenStore.server.raft: vote granted: from=66837bdf-3cec-10cc-5868-2f5f86e40622 term=2 tally=1 writer.go:29: 2021-01-29T19:34:10.328Z [INFO] TestAgent_TokenStore.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:10.329Z [INFO] TestAgent_TokenStore.server.raft: entering leader state: leader="Node at 127.0.0.1:30113 [Leader]" writer.go:29: 2021-01-29T19:34:10.330Z [INFO] TestAgent_TokenStore.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:10.331Z [INFO] TestAgent_TokenStore.server: New leader elected: payload=Node-66837bdf-3cec-10cc-5868-2f5f86e40622 writer.go:29: 2021-01-29T19:34:10.333Z [DEBUG] TestAgent_TokenStore.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30113 writer.go:29: 2021-01-29T19:34:10.343Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:10.354Z [INFO] TestAgent_TokenStore: Synced node info writer.go:29: 2021-01-29T19:34:10.355Z [DEBUG] TestAgent_TokenStore: Node info in sync writer.go:29: 2021-01-29T19:34:10.367Z [INFO] TestAgent_TokenStore.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:10.369Z [INFO] TestAgent_TokenStore.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.371Z [DEBUG] TestAgent_TokenStore.server: Skipping self join check for node since the cluster is too small: node=Node-66837bdf-3cec-10cc-5868-2f5f86e40622 writer.go:29: 2021-01-29T19:34:10.383Z [INFO] TestAgent_TokenStore.server: member joined, marking health alive: member=Node-66837bdf-3cec-10cc-5868-2f5f86e40622 === CONT TestAgent_IndexChurn/no_tags writer.go:29: 2021-01-29T19:34:10.464Z [INFO] TestAgent_IndexChurn/no_tags: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:10.465Z [INFO] TestAgent_IndexChurn/no_tags: Endpoints down === RUN TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:10.475Z [WARN] TestAgent_IndexChurn/with_tags: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:10.476Z [DEBUG] TestAgent_IndexChurn/with_tags.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:10.477Z [DEBUG] TestAgent_IndexChurn/with_tags.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:10.497Z [INFO] TestAgent_IndexChurn/with_tags.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cbabdc0e-eccc-af84-b0c8-cacd36315c7f Address:127.0.0.1:30119}]" writer.go:29: 2021-01-29T19:34:10.499Z [INFO] TestAgent_IndexChurn/with_tags.server.raft: entering follower state: follower="Node at 127.0.0.1:30119 [Follower]" leader= writer.go:29: 2021-01-29T19:34:10.504Z [INFO] TestAgent_IndexChurn/with_tags.server.serf.wan: serf: EventMemberJoin: Node-cbabdc0e-eccc-af84-b0c8-cacd36315c7f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.508Z [INFO] TestAgent_IndexChurn/with_tags.server.serf.lan: serf: EventMemberJoin: Node-cbabdc0e-eccc-af84-b0c8-cacd36315c7f 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.511Z [INFO] TestAgent_IndexChurn/with_tags.server: Adding LAN server: server="Node-cbabdc0e-eccc-af84-b0c8-cacd36315c7f (Addr: tcp/127.0.0.1:30119) (DC: dc1)" === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.511Z [INFO] TestAgent_TokenStore: Requesting shutdown === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:10.513Z [INFO] TestAgent_IndexChurn/with_tags.server: Handled event for server in area: event=member-join server=Node-cbabdc0e-eccc-af84-b0c8-cacd36315c7f.dc1 area=wan === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.513Z [INFO] TestAgent_TokenStore.server: shutting down server writer.go:29: 2021-01-29T19:34:10.515Z [DEBUG] TestAgent_TokenStore.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.516Z [WARN] TestAgent_TokenStore.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:10.517Z [INFO] TestAgent_IndexChurn/with_tags: Started DNS server: address=127.0.0.1:30114 network=tcp === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.517Z [DEBUG] TestAgent_TokenStore.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.520Z [WARN] TestAgent_TokenStore.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.523Z [INFO] TestAgent_TokenStore.server.router.manager: shutting down === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:10.521Z [INFO] TestAgent_IndexChurn/with_tags: Started DNS server: address=127.0.0.1:30114 network=udp === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.523Z [INFO] TestAgent_TokenStore: consul server down writer.go:29: 2021-01-29T19:34:10.527Z [INFO] TestAgent_TokenStore: shutdown complete writer.go:29: 2021-01-29T19:34:10.528Z [INFO] TestAgent_TokenStore: Stopping server: protocol=DNS address=127.0.0.1:30108 network=tcp === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:10.528Z [INFO] TestAgent_IndexChurn/with_tags: Started HTTP server: address=127.0.0.1:30115 network=tcp === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.529Z [INFO] TestAgent_TokenStore: Stopping server: protocol=DNS address=127.0.0.1:30108 network=udp === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:10.529Z [INFO] TestAgent_IndexChurn/with_tags: started state syncer === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:10.530Z [INFO] TestAgent_TokenStore: Stopping server: protocol=HTTP address=127.0.0.1:30109 network=tcp === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:10.549Z [WARN] TestAgent_IndexChurn/with_tags.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:10.550Z [INFO] TestAgent_IndexChurn/with_tags.server.raft: entering candidate state: node="Node at 127.0.0.1:30119 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:10.552Z [DEBUG] TestAgent_IndexChurn/with_tags.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:10.557Z [DEBUG] TestAgent_IndexChurn/with_tags.server.raft: vote granted: from=cbabdc0e-eccc-af84-b0c8-cacd36315c7f term=2 tally=1 writer.go:29: 2021-01-29T19:34:10.582Z [INFO] TestAgent_IndexChurn/with_tags.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:10.583Z [INFO] TestAgent_IndexChurn/with_tags.server.raft: entering leader state: leader="Node at 127.0.0.1:30119 [Leader]" writer.go:29: 2021-01-29T19:34:10.584Z [INFO] TestAgent_IndexChurn/with_tags.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:10.585Z [INFO] TestAgent_IndexChurn/with_tags.server: New leader elected: payload=Node-cbabdc0e-eccc-af84-b0c8-cacd36315c7f writer.go:29: 2021-01-29T19:34:10.588Z [DEBUG] TestAgent_IndexChurn/with_tags.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30119 writer.go:29: 2021-01-29T19:34:10.603Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:10.633Z [INFO] TestAgent_IndexChurn/with_tags.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:10.636Z [INFO] TestAgent_IndexChurn/with_tags.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.638Z [DEBUG] TestAgent_IndexChurn/with_tags.server: Skipping self join check for node since the cluster is too small: node=Node-cbabdc0e-eccc-af84-b0c8-cacd36315c7f writer.go:29: 2021-01-29T19:34:10.647Z [INFO] TestAgent_IndexChurn/with_tags.server: member joined, marking health alive: member=Node-cbabdc0e-eccc-af84-b0c8-cacd36315c7f writer.go:29: 2021-01-29T19:34:10.661Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:10.664Z [INFO] TestAgent_IndexChurn/with_tags: Synced node info writer.go:29: 2021-01-29T19:34:10.666Z [INFO] TestAgent_IndexChurn/with_tags: Synced service: service=redis writer.go:29: 2021-01-29T19:34:10.682Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:10.693Z [INFO] TestAgent_IndexChurn/with_tags: Synced check: check=node-check writer.go:29: 2021-01-29T19:34:10.732Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:10.733Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync writer.go:29: 2021-01-29T19:34:10.734Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:10.735Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:10.736Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:10.737Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync writer.go:29: 2021-01-29T19:34:10.737Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:10.738Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:10.739Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check === CONT TestAgent_ReconnectConfigWanDisabled writer.go:29: 2021-01-29T19:34:10.790Z [INFO] TestAgent_ReconnectConfigWanDisabled: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:10.792Z [INFO] TestAgent_ReconnectConfigWanDisabled: Endpoints down --- PASS: TestAgent_ReconnectConfigWanDisabled (0.91s) === CONT TestAgent_RPCPing === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.817Z [INFO] TestAgent_ReconnectConfigSettings: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:10.818Z [INFO] TestAgent_ReconnectConfigSettings: Endpoints down writer.go:29: 2021-01-29T19:34:10.830Z [WARN] TestAgent_ReconnectConfigSettings: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:10.831Z [DEBUG] TestAgent_ReconnectConfigSettings.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:10.833Z [DEBUG] TestAgent_ReconnectConfigSettings.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.834Z [WARN] TestAgent_RPCPing: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:10.835Z [DEBUG] TestAgent_RPCPing.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:10.838Z [DEBUG] TestAgent_RPCPing.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:10.843Z [INFO] TestAgent_RPCPing.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ed8a88e1-98e4-6f55-d076-707316411b7d Address:127.0.0.1:30125}]" writer.go:29: 2021-01-29T19:34:10.846Z [INFO] TestAgent_RPCPing.server.raft: entering follower state: follower="Node at 127.0.0.1:30125 [Follower]" leader= === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.865Z [INFO] TestAgent_ReconnectConfigSettings.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8932e132-c6c5-96dd-89cc-7c4e35462c7d Address:127.0.0.1:30131}]" writer.go:29: 2021-01-29T19:34:10.867Z [INFO] TestAgent_ReconnectConfigSettings.server.raft: entering follower state: follower="Node at 127.0.0.1:30131 [Follower]" leader= writer.go:29: 2021-01-29T19:34:10.868Z [INFO] TestAgent_ReconnectConfigSettings.server.serf.wan: serf: EventMemberJoin: Node-8932e132-c6c5-96dd-89cc-7c4e35462c7d.dc1 127.0.0.1 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.846Z [INFO] TestAgent_RPCPing.server.serf.wan: serf: EventMemberJoin: Node-ed8a88e1-98e4-6f55-d076-707316411b7d.dc1 127.0.0.1 === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.871Z [INFO] TestAgent_ReconnectConfigSettings.server.serf.lan: serf: EventMemberJoin: Node-8932e132-c6c5-96dd-89cc-7c4e35462c7d 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.873Z [INFO] TestAgent_ReconnectConfigSettings.server: Handled event for server in area: event=member-join server=Node-8932e132-c6c5-96dd-89cc-7c4e35462c7d.dc1 area=wan writer.go:29: 2021-01-29T19:34:10.874Z [INFO] TestAgent_ReconnectConfigSettings: Started DNS server: address=127.0.0.1:30126 network=udp writer.go:29: 2021-01-29T19:34:10.875Z [INFO] TestAgent_ReconnectConfigSettings: Started DNS server: address=127.0.0.1:30126 network=tcp === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.875Z [INFO] TestAgent_RPCPing.server.serf.lan: serf: EventMemberJoin: Node-ed8a88e1-98e4-6f55-d076-707316411b7d 127.0.0.1 === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.874Z [INFO] TestAgent_ReconnectConfigSettings.server: Adding LAN server: server="Node-8932e132-c6c5-96dd-89cc-7c4e35462c7d (Addr: tcp/127.0.0.1:30131) (DC: dc1)" === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.878Z [INFO] TestAgent_RPCPing.server: Handled event for server in area: event=member-join server=Node-ed8a88e1-98e4-6f55-d076-707316411b7d.dc1 area=wan === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.878Z [INFO] TestAgent_ReconnectConfigSettings: Started HTTP server: address=127.0.0.1:30127 network=tcp === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.879Z [INFO] TestAgent_RPCPing.server: Adding LAN server: server="Node-ed8a88e1-98e4-6f55-d076-707316411b7d (Addr: tcp/127.0.0.1:30125) (DC: dc1)" === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.879Z [INFO] TestAgent_ReconnectConfigSettings: started state syncer === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.879Z [INFO] TestAgent_RPCPing: Started DNS server: address=127.0.0.1:30120 network=udp writer.go:29: 2021-01-29T19:34:10.883Z [INFO] TestAgent_RPCPing: Started DNS server: address=127.0.0.1:30120 network=tcp writer.go:29: 2021-01-29T19:34:10.885Z [INFO] TestAgent_RPCPing: Started HTTP server: address=127.0.0.1:30121 network=tcp writer.go:29: 2021-01-29T19:34:10.891Z [INFO] TestAgent_RPCPing: started state syncer writer.go:29: 2021-01-29T19:34:10.900Z [WARN] TestAgent_RPCPing.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:10.901Z [INFO] TestAgent_RPCPing.server.raft: entering candidate state: node="Node at 127.0.0.1:30125 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:10.904Z [DEBUG] TestAgent_RPCPing.server.raft: votes: needed=1 === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.906Z [WARN] TestAgent_ReconnectConfigSettings.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.905Z [DEBUG] TestAgent_RPCPing.server.raft: vote granted: from=ed8a88e1-98e4-6f55-d076-707316411b7d term=2 tally=1 === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.907Z [INFO] TestAgent_ReconnectConfigSettings.server.raft: entering candidate state: node="Node at 127.0.0.1:30131 [Candidate]" term=2 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.907Z [INFO] TestAgent_RPCPing.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:10.908Z [INFO] TestAgent_RPCPing.server.raft: entering leader state: leader="Node at 127.0.0.1:30125 [Leader]" writer.go:29: 2021-01-29T19:34:10.910Z [INFO] TestAgent_RPCPing.server: cluster leadership acquired === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.911Z [DEBUG] TestAgent_ReconnectConfigSettings.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:10.912Z [DEBUG] TestAgent_ReconnectConfigSettings.server.raft: vote granted: from=8932e132-c6c5-96dd-89cc-7c4e35462c7d term=2 tally=1 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.910Z [INFO] TestAgent_RPCPing.server: New leader elected: payload=Node-ed8a88e1-98e4-6f55-d076-707316411b7d === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.913Z [INFO] TestAgent_ReconnectConfigSettings.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:10.914Z [INFO] TestAgent_ReconnectConfigSettings.server.raft: entering leader state: leader="Node at 127.0.0.1:30131 [Leader]" writer.go:29: 2021-01-29T19:34:10.915Z [INFO] TestAgent_ReconnectConfigSettings.server: cluster leadership acquired === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.911Z [DEBUG] TestAgent_RPCPing.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30125 === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:10.917Z [INFO] TestAgent_ReconnectConfigSettings.server: New leader elected: payload=Node-8932e132-c6c5-96dd-89cc-7c4e35462c7d writer.go:29: 2021-01-29T19:34:10.925Z [DEBUG] TestAgent_ReconnectConfigSettings.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30131 writer.go:29: 2021-01-29T19:34:10.928Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:10.932Z [INFO] TestAgent_ReconnectConfigSettings.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:10.933Z [INFO] TestAgent_ReconnectConfigSettings.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.934Z [DEBUG] TestAgent_ReconnectConfigSettings.server: Skipping self join check for node since the cluster is too small: node=Node-8932e132-c6c5-96dd-89cc-7c4e35462c7d writer.go:29: 2021-01-29T19:34:10.935Z [INFO] TestAgent_ReconnectConfigSettings.server: member joined, marking health alive: member=Node-8932e132-c6c5-96dd-89cc-7c4e35462c7d === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:10.935Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:10.957Z [INFO] TestAgent_RPCPing.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:10.961Z [INFO] TestAgent_RPCPing.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.965Z [DEBUG] TestAgent_RPCPing.server: Skipping self join check for node since the cluster is too small: node=Node-ed8a88e1-98e4-6f55-d076-707316411b7d writer.go:29: 2021-01-29T19:34:10.968Z [INFO] TestAgent_RPCPing.server: member joined, marking health alive: member=Node-ed8a88e1-98e4-6f55-d076-707316411b7d === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:11.018Z [DEBUG] TestAgent_ReconnectConfigSettings: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:11.021Z [INFO] TestAgent_ReconnectConfigSettings: Synced node info writer.go:29: 2021-01-29T19:34:11.024Z [DEBUG] TestAgent_ReconnectConfigSettings: Node info in sync writer.go:29: 2021-01-29T19:34:11.024Z [INFO] TestAgent_ReconnectConfigSettings: Requesting shutdown writer.go:29: 2021-01-29T19:34:11.026Z [INFO] TestAgent_ReconnectConfigSettings.server: shutting down server writer.go:29: 2021-01-29T19:34:11.027Z [DEBUG] TestAgent_ReconnectConfigSettings.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_TokenStore writer.go:29: 2021-01-29T19:34:11.031Z [INFO] TestAgent_TokenStore: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:11.032Z [INFO] TestAgent_TokenStore: Endpoints down --- PASS: TestAgent_TokenStore (0.82s) === CONT TestAgent_StartStop === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:11.028Z [WARN] TestAgent_ReconnectConfigSettings.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:11.028Z [DEBUG] TestAgent_ReconnectConfigSettings.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:11.045Z [WARN] TestAgent_StartStop: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:11.046Z [DEBUG] TestAgent_StartStop.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:11.047Z [DEBUG] TestAgent_StartStop.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:11.043Z [WARN] TestAgent_ReconnectConfigSettings.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.077Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:11.079Z [INFO] TestAgent_ReconnectConfigSettings.server.router.manager: shutting down === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.079Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:11.081Z [INFO] TestAgent_ReconnectConfigSettings: consul server down writer.go:29: 2021-01-29T19:34:11.081Z [INFO] TestAgent_ReconnectConfigSettings: shutdown complete writer.go:29: 2021-01-29T19:34:11.082Z [INFO] TestAgent_ReconnectConfigSettings: Stopping server: protocol=DNS address=127.0.0.1:30126 network=tcp === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.081Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:11.083Z [INFO] TestAgent_ReconnectConfigSettings: Stopping server: protocol=DNS address=127.0.0.1:30126 network=udp writer.go:29: 2021-01-29T19:34:11.084Z [INFO] TestAgent_ReconnectConfigSettings: Stopping server: protocol=HTTP address=127.0.0.1:30127 network=tcp === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.083Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:11.086Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:11.102Z [INFO] TestAgent_StartStop.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:33dc3d00-b7b7-13ad-16dd-3c5ffec4e086 Address:127.0.0.1:30137}]" writer.go:29: 2021-01-29T19:34:11.105Z [INFO] TestAgent_StartStop.server.serf.wan: serf: EventMemberJoin: Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:11.109Z [INFO] TestAgent_StartStop.server.serf.lan: serf: EventMemberJoin: Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086 127.0.0.1 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:11.112Z [DEBUG] TestAgent_RPCPing: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:11.111Z [INFO] TestAgent_StartStop.server.raft: entering follower state: follower="Node at 127.0.0.1:30137 [Follower]" leader= writer.go:29: 2021-01-29T19:34:11.119Z [INFO] TestAgent_StartStop.server: Adding LAN server: server="Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086 (Addr: tcp/127.0.0.1:30137) (DC: dc1)" === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:11.122Z [INFO] TestAgent_RPCPing: Synced node info === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:11.123Z [INFO] TestAgent_StartStop.server: Handled event for server in area: event=member-join server=Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086.dc1 area=wan writer.go:29: 2021-01-29T19:34:11.126Z [INFO] TestAgent_StartStop: Started DNS server: address=127.0.0.1:30132 network=tcp writer.go:29: 2021-01-29T19:34:11.130Z [INFO] TestAgent_StartStop: Started DNS server: address=127.0.0.1:30132 network=udp writer.go:29: 2021-01-29T19:34:11.140Z [INFO] TestAgent_StartStop: Started HTTP server: address=127.0.0.1:30133 network=tcp writer.go:29: 2021-01-29T19:34:11.142Z [INFO] TestAgent_StartStop: started state syncer writer.go:29: 2021-01-29T19:34:11.172Z [WARN] TestAgent_StartStop.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:11.178Z [INFO] TestAgent_StartStop.server.raft: entering candidate state: node="Node at 127.0.0.1:30137 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:11.184Z [DEBUG] TestAgent_StartStop.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:11.186Z [DEBUG] TestAgent_StartStop.server.raft: vote granted: from=33dc3d00-b7b7-13ad-16dd-3c5ffec4e086 term=2 tally=1 writer.go:29: 2021-01-29T19:34:11.187Z [INFO] TestAgent_StartStop.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:11.188Z [INFO] TestAgent_StartStop.server.raft: entering leader state: leader="Node at 127.0.0.1:30137 [Leader]" writer.go:29: 2021-01-29T19:34:11.190Z [INFO] TestAgent_StartStop.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:11.191Z [INFO] TestAgent_StartStop.server: New leader elected: payload=Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086 writer.go:29: 2021-01-29T19:34:11.203Z [DEBUG] TestAgent_StartStop.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30137 writer.go:29: 2021-01-29T19:34:11.217Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.201Z [DEBUG] TestAgent_IndexChurn/with_tags: Registered node: node="Node-level check" writer.go:29: 2021-01-29T19:34:11.223Z [DEBUG] TestAgent_IndexChurn/with_tags: Registered node: node="Serf Health Status" writer.go:29: 2021-01-29T19:34:11.224Z [DEBUG] TestAgent_IndexChurn/with_tags: Registered node: node="Service-level check" writer.go:29: 2021-01-29T19:34:11.225Z [INFO] TestAgent_IndexChurn/with_tags: Sync in progress: iteration=1 writer.go:29: 2021-01-29T19:34:11.226Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:11.227Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync writer.go:29: 2021-01-29T19:34:11.228Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:11.229Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:11.229Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:11.230Z [INFO] TestAgent_IndexChurn/with_tags: Sync in progress: iteration=2 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:11.236Z [INFO] TestAgent_RPCPing: Requesting shutdown writer.go:29: 2021-01-29T19:34:11.252Z [INFO] TestAgent_RPCPing.server: shutting down server writer.go:29: 2021-01-29T19:34:11.253Z [DEBUG] TestAgent_RPCPing.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:11.255Z [WARN] TestAgent_RPCPing.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:11.257Z [DEBUG] TestAgent_RPCPing.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.261Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:11.263Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync writer.go:29: 2021-01-29T19:34:11.265Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:11.267Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:11.268Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:11.270Z [INFO] TestAgent_IndexChurn/with_tags: Sync in progress: iteration=3 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:11.261Z [WARN] TestAgent_RPCPing.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.272Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:11.275Z [INFO] TestAgent_StartStop: Synced node info === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.274Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:11.276Z [DEBUG] TestAgent_StartStop: Node info in sync writer.go:29: 2021-01-29T19:34:11.277Z [INFO] TestAgent_StartStop.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:11.274Z [INFO] TestAgent_RPCPing.server.router.manager: shutting down === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.276Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:11.277Z [INFO] TestAgent_StartStop.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:11.278Z [DEBUG] TestAgent_StartStop.server: Skipping self join check for node since the cluster is too small: node=Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:11.278Z [INFO] TestAgent_RPCPing: consul server down === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.278Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:11.279Z [INFO] TestAgent_StartStop.server: member joined, marking health alive: member=Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:11.280Z [INFO] TestAgent_RPCPing: shutdown complete === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.280Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:11.282Z [INFO] TestAgent_RPCPing: Stopping server: protocol=DNS address=127.0.0.1:30120 network=tcp writer.go:29: 2021-01-29T19:34:11.283Z [INFO] TestAgent_RPCPing: Stopping server: protocol=DNS address=127.0.0.1:30120 network=udp === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.282Z [INFO] TestAgent_IndexChurn/with_tags: Sync in progress: iteration=4 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:11.284Z [INFO] TestAgent_RPCPing: Stopping server: protocol=HTTP address=127.0.0.1:30121 network=tcp === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.284Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:11.286Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync writer.go:29: 2021-01-29T19:34:11.288Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:11.290Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:11.292Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:11.294Z [INFO] TestAgent_IndexChurn/with_tags: Sync in progress: iteration=5 writer.go:29: 2021-01-29T19:34:11.296Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:11.298Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync writer.go:29: 2021-01-29T19:34:11.300Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:11.301Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:11.303Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:11.306Z [INFO] TestAgent_IndexChurn/with_tags: Sync in progress: iteration=6 writer.go:29: 2021-01-29T19:34:11.309Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:11.311Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:11.312Z [INFO] TestAgent_StartStop.server: server starting leave writer.go:29: 2021-01-29T19:34:11.313Z [INFO] TestAgent_StartStop.server.serf.wan: serf: EventMemberLeave: Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086.dc1 127.0.0.1 === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.313Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:11.315Z [INFO] TestAgent_StartStop.server: Handled event for server in area: event=member-leave server=Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086.dc1 area=wan writer.go:29: 2021-01-29T19:34:11.315Z [INFO] TestAgent_StartStop.server.router.manager: shutting down === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.315Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:11.317Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:11.318Z [INFO] TestAgent_IndexChurn/with_tags: Sync in progress: iteration=7 writer.go:29: 2021-01-29T19:34:11.319Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:11.320Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync writer.go:29: 2021-01-29T19:34:11.320Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:11.322Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:11.322Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:11.323Z [INFO] TestAgent_IndexChurn/with_tags: Sync in progress: iteration=8 writer.go:29: 2021-01-29T19:34:11.324Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:11.325Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync writer.go:29: 2021-01-29T19:34:11.333Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:11.334Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:11.335Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:11.337Z [INFO] TestAgent_IndexChurn/with_tags: Sync in progress: iteration=9 writer.go:29: 2021-01-29T19:34:11.339Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:11.341Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync writer.go:29: 2021-01-29T19:34:11.344Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:11.346Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:11.348Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:11.350Z [INFO] TestAgent_IndexChurn/with_tags: Sync in progress: iteration=10 writer.go:29: 2021-01-29T19:34:11.352Z [DEBUG] TestAgent_IndexChurn/with_tags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:11.354Z [DEBUG] TestAgent_IndexChurn/with_tags: Node info in sync writer.go:29: 2021-01-29T19:34:11.356Z [DEBUG] TestAgent_IndexChurn/with_tags: Service in sync: service=redis writer.go:29: 2021-01-29T19:34:11.359Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=redis-check writer.go:29: 2021-01-29T19:34:11.361Z [DEBUG] TestAgent_IndexChurn/with_tags: Check in sync: check=node-check writer.go:29: 2021-01-29T19:34:11.363Z [INFO] TestAgent_IndexChurn/with_tags: Requesting shutdown writer.go:29: 2021-01-29T19:34:11.366Z [INFO] TestAgent_IndexChurn/with_tags.server: shutting down server writer.go:29: 2021-01-29T19:34:11.367Z [DEBUG] TestAgent_IndexChurn/with_tags.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:11.369Z [WARN] TestAgent_IndexChurn/with_tags.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:11.369Z [DEBUG] TestAgent_IndexChurn/with_tags.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:11.372Z [WARN] TestAgent_IndexChurn/with_tags.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:11.375Z [INFO] TestAgent_IndexChurn/with_tags.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:11.376Z [INFO] TestAgent_IndexChurn/with_tags: consul server down writer.go:29: 2021-01-29T19:34:11.377Z [INFO] TestAgent_IndexChurn/with_tags: shutdown complete writer.go:29: 2021-01-29T19:34:11.378Z [INFO] TestAgent_IndexChurn/with_tags: Stopping server: protocol=DNS address=127.0.0.1:30114 network=tcp writer.go:29: 2021-01-29T19:34:11.379Z [INFO] TestAgent_IndexChurn/with_tags: Stopping server: protocol=DNS address=127.0.0.1:30114 network=udp writer.go:29: 2021-01-29T19:34:11.391Z [INFO] TestAgent_IndexChurn/with_tags: Stopping server: protocol=HTTP address=127.0.0.1:30115 network=tcp === CONT TestAgent_ReconnectConfigSettings writer.go:29: 2021-01-29T19:34:11.585Z [INFO] TestAgent_ReconnectConfigSettings: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:11.586Z [INFO] TestAgent_ReconnectConfigSettings: Endpoints down --- PASS: TestAgent_ReconnectConfigSettings (1.52s) === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:11.612Z [WARN] TestAgent_Services_ExposeConfig: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:11.613Z [DEBUG] TestAgent_Services_ExposeConfig.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:11.616Z [DEBUG] TestAgent_Services_ExposeConfig.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:11.623Z [INFO] TestAgent_Services_ExposeConfig.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:195d5cc3-116a-f147-9b6f-abf5833bbd95 Address:127.0.0.1:30143}]" writer.go:29: 2021-01-29T19:34:11.626Z [INFO] TestAgent_Services_ExposeConfig.server.raft: entering follower state: follower="Node at 127.0.0.1:30143 [Follower]" leader= writer.go:29: 2021-01-29T19:34:11.630Z [INFO] TestAgent_Services_ExposeConfig.server.serf.wan: serf: EventMemberJoin: Node-195d5cc3-116a-f147-9b6f-abf5833bbd95.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:11.633Z [INFO] TestAgent_Services_ExposeConfig.server.serf.lan: serf: EventMemberJoin: Node-195d5cc3-116a-f147-9b6f-abf5833bbd95 127.0.0.1 writer.go:29: 2021-01-29T19:34:11.637Z [INFO] TestAgent_Services_ExposeConfig.server: Handled event for server in area: event=member-join server=Node-195d5cc3-116a-f147-9b6f-abf5833bbd95.dc1 area=wan writer.go:29: 2021-01-29T19:34:11.637Z [INFO] TestAgent_Services_ExposeConfig.server: Adding LAN server: server="Node-195d5cc3-116a-f147-9b6f-abf5833bbd95 (Addr: tcp/127.0.0.1:30143) (DC: dc1)" writer.go:29: 2021-01-29T19:34:11.644Z [INFO] TestAgent_Services_ExposeConfig: Started DNS server: address=127.0.0.1:30138 network=udp writer.go:29: 2021-01-29T19:34:11.646Z [INFO] TestAgent_Services_ExposeConfig: Started DNS server: address=127.0.0.1:30138 network=tcp writer.go:29: 2021-01-29T19:34:11.649Z [INFO] TestAgent_Services_ExposeConfig: Started HTTP server: address=127.0.0.1:30139 network=tcp writer.go:29: 2021-01-29T19:34:11.650Z [INFO] TestAgent_Services_ExposeConfig: started state syncer writer.go:29: 2021-01-29T19:34:11.686Z [WARN] TestAgent_Services_ExposeConfig.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:11.688Z [INFO] TestAgent_Services_ExposeConfig.server.raft: entering candidate state: node="Node at 127.0.0.1:30143 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:11.690Z [DEBUG] TestAgent_Services_ExposeConfig.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:11.692Z [DEBUG] TestAgent_Services_ExposeConfig.server.raft: vote granted: from=195d5cc3-116a-f147-9b6f-abf5833bbd95 term=2 tally=1 writer.go:29: 2021-01-29T19:34:11.693Z [INFO] TestAgent_Services_ExposeConfig.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:11.695Z [INFO] TestAgent_Services_ExposeConfig.server.raft: entering leader state: leader="Node at 127.0.0.1:30143 [Leader]" writer.go:29: 2021-01-29T19:34:11.696Z [INFO] TestAgent_Services_ExposeConfig.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:11.699Z [INFO] TestAgent_Services_ExposeConfig.server: New leader elected: payload=Node-195d5cc3-116a-f147-9b6f-abf5833bbd95 writer.go:29: 2021-01-29T19:34:11.702Z [DEBUG] TestAgent_Services_ExposeConfig.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30143 writer.go:29: 2021-01-29T19:34:11.712Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:11.719Z [INFO] TestAgent_Services_ExposeConfig.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:11.721Z [INFO] TestAgent_Services_ExposeConfig.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:11.723Z [DEBUG] TestAgent_Services_ExposeConfig.server: Skipping self join check for node since the cluster is too small: node=Node-195d5cc3-116a-f147-9b6f-abf5833bbd95 writer.go:29: 2021-01-29T19:34:11.726Z [INFO] TestAgent_Services_ExposeConfig.server: member joined, marking health alive: member=Node-195d5cc3-116a-f147-9b6f-abf5833bbd95 === CONT TestAgent_RPCPing writer.go:29: 2021-01-29T19:34:11.784Z [INFO] TestAgent_RPCPing: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:11.786Z [INFO] TestAgent_RPCPing: Endpoints down --- PASS: TestAgent_RPCPing (0.99s) === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:11.816Z [WARN] TestAgent_HostBadACL: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:11.820Z [WARN] TestAgent_HostBadACL: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:11.823Z [DEBUG] TestAgent_HostBadACL.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:11.825Z [DEBUG] TestAgent_HostBadACL.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:11.846Z [INFO] TestAgent_HostBadACL.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a437185a-506b-084c-38a4-65b9ae13c79d Address:127.0.0.1:30149}]" writer.go:29: 2021-01-29T19:34:11.849Z [INFO] TestAgent_HostBadACL.server.raft: entering follower state: follower="Node at 127.0.0.1:30149 [Follower]" leader= writer.go:29: 2021-01-29T19:34:11.854Z [INFO] TestAgent_HostBadACL.server.serf.wan: serf: EventMemberJoin: Node-a437185a-506b-084c-38a4-65b9ae13c79d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:11.868Z [INFO] TestAgent_HostBadACL.server.serf.lan: serf: EventMemberJoin: Node-a437185a-506b-084c-38a4-65b9ae13c79d 127.0.0.1 writer.go:29: 2021-01-29T19:34:11.876Z [INFO] TestAgent_HostBadACL.server: Handled event for server in area: event=member-join server=Node-a437185a-506b-084c-38a4-65b9ae13c79d.dc1 area=wan writer.go:29: 2021-01-29T19:34:11.877Z [INFO] TestAgent_HostBadACL.server: Adding LAN server: server="Node-a437185a-506b-084c-38a4-65b9ae13c79d (Addr: tcp/127.0.0.1:30149) (DC: dc1)" writer.go:29: 2021-01-29T19:34:11.881Z [INFO] TestAgent_HostBadACL: Started DNS server: address=127.0.0.1:30144 network=udp writer.go:29: 2021-01-29T19:34:11.883Z [INFO] TestAgent_HostBadACL: Started DNS server: address=127.0.0.1:30144 network=tcp writer.go:29: 2021-01-29T19:34:11.886Z [INFO] TestAgent_HostBadACL: Started HTTP server: address=127.0.0.1:30145 network=tcp writer.go:29: 2021-01-29T19:34:11.887Z [INFO] TestAgent_HostBadACL: started state syncer writer.go:29: 2021-01-29T19:34:11.888Z [WARN] TestAgent_HostBadACL.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:11.890Z [INFO] TestAgent_HostBadACL.server.raft: entering candidate state: node="Node at 127.0.0.1:30149 [Candidate]" term=2 === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.893Z [INFO] TestAgent_IndexChurn/with_tags: Waiting for endpoints to shut down === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:11.892Z [DEBUG] TestAgent_HostBadACL.server.raft: votes: needed=1 === CONT TestAgent_IndexChurn/with_tags writer.go:29: 2021-01-29T19:34:11.894Z [INFO] TestAgent_IndexChurn/with_tags: Endpoints down --- PASS: TestAgent_IndexChurn (2.77s) --- PASS: TestAgent_IndexChurn/no_tags (1.34s) --- PASS: TestAgent_IndexChurn/with_tags (1.43s) === CONT TestAgent_Host === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:11.895Z [DEBUG] TestAgent_HostBadACL.server.raft: vote granted: from=a437185a-506b-084c-38a4-65b9ae13c79d term=2 tally=1 writer.go:29: 2021-01-29T19:34:11.897Z [INFO] TestAgent_HostBadACL.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:11.899Z [INFO] TestAgent_HostBadACL.server.raft: entering leader state: leader="Node at 127.0.0.1:30149 [Leader]" writer.go:29: 2021-01-29T19:34:11.903Z [INFO] TestAgent_HostBadACL.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:11.905Z [INFO] TestAgent_HostBadACL.server: New leader elected: payload=Node-a437185a-506b-084c-38a4-65b9ae13c79d writer.go:29: 2021-01-29T19:34:11.908Z [INFO] TestAgent_HostBadACL.server: initializing acls writer.go:29: 2021-01-29T19:34:11.913Z [INFO] TestAgent_HostBadACL.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:11.916Z [WARN] TestAgent_HostBadACL.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:11.917Z [WARN] TestAgent_Host: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:11.919Z [WARN] TestAgent_Host: bootstrap = true: do not enable unless necessary === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:11.919Z [INFO] TestAgent_HostBadACL.server: Bootstrapped ACL master token from configuration === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:11.921Z [DEBUG] TestAgent_Host.tlsutil: Update: version=1 === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:11.923Z [INFO] TestAgent_HostBadACL.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:11.925Z [INFO] TestAgent_HostBadACL.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:11.926Z [INFO] TestAgent_HostBadACL.leader: started routine: routine="acl token reaping" === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:11.926Z [DEBUG] TestAgent_Host.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:11.928Z [INFO] TestAgent_HostBadACL.server.serf.lan: serf: EventMemberUpdate: Node-a437185a-506b-084c-38a4-65b9ae13c79d writer.go:29: 2021-01-29T19:34:11.932Z [INFO] TestAgent_HostBadACL.server: Updating LAN server: server="Node-a437185a-506b-084c-38a4-65b9ae13c79d (Addr: tcp/127.0.0.1:30149) (DC: dc1)" writer.go:29: 2021-01-29T19:34:11.932Z [INFO] TestAgent_HostBadACL.server.serf.wan: serf: EventMemberUpdate: Node-a437185a-506b-084c-38a4-65b9ae13c79d.dc1 writer.go:29: 2021-01-29T19:34:11.934Z [INFO] TestAgent_HostBadACL.server: Handled event for server in area: event=member-update server=Node-a437185a-506b-084c-38a4-65b9ae13c79d.dc1 area=wan writer.go:29: 2021-01-29T19:34:11.937Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:11.938Z [INFO] TestAgent_Host.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:adcef449-434e-6035-7f86-0b42049d0250 Address:127.0.0.1:30167}]" writer.go:29: 2021-01-29T19:34:11.941Z [INFO] TestAgent_Host.server.raft: entering follower state: follower="Node at 127.0.0.1:30167 [Follower]" leader= writer.go:29: 2021-01-29T19:34:11.944Z [INFO] TestAgent_Host.server.serf.wan: serf: EventMemberJoin: Node-adcef449-434e-6035-7f86-0b42049d0250.dc1 127.0.0.1 === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:11.953Z [INFO] TestAgent_HostBadACL.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:11.952Z [INFO] TestAgent_Host.server.serf.lan: serf: EventMemberJoin: Node-adcef449-434e-6035-7f86-0b42049d0250 127.0.0.1 === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:11.955Z [INFO] TestAgent_HostBadACL.leader: started routine: routine="CA root pruning" === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:11.956Z [INFO] TestAgent_Host.server: Handled event for server in area: event=member-join server=Node-adcef449-434e-6035-7f86-0b42049d0250.dc1 area=wan === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:11.956Z [DEBUG] TestAgent_HostBadACL.server: Skipping self join check for node since the cluster is too small: node=Node-a437185a-506b-084c-38a4-65b9ae13c79d === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:11.957Z [INFO] TestAgent_Host.server: Adding LAN server: server="Node-adcef449-434e-6035-7f86-0b42049d0250 (Addr: tcp/127.0.0.1:30167) (DC: dc1)" === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:11.958Z [INFO] TestAgent_HostBadACL.server: member joined, marking health alive: member=Node-a437185a-506b-084c-38a4-65b9ae13c79d writer.go:29: 2021-01-29T19:34:11.960Z [DEBUG] TestAgent_HostBadACL.server: Skipping self join check for node since the cluster is too small: node=Node-a437185a-506b-084c-38a4-65b9ae13c79d === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:11.961Z [INFO] TestAgent_Host: Started DNS server: address=127.0.0.1:30162 network=tcp writer.go:29: 2021-01-29T19:34:11.963Z [INFO] TestAgent_Host: Started DNS server: address=127.0.0.1:30162 network=udp writer.go:29: 2021-01-29T19:34:11.965Z [INFO] TestAgent_Host: Started HTTP server: address=127.0.0.1:30163 network=tcp writer.go:29: 2021-01-29T19:34:11.968Z [INFO] TestAgent_Host: started state syncer writer.go:29: 2021-01-29T19:34:12.009Z [DEBUG] TestAgent_Host.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.008Z [ERROR] TestAgent_Services_ExposeConfig.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.009Z [WARN] TestAgent_Host.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:12.010Z [INFO] TestAgent_Host.server.raft: entering candidate state: node="Node at 127.0.0.1:30167 [Candidate]" term=2 === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.010Z [ERROR] TestAgent_Services_ExposeConfig.proxycfg: watch error: id=leaf error="error filling agent cache: URI must be either service or agent" writer.go:29: 2021-01-29T19:34:12.012Z [ERROR] TestAgent_Services_ExposeConfig.proxycfg: watch error: id=leaf error="error filling agent cache: URI must be either service or agent" === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.015Z [DEBUG] TestAgent_Host.server.raft: votes: needed=1 === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.016Z [ERROR] TestAgent_Services_ExposeConfig.proxycfg: watch error: id=leaf error="error filling agent cache: URI must be either service or agent" writer.go:29: 2021-01-29T19:34:12.017Z [ERROR] TestAgent_Services_ExposeConfig.proxycfg: watch error: id=leaf error="error filling agent cache: URI must be either service or agent" === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.016Z [DEBUG] TestAgent_Host.server.raft: vote granted: from=adcef449-434e-6035-7f86-0b42049d0250 term=2 tally=1 === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.013Z [INFO] TestAgent_Services_ExposeConfig: Requesting shutdown writer.go:29: 2021-01-29T19:34:12.021Z [INFO] TestAgent_Services_ExposeConfig.server: shutting down server writer.go:29: 2021-01-29T19:34:12.022Z [DEBUG] TestAgent_Services_ExposeConfig.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:12.023Z [ERROR] TestAgent_Services_ExposeConfig.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.019Z [INFO] TestAgent_Host.server.raft: election won: tally=1 === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.024Z [WARN] TestAgent_Services_ExposeConfig.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:12.024Z [DEBUG] TestAgent_Services_ExposeConfig.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.026Z [INFO] TestAgent_Host.server.raft: entering leader state: leader="Node at 127.0.0.1:30167 [Leader]" writer.go:29: 2021-01-29T19:34:12.030Z [INFO] TestAgent_Host.server: cluster leadership acquired === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.029Z [WARN] TestAgent_Services_ExposeConfig.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.031Z [INFO] TestAgent_Host.server: New leader elected: payload=Node-adcef449-434e-6035-7f86-0b42049d0250 === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.032Z [INFO] TestAgent_Services_ExposeConfig.server.router.manager: shutting down === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.043Z [INFO] TestAgent_Host.server: initializing acls === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.044Z [INFO] TestAgent_Services_ExposeConfig: consul server down === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.046Z [INFO] TestAgent_Host.server: Created ACL 'global-management' policy === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.046Z [INFO] TestAgent_Services_ExposeConfig: shutdown complete === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.046Z [WARN] TestAgent_Host.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:12.048Z [INFO] TestAgent_Host.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:12.050Z [INFO] TestAgent_Host.server: Created ACL anonymous token from configuration === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.047Z [INFO] TestAgent_Services_ExposeConfig: Stopping server: protocol=DNS address=127.0.0.1:30138 network=tcp === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.051Z [INFO] TestAgent_Host.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.053Z [INFO] TestAgent_Services_ExposeConfig: Stopping server: protocol=DNS address=127.0.0.1:30138 network=udp === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.054Z [INFO] TestAgent_Host.leader: started routine: routine="acl token reaping" === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.056Z [INFO] TestAgent_Services_ExposeConfig: Stopping server: protocol=HTTP address=127.0.0.1:30139 network=tcp === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.058Z [INFO] TestAgent_Host.server.serf.lan: serf: EventMemberUpdate: Node-adcef449-434e-6035-7f86-0b42049d0250 writer.go:29: 2021-01-29T19:34:12.064Z [INFO] TestAgent_Host.server: Updating LAN server: server="Node-adcef449-434e-6035-7f86-0b42049d0250 (Addr: tcp/127.0.0.1:30167) (DC: dc1)" === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:12.063Z [DEBUG] TestAgent_StartStop: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.064Z [INFO] TestAgent_Host.server.serf.wan: serf: EventMemberUpdate: Node-adcef449-434e-6035-7f86-0b42049d0250.dc1 === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:12.065Z [ERROR] TestAgent_HostBadACL.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.068Z [INFO] TestAgent_Host.server: Handled event for server in area: event=member-update server=Node-adcef449-434e-6035-7f86-0b42049d0250.dc1 area=wan === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:12.066Z [DEBUG] TestAgent_StartStop: Node info in sync === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.070Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:12.085Z [INFO] TestAgent_Host.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:12.087Z [INFO] TestAgent_Host.leader: started routine: routine="CA root pruning" === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:12.102Z [DEBUG] TestAgent_HostBadACL.acl: dropping node from result due to ACLs: node=Node-a437185a-506b-084c-38a4-65b9ae13c79d writer.go:29: 2021-01-29T19:34:12.105Z [DEBUG] TestAgent_HostBadACL.acl: dropping node from result due to ACLs: node=Node-a437185a-506b-084c-38a4-65b9ae13c79d writer.go:29: 2021-01-29T19:34:12.108Z [INFO] TestAgent_HostBadACL: Requesting shutdown writer.go:29: 2021-01-29T19:34:12.109Z [INFO] TestAgent_HostBadACL.server: shutting down server writer.go:29: 2021-01-29T19:34:12.110Z [DEBUG] TestAgent_HostBadACL.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:12.111Z [DEBUG] TestAgent_HostBadACL.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:12.111Z [DEBUG] TestAgent_HostBadACL.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:12.111Z [DEBUG] TestAgent_HostBadACL.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:12.111Z [DEBUG] TestAgent_HostBadACL.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:12.113Z [DEBUG] TestAgent_HostBadACL.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:12.113Z [WARN] TestAgent_HostBadACL.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.091Z [DEBUG] TestAgent_Host.server: Skipping self join check for node since the cluster is too small: node=Node-adcef449-434e-6035-7f86-0b42049d0250 === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:12.119Z [WARN] TestAgent_HostBadACL.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.117Z [INFO] TestAgent_Host.server: member joined, marking health alive: member=Node-adcef449-434e-6035-7f86-0b42049d0250 === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:12.120Z [INFO] TestAgent_HostBadACL.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:12.121Z [INFO] TestAgent_HostBadACL: consul server down writer.go:29: 2021-01-29T19:34:12.122Z [INFO] TestAgent_HostBadACL: shutdown complete writer.go:29: 2021-01-29T19:34:12.123Z [INFO] TestAgent_HostBadACL: Stopping server: protocol=DNS address=127.0.0.1:30144 network=tcp writer.go:29: 2021-01-29T19:34:12.124Z [INFO] TestAgent_HostBadACL: Stopping server: protocol=DNS address=127.0.0.1:30144 network=udp === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.122Z [DEBUG] TestAgent_Host.server: Skipping self join check for node since the cluster is too small: node=Node-adcef449-434e-6035-7f86-0b42049d0250 === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:12.125Z [INFO] TestAgent_HostBadACL: Stopping server: protocol=HTTP address=127.0.0.1:30145 network=tcp === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.198Z [INFO] TestAgent_Host: Requesting shutdown writer.go:29: 2021-01-29T19:34:12.199Z [INFO] TestAgent_Host.server: shutting down server writer.go:29: 2021-01-29T19:34:12.201Z [DEBUG] TestAgent_Host.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:12.202Z [DEBUG] TestAgent_Host.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:12.201Z [ERROR] TestAgent_Host.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:12.202Z [DEBUG] TestAgent_Host.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:12.204Z [DEBUG] TestAgent_Host.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:12.204Z [DEBUG] TestAgent_Host.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:12.208Z [WARN] TestAgent_Host.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:12.208Z [DEBUG] TestAgent_Host.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:12.222Z [WARN] TestAgent_Host.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:12.225Z [INFO] TestAgent_Host.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:12.225Z [INFO] TestAgent_Host: consul server down writer.go:29: 2021-01-29T19:34:12.226Z [INFO] TestAgent_Host: shutdown complete writer.go:29: 2021-01-29T19:34:12.227Z [INFO] TestAgent_Host: Stopping server: protocol=DNS address=127.0.0.1:30162 network=tcp writer.go:29: 2021-01-29T19:34:12.228Z [INFO] TestAgent_Host: Stopping server: protocol=DNS address=127.0.0.1:30162 network=udp writer.go:29: 2021-01-29T19:34:12.230Z [INFO] TestAgent_Host: Stopping server: protocol=HTTP address=127.0.0.1:30163 network=tcp === CONT TestAgent_Services_ExposeConfig writer.go:29: 2021-01-29T19:34:12.558Z [INFO] TestAgent_Services_ExposeConfig: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:12.561Z [INFO] TestAgent_Services_ExposeConfig: Endpoints down --- PASS: TestAgent_Services_ExposeConfig (0.97s) === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.570Z [WARN] TestAgentConnectAuthorize_defaultAllow: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:12.571Z [WARN] TestAgentConnectAuthorize_defaultAllow: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:12.581Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:12.586Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:12.617Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:746c73f1-7446-8845-98b0-57769cad6e6e Address:127.0.0.1:30155}]" writer.go:29: 2021-01-29T19:34:12.619Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.raft: entering follower state: follower="Node at 127.0.0.1:30155 [Follower]" leader= writer.go:29: 2021-01-29T19:34:12.620Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.serf.wan: serf: EventMemberJoin: Node-746c73f1-7446-8845-98b0-57769cad6e6e.dc1 127.0.0.1 === CONT TestAgent_HostBadACL writer.go:29: 2021-01-29T19:34:12.625Z [INFO] TestAgent_HostBadACL: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:12.627Z [INFO] TestAgent_HostBadACL: Endpoints down --- PASS: TestAgent_HostBadACL (0.84s) === CONT TestAgentConnectAuthorize_defaultDeny === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.628Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.serf.lan: serf: EventMemberJoin: Node-746c73f1-7446-8845-98b0-57769cad6e6e 127.0.0.1 writer.go:29: 2021-01-29T19:34:12.634Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: Adding LAN server: server="Node-746c73f1-7446-8845-98b0-57769cad6e6e (Addr: tcp/127.0.0.1:30155) (DC: dc1)" writer.go:29: 2021-01-29T19:34:12.634Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: Handled event for server in area: event=member-join server=Node-746c73f1-7446-8845-98b0-57769cad6e6e.dc1 area=wan writer.go:29: 2021-01-29T19:34:12.637Z [INFO] TestAgentConnectAuthorize_defaultAllow: Started DNS server: address=127.0.0.1:30150 network=tcp writer.go:29: 2021-01-29T19:34:12.639Z [INFO] TestAgentConnectAuthorize_defaultAllow: Started DNS server: address=127.0.0.1:30150 network=udp writer.go:29: 2021-01-29T19:34:12.642Z [INFO] TestAgentConnectAuthorize_defaultAllow: Started HTTP server: address=127.0.0.1:30151 network=tcp writer.go:29: 2021-01-29T19:34:12.644Z [INFO] TestAgentConnectAuthorize_defaultAllow: started state syncer === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.649Z [WARN] TestAgentConnectAuthorize_defaultDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:12.650Z [WARN] TestAgentConnectAuthorize_defaultDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:12.653Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:12.657Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:12.663Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:300e0834-43a3-6e0f-c550-a2b9d86792c0 Address:127.0.0.1:30161}]" writer.go:29: 2021-01-29T19:34:12.665Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.serf.wan: serf: EventMemberJoin: Node-300e0834-43a3-6e0f-c550-a2b9d86792c0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:12.665Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30161 [Follower]" leader= === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.673Z [WARN] TestAgentConnectAuthorize_defaultAllow.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:12.674Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.raft: entering candidate state: node="Node at 127.0.0.1:30155 [Candidate]" term=2 === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.670Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.serf.lan: serf: EventMemberJoin: Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 127.0.0.1 === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.677Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.server.raft: votes: needed=1 === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.678Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Adding LAN server: server="Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 (Addr: tcp/127.0.0.1:30161) (DC: dc1)" === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.678Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.server.raft: vote granted: from=746c73f1-7446-8845-98b0-57769cad6e6e term=2 tally=1 === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.679Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Handled event for server in area: event=member-join server=Node-300e0834-43a3-6e0f-c550-a2b9d86792c0.dc1 area=wan === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.679Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:12.680Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.raft: entering leader state: leader="Node at 127.0.0.1:30155 [Leader]" writer.go:29: 2021-01-29T19:34:12.682Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:12.686Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: initializing acls writer.go:29: 2021-01-29T19:34:12.683Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: New leader elected: payload=Node-746c73f1-7446-8845-98b0-57769cad6e6e writer.go:29: 2021-01-29T19:34:12.684Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: initializing acls === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.680Z [INFO] TestAgentConnectAuthorize_defaultDeny: Started DNS server: address=127.0.0.1:30156 network=udp === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.692Z [WARN] TestAgentConnectAuthorize_defaultAllow.server: Configuring a non-UUID master token is deprecated === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.694Z [INFO] TestAgentConnectAuthorize_defaultDeny: Started DNS server: address=127.0.0.1:30156 network=tcp === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.690Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: Created ACL 'global-management' policy === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.699Z [INFO] TestAgentConnectAuthorize_defaultDeny: Started HTTP server: address=127.0.0.1:30157 network=tcp === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.698Z [WARN] TestAgentConnectAuthorize_defaultAllow.server: Configuring a non-UUID master token is deprecated === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.700Z [INFO] TestAgentConnectAuthorize_defaultDeny: started state syncer === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.696Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:12.709Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:12.713Z [INFO] TestAgentConnectAuthorize_defaultAllow.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:12.710Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:12.714Z [INFO] TestAgentConnectAuthorize_defaultAllow.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:12.718Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.server: transitioning out of legacy ACL mode === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.721Z [WARN] TestAgentConnectAuthorize_defaultDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:12.722Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30161 [Candidate]" term=2 === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.718Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.serf.lan: serf: EventMemberUpdate: Node-746c73f1-7446-8845-98b0-57769cad6e6e === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.724Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.server.raft: votes: needed=1 === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.724Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: Updating LAN server: server="Node-746c73f1-7446-8845-98b0-57769cad6e6e (Addr: tcp/127.0.0.1:30155) (DC: dc1)" === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.725Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.server.raft: vote granted: from=300e0834-43a3-6e0f-c550-a2b9d86792c0 term=2 tally=1 writer.go:29: 2021-01-29T19:34:12.726Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.raft: election won: tally=1 === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.724Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.serf.lan: serf: EventMemberUpdate: Node-746c73f1-7446-8845-98b0-57769cad6e6e === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.726Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30161 [Leader]" === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.726Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: Updating LAN server: server="Node-746c73f1-7446-8845-98b0-57769cad6e6e (Addr: tcp/127.0.0.1:30155) (DC: dc1)" writer.go:29: 2021-01-29T19:34:12.725Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.serf.wan: serf: EventMemberUpdate: Node-746c73f1-7446-8845-98b0-57769cad6e6e.dc1 writer.go:29: 2021-01-29T19:34:12.728Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: Handled event for server in area: event=member-update server=Node-746c73f1-7446-8845-98b0-57769cad6e6e.dc1 area=wan === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.727Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: cluster leadership acquired === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.729Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.serf.wan: serf: EventMemberUpdate: Node-746c73f1-7446-8845-98b0-57769cad6e6e.dc1 writer.go:29: 2021-01-29T19:34:12.730Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: Handled event for server in area: event=member-update server=Node-746c73f1-7446-8845-98b0-57769cad6e6e.dc1 area=wan writer.go:29: 2021-01-29T19:34:12.732Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_Host writer.go:29: 2021-01-29T19:34:12.731Z [INFO] TestAgent_Host: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:12.733Z [INFO] TestAgent_Host: Endpoints down --- PASS: TestAgent_Host (0.84s) === CONT TestAgentConnectAuthorize_serviceWrite === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.736Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:12.737Z [INFO] TestAgentConnectAuthorize_defaultAllow.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:12.738Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.server: Skipping self join check for node since the cluster is too small: node=Node-746c73f1-7446-8845-98b0-57769cad6e6e writer.go:29: 2021-01-29T19:34:12.738Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: member joined, marking health alive: member=Node-746c73f1-7446-8845-98b0-57769cad6e6e writer.go:29: 2021-01-29T19:34:12.740Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.server: Skipping self join check for node since the cluster is too small: node=Node-746c73f1-7446-8845-98b0-57769cad6e6e writer.go:29: 2021-01-29T19:34:12.741Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.server: Skipping self join check for node since the cluster is too small: node=Node-746c73f1-7446-8845-98b0-57769cad6e6e === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.728Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: New leader elected: payload=Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 writer.go:29: 2021-01-29T19:34:12.728Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:12.730Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:12.750Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:12.754Z [WARN] TestAgentConnectAuthorize_defaultDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.758Z [WARN] TestAgentConnectAuthorize_serviceWrite: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.756Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Created ACL 'global-management' policy === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.759Z [WARN] TestAgentConnectAuthorize_serviceWrite: bootstrap = true: do not enable unless necessary === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.759Z [WARN] TestAgentConnectAuthorize_defaultDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.760Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.tlsutil: Update: version=1 === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.759Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Bootstrapped ACL master token from configuration === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.762Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.763Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Created ACL anonymous token from configuration === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.766Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6928a9e4-e412-d5d8-9b1a-11e8b0a4082d Address:127.0.0.1:30173}]" === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.764Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Created ACL anonymous token from configuration === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.767Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.raft: entering follower state: follower="Node at 127.0.0.1:30173 [Follower]" leader= === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.765Z [INFO] TestAgentConnectAuthorize_defaultDeny.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.769Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.serf.wan: serf: EventMemberJoin: Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d.dc1 127.0.0.1 === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.769Z [INFO] TestAgentConnectAuthorize_defaultDeny.leader: started routine: routine="acl token reaping" === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.773Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.serf.lan: serf: EventMemberJoin: Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d 127.0.0.1 === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.772Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.server: transitioning out of legacy ACL mode === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.775Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: Adding LAN server: server="Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d (Addr: tcp/127.0.0.1:30173) (DC: dc1)" writer.go:29: 2021-01-29T19:34:12.776Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: Handled event for server in area: event=member-join server=Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d.dc1 area=wan writer.go:29: 2021-01-29T19:34:12.778Z [INFO] TestAgentConnectAuthorize_serviceWrite: Started DNS server: address=127.0.0.1:30168 network=tcp === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.773Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.serf.lan: serf: EventMemberUpdate: Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 writer.go:29: 2021-01-29T19:34:12.781Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Updating LAN server: server="Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 (Addr: tcp/127.0.0.1:30161) (DC: dc1)" writer.go:29: 2021-01-29T19:34:12.781Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.serf.lan: serf: EventMemberUpdate: Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 writer.go:29: 2021-01-29T19:34:12.783Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Updating LAN server: server="Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 (Addr: tcp/127.0.0.1:30161) (DC: dc1)" writer.go:29: 2021-01-29T19:34:12.781Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.serf.wan: serf: EventMemberUpdate: Node-300e0834-43a3-6e0f-c550-a2b9d86792c0.dc1 writer.go:29: 2021-01-29T19:34:12.785Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Handled event for server in area: event=member-update server=Node-300e0834-43a3-6e0f-c550-a2b9d86792c0.dc1 area=wan === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.780Z [INFO] TestAgentConnectAuthorize_serviceWrite: Started DNS server: address=127.0.0.1:30168 network=udp === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.785Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.serf.wan: serf: EventMemberUpdate: Node-300e0834-43a3-6e0f-c550-a2b9d86792c0.dc1 writer.go:29: 2021-01-29T19:34:12.786Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:12.787Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: Handled event for server in area: event=member-update server=Node-300e0834-43a3-6e0f-c550-a2b9d86792c0.dc1 area=wan writer.go:29: 2021-01-29T19:34:12.791Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:12.793Z [INFO] TestAgentConnectAuthorize_defaultDeny.leader: started routine: routine="CA root pruning" === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.792Z [INFO] TestAgentConnectAuthorize_serviceWrite: Started HTTP server: address=127.0.0.1:30169 network=tcp writer.go:29: 2021-01-29T19:34:12.794Z [INFO] TestAgentConnectAuthorize_serviceWrite: started state syncer === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:12.794Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.server: Skipping self join check for node since the cluster is too small: node=Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 writer.go:29: 2021-01-29T19:34:12.796Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: member joined, marking health alive: member=Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 writer.go:29: 2021-01-29T19:34:12.798Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.server: Skipping self join check for node since the cluster is too small: node=Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 writer.go:29: 2021-01-29T19:34:12.800Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.server: Skipping self join check for node since the cluster is too small: node=Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.805Z [WARN] TestAgentConnectAuthorize_serviceWrite.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:12.806Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.raft: entering candidate state: node="Node at 127.0.0.1:30173 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:12.808Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:12.809Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.server.raft: vote granted: from=6928a9e4-e412-d5d8-9b1a-11e8b0a4082d term=2 tally=1 writer.go:29: 2021-01-29T19:34:12.810Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:12.811Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.raft: entering leader state: leader="Node at 127.0.0.1:30173 [Leader]" writer.go:29: 2021-01-29T19:34:12.813Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:12.814Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: New leader elected: payload=Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d writer.go:29: 2021-01-29T19:34:12.815Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: initializing acls writer.go:29: 2021-01-29T19:34:12.821Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:12.823Z [WARN] TestAgentConnectAuthorize_serviceWrite.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:12.826Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:12.827Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: initializing acls writer.go:29: 2021-01-29T19:34:12.830Z [WARN] TestAgentConnectAuthorize_serviceWrite.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:12.828Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:12.832Z [INFO] TestAgentConnectAuthorize_serviceWrite.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:12.834Z [INFO] TestAgentConnectAuthorize_serviceWrite.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:12.835Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:34:12.836Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.serf.lan: serf: EventMemberUpdate: Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d writer.go:29: 2021-01-29T19:34:12.837Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.serf.wan: serf: EventMemberUpdate: Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d.dc1 writer.go:29: 2021-01-29T19:34:12.839Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: Handled event for server in area: event=member-update server=Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d.dc1 area=wan writer.go:29: 2021-01-29T19:34:12.838Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.serf.lan: serf: EventMemberUpdate: Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d writer.go:29: 2021-01-29T19:34:12.838Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: Updating LAN server: server="Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d (Addr: tcp/127.0.0.1:30173) (DC: dc1)" writer.go:29: 2021-01-29T19:34:12.847Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: Updating LAN server: server="Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d (Addr: tcp/127.0.0.1:30173) (DC: dc1)" writer.go:29: 2021-01-29T19:34:12.848Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:12.851Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.serf.wan: serf: EventMemberUpdate: Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d.dc1 writer.go:29: 2021-01-29T19:34:12.858Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: Handled event for server in area: event=member-update server=Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d.dc1 area=wan writer.go:29: 2021-01-29T19:34:12.855Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:12.861Z [INFO] TestAgentConnectAuthorize_serviceWrite.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:12.863Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.server: Skipping self join check for node since the cluster is too small: node=Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d writer.go:29: 2021-01-29T19:34:12.864Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: member joined, marking health alive: member=Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.867Z [DEBUG] TestAgentConnectAuthorize_defaultAllow: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:12.867Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.server: Skipping self join check for node since the cluster is too small: node=Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d writer.go:29: 2021-01-29T19:34:12.868Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.server: Skipping self join check for node since the cluster is too small: node=Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:12.869Z [INFO] TestAgentConnectAuthorize_defaultAllow: Synced node info writer.go:29: 2021-01-29T19:34:12.944Z [INFO] TestAgentConnectAuthorize_defaultAllow: Requesting shutdown writer.go:29: 2021-01-29T19:34:12.945Z [INFO] TestAgentConnectAuthorize_defaultAllow.server: shutting down server writer.go:29: 2021-01-29T19:34:12.947Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:12.948Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:12.950Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:12.949Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:12.951Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:12.951Z [WARN] TestAgentConnectAuthorize_defaultAllow.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:12.951Z [DEBUG] TestAgentConnectAuthorize_defaultAllow.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:12.957Z [WARN] TestAgentConnectAuthorize_defaultAllow.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:12.962Z [INFO] TestAgentConnectAuthorize_defaultAllow.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:12.963Z [INFO] TestAgentConnectAuthorize_defaultAllow: consul server down writer.go:29: 2021-01-29T19:34:12.966Z [INFO] TestAgentConnectAuthorize_defaultAllow: shutdown complete writer.go:29: 2021-01-29T19:34:12.967Z [INFO] TestAgentConnectAuthorize_defaultAllow: Stopping server: protocol=DNS address=127.0.0.1:30150 network=tcp writer.go:29: 2021-01-29T19:34:12.968Z [INFO] TestAgentConnectAuthorize_defaultAllow: Stopping server: protocol=DNS address=127.0.0.1:30150 network=udp writer.go:29: 2021-01-29T19:34:12.970Z [INFO] TestAgentConnectAuthorize_defaultAllow: Stopping server: protocol=HTTP address=127.0.0.1:30151 network=tcp === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:13.008Z [DEBUG] TestAgentConnectAuthorize_serviceWrite: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:13.016Z [INFO] TestAgentConnectAuthorize_serviceWrite: Synced node info === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:13.049Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.acl: dropping node from result due to ACLs: node=Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 writer.go:29: 2021-01-29T19:34:13.054Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.acl: dropping node from result due to ACLs: node=Node-300e0834-43a3-6e0f-c550-a2b9d86792c0 writer.go:29: 2021-01-29T19:34:13.065Z [INFO] TestAgentConnectAuthorize_defaultDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:13.067Z [INFO] TestAgentConnectAuthorize_defaultDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:13.068Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:13.070Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:13.071Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:13.073Z [WARN] TestAgentConnectAuthorize_defaultDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:13.068Z [ERROR] TestAgentConnectAuthorize_defaultDeny.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:13.076Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:13.076Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:13.076Z [DEBUG] TestAgentConnectAuthorize_defaultDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:13.077Z [WARN] TestAgentConnectAuthorize_defaultDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:13.083Z [INFO] TestAgentConnectAuthorize_defaultDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:13.084Z [INFO] TestAgentConnectAuthorize_defaultDeny: consul server down writer.go:29: 2021-01-29T19:34:13.085Z [INFO] TestAgentConnectAuthorize_defaultDeny: shutdown complete writer.go:29: 2021-01-29T19:34:13.086Z [INFO] TestAgentConnectAuthorize_defaultDeny: Stopping server: protocol=DNS address=127.0.0.1:30156 network=tcp writer.go:29: 2021-01-29T19:34:13.087Z [INFO] TestAgentConnectAuthorize_defaultDeny: Stopping server: protocol=DNS address=127.0.0.1:30156 network=udp writer.go:29: 2021-01-29T19:34:13.088Z [INFO] TestAgentConnectAuthorize_defaultDeny: Stopping server: protocol=HTTP address=127.0.0.1:30157 network=tcp === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:13.170Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.acl: dropping node from result due to ACLs: node=Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d writer.go:29: 2021-01-29T19:34:13.175Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.acl: dropping node from result due to ACLs: node=Node-6928a9e4-e412-d5d8-9b1a-11e8b0a4082d writer.go:29: 2021-01-29T19:34:13.183Z [INFO] TestAgentConnectAuthorize_serviceWrite: Requesting shutdown writer.go:29: 2021-01-29T19:34:13.184Z [INFO] TestAgentConnectAuthorize_serviceWrite.server: shutting down server writer.go:29: 2021-01-29T19:34:13.186Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:13.187Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:13.188Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:13.187Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:13.188Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:13.190Z [DEBUG] TestAgentConnectAuthorize_serviceWrite.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:13.190Z [WARN] TestAgentConnectAuthorize_serviceWrite.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:13.198Z [WARN] TestAgentConnectAuthorize_serviceWrite.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:13.200Z [INFO] TestAgentConnectAuthorize_serviceWrite.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:13.200Z [INFO] TestAgentConnectAuthorize_serviceWrite: consul server down writer.go:29: 2021-01-29T19:34:13.201Z [INFO] TestAgentConnectAuthorize_serviceWrite: shutdown complete writer.go:29: 2021-01-29T19:34:13.202Z [INFO] TestAgentConnectAuthorize_serviceWrite: Stopping server: protocol=DNS address=127.0.0.1:30168 network=tcp writer.go:29: 2021-01-29T19:34:13.203Z [INFO] TestAgentConnectAuthorize_serviceWrite: Stopping server: protocol=DNS address=127.0.0.1:30168 network=udp writer.go:29: 2021-01-29T19:34:13.204Z [INFO] TestAgentConnectAuthorize_serviceWrite: Stopping server: protocol=HTTP address=127.0.0.1:30169 network=tcp === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:13.226Z [DEBUG] TestAgent_StartStop.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentConnectAuthorize_defaultAllow writer.go:29: 2021-01-29T19:34:13.472Z [INFO] TestAgentConnectAuthorize_defaultAllow: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:13.474Z [INFO] TestAgentConnectAuthorize_defaultAllow: Endpoints down --- PASS: TestAgentConnectAuthorize_defaultAllow (0.91s) === CONT TestAgentConnectAuthorize_denyWildcard writer.go:29: 2021-01-29T19:34:13.485Z [WARN] TestAgentConnectAuthorize_denyWildcard: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:13.486Z [DEBUG] TestAgentConnectAuthorize_denyWildcard.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:13.488Z [DEBUG] TestAgentConnectAuthorize_denyWildcard.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:13.491Z [INFO] TestAgentConnectAuthorize_denyWildcard.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:982f9adb-3c93-7a25-7a28-1ddcba576eac Address:127.0.0.1:30179}]" writer.go:29: 2021-01-29T19:34:13.493Z [INFO] TestAgentConnectAuthorize_denyWildcard.server.raft: entering follower state: follower="Node at 127.0.0.1:30179 [Follower]" leader= writer.go:29: 2021-01-29T19:34:13.496Z [INFO] TestAgentConnectAuthorize_denyWildcard.server.serf.wan: serf: EventMemberJoin: Node-982f9adb-3c93-7a25-7a28-1ddcba576eac.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:13.500Z [INFO] TestAgentConnectAuthorize_denyWildcard.server.serf.lan: serf: EventMemberJoin: Node-982f9adb-3c93-7a25-7a28-1ddcba576eac 127.0.0.1 writer.go:29: 2021-01-29T19:34:13.504Z [INFO] TestAgentConnectAuthorize_denyWildcard.server: Adding LAN server: server="Node-982f9adb-3c93-7a25-7a28-1ddcba576eac (Addr: tcp/127.0.0.1:30179) (DC: dc1)" writer.go:29: 2021-01-29T19:34:13.505Z [INFO] TestAgentConnectAuthorize_denyWildcard.server: Handled event for server in area: event=member-join server=Node-982f9adb-3c93-7a25-7a28-1ddcba576eac.dc1 area=wan writer.go:29: 2021-01-29T19:34:13.507Z [INFO] TestAgentConnectAuthorize_denyWildcard: Started DNS server: address=127.0.0.1:30174 network=tcp writer.go:29: 2021-01-29T19:34:13.509Z [INFO] TestAgentConnectAuthorize_denyWildcard: Started DNS server: address=127.0.0.1:30174 network=udp writer.go:29: 2021-01-29T19:34:13.511Z [INFO] TestAgentConnectAuthorize_denyWildcard: Started HTTP server: address=127.0.0.1:30175 network=tcp writer.go:29: 2021-01-29T19:34:13.511Z [INFO] TestAgentConnectAuthorize_denyWildcard: started state syncer writer.go:29: 2021-01-29T19:34:13.544Z [WARN] TestAgentConnectAuthorize_denyWildcard.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:13.545Z [INFO] TestAgentConnectAuthorize_denyWildcard.server.raft: entering candidate state: node="Node at 127.0.0.1:30179 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:13.549Z [DEBUG] TestAgentConnectAuthorize_denyWildcard.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:13.552Z [DEBUG] TestAgentConnectAuthorize_denyWildcard.server.raft: vote granted: from=982f9adb-3c93-7a25-7a28-1ddcba576eac term=2 tally=1 writer.go:29: 2021-01-29T19:34:13.554Z [INFO] TestAgentConnectAuthorize_denyWildcard.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:13.557Z [INFO] TestAgentConnectAuthorize_denyWildcard.server.raft: entering leader state: leader="Node at 127.0.0.1:30179 [Leader]" writer.go:29: 2021-01-29T19:34:13.562Z [INFO] TestAgentConnectAuthorize_denyWildcard.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:13.562Z [INFO] TestAgentConnectAuthorize_denyWildcard.server: New leader elected: payload=Node-982f9adb-3c93-7a25-7a28-1ddcba576eac writer.go:29: 2021-01-29T19:34:13.567Z [DEBUG] TestAgentConnectAuthorize_denyWildcard.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30179 writer.go:29: 2021-01-29T19:34:13.577Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:13.587Z [INFO] TestAgentConnectAuthorize_denyWildcard.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgentConnectAuthorize_defaultDeny writer.go:29: 2021-01-29T19:34:13.589Z [INFO] TestAgentConnectAuthorize_defaultDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:13.590Z [INFO] TestAgentConnectAuthorize_defaultDeny: Endpoints down --- PASS: TestAgentConnectAuthorize_defaultDeny (0.96s) === CONT TestAgentConnectAuthorize_allowTrustDomain === CONT TestAgentConnectAuthorize_denyWildcard writer.go:29: 2021-01-29T19:34:13.590Z [INFO] TestAgentConnectAuthorize_denyWildcard.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:13.594Z [DEBUG] TestAgentConnectAuthorize_denyWildcard.server: Skipping self join check for node since the cluster is too small: node=Node-982f9adb-3c93-7a25-7a28-1ddcba576eac writer.go:29: 2021-01-29T19:34:13.597Z [INFO] TestAgentConnectAuthorize_denyWildcard.server: member joined, marking health alive: member=Node-982f9adb-3c93-7a25-7a28-1ddcba576eac === CONT TestAgentConnectAuthorize_allowTrustDomain writer.go:29: 2021-01-29T19:34:13.602Z [WARN] TestAgentConnectAuthorize_allowTrustDomain: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:13.602Z [DEBUG] TestAgentConnectAuthorize_allowTrustDomain.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:13.604Z [DEBUG] TestAgentConnectAuthorize_allowTrustDomain.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentConnectAuthorize_denyWildcard writer.go:29: 2021-01-29T19:34:13.613Z [DEBUG] TestAgentConnectAuthorize_denyWildcard: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgentConnectAuthorize_allowTrustDomain writer.go:29: 2021-01-29T19:34:13.616Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6cc42b5a-7f01-5f1d-a18f-9560f3d59ac4 Address:127.0.0.1:30191}]" === CONT TestAgentConnectAuthorize_denyWildcard writer.go:29: 2021-01-29T19:34:13.617Z [INFO] TestAgentConnectAuthorize_denyWildcard: Synced node info === CONT TestAgentConnectAuthorize_allowTrustDomain writer.go:29: 2021-01-29T19:34:13.618Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server.serf.wan: serf: EventMemberJoin: Node-6cc42b5a-7f01-5f1d-a18f-9560f3d59ac4.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:13.618Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server.raft: entering follower state: follower="Node at 127.0.0.1:30191 [Follower]" leader= writer.go:29: 2021-01-29T19:34:13.621Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server.serf.lan: serf: EventMemberJoin: Node-6cc42b5a-7f01-5f1d-a18f-9560f3d59ac4 127.0.0.1 writer.go:29: 2021-01-29T19:34:13.625Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server: Adding LAN server: server="Node-6cc42b5a-7f01-5f1d-a18f-9560f3d59ac4 (Addr: tcp/127.0.0.1:30191) (DC: dc1)" writer.go:29: 2021-01-29T19:34:13.625Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server: Handled event for server in area: event=member-join server=Node-6cc42b5a-7f01-5f1d-a18f-9560f3d59ac4.dc1 area=wan writer.go:29: 2021-01-29T19:34:13.627Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: Started DNS server: address=127.0.0.1:30186 network=tcp writer.go:29: 2021-01-29T19:34:13.628Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: Started DNS server: address=127.0.0.1:30186 network=udp writer.go:29: 2021-01-29T19:34:13.630Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: Started HTTP server: address=127.0.0.1:30187 network=tcp writer.go:29: 2021-01-29T19:34:13.632Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: started state syncer writer.go:29: 2021-01-29T19:34:13.681Z [WARN] TestAgentConnectAuthorize_allowTrustDomain.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:13.682Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server.raft: entering candidate state: node="Node at 127.0.0.1:30191 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:13.684Z [DEBUG] TestAgentConnectAuthorize_allowTrustDomain.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:13.684Z [DEBUG] TestAgentConnectAuthorize_allowTrustDomain.server.raft: vote granted: from=6cc42b5a-7f01-5f1d-a18f-9560f3d59ac4 term=2 tally=1 writer.go:29: 2021-01-29T19:34:13.686Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:13.688Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server.raft: entering leader state: leader="Node at 127.0.0.1:30191 [Leader]" writer.go:29: 2021-01-29T19:34:13.689Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:13.690Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server: New leader elected: payload=Node-6cc42b5a-7f01-5f1d-a18f-9560f3d59ac4 writer.go:29: 2021-01-29T19:34:13.691Z [DEBUG] TestAgentConnectAuthorize_allowTrustDomain.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30191 writer.go:29: 2021-01-29T19:34:13.697Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:13.704Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:13.706Z [INFO] TestAgentConnectAuthorize_serviceWrite: Waiting for endpoints to shut down === CONT TestAgentConnectAuthorize_allowTrustDomain writer.go:29: 2021-01-29T19:34:13.706Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.leader: started routine: routine="CA root pruning" === CONT TestAgentConnectAuthorize_serviceWrite writer.go:29: 2021-01-29T19:34:13.707Z [INFO] TestAgentConnectAuthorize_serviceWrite: Endpoints down --- PASS: TestAgentConnectAuthorize_serviceWrite (0.97s) === CONT TestAgentConnectAuthorize_deny === CONT TestAgentConnectAuthorize_allowTrustDomain writer.go:29: 2021-01-29T19:34:13.707Z [DEBUG] TestAgentConnectAuthorize_allowTrustDomain.server: Skipping self join check for node since the cluster is too small: node=Node-6cc42b5a-7f01-5f1d-a18f-9560f3d59ac4 writer.go:29: 2021-01-29T19:34:13.709Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server: member joined, marking health alive: member=Node-6cc42b5a-7f01-5f1d-a18f-9560f3d59ac4 === CONT TestAgentConnectAuthorize_deny writer.go:29: 2021-01-29T19:34:13.716Z [WARN] TestAgentConnectAuthorize_deny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:13.717Z [DEBUG] TestAgentConnectAuthorize_deny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:13.718Z [DEBUG] TestAgentConnectAuthorize_deny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:13.723Z [INFO] TestAgentConnectAuthorize_deny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fab0c5ff-13e0-9c3e-9dbc-7066e3265d3b Address:127.0.0.1:30185}]" writer.go:29: 2021-01-29T19:34:13.726Z [INFO] TestAgentConnectAuthorize_deny.server.raft: entering follower state: follower="Node at 127.0.0.1:30185 [Follower]" leader= writer.go:29: 2021-01-29T19:34:13.727Z [INFO] TestAgentConnectAuthorize_deny.server.serf.wan: serf: EventMemberJoin: Node-fab0c5ff-13e0-9c3e-9dbc-7066e3265d3b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:13.730Z [INFO] TestAgentConnectAuthorize_deny.server.serf.lan: serf: EventMemberJoin: Node-fab0c5ff-13e0-9c3e-9dbc-7066e3265d3b 127.0.0.1 writer.go:29: 2021-01-29T19:34:13.732Z [INFO] TestAgentConnectAuthorize_deny.server: Handled event for server in area: event=member-join server=Node-fab0c5ff-13e0-9c3e-9dbc-7066e3265d3b.dc1 area=wan writer.go:29: 2021-01-29T19:34:13.732Z [INFO] TestAgentConnectAuthorize_deny.server: Adding LAN server: server="Node-fab0c5ff-13e0-9c3e-9dbc-7066e3265d3b (Addr: tcp/127.0.0.1:30185) (DC: dc1)" writer.go:29: 2021-01-29T19:34:13.733Z [INFO] TestAgentConnectAuthorize_deny: Started DNS server: address=127.0.0.1:30180 network=udp writer.go:29: 2021-01-29T19:34:13.737Z [INFO] TestAgentConnectAuthorize_deny: Started DNS server: address=127.0.0.1:30180 network=tcp writer.go:29: 2021-01-29T19:34:13.739Z [INFO] TestAgentConnectAuthorize_deny: Started HTTP server: address=127.0.0.1:30181 network=tcp writer.go:29: 2021-01-29T19:34:13.741Z [INFO] TestAgentConnectAuthorize_deny: started state syncer writer.go:29: 2021-01-29T19:34:13.765Z [WARN] TestAgentConnectAuthorize_deny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:13.768Z [INFO] TestAgentConnectAuthorize_deny.server.raft: entering candidate state: node="Node at 127.0.0.1:30185 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:13.772Z [DEBUG] TestAgentConnectAuthorize_deny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:13.774Z [DEBUG] TestAgentConnectAuthorize_deny.server.raft: vote granted: from=fab0c5ff-13e0-9c3e-9dbc-7066e3265d3b term=2 tally=1 writer.go:29: 2021-01-29T19:34:13.777Z [INFO] TestAgentConnectAuthorize_deny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:13.779Z [INFO] TestAgentConnectAuthorize_deny.server.raft: entering leader state: leader="Node at 127.0.0.1:30185 [Leader]" writer.go:29: 2021-01-29T19:34:13.781Z [INFO] TestAgentConnectAuthorize_deny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:13.784Z [INFO] TestAgentConnectAuthorize_deny.server: New leader elected: payload=Node-fab0c5ff-13e0-9c3e-9dbc-7066e3265d3b writer.go:29: 2021-01-29T19:34:13.788Z [DEBUG] TestAgentConnectAuthorize_deny.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30185 writer.go:29: 2021-01-29T19:34:13.796Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:13.804Z [INFO] TestAgentConnectAuthorize_deny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:13.806Z [INFO] TestAgentConnectAuthorize_deny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:13.809Z [DEBUG] TestAgentConnectAuthorize_deny.server: Skipping self join check for node since the cluster is too small: node=Node-fab0c5ff-13e0-9c3e-9dbc-7066e3265d3b writer.go:29: 2021-01-29T19:34:13.811Z [INFO] TestAgentConnectAuthorize_deny.server: member joined, marking health alive: member=Node-fab0c5ff-13e0-9c3e-9dbc-7066e3265d3b === CONT TestAgentConnectAuthorize_denyWildcard writer.go:29: 2021-01-29T19:34:13.899Z [INFO] TestAgentConnectAuthorize_denyWildcard: Requesting shutdown writer.go:29: 2021-01-29T19:34:13.901Z [INFO] TestAgentConnectAuthorize_denyWildcard.server: shutting down server writer.go:29: 2021-01-29T19:34:13.903Z [DEBUG] TestAgentConnectAuthorize_denyWildcard.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:13.905Z [WARN] TestAgentConnectAuthorize_denyWildcard.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:13.905Z [DEBUG] TestAgentConnectAuthorize_denyWildcard.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:13.909Z [WARN] TestAgentConnectAuthorize_denyWildcard.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:13.911Z [INFO] TestAgentConnectAuthorize_denyWildcard.server.router.manager: shutting down === CONT TestAgentConnectAuthorize_allowTrustDomain writer.go:29: 2021-01-29T19:34:13.914Z [DEBUG] TestAgentConnectAuthorize_allowTrustDomain: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:13.919Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: Synced node info === CONT TestAgentConnectAuthorize_denyWildcard writer.go:29: 2021-01-29T19:34:13.912Z [INFO] TestAgentConnectAuthorize_denyWildcard: consul server down === CONT TestAgentConnectAuthorize_allowTrustDomain writer.go:29: 2021-01-29T19:34:13.920Z [DEBUG] TestAgentConnectAuthorize_allowTrustDomain: Node info in sync === CONT TestAgentConnectAuthorize_denyWildcard writer.go:29: 2021-01-29T19:34:13.920Z [INFO] TestAgentConnectAuthorize_denyWildcard: shutdown complete writer.go:29: 2021-01-29T19:34:13.922Z [INFO] TestAgentConnectAuthorize_denyWildcard: Stopping server: protocol=DNS address=127.0.0.1:30174 network=tcp writer.go:29: 2021-01-29T19:34:13.924Z [INFO] TestAgentConnectAuthorize_denyWildcard: Stopping server: protocol=DNS address=127.0.0.1:30174 network=udp writer.go:29: 2021-01-29T19:34:13.927Z [INFO] TestAgentConnectAuthorize_denyWildcard: Stopping server: protocol=HTTP address=127.0.0.1:30175 network=tcp === CONT TestAgentConnectAuthorize_allowTrustDomain writer.go:29: 2021-01-29T19:34:14.045Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: Requesting shutdown writer.go:29: 2021-01-29T19:34:14.047Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server: shutting down server writer.go:29: 2021-01-29T19:34:14.049Z [DEBUG] TestAgentConnectAuthorize_allowTrustDomain.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.051Z [DEBUG] TestAgentConnectAuthorize_allowTrustDomain.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.053Z [WARN] TestAgentConnectAuthorize_allowTrustDomain.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:14.060Z [WARN] TestAgentConnectAuthorize_allowTrustDomain.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:14.062Z [INFO] TestAgentConnectAuthorize_allowTrustDomain.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:14.063Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: consul server down writer.go:29: 2021-01-29T19:34:14.064Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: shutdown complete writer.go:29: 2021-01-29T19:34:14.065Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: Stopping server: protocol=DNS address=127.0.0.1:30186 network=tcp writer.go:29: 2021-01-29T19:34:14.065Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: Stopping server: protocol=DNS address=127.0.0.1:30186 network=udp writer.go:29: 2021-01-29T19:34:14.066Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: Stopping server: protocol=HTTP address=127.0.0.1:30187 network=tcp === CONT TestAgentConnectAuthorize_deny writer.go:29: 2021-01-29T19:34:14.126Z [INFO] TestAgentConnectAuthorize_deny: Requesting shutdown writer.go:29: 2021-01-29T19:34:14.127Z [INFO] TestAgentConnectAuthorize_deny.server: shutting down server writer.go:29: 2021-01-29T19:34:14.128Z [DEBUG] TestAgentConnectAuthorize_deny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.129Z [WARN] TestAgentConnectAuthorize_deny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:14.130Z [ERROR] TestAgentConnectAuthorize_deny.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:14.131Z [DEBUG] TestAgentConnectAuthorize_deny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.136Z [WARN] TestAgentConnectAuthorize_deny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:14.138Z [INFO] TestAgentConnectAuthorize_deny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:14.139Z [INFO] TestAgentConnectAuthorize_deny: consul server down writer.go:29: 2021-01-29T19:34:14.140Z [INFO] TestAgentConnectAuthorize_deny: shutdown complete writer.go:29: 2021-01-29T19:34:14.141Z [INFO] TestAgentConnectAuthorize_deny: Stopping server: protocol=DNS address=127.0.0.1:30180 network=tcp writer.go:29: 2021-01-29T19:34:14.142Z [INFO] TestAgentConnectAuthorize_deny: Stopping server: protocol=DNS address=127.0.0.1:30180 network=udp writer.go:29: 2021-01-29T19:34:14.143Z [INFO] TestAgentConnectAuthorize_deny: Stopping server: protocol=HTTP address=127.0.0.1:30181 network=tcp === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:14.315Z [INFO] TestAgent_StartStop.server.serf.lan: serf: EventMemberLeave: Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086 127.0.0.1 writer.go:29: 2021-01-29T19:34:14.317Z [INFO] TestAgent_StartStop.server: Removing LAN server: server="Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086 (Addr: tcp/127.0.0.1:30137) (DC: dc1)" writer.go:29: 2021-01-29T19:34:14.317Z [WARN] TestAgent_StartStop.server: deregistering self should be done by follower: name=Node-33dc3d00-b7b7-13ad-16dd-3c5ffec4e086 === CONT TestAgentConnectAuthorize_denyWildcard writer.go:29: 2021-01-29T19:34:14.431Z [INFO] TestAgentConnectAuthorize_denyWildcard: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:14.433Z [INFO] TestAgentConnectAuthorize_denyWildcard: Endpoints down --- PASS: TestAgentConnectAuthorize_denyWildcard (0.96s) === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.469Z [WARN] TestAgentConnectAuthorize_allow: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:14.474Z [DEBUG] TestAgentConnectAuthorize_allow.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:14.494Z [DEBUG] TestAgentConnectAuthorize_allow.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:14.508Z [INFO] TestAgentConnectAuthorize_allow.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a6a767a3-f1d2-6a69-b2a3-3bfd8662b95c Address:127.0.0.1:30197}]" writer.go:29: 2021-01-29T19:34:14.512Z [INFO] TestAgentConnectAuthorize_allow.server.raft: entering follower state: follower="Node at 127.0.0.1:30197 [Follower]" leader= writer.go:29: 2021-01-29T19:34:14.520Z [INFO] TestAgentConnectAuthorize_allow.server.serf.wan: serf: EventMemberJoin: Node-a6a767a3-f1d2-6a69-b2a3-3bfd8662b95c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:14.539Z [INFO] TestAgentConnectAuthorize_allow.server.serf.lan: serf: EventMemberJoin: Node-a6a767a3-f1d2-6a69-b2a3-3bfd8662b95c 127.0.0.1 writer.go:29: 2021-01-29T19:34:14.552Z [INFO] TestAgentConnectAuthorize_allow.server: Handled event for server in area: event=member-join server=Node-a6a767a3-f1d2-6a69-b2a3-3bfd8662b95c.dc1 area=wan writer.go:29: 2021-01-29T19:34:14.553Z [INFO] TestAgentConnectAuthorize_allow.server: Adding LAN server: server="Node-a6a767a3-f1d2-6a69-b2a3-3bfd8662b95c (Addr: tcp/127.0.0.1:30197) (DC: dc1)" writer.go:29: 2021-01-29T19:34:14.553Z [INFO] TestAgentConnectAuthorize_allow: Started DNS server: address=127.0.0.1:30192 network=udp writer.go:29: 2021-01-29T19:34:14.556Z [INFO] TestAgentConnectAuthorize_allow: Started DNS server: address=127.0.0.1:30192 network=tcp writer.go:29: 2021-01-29T19:34:14.559Z [INFO] TestAgentConnectAuthorize_allow: Started HTTP server: address=127.0.0.1:30193 network=tcp === CONT TestAgentConnectAuthorize_allowTrustDomain writer.go:29: 2021-01-29T19:34:14.567Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: Waiting for endpoints to shut down === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.560Z [INFO] TestAgentConnectAuthorize_allow: started state syncer === CONT TestAgentConnectAuthorize_allowTrustDomain writer.go:29: 2021-01-29T19:34:14.568Z [INFO] TestAgentConnectAuthorize_allowTrustDomain: Endpoints down --- PASS: TestAgentConnectAuthorize_allowTrustDomain (0.98s) === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.577Z [WARN] TestAgentConnectAuthorize_idNotService: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:14.578Z [DEBUG] TestAgentConnectAuthorize_idNotService.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:14.580Z [DEBUG] TestAgentConnectAuthorize_idNotService.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.578Z [WARN] TestAgentConnectAuthorize_allow.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:14.581Z [INFO] TestAgentConnectAuthorize_allow.server.raft: entering candidate state: node="Node at 127.0.0.1:30197 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:14.592Z [DEBUG] TestAgentConnectAuthorize_allow.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:14.593Z [DEBUG] TestAgentConnectAuthorize_allow.server.raft: vote granted: from=a6a767a3-f1d2-6a69-b2a3-3bfd8662b95c term=2 tally=1 writer.go:29: 2021-01-29T19:34:14.595Z [INFO] TestAgentConnectAuthorize_allow.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:14.597Z [INFO] TestAgentConnectAuthorize_allow.server.raft: entering leader state: leader="Node at 127.0.0.1:30197 [Leader]" === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.597Z [INFO] TestAgentConnectAuthorize_idNotService.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d5168af2-b295-b87c-ab7c-dcb2b44d461b Address:127.0.0.1:30203}]" === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.598Z [INFO] TestAgentConnectAuthorize_allow.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:14.599Z [INFO] TestAgentConnectAuthorize_allow.server: New leader elected: payload=Node-a6a767a3-f1d2-6a69-b2a3-3bfd8662b95c === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.600Z [INFO] TestAgentConnectAuthorize_idNotService.server.raft: entering follower state: follower="Node at 127.0.0.1:30203 [Follower]" leader= === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.603Z [DEBUG] TestAgentConnectAuthorize_allow.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30197 writer.go:29: 2021-01-29T19:34:14.608Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.605Z [INFO] TestAgentConnectAuthorize_idNotService.server.serf.wan: serf: EventMemberJoin: Node-d5168af2-b295-b87c-ab7c-dcb2b44d461b.dc1 127.0.0.1 === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.615Z [INFO] TestAgentConnectAuthorize_allow.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:14.617Z [INFO] TestAgentConnectAuthorize_allow.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.619Z [DEBUG] TestAgentConnectAuthorize_allow.server: Skipping self join check for node since the cluster is too small: node=Node-a6a767a3-f1d2-6a69-b2a3-3bfd8662b95c writer.go:29: 2021-01-29T19:34:14.620Z [INFO] TestAgentConnectAuthorize_allow.server: member joined, marking health alive: member=Node-a6a767a3-f1d2-6a69-b2a3-3bfd8662b95c === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.627Z [INFO] TestAgentConnectAuthorize_idNotService.server.serf.lan: serf: EventMemberJoin: Node-d5168af2-b295-b87c-ab7c-dcb2b44d461b 127.0.0.1 writer.go:29: 2021-01-29T19:34:14.641Z [INFO] TestAgentConnectAuthorize_idNotService.server: Adding LAN server: server="Node-d5168af2-b295-b87c-ab7c-dcb2b44d461b (Addr: tcp/127.0.0.1:30203) (DC: dc1)" writer.go:29: 2021-01-29T19:34:14.643Z [INFO] TestAgentConnectAuthorize_idNotService.server: Handled event for server in area: event=member-join server=Node-d5168af2-b295-b87c-ab7c-dcb2b44d461b.dc1 area=wan === CONT TestAgentConnectAuthorize_deny writer.go:29: 2021-01-29T19:34:14.644Z [INFO] TestAgentConnectAuthorize_deny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:14.646Z [INFO] TestAgentConnectAuthorize_deny: Endpoints down --- PASS: TestAgentConnectAuthorize_deny (0.94s) === CONT TestAgentConnectAuthorize_idInvalidFormat === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.647Z [INFO] TestAgentConnectAuthorize_idNotService: Started DNS server: address=127.0.0.1:30198 network=udp writer.go:29: 2021-01-29T19:34:14.650Z [INFO] TestAgentConnectAuthorize_idNotService: Started DNS server: address=127.0.0.1:30198 network=tcp writer.go:29: 2021-01-29T19:34:14.655Z [INFO] TestAgentConnectAuthorize_idNotService: Started HTTP server: address=127.0.0.1:30199 network=tcp writer.go:29: 2021-01-29T19:34:14.672Z [INFO] TestAgentConnectAuthorize_idNotService: started state syncer writer.go:29: 2021-01-29T19:34:14.682Z [WARN] TestAgentConnectAuthorize_idNotService.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:14.683Z [INFO] TestAgentConnectAuthorize_idNotService.server.raft: entering candidate state: node="Node at 127.0.0.1:30203 [Candidate]" term=2 === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.695Z [WARN] TestAgentConnectAuthorize_idInvalidFormat: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:14.698Z [DEBUG] TestAgentConnectAuthorize_idInvalidFormat.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:14.707Z [DEBUG] TestAgentConnectAuthorize_idInvalidFormat.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.709Z [DEBUG] TestAgentConnectAuthorize_idNotService.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:14.742Z [DEBUG] TestAgentConnectAuthorize_idNotService.server.raft: vote granted: from=d5168af2-b295-b87c-ab7c-dcb2b44d461b term=2 tally=1 writer.go:29: 2021-01-29T19:34:14.743Z [INFO] TestAgentConnectAuthorize_idNotService.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:14.744Z [INFO] TestAgentConnectAuthorize_idNotService.server.raft: entering leader state: leader="Node at 127.0.0.1:30203 [Leader]" writer.go:29: 2021-01-29T19:34:14.746Z [INFO] TestAgentConnectAuthorize_idNotService.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:14.748Z [INFO] TestAgentConnectAuthorize_idNotService.server: New leader elected: payload=Node-d5168af2-b295-b87c-ab7c-dcb2b44d461b writer.go:29: 2021-01-29T19:34:14.753Z [DEBUG] TestAgentConnectAuthorize_idNotService.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30203 === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.770Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:517f173d-3674-43ee-f7da-dd9b4f7c6e9a Address:127.0.0.1:30209}]" writer.go:29: 2021-01-29T19:34:14.774Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server.serf.wan: serf: EventMemberJoin: Node-517f173d-3674-43ee-f7da-dd9b4f7c6e9a.dc1 127.0.0.1 === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.796Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.779Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server.serf.lan: serf: EventMemberJoin: Node-517f173d-3674-43ee-f7da-dd9b4f7c6e9a 127.0.0.1 writer.go:29: 2021-01-29T19:34:14.801Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server.raft: entering follower state: follower="Node at 127.0.0.1:30209 [Follower]" leader= writer.go:29: 2021-01-29T19:34:14.827Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: Started DNS server: address=127.0.0.1:30204 network=udp writer.go:29: 2021-01-29T19:34:14.835Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server: Adding LAN server: server="Node-517f173d-3674-43ee-f7da-dd9b4f7c6e9a (Addr: tcp/127.0.0.1:30209) (DC: dc1)" writer.go:29: 2021-01-29T19:34:14.846Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: Started DNS server: address=127.0.0.1:30204 network=tcp === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.850Z [INFO] TestAgentConnectAuthorize_idNotService.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.847Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server: Handled event for server in area: event=member-join server=Node-517f173d-3674-43ee-f7da-dd9b4f7c6e9a.dc1 area=wan === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.851Z [INFO] TestAgentConnectAuthorize_idNotService.leader: started routine: routine="CA root pruning" === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.850Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: Started HTTP server: address=127.0.0.1:30205 network=tcp === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.854Z [DEBUG] TestAgentConnectAuthorize_idNotService.server: Skipping self join check for node since the cluster is too small: node=Node-d5168af2-b295-b87c-ab7c-dcb2b44d461b === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.856Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: started state syncer === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.856Z [INFO] TestAgentConnectAuthorize_idNotService.server: member joined, marking health alive: member=Node-d5168af2-b295-b87c-ab7c-dcb2b44d461b === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.852Z [DEBUG] TestAgentConnectAuthorize_allow: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:14.887Z [INFO] TestAgentConnectAuthorize_allow: Synced node info writer.go:29: 2021-01-29T19:34:14.890Z [DEBUG] TestAgentConnectAuthorize_allow: Node info in sync === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.901Z [WARN] TestAgentConnectAuthorize_idInvalidFormat.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.921Z [INFO] TestAgentConnectAuthorize_idNotService: Requesting shutdown === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.932Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server.raft: entering candidate state: node="Node at 127.0.0.1:30209 [Candidate]" term=2 === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.933Z [INFO] TestAgentConnectAuthorize_idNotService.server: shutting down server writer.go:29: 2021-01-29T19:34:14.934Z [DEBUG] TestAgentConnectAuthorize_idNotService.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.935Z [WARN] TestAgentConnectAuthorize_idNotService.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:14.934Z [ERROR] TestAgentConnectAuthorize_idNotService.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:14.937Z [DEBUG] TestAgentConnectAuthorize_idNotService.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.939Z [WARN] TestAgentConnectAuthorize_idNotService.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.941Z [INFO] TestAgentConnectAuthorize_allow: Requesting shutdown writer.go:29: 2021-01-29T19:34:14.942Z [INFO] TestAgentConnectAuthorize_allow.server: shutting down server writer.go:29: 2021-01-29T19:34:14.943Z [DEBUG] TestAgentConnectAuthorize_allow.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.944Z [WARN] TestAgentConnectAuthorize_allow.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:14.945Z [DEBUG] TestAgentConnectAuthorize_allow.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.947Z [WARN] TestAgentConnectAuthorize_allow.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.946Z [INFO] TestAgentConnectAuthorize_idNotService.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:14.950Z [INFO] TestAgentConnectAuthorize_idNotService: consul server down === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.958Z [INFO] TestAgentConnectAuthorize_allow.server.router.manager: shutting down === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.958Z [INFO] TestAgentConnectAuthorize_idNotService: shutdown complete === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.962Z [DEBUG] TestAgentConnectAuthorize_idInvalidFormat.server.raft: votes: needed=1 === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.959Z [INFO] TestAgentConnectAuthorize_allow: consul server down === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.963Z [DEBUG] TestAgentConnectAuthorize_idInvalidFormat.server.raft: vote granted: from=517f173d-3674-43ee-f7da-dd9b4f7c6e9a term=2 tally=1 writer.go:29: 2021-01-29T19:34:14.963Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:14.964Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server.raft: entering leader state: leader="Node at 127.0.0.1:30209 [Leader]" === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.963Z [INFO] TestAgentConnectAuthorize_allow: shutdown complete === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.962Z [INFO] TestAgentConnectAuthorize_idNotService: Stopping server: protocol=DNS address=127.0.0.1:30198 network=tcp === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.965Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:14.967Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server: New leader elected: payload=Node-517f173d-3674-43ee-f7da-dd9b4f7c6e9a === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.965Z [INFO] TestAgentConnectAuthorize_allow: Stopping server: protocol=DNS address=127.0.0.1:30192 network=tcp === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.966Z [INFO] TestAgentConnectAuthorize_idNotService: Stopping server: protocol=DNS address=127.0.0.1:30198 network=udp === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.968Z [INFO] TestAgentConnectAuthorize_allow: Stopping server: protocol=DNS address=127.0.0.1:30192 network=udp === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:14.969Z [INFO] TestAgentConnectAuthorize_idNotService: Stopping server: protocol=HTTP address=127.0.0.1:30199 network=tcp === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:14.969Z [INFO] TestAgentConnectAuthorize_allow: Stopping server: protocol=HTTP address=127.0.0.1:30193 network=tcp === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:14.970Z [DEBUG] TestAgentConnectAuthorize_idInvalidFormat.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30209 writer.go:29: 2021-01-29T19:34:14.971Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: Synced node info writer.go:29: 2021-01-29T19:34:14.973Z [DEBUG] TestAgentConnectAuthorize_idInvalidFormat: Node info in sync writer.go:29: 2021-01-29T19:34:14.982Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:14.994Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:14.995Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.996Z [DEBUG] TestAgentConnectAuthorize_idInvalidFormat.server: Skipping self join check for node since the cluster is too small: node=Node-517f173d-3674-43ee-f7da-dd9b4f7c6e9a writer.go:29: 2021-01-29T19:34:14.997Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server: member joined, marking health alive: member=Node-517f173d-3674-43ee-f7da-dd9b4f7c6e9a === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:15.215Z [ERROR] TestAgent_StartStop.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:15.289Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: Requesting shutdown writer.go:29: 2021-01-29T19:34:15.291Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server: shutting down server writer.go:29: 2021-01-29T19:34:15.292Z [DEBUG] TestAgentConnectAuthorize_idInvalidFormat.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.294Z [DEBUG] TestAgentConnectAuthorize_idInvalidFormat.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.295Z [WARN] TestAgentConnectAuthorize_idInvalidFormat.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:15.298Z [WARN] TestAgentConnectAuthorize_idInvalidFormat.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:15.300Z [INFO] TestAgentConnectAuthorize_idInvalidFormat.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:15.300Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: consul server down writer.go:29: 2021-01-29T19:34:15.303Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: shutdown complete writer.go:29: 2021-01-29T19:34:15.304Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: Stopping server: protocol=DNS address=127.0.0.1:30204 network=tcp writer.go:29: 2021-01-29T19:34:15.306Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: Stopping server: protocol=DNS address=127.0.0.1:30204 network=udp writer.go:29: 2021-01-29T19:34:15.308Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: Stopping server: protocol=HTTP address=127.0.0.1:30205 network=tcp === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:15.470Z [INFO] TestAgentConnectAuthorize_allow: Waiting for endpoints to shut down === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:15.470Z [INFO] TestAgentConnectAuthorize_idNotService: Waiting for endpoints to shut down === CONT TestAgentConnectAuthorize_allow writer.go:29: 2021-01-29T19:34:15.471Z [INFO] TestAgentConnectAuthorize_allow: Endpoints down --- PASS: TestAgentConnectAuthorize_allow (1.04s) === CONT TestAgentConnectAuthorize_noTarget === CONT TestAgentConnectAuthorize_idNotService writer.go:29: 2021-01-29T19:34:15.472Z [INFO] TestAgentConnectAuthorize_idNotService: Endpoints down --- PASS: TestAgentConnectAuthorize_idNotService (0.91s) === CONT TestAgentConnectAuthorize_badBody === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.480Z [WARN] TestAgentConnectAuthorize_noTarget: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:15.481Z [DEBUG] TestAgentConnectAuthorize_noTarget.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:15.494Z [DEBUG] TestAgentConnectAuthorize_noTarget.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.513Z [WARN] TestAgentConnectAuthorize_badBody: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:15.515Z [DEBUG] TestAgentConnectAuthorize_badBody.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:15.517Z [DEBUG] TestAgentConnectAuthorize_badBody.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.526Z [INFO] TestAgentConnectAuthorize_noTarget.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a8190e33-304d-f6df-1f75-3c820f7f469c Address:127.0.0.1:30233}]" writer.go:29: 2021-01-29T19:34:15.529Z [INFO] TestAgentConnectAuthorize_noTarget.server.serf.wan: serf: EventMemberJoin: Node-a8190e33-304d-f6df-1f75-3c820f7f469c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:15.533Z [INFO] TestAgentConnectAuthorize_noTarget.server.raft: entering follower state: follower="Node at 127.0.0.1:30233 [Follower]" leader= === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.548Z [INFO] TestAgentConnectAuthorize_badBody.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9b1c9bdb-2109-9d29-a547-fb052a8ec197 Address:127.0.0.1:30227}]" writer.go:29: 2021-01-29T19:34:15.551Z [INFO] TestAgentConnectAuthorize_badBody.server.raft: entering follower state: follower="Node at 127.0.0.1:30227 [Follower]" leader= writer.go:29: 2021-01-29T19:34:15.551Z [INFO] TestAgentConnectAuthorize_badBody.server.serf.wan: serf: EventMemberJoin: Node-9b1c9bdb-2109-9d29-a547-fb052a8ec197.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:15.563Z [INFO] TestAgentConnectAuthorize_badBody.server.serf.lan: serf: EventMemberJoin: Node-9b1c9bdb-2109-9d29-a547-fb052a8ec197 127.0.0.1 writer.go:29: 2021-01-29T19:34:15.565Z [INFO] TestAgentConnectAuthorize_badBody.server: Adding LAN server: server="Node-9b1c9bdb-2109-9d29-a547-fb052a8ec197 (Addr: tcp/127.0.0.1:30227) (DC: dc1)" writer.go:29: 2021-01-29T19:34:15.566Z [INFO] TestAgentConnectAuthorize_badBody.server: Handled event for server in area: event=member-join server=Node-9b1c9bdb-2109-9d29-a547-fb052a8ec197.dc1 area=wan writer.go:29: 2021-01-29T19:34:15.569Z [INFO] TestAgentConnectAuthorize_badBody: Started DNS server: address=127.0.0.1:30222 network=tcp === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.570Z [INFO] TestAgentConnectAuthorize_noTarget.server.serf.lan: serf: EventMemberJoin: Node-a8190e33-304d-f6df-1f75-3c820f7f469c 127.0.0.1 writer.go:29: 2021-01-29T19:34:15.580Z [WARN] TestAgentConnectAuthorize_noTarget.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:15.580Z [INFO] TestAgentConnectAuthorize_noTarget.server.raft: entering candidate state: node="Node at 127.0.0.1:30233 [Candidate]" term=2 === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.573Z [INFO] TestAgentConnectAuthorize_badBody: Started DNS server: address=127.0.0.1:30222 network=udp writer.go:29: 2021-01-29T19:34:15.586Z [INFO] TestAgentConnectAuthorize_badBody: Started HTTP server: address=127.0.0.1:30223 network=tcp writer.go:29: 2021-01-29T19:34:15.587Z [INFO] TestAgentConnectAuthorize_badBody: started state syncer === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.584Z [INFO] TestAgentConnectAuthorize_noTarget.server: Handled event for server in area: event=member-join server=Node-a8190e33-304d-f6df-1f75-3c820f7f469c.dc1 area=wan === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.602Z [WARN] TestAgentConnectAuthorize_badBody.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:15.604Z [INFO] TestAgentConnectAuthorize_badBody.server.raft: entering candidate state: node="Node at 127.0.0.1:30227 [Candidate]" term=2 === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.584Z [INFO] TestAgentConnectAuthorize_noTarget.server: Adding LAN server: server="Node-a8190e33-304d-f6df-1f75-3c820f7f469c (Addr: tcp/127.0.0.1:30233) (DC: dc1)" === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.606Z [DEBUG] TestAgentConnectAuthorize_badBody.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:15.608Z [DEBUG] TestAgentConnectAuthorize_badBody.server.raft: vote granted: from=9b1c9bdb-2109-9d29-a547-fb052a8ec197 term=2 tally=1 === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.586Z [WARN] TestAgentConnectAuthorize_noTarget.server.raft: unable to get address for sever, using fallback address: id=a8190e33-304d-f6df-1f75-3c820f7f469c fallback=127.0.0.1:30233 error="Could not find address for server id a8190e33-304d-f6df-1f75-3c820f7f469c" === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.609Z [INFO] TestAgentConnectAuthorize_badBody.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:15.611Z [INFO] TestAgentConnectAuthorize_badBody.server.raft: entering leader state: leader="Node at 127.0.0.1:30227 [Leader]" writer.go:29: 2021-01-29T19:34:15.613Z [INFO] TestAgentConnectAuthorize_badBody.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:15.615Z [INFO] TestAgentConnectAuthorize_badBody.server: New leader elected: payload=Node-9b1c9bdb-2109-9d29-a547-fb052a8ec197 === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.586Z [INFO] TestAgentConnectAuthorize_noTarget: Started DNS server: address=127.0.0.1:30228 network=tcp writer.go:29: 2021-01-29T19:34:15.620Z [DEBUG] TestAgentConnectAuthorize_noTarget.server.raft: votes: needed=1 === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.631Z [DEBUG] TestAgentConnectAuthorize_badBody.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30227 === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.632Z [DEBUG] TestAgentConnectAuthorize_noTarget.server.raft: vote granted: from=a8190e33-304d-f6df-1f75-3c820f7f469c term=2 tally=1 === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.643Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.630Z [INFO] TestAgentConnectAuthorize_noTarget: Started DNS server: address=127.0.0.1:30228 network=udp writer.go:29: 2021-01-29T19:34:15.643Z [INFO] TestAgentConnectAuthorize_noTarget.server.raft: election won: tally=1 === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.648Z [INFO] TestAgentConnectAuthorize_badBody.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:15.649Z [INFO] TestAgentConnectAuthorize_badBody.leader: started routine: routine="CA root pruning" === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.646Z [INFO] TestAgentConnectAuthorize_noTarget.server.raft: entering leader state: leader="Node at 127.0.0.1:30233 [Leader]" === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.650Z [DEBUG] TestAgentConnectAuthorize_badBody.server: Skipping self join check for node since the cluster is too small: node=Node-9b1c9bdb-2109-9d29-a547-fb052a8ec197 === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.646Z [INFO] TestAgentConnectAuthorize_noTarget: Started HTTP server: address=127.0.0.1:30229 network=tcp === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.654Z [INFO] TestAgentConnectAuthorize_badBody.server: member joined, marking health alive: member=Node-9b1c9bdb-2109-9d29-a547-fb052a8ec197 === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:15.653Z [INFO] TestAgentConnectAuthorize_noTarget.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:15.654Z [INFO] TestAgentConnectAuthorize_noTarget.server: New leader elected: payload=Node-a8190e33-304d-f6df-1f75-3c820f7f469c writer.go:29: 2021-01-29T19:34:15.654Z [INFO] TestAgentConnectAuthorize_noTarget: started state syncer writer.go:29: 2021-01-29T19:34:15.656Z [DEBUG] TestAgentConnectAuthorize_noTarget.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30233 writer.go:29: 2021-01-29T19:34:15.657Z [INFO] TestAgentConnectAuthorize_noTarget: Synced node info writer.go:29: 2021-01-29T19:34:15.664Z [DEBUG] TestAgentConnectAuthorize_noTarget: Node info in sync writer.go:29: 2021-01-29T19:34:15.666Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:15.688Z [INFO] TestAgentConnectAuthorize_noTarget.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:15.708Z [INFO] TestAgentConnectAuthorize_noTarget.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.715Z [DEBUG] TestAgentConnectAuthorize_noTarget.server: Skipping self join check for node since the cluster is too small: node=Node-a8190e33-304d-f6df-1f75-3c820f7f469c writer.go:29: 2021-01-29T19:34:15.727Z [INFO] TestAgentConnectAuthorize_noTarget.server: member joined, marking health alive: member=Node-a8190e33-304d-f6df-1f75-3c820f7f469c writer.go:29: 2021-01-29T19:34:15.761Z [INFO] TestAgentConnectAuthorize_noTarget: Requesting shutdown writer.go:29: 2021-01-29T19:34:15.763Z [INFO] TestAgentConnectAuthorize_noTarget.server: shutting down server writer.go:29: 2021-01-29T19:34:15.764Z [DEBUG] TestAgentConnectAuthorize_noTarget.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.764Z [WARN] TestAgentConnectAuthorize_noTarget.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:15.764Z [DEBUG] TestAgentConnectAuthorize_noTarget.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.767Z [WARN] TestAgentConnectAuthorize_noTarget.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:15.770Z [INFO] TestAgentConnectAuthorize_noTarget.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:15.771Z [INFO] TestAgentConnectAuthorize_noTarget: consul server down writer.go:29: 2021-01-29T19:34:15.772Z [INFO] TestAgentConnectAuthorize_noTarget: shutdown complete writer.go:29: 2021-01-29T19:34:15.772Z [INFO] TestAgentConnectAuthorize_noTarget: Stopping server: protocol=DNS address=127.0.0.1:30228 network=tcp writer.go:29: 2021-01-29T19:34:15.774Z [INFO] TestAgentConnectAuthorize_noTarget: Stopping server: protocol=DNS address=127.0.0.1:30228 network=udp writer.go:29: 2021-01-29T19:34:15.774Z [INFO] TestAgentConnectAuthorize_noTarget: Stopping server: protocol=HTTP address=127.0.0.1:30229 network=tcp === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.798Z [DEBUG] TestAgentConnectAuthorize_badBody: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:15.807Z [INFO] TestAgentConnectAuthorize_badBody: Synced node info === CONT TestAgentConnectAuthorize_idInvalidFormat writer.go:29: 2021-01-29T19:34:15.810Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:15.811Z [INFO] TestAgentConnectAuthorize_idInvalidFormat: Endpoints down --- PASS: TestAgentConnectAuthorize_idInvalidFormat (1.16s) === CONT TestAgentConnectCALeafCert_secondaryDC_good === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.809Z [DEBUG] TestAgentConnectAuthorize_badBody: Node info in sync === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:15.819Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc1: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:15.820Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:15.821Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:15.828Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b5dd19e0-b11c-5447-7bfc-308fdb5df756 Address:127.0.0.1:30215}]" writer.go:29: 2021-01-29T19:34:15.830Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.raft: entering follower state: follower="Node at 127.0.0.1:30215 [Follower]" leader= writer.go:29: 2021-01-29T19:34:15.830Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.serf.wan: serf: EventMemberJoin: Node-b5dd19e0-b11c-5447-7bfc-308fdb5df756.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:15.834Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.serf.lan: serf: EventMemberJoin: Node-b5dd19e0-b11c-5447-7bfc-308fdb5df756 127.0.0.1 writer.go:29: 2021-01-29T19:34:15.837Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server: Adding LAN server: server="Node-b5dd19e0-b11c-5447-7bfc-308fdb5df756 (Addr: tcp/127.0.0.1:30215) (DC: dc1)" writer.go:29: 2021-01-29T19:34:15.837Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server: Handled event for server in area: event=member-join server=Node-b5dd19e0-b11c-5447-7bfc-308fdb5df756.dc1 area=wan writer.go:29: 2021-01-29T19:34:15.840Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Started DNS server: address=127.0.0.1:30210 network=tcp writer.go:29: 2021-01-29T19:34:15.846Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Started DNS server: address=127.0.0.1:30210 network=udp writer.go:29: 2021-01-29T19:34:15.849Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Started HTTP server: address=127.0.0.1:30211 network=tcp writer.go:29: 2021-01-29T19:34:15.850Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: started state syncer writer.go:29: 2021-01-29T19:34:15.868Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:15.869Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.raft: entering candidate state: node="Node at 127.0.0.1:30215 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:15.873Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:15.874Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.raft: vote granted: from=b5dd19e0-b11c-5447-7bfc-308fdb5df756 term=2 tally=1 writer.go:29: 2021-01-29T19:34:15.875Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:15.876Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.raft: entering leader state: leader="Node at 127.0.0.1:30215 [Leader]" writer.go:29: 2021-01-29T19:34:15.882Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:15.883Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server: New leader elected: payload=Node-b5dd19e0-b11c-5447-7bfc-308fdb5df756 writer.go:29: 2021-01-29T19:34:15.884Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30215 writer.go:29: 2021-01-29T19:34:15.888Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:15.893Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:15.894Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.895Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server: Skipping self join check for node since the cluster is too small: node=Node-b5dd19e0-b11c-5447-7bfc-308fdb5df756 writer.go:29: 2021-01-29T19:34:15.896Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server: member joined, marking health alive: member=Node-b5dd19e0-b11c-5447-7bfc-308fdb5df756 === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:15.904Z [INFO] TestAgentConnectAuthorize_badBody: Requesting shutdown writer.go:29: 2021-01-29T19:34:15.905Z [INFO] TestAgentConnectAuthorize_badBody.server: shutting down server writer.go:29: 2021-01-29T19:34:15.906Z [DEBUG] TestAgentConnectAuthorize_badBody.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.906Z [WARN] TestAgentConnectAuthorize_badBody.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:15.908Z [DEBUG] TestAgentConnectAuthorize_badBody.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.914Z [WARN] TestAgentConnectAuthorize_badBody.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:15.916Z [INFO] TestAgentConnectAuthorize_badBody.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:15.916Z [INFO] TestAgentConnectAuthorize_badBody: consul server down writer.go:29: 2021-01-29T19:34:15.919Z [INFO] TestAgentConnectAuthorize_badBody: shutdown complete writer.go:29: 2021-01-29T19:34:15.919Z [INFO] TestAgentConnectAuthorize_badBody: Stopping server: protocol=DNS address=127.0.0.1:30222 network=tcp writer.go:29: 2021-01-29T19:34:15.920Z [INFO] TestAgentConnectAuthorize_badBody: Stopping server: protocol=DNS address=127.0.0.1:30222 network=udp writer.go:29: 2021-01-29T19:34:15.921Z [INFO] TestAgentConnectAuthorize_badBody: Stopping server: protocol=HTTP address=127.0.0.1:30223 network=tcp === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:15.948Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:15.950Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Synced node info writer.go:29: 2021-01-29T19:34:15.952Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Node info in sync writer.go:29: 2021-01-29T19:34:15.982Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc2: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:15.983Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:15.985Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:15.988Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a2e1d71c-7edc-cbd3-a57d-9d3497079211 Address:127.0.0.1:30239}]" writer.go:29: 2021-01-29T19:34:15.989Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.raft: entering follower state: follower="Node at 127.0.0.1:30239 [Follower]" leader= writer.go:29: 2021-01-29T19:34:15.990Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.serf.wan: serf: EventMemberJoin: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 127.0.0.1 writer.go:29: 2021-01-29T19:34:15.993Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.serf.lan: serf: EventMemberJoin: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211 127.0.0.1 writer.go:29: 2021-01-29T19:34:15.996Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server: Handled event for server in area: event=member-join server=Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 area=wan writer.go:29: 2021-01-29T19:34:15.996Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server: Adding LAN server: server="Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211 (Addr: tcp/127.0.0.1:30239) (DC: dc2)" writer.go:29: 2021-01-29T19:34:15.999Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Started DNS server: address=127.0.0.1:30234 network=tcp writer.go:29: 2021-01-29T19:34:16.001Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Started DNS server: address=127.0.0.1:30234 network=udp writer.go:29: 2021-01-29T19:34:16.002Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Started HTTP server: address=127.0.0.1:30235 network=tcp writer.go:29: 2021-01-29T19:34:16.003Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: started state syncer writer.go:29: 2021-01-29T19:34:16.060Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:16.061Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.raft: entering candidate state: node="Node at 127.0.0.1:30239 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:16.064Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:16.065Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.raft: vote granted: from=a2e1d71c-7edc-cbd3-a57d-9d3497079211 term=2 tally=1 writer.go:29: 2021-01-29T19:34:16.066Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:16.067Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.raft: entering leader state: leader="Node at 127.0.0.1:30239 [Leader]" writer.go:29: 2021-01-29T19:34:16.068Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:16.069Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server: New leader elected: payload=Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211 writer.go:29: 2021-01-29T19:34:16.070Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server: Cannot upgrade to new ACLs, servers in acl datacenter are not yet upgraded: ACLDatacenter=dc1 mode=1 found=false writer.go:29: 2021-01-29T19:34:16.075Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: primary datacenter is configured but unreachable - deferring initialization of the secondary datacenter CA writer.go:29: 2021-01-29T19:34:16.077Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: started routine: routine="config entry replication" writer.go:29: 2021-01-29T19:34:16.078Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: started routine: routine="secondary CA roots watch" writer.go:29: 2021-01-29T19:34:16.078Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ConfigEntry.ListAll writer.go:29: 2021-01-29T19:34:16.078Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: starting Connect CA root replication from primary datacenter: primary=dc1 writer.go:29: 2021-01-29T19:34:16.084Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ConnectCA.Roots writer.go:29: 2021-01-29T19:34:16.087Z [ERROR] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: CA root replication failed, will retry: routine="secondary CA roots watch" error="Error retrieving the primary datacenter's roots: No path to datacenter" writer.go:29: 2021-01-29T19:34:16.085Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: starting Connect intention replication from primary datacenter: primary=dc1 writer.go:29: 2021-01-29T19:34:16.085Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: started routine: routine="intention replication" writer.go:29: 2021-01-29T19:34:16.093Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: started routine: routine="secondary cert renew watch" writer.go:29: 2021-01-29T19:34:16.092Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=Intention.List writer.go:29: 2021-01-29T19:34:16.097Z [ERROR] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: error replicating intentions: routine="intention replication" error="No path to datacenter" writer.go:29: 2021-01-29T19:34:16.094Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:16.099Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server: Skipping self join check for node since the cluster is too small: node=Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211 writer.go:29: 2021-01-29T19:34:16.101Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server: member joined, marking health alive: member=Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211 writer.go:29: 2021-01-29T19:34:16.101Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Synced node info === CONT TestAgentConnectAuthorize_noTarget writer.go:29: 2021-01-29T19:34:16.275Z [INFO] TestAgentConnectAuthorize_noTarget: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:16.277Z [INFO] TestAgentConnectAuthorize_noTarget: Endpoints down --- PASS: TestAgentConnectAuthorize_noTarget (0.81s) === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.287Z [WARN] TestAgentConnectCALeafCert_aclServiceReadDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:16.288Z [WARN] TestAgentConnectCALeafCert_aclServiceReadDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:16.289Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:16.290Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:16.296Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1e0a0cdd-5f25-bd39-64e5-676059ddbab3 Address:127.0.0.1:30245}]" writer.go:29: 2021-01-29T19:34:16.298Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30245 [Follower]" leader= writer.go:29: 2021-01-29T19:34:16.299Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.serf.wan: serf: EventMemberJoin: Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:16.307Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.serf.lan: serf: EventMemberJoin: Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3 127.0.0.1 writer.go:29: 2021-01-29T19:34:16.311Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Handled event for server in area: event=member-join server=Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3.dc1 area=wan writer.go:29: 2021-01-29T19:34:16.311Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Adding LAN server: server="Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3 (Addr: tcp/127.0.0.1:30245) (DC: dc1)" writer.go:29: 2021-01-29T19:34:16.312Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Started DNS server: address=127.0.0.1:30240 network=tcp writer.go:29: 2021-01-29T19:34:16.314Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Started DNS server: address=127.0.0.1:30240 network=udp writer.go:29: 2021-01-29T19:34:16.315Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Started HTTP server: address=127.0.0.1:30241 network=tcp writer.go:29: 2021-01-29T19:34:16.316Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: started state syncer writer.go:29: 2021-01-29T19:34:16.361Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:16.366Z [WARN] TestAgentConnectCALeafCert_aclServiceReadDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:16.368Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30245 [Candidate]" term=2 === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.364Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: (WAN) joining: wan_addresses=[127.0.0.1:30214] === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.371Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:16.373Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.server.raft: vote granted: from=1e0a0cdd-5f25-bd39-64e5-676059ddbab3 term=2 tally=1 writer.go:29: 2021-01-29T19:34:16.374Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:16.376Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30245 [Leader]" writer.go:29: 2021-01-29T19:34:16.378Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:16.378Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: New leader elected: payload=Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3 === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.373Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.memberlist.wan: memberlist: Initiating push/pull sync with: 127.0.0.1:30214 === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.382Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: initializing acls === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.373Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.memberlist.wan: memberlist: Stream connection from=127.0.0.1:52188 === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.394Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:16.396Z [WARN] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.395Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.serf.wan: serf: EventMemberJoin: Node-b5dd19e0-b11c-5447-7bfc-308fdb5df756.dc1 127.0.0.1 === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.398Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Bootstrapped ACL master token from configuration === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.398Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server: Handled event for server in area: event=member-join server=Node-b5dd19e0-b11c-5447-7bfc-308fdb5df756.dc1 area=wan writer.go:29: 2021-01-29T19:34:16.400Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: (WAN) joined: number_of_nodes=1 === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.412Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Created ACL anonymous token from configuration === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.408Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.serf.wan: serf: EventMemberJoin: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 127.0.0.1 === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.414Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:16.415Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.leader: started routine: routine="acl token reaping" === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.414Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server: Handled event for server in area: event=member-join server=Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 area=wan === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.415Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.serf.lan: serf: EventMemberUpdate: Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3 writer.go:29: 2021-01-29T19:34:16.417Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.serf.wan: serf: EventMemberUpdate: Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3.dc1 === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.418Z [DEBUG] connect.ca.consul: consul CA provider configured: id=a7:23:52:b1:78:8b:dc:e7:e1:99:7f:c2:c7:cb:73:30:bd:66:31:62:6e:fb:71:4d:ca:53:89:b0:a6:2f:e4:0f is_primary=true === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.417Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Updating LAN server: server="Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3 (Addr: tcp/127.0.0.1:30245) (DC: dc1)" writer.go:29: 2021-01-29T19:34:16.419Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Handled event for server in area: event=member-update server=Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3.dc1 area=wan writer.go:29: 2021-01-29T19:34:16.419Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:16.422Z [INFO] TestAgentConnectAuthorize_badBody: Waiting for endpoints to shut down === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.422Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.connect: CA rotated to new root under provider: provider=consul === CONT TestAgentConnectAuthorize_badBody writer.go:29: 2021-01-29T19:34:16.423Z [INFO] TestAgentConnectAuthorize_badBody: Endpoints down --- PASS: TestAgentConnectAuthorize_badBody (0.95s) === CONT TestAgentConnectCALeafCert_aclServiceWrite === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.427Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:16.428Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:16.432Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Skipping self join check for node since the cluster is too small: node=Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3 === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.446Z [WARN] TestAgentConnectCALeafCert_aclServiceWrite: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:16.446Z [WARN] TestAgentConnectCALeafCert_aclServiceWrite: bootstrap = true: do not enable unless necessary === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.438Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: member joined, marking health alive: member=Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3 === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.447Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.tlsutil: Update: version=1 === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.449Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.server: Skipping self join check for node since the cluster is too small: node=Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3 === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.450Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.464Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:16.471Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.acl: dropping node from result due to ACLs: node=Node-1e0a0cdd-5f25-bd39-64e5-676059ddbab3 === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.457Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b29c1383-20ab-9149-b882-26676bc59020 Address:127.0.0.1:30257}]" === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.476Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Synced node info writer.go:29: 2021-01-29T19:34:16.479Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny: Node info in sync === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.475Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.raft: entering follower state: follower="Node at 127.0.0.1:30257 [Follower]" leader= === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.483Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Synced service: service=test-id === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.483Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.serf.wan: serf: EventMemberJoin: Node-b29c1383-20ab-9149-b882-26676bc59020.dc1 127.0.0.1 === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.485Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny: Check in sync: check=service:test-id === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.488Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.serf.lan: serf: EventMemberJoin: Node-b29c1383-20ab-9149-b882-26676bc59020 127.0.0.1 writer.go:29: 2021-01-29T19:34:16.490Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: Handled event for server in area: event=member-join server=Node-b29c1383-20ab-9149-b882-26676bc59020.dc1 area=wan writer.go:29: 2021-01-29T19:34:16.491Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Started DNS server: address=127.0.0.1:30252 network=tcp writer.go:29: 2021-01-29T19:34:16.492Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Started DNS server: address=127.0.0.1:30252 network=udp writer.go:29: 2021-01-29T19:34:16.494Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Started HTTP server: address=127.0.0.1:30253 network=tcp writer.go:29: 2021-01-29T19:34:16.495Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: started state syncer === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.494Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.serf.wan: serf: messageJoinType: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.491Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: Adding LAN server: server="Node-b29c1383-20ab-9149-b882-26676bc59020 (Addr: tcp/127.0.0.1:30257) (DC: dc1)" === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:16.499Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:16.500Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:16.506Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:16.511Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:16.512Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:16.512Z [WARN] TestAgentConnectCALeafCert_aclServiceReadDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:16.511Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:16.512Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:16.513Z [DEBUG] TestAgentConnectCALeafCert_aclServiceReadDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:16.514Z [WARN] TestAgentConnectCALeafCert_aclServiceReadDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:16.518Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:16.519Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: consul server down writer.go:29: 2021-01-29T19:34:16.520Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: shutdown complete writer.go:29: 2021-01-29T19:34:16.521Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Stopping server: protocol=DNS address=127.0.0.1:30240 network=tcp writer.go:29: 2021-01-29T19:34:16.522Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Stopping server: protocol=DNS address=127.0.0.1:30240 network=udp writer.go:29: 2021-01-29T19:34:16.523Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Stopping server: protocol=HTTP address=127.0.0.1:30241 network=tcp === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.550Z [WARN] TestAgentConnectCALeafCert_aclServiceWrite.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:16.551Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.raft: entering candidate state: node="Node at 127.0.0.1:30257 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:16.553Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:16.554Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:16.555Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.server.raft: vote granted: from=b29c1383-20ab-9149-b882-26676bc59020 term=2 tally=1 writer.go:29: 2021-01-29T19:34:16.556Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:16.557Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.raft: entering leader state: leader="Node at 127.0.0.1:30257 [Leader]" writer.go:29: 2021-01-29T19:34:16.558Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:16.558Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: New leader elected: payload=Node-b29c1383-20ab-9149-b882-26676bc59020 writer.go:29: 2021-01-29T19:34:16.560Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: initializing acls writer.go:29: 2021-01-29T19:34:16.563Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:16.563Z [WARN] TestAgentConnectCALeafCert_aclServiceWrite.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:16.565Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:16.570Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:16.588Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:16.590Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:16.592Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.serf.lan: serf: EventMemberUpdate: Node-b29c1383-20ab-9149-b882-26676bc59020 writer.go:29: 2021-01-29T19:34:16.594Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.serf.wan: serf: EventMemberUpdate: Node-b29c1383-20ab-9149-b882-26676bc59020.dc1 writer.go:29: 2021-01-29T19:34:16.594Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: Updating LAN server: server="Node-b29c1383-20ab-9149-b882-26676bc59020 (Addr: tcp/127.0.0.1:30257) (DC: dc1)" writer.go:29: 2021-01-29T19:34:16.596Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: Handled event for server in area: event=member-update server=Node-b29c1383-20ab-9149-b882-26676bc59020.dc1 area=wan writer.go:29: 2021-01-29T19:34:16.600Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:16.604Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:16.605Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:16.607Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.server: Skipping self join check for node since the cluster is too small: node=Node-b29c1383-20ab-9149-b882-26676bc59020 writer.go:29: 2021-01-29T19:34:16.608Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: member joined, marking health alive: member=Node-b29c1383-20ab-9149-b882-26676bc59020 writer.go:29: 2021-01-29T19:34:16.611Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.server: Skipping self join check for node since the cluster is too small: node=Node-b29c1383-20ab-9149-b882-26676bc59020 writer.go:29: 2021-01-29T19:34:16.752Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:16.755Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Synced node info writer.go:29: 2021-01-29T19:34:16.758Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite: Node info in sync === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.836Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.serf.wan: serf: messageJoinType: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.908Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.acl: dropping node from result due to ACLs: node=Node-b29c1383-20ab-9149-b882-26676bc59020 writer.go:29: 2021-01-29T19:34:16.913Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.acl: dropping node from result due to ACLs: node=Node-b29c1383-20ab-9149-b882-26676bc59020 writer.go:29: 2021-01-29T19:34:16.927Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite: Node info in sync writer.go:29: 2021-01-29T19:34:16.930Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Synced service: service=test-id writer.go:29: 2021-01-29T19:34:16.934Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite: Check in sync: check=service:test-id writer.go:29: 2021-01-29T19:34:16.950Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Requesting shutdown writer.go:29: 2021-01-29T19:34:16.952Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server: shutting down server writer.go:29: 2021-01-29T19:34:16.954Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:16.955Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:16.957Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:16.955Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:16.957Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:16.960Z [DEBUG] TestAgentConnectCALeafCert_aclServiceWrite.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:16.960Z [WARN] TestAgentConnectCALeafCert_aclServiceWrite.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:16.968Z [WARN] TestAgentConnectCALeafCert_aclServiceWrite.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:16.984Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:16.984Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: consul server down writer.go:29: 2021-01-29T19:34:16.988Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: shutdown complete writer.go:29: 2021-01-29T19:34:16.989Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Stopping server: protocol=DNS address=127.0.0.1:30252 network=tcp === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:16.993Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.serf.wan: serf: messageJoinType: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:16.991Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Stopping server: protocol=DNS address=127.0.0.1:30252 network=udp writer.go:29: 2021-01-29T19:34:16.995Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Stopping server: protocol=HTTP address=127.0.0.1:30253 network=tcp === CONT TestAgentConnectCALeafCert_aclServiceReadDeny writer.go:29: 2021-01-29T19:34:17.024Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:17.026Z [INFO] TestAgentConnectCALeafCert_aclServiceReadDeny: Endpoints down --- PASS: TestAgentConnectCALeafCert_aclServiceReadDeny (0.75s) === CONT TestAgentConnectCALeafCert_aclDefaultDeny writer.go:29: 2021-01-29T19:34:17.040Z [WARN] TestAgentConnectCALeafCert_aclDefaultDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:17.042Z [WARN] TestAgentConnectCALeafCert_aclDefaultDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:17.044Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:17.047Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:17.053Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8bbbbbdf-b596-2b11-d54c-8107c86e3e44 Address:127.0.0.1:30263}]" writer.go:29: 2021-01-29T19:34:17.056Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30263 [Follower]" leader= writer.go:29: 2021-01-29T19:34:17.056Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.serf.wan: serf: EventMemberJoin: Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:17.063Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.serf.lan: serf: EventMemberJoin: Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44 127.0.0.1 writer.go:29: 2021-01-29T19:34:17.069Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: Adding LAN server: server="Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44 (Addr: tcp/127.0.0.1:30263) (DC: dc1)" writer.go:29: 2021-01-29T19:34:17.071Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: Handled event for server in area: event=member-join server=Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44.dc1 area=wan writer.go:29: 2021-01-29T19:34:17.093Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Started DNS server: address=127.0.0.1:30258 network=udp writer.go:29: 2021-01-29T19:34:17.096Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Started DNS server: address=127.0.0.1:30258 network=tcp writer.go:29: 2021-01-29T19:34:17.099Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Started HTTP server: address=127.0.0.1:30259 network=tcp === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:17.102Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:17.104Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Node info in sync === CONT TestAgentConnectCALeafCert_aclDefaultDeny writer.go:29: 2021-01-29T19:34:17.101Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: started state syncer writer.go:29: 2021-01-29T19:34:17.124Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:17.125Z [WARN] TestAgentConnectCALeafCert_aclDefaultDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:17.131Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30263 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:17.136Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:17.138Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.server.raft: vote granted: from=8bbbbbdf-b596-2b11-d54c-8107c86e3e44 term=2 tally=1 writer.go:29: 2021-01-29T19:34:17.141Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:17.143Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30263 [Leader]" writer.go:29: 2021-01-29T19:34:17.145Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:17.145Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: New leader elected: payload=Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44 writer.go:29: 2021-01-29T19:34:17.146Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: initializing acls === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:17.142Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentConnectCALeafCert_aclDefaultDeny writer.go:29: 2021-01-29T19:34:17.152Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:17.153Z [WARN] TestAgentConnectCALeafCert_aclDefaultDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:17.156Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.replication.config_entry: finished fetching config entries: amount=0 writer.go:29: 2021-01-29T19:34:17.157Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.replication.config_entry: Config Entry replication: local=0 remote=0 writer.go:29: 2021-01-29T19:34:17.158Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.replication.config_entry: Config Entry replication: deletions=0 updates=0 writer.go:29: 2021-01-29T19:34:17.159Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.replication.config_entry: replication completed through remote index: index=1 === CONT TestAgentConnectCALeafCert_aclDefaultDeny writer.go:29: 2021-01-29T19:34:17.166Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:17.168Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:17.169Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:17.170Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:17.171Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.serf.lan: serf: EventMemberUpdate: Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44 writer.go:29: 2021-01-29T19:34:17.172Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.serf.wan: serf: EventMemberUpdate: Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44.dc1 writer.go:29: 2021-01-29T19:34:17.174Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: Updating LAN server: server="Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44 (Addr: tcp/127.0.0.1:30263) (DC: dc1)" writer.go:29: 2021-01-29T19:34:17.174Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: Handled event for server in area: event=member-update server=Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44.dc1 area=wan writer.go:29: 2021-01-29T19:34:17.200Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:17.205Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:17.208Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:17.208Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.server: Skipping self join check for node since the cluster is too small: node=Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44 writer.go:29: 2021-01-29T19:34:17.209Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: member joined, marking health alive: member=Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44 writer.go:29: 2021-01-29T19:34:17.212Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.server: Skipping self join check for node since the cluster is too small: node=Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44 === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:17.215Z [ERROR] TestAgent_StartStop.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" writer.go:29: 2021-01-29T19:34:17.322Z [INFO] TestAgent_StartStop.server: Waiting to drain RPC traffic: drain_time=5s === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:17.333Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.serf.wan: serf: messageJoinType: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 === CONT TestAgentConnectCALeafCert_aclDefaultDeny writer.go:29: 2021-01-29T19:34:17.409Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.acl: dropping node from result due to ACLs: node=Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44 writer.go:29: 2021-01-29T19:34:17.412Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.acl: dropping node from result due to ACLs: node=Node-8bbbbbdf-b596-2b11-d54c-8107c86e3e44 writer.go:29: 2021-01-29T19:34:17.430Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Synced node info writer.go:29: 2021-01-29T19:34:17.433Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Synced service: service=test-id writer.go:29: 2021-01-29T19:34:17.435Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny: Check in sync: check=service:test-id writer.go:29: 2021-01-29T19:34:17.447Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:17.449Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:17.450Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:17.451Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:17.452Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:17.453Z [WARN] TestAgentConnectCALeafCert_aclDefaultDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:17.451Z [ERROR] TestAgentConnectCALeafCert_aclDefaultDeny.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:17.453Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:17.451Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:17.452Z [DEBUG] TestAgentConnectCALeafCert_aclDefaultDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:17.455Z [WARN] TestAgentConnectCALeafCert_aclDefaultDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:17.473Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:17.473Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: consul server down writer.go:29: 2021-01-29T19:34:17.477Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: shutdown complete writer.go:29: 2021-01-29T19:34:17.478Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Stopping server: protocol=DNS address=127.0.0.1:30258 network=tcp writer.go:29: 2021-01-29T19:34:17.480Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Stopping server: protocol=DNS address=127.0.0.1:30258 network=udp writer.go:29: 2021-01-29T19:34:17.482Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Stopping server: protocol=HTTP address=127.0.0.1:30259 network=tcp === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:17.493Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.serf.wan: serf: messageJoinType: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 === CONT TestAgentConnectCALeafCert_aclServiceWrite writer.go:29: 2021-01-29T19:34:17.497Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:17.498Z [INFO] TestAgentConnectCALeafCert_aclServiceWrite: Endpoints down --- PASS: TestAgentConnectCALeafCert_aclServiceWrite (1.08s) === CONT TestAgentConnectCARoots_list writer.go:29: 2021-01-29T19:34:17.516Z [WARN] TestAgentConnectCARoots_list: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:17.517Z [DEBUG] TestAgentConnectCARoots_list.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:17.519Z [DEBUG] TestAgentConnectCARoots_list.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:17.527Z [INFO] TestAgentConnectCARoots_list.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2e20ae53-1637-30f9-9fb4-9c3b8217d8a0 Address:127.0.0.1:30269}]" writer.go:29: 2021-01-29T19:34:17.532Z [INFO] TestAgentConnectCARoots_list.server.raft: entering follower state: follower="Node at 127.0.0.1:30269 [Follower]" leader= writer.go:29: 2021-01-29T19:34:17.537Z [INFO] TestAgentConnectCARoots_list.server.serf.wan: serf: EventMemberJoin: Node-2e20ae53-1637-30f9-9fb4-9c3b8217d8a0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:17.541Z [INFO] TestAgentConnectCARoots_list.server.serf.lan: serf: EventMemberJoin: Node-2e20ae53-1637-30f9-9fb4-9c3b8217d8a0 127.0.0.1 writer.go:29: 2021-01-29T19:34:17.545Z [INFO] TestAgentConnectCARoots_list.server: Handled event for server in area: event=member-join server=Node-2e20ae53-1637-30f9-9fb4-9c3b8217d8a0.dc1 area=wan writer.go:29: 2021-01-29T19:34:17.546Z [INFO] TestAgentConnectCARoots_list.server: Adding LAN server: server="Node-2e20ae53-1637-30f9-9fb4-9c3b8217d8a0 (Addr: tcp/127.0.0.1:30269) (DC: dc1)" writer.go:29: 2021-01-29T19:34:17.550Z [INFO] TestAgentConnectCARoots_list: Started DNS server: address=127.0.0.1:30264 network=tcp writer.go:29: 2021-01-29T19:34:17.554Z [INFO] TestAgentConnectCARoots_list: Started DNS server: address=127.0.0.1:30264 network=udp writer.go:29: 2021-01-29T19:34:17.570Z [INFO] TestAgentConnectCARoots_list: Started HTTP server: address=127.0.0.1:30265 network=tcp writer.go:29: 2021-01-29T19:34:17.572Z [INFO] TestAgentConnectCARoots_list: started state syncer writer.go:29: 2021-01-29T19:34:17.577Z [WARN] TestAgentConnectCARoots_list.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:17.579Z [INFO] TestAgentConnectCARoots_list.server.raft: entering candidate state: node="Node at 127.0.0.1:30269 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:17.582Z [DEBUG] TestAgentConnectCARoots_list.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:17.583Z [DEBUG] TestAgentConnectCARoots_list.server.raft: vote granted: from=2e20ae53-1637-30f9-9fb4-9c3b8217d8a0 term=2 tally=1 writer.go:29: 2021-01-29T19:34:17.585Z [INFO] TestAgentConnectCARoots_list.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:17.587Z [INFO] TestAgentConnectCARoots_list.server.raft: entering leader state: leader="Node at 127.0.0.1:30269 [Leader]" writer.go:29: 2021-01-29T19:34:17.588Z [INFO] TestAgentConnectCARoots_list.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:17.590Z [INFO] TestAgentConnectCARoots_list.server: New leader elected: payload=Node-2e20ae53-1637-30f9-9fb4-9c3b8217d8a0 writer.go:29: 2021-01-29T19:34:17.591Z [DEBUG] TestAgentConnectCARoots_list.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30269 writer.go:29: 2021-01-29T19:34:17.596Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:17.622Z [INFO] TestAgentConnectCARoots_list.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:17.623Z [INFO] TestAgentConnectCARoots_list.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:17.624Z [DEBUG] TestAgentConnectCARoots_list.server: Skipping self join check for node since the cluster is too small: node=Node-2e20ae53-1637-30f9-9fb4-9c3b8217d8a0 writer.go:29: 2021-01-29T19:34:17.624Z [INFO] TestAgentConnectCARoots_list.server: member joined, marking health alive: member=Node-2e20ae53-1637-30f9-9fb4-9c3b8217d8a0 writer.go:29: 2021-01-29T19:34:17.759Z [DEBUG] TestAgentConnectCARoots_list: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:17.764Z [INFO] TestAgentConnectCARoots_list: Synced node info === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:17.833Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.serf.wan: serf: messageJoinType: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 writer.go:29: 2021-01-29T19:34:17.888Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentConnectCARoots_list writer.go:29: 2021-01-29T19:34:17.901Z [DEBUG] connect.ca.consul: consul CA provider configured: id=b3:75:da:bd:9f:ba:40:74:76:2f:d3:5a:6b:4a:5c:d9:cd:c4:aa:8d:a8:b4:32:7b:e9:29:07:fa:ec:bd:13:0a is_primary=true writer.go:29: 2021-01-29T19:34:17.919Z [INFO] TestAgentConnectCARoots_list.server.connect: CA rotated to new root under provider: provider=consul writer.go:29: 2021-01-29T19:34:17.928Z [DEBUG] connect.ca.consul: consul CA provider configured: id=62:bb:b0:d3:ba:57:1d:ab:13:d6:5a:2f:63:03:c5:cc:58:ad:07:82:11:9c:f9:1e:bd:e1:98:3a:89:a1:b2:f9 is_primary=true writer.go:29: 2021-01-29T19:34:17.942Z [INFO] TestAgentConnectCARoots_list.server.connect: CA rotated to new root under provider: provider=consul writer.go:29: 2021-01-29T19:34:17.944Z [INFO] TestAgentConnectCARoots_list: Requesting shutdown writer.go:29: 2021-01-29T19:34:17.944Z [INFO] TestAgentConnectCARoots_list.server: shutting down server writer.go:29: 2021-01-29T19:34:17.946Z [DEBUG] TestAgentConnectCARoots_list.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:17.948Z [WARN] TestAgentConnectCARoots_list.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:17.950Z [DEBUG] TestAgentConnectCARoots_list.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:17.955Z [WARN] TestAgentConnectCARoots_list.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:17.967Z [INFO] TestAgentConnectCARoots_list.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:17.972Z [INFO] TestAgentConnectCARoots_list: consul server down writer.go:29: 2021-01-29T19:34:17.975Z [INFO] TestAgentConnectCARoots_list: shutdown complete writer.go:29: 2021-01-29T19:34:17.977Z [INFO] TestAgentConnectCARoots_list: Stopping server: protocol=DNS address=127.0.0.1:30264 network=tcp writer.go:29: 2021-01-29T19:34:17.979Z [INFO] TestAgentConnectCARoots_list: Stopping server: protocol=DNS address=127.0.0.1:30264 network=udp writer.go:29: 2021-01-29T19:34:17.981Z [INFO] TestAgentConnectCARoots_list: Stopping server: protocol=HTTP address=127.0.0.1:30265 network=tcp === CONT TestAgentConnectCALeafCert_aclDefaultDeny writer.go:29: 2021-01-29T19:34:17.984Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:17.986Z [INFO] TestAgentConnectCALeafCert_aclDefaultDeny: Endpoints down --- PASS: TestAgentConnectCALeafCert_aclDefaultDeny (0.96s) === CONT TestAgentConnectCARoots_empty === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:17.994Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.serf.wan: serf: messageJoinType: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 === CONT TestAgentConnectCARoots_empty writer.go:29: 2021-01-29T19:34:18.023Z [WARN] TestAgentConnectCARoots_empty: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:18.025Z [DEBUG] TestAgentConnectCARoots_empty.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:18.027Z [DEBUG] TestAgentConnectCARoots_empty.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:18.046Z [INFO] TestAgentConnectCARoots_empty.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:62bcc901-bf2d-174e-fd92-5e75cbe15efa Address:127.0.0.1:30275}]" writer.go:29: 2021-01-29T19:34:18.047Z [INFO] TestAgentConnectCARoots_empty.server.serf.wan: serf: EventMemberJoin: Node-62bcc901-bf2d-174e-fd92-5e75cbe15efa.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.048Z [INFO] TestAgentConnectCARoots_empty.server.raft: entering follower state: follower="Node at 127.0.0.1:30275 [Follower]" leader= writer.go:29: 2021-01-29T19:34:18.050Z [INFO] TestAgentConnectCARoots_empty.server.serf.lan: serf: EventMemberJoin: Node-62bcc901-bf2d-174e-fd92-5e75cbe15efa 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.056Z [INFO] TestAgentConnectCARoots_empty.server: Handled event for server in area: event=member-join server=Node-62bcc901-bf2d-174e-fd92-5e75cbe15efa.dc1 area=wan writer.go:29: 2021-01-29T19:34:18.056Z [INFO] TestAgentConnectCARoots_empty.server: Adding LAN server: server="Node-62bcc901-bf2d-174e-fd92-5e75cbe15efa (Addr: tcp/127.0.0.1:30275) (DC: dc1)" writer.go:29: 2021-01-29T19:34:18.065Z [INFO] TestAgentConnectCARoots_empty: Started DNS server: address=127.0.0.1:30270 network=udp writer.go:29: 2021-01-29T19:34:18.066Z [INFO] TestAgentConnectCARoots_empty: Started DNS server: address=127.0.0.1:30270 network=tcp writer.go:29: 2021-01-29T19:34:18.070Z [INFO] TestAgentConnectCARoots_empty: Started HTTP server: address=127.0.0.1:30271 network=tcp writer.go:29: 2021-01-29T19:34:18.075Z [INFO] TestAgentConnectCARoots_empty: started state syncer === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:18.075Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:18.095Z [DEBUG] connect.ca.consul: consul CA provider configured: id=ad:4a:c6:ab:ef:63:c9:60:1a:51:7f:19:62:e3:e9:d9:0e:76:55:10:6e:74:24:69:28:a1:6c:b8:b9:8f:fd:89 is_primary=false writer.go:29: 2021-01-29T19:34:18.106Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: received new intermediate certificate from primary datacenter writer.go:29: 2021-01-29T19:34:18.109Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: updated root certificates from primary datacenter === CONT TestAgentConnectCARoots_empty writer.go:29: 2021-01-29T19:34:18.111Z [WARN] TestAgentConnectCARoots_empty.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:18.111Z [INFO] TestAgentConnectCARoots_empty.server.raft: entering candidate state: node="Node at 127.0.0.1:30275 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:18.113Z [DEBUG] TestAgentConnectCARoots_empty.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:18.114Z [DEBUG] TestAgentConnectCARoots_empty.server.raft: vote granted: from=62bcc901-bf2d-174e-fd92-5e75cbe15efa term=2 tally=1 writer.go:29: 2021-01-29T19:34:18.115Z [INFO] TestAgentConnectCARoots_empty.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:18.116Z [INFO] TestAgentConnectCARoots_empty.server.raft: entering leader state: leader="Node at 127.0.0.1:30275 [Leader]" writer.go:29: 2021-01-29T19:34:18.117Z [INFO] TestAgentConnectCARoots_empty.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:18.117Z [INFO] TestAgentConnectCARoots_empty.server: New leader elected: payload=Node-62bcc901-bf2d-174e-fd92-5e75cbe15efa writer.go:29: 2021-01-29T19:34:18.119Z [DEBUG] TestAgentConnectCARoots_empty.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30275 writer.go:29: 2021-01-29T19:34:18.127Z [INFO] TestAgentConnectCARoots_empty.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:18.128Z [DEBUG] TestAgentConnectCARoots_empty.server: Skipping self join check for node since the cluster is too small: node=Node-62bcc901-bf2d-174e-fd92-5e75cbe15efa writer.go:29: 2021-01-29T19:34:18.129Z [INFO] TestAgentConnectCARoots_empty.server: member joined, marking health alive: member=Node-62bcc901-bf2d-174e-fd92-5e75cbe15efa === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:18.133Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Node info in sync writer.go:29: 2021-01-29T19:34:18.134Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Synced service: service=foo writer.go:29: 2021-01-29T19:34:18.135Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Check in sync: check=service:foo === CONT TestAgentConnectCARoots_empty writer.go:29: 2021-01-29T19:34:18.166Z [DEBUG] TestAgentConnectCARoots_empty: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:18.168Z [INFO] TestAgentConnectCARoots_empty: Synced node info writer.go:29: 2021-01-29T19:34:18.185Z [INFO] TestAgentConnectCARoots_empty: Requesting shutdown writer.go:29: 2021-01-29T19:34:18.186Z [INFO] TestAgentConnectCARoots_empty.server: shutting down server writer.go:29: 2021-01-29T19:34:18.187Z [DEBUG] TestAgentConnectCARoots_empty.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:18.188Z [WARN] TestAgentConnectCARoots_empty.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:18.188Z [DEBUG] TestAgentConnectCARoots_empty.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:18.192Z [WARN] TestAgentConnectCARoots_empty.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:18.195Z [INFO] TestAgentConnectCARoots_empty.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:18.196Z [INFO] TestAgentConnectCARoots_empty: consul server down writer.go:29: 2021-01-29T19:34:18.197Z [INFO] TestAgentConnectCARoots_empty: shutdown complete writer.go:29: 2021-01-29T19:34:18.198Z [INFO] TestAgentConnectCARoots_empty: Stopping server: protocol=DNS address=127.0.0.1:30270 network=tcp writer.go:29: 2021-01-29T19:34:18.199Z [INFO] TestAgentConnectCARoots_empty: Stopping server: protocol=DNS address=127.0.0.1:30270 network=udp writer.go:29: 2021-01-29T19:34:18.201Z [INFO] TestAgentConnectCARoots_empty: Stopping server: protocol=HTTP address=127.0.0.1:30271 network=tcp === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:18.333Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.serf.wan: serf: messageJoinType: Node-a2e1d71c-7edc-cbd3-a57d-9d3497079211.dc2 writer.go:29: 2021-01-29T19:34:18.348Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:18.351Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Synced node info writer.go:29: 2021-01-29T19:34:18.353Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Service in sync: service=foo writer.go:29: 2021-01-29T19:34:18.354Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Check in sync: check=service:foo writer.go:29: 2021-01-29T19:34:18.355Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Node info in sync writer.go:29: 2021-01-29T19:34:18.356Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Service in sync: service=foo writer.go:29: 2021-01-29T19:34:18.357Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Check in sync: check=service:foo === CONT TestAgentConnectCARoots_list writer.go:29: 2021-01-29T19:34:18.482Z [INFO] TestAgentConnectCARoots_list: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:18.484Z [INFO] TestAgentConnectCARoots_list: Endpoints down --- PASS: TestAgentConnectCARoots_list (0.99s) === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:18.569Z [WARN] TestAgent_Token: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:18.570Z [WARN] TestAgent_Token: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:18.572Z [DEBUG] TestAgent_Token.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:18.575Z [DEBUG] TestAgent_Token.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:18.598Z [INFO] TestAgent_Token.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3af00ac6-d252-b4ef-34d9-ff9205043064 Address:127.0.0.1:30287}]" writer.go:29: 2021-01-29T19:34:18.607Z [INFO] TestAgent_Token.server.raft: entering follower state: follower="Node at 127.0.0.1:30287 [Follower]" leader= writer.go:29: 2021-01-29T19:34:18.609Z [INFO] TestAgent_Token.server.serf.wan: serf: EventMemberJoin: Node-3af00ac6-d252-b4ef-34d9-ff9205043064.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.646Z [INFO] TestAgent_Token.server.serf.lan: serf: EventMemberJoin: Node-3af00ac6-d252-b4ef-34d9-ff9205043064 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.653Z [INFO] TestAgent_Token: Started DNS server: address=127.0.0.1:30282 network=udp writer.go:29: 2021-01-29T19:34:18.660Z [INFO] TestAgent_Token: Started DNS server: address=127.0.0.1:30282 network=tcp writer.go:29: 2021-01-29T19:34:18.666Z [INFO] TestAgent_Token: Started HTTP server: address=127.0.0.1:30283 network=tcp writer.go:29: 2021-01-29T19:34:18.657Z [INFO] TestAgent_Token.server: Handled event for server in area: event=member-join server=Node-3af00ac6-d252-b4ef-34d9-ff9205043064.dc1 area=wan writer.go:29: 2021-01-29T19:34:18.657Z [INFO] TestAgent_Token.server: Adding LAN server: server="Node-3af00ac6-d252-b4ef-34d9-ff9205043064 (Addr: tcp/127.0.0.1:30287) (DC: dc1)" writer.go:29: 2021-01-29T19:34:18.690Z [INFO] TestAgent_Token: started state syncer writer.go:29: 2021-01-29T19:34:18.691Z [WARN] TestAgent_Token.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:18.696Z [INFO] TestAgent_Token.server.raft: entering candidate state: node="Node at 127.0.0.1:30287 [Candidate]" term=2 === CONT TestAgentConnectCARoots_empty writer.go:29: 2021-01-29T19:34:18.704Z [INFO] TestAgentConnectCARoots_empty: Waiting for endpoints to shut down === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:18.713Z [DEBUG] TestAgent_Token.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAgentConnectCARoots_empty writer.go:29: 2021-01-29T19:34:18.712Z [INFO] TestAgentConnectCARoots_empty: Endpoints down --- PASS: TestAgentConnectCARoots_empty (0.73s) === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:18.717Z [DEBUG] TestAgent_Token.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:18.718Z [DEBUG] TestAgent_Token.server.raft: vote granted: from=3af00ac6-d252-b4ef-34d9-ff9205043064 term=2 tally=1 writer.go:29: 2021-01-29T19:34:18.719Z [INFO] TestAgent_Token.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:18.720Z [INFO] TestAgent_Token.server.raft: entering leader state: leader="Node at 127.0.0.1:30287 [Leader]" === CONT TestAgent_TokenTriggersFullSync === RUN TestAgent_TokenTriggersFullSync/acl_agent_token === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:18.721Z [INFO] TestAgent_Token.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:18.724Z [INFO] TestAgent_Token.server: New leader elected: payload=Node-3af00ac6-d252-b4ef-34d9-ff9205043064 writer.go:29: 2021-01-29T19:34:18.736Z [INFO] TestAgent_Token.server: initializing acls writer.go:29: 2021-01-29T19:34:18.751Z [INFO] TestAgent_Token.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:18.752Z [WARN] TestAgent_Token.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:18.756Z [INFO] TestAgent_Token.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:18.759Z [INFO] TestAgent_Token.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:18.772Z [INFO] TestAgent_Token.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:18.773Z [INFO] TestAgent_Token.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:18.776Z [INFO] TestAgent_Token.server.serf.lan: serf: EventMemberUpdate: Node-3af00ac6-d252-b4ef-34d9-ff9205043064 writer.go:29: 2021-01-29T19:34:18.778Z [INFO] TestAgent_Token.server.serf.wan: serf: EventMemberUpdate: Node-3af00ac6-d252-b4ef-34d9-ff9205043064.dc1 writer.go:29: 2021-01-29T19:34:18.782Z [INFO] TestAgent_Token.server: Updating LAN server: server="Node-3af00ac6-d252-b4ef-34d9-ff9205043064 (Addr: tcp/127.0.0.1:30287) (DC: dc1)" writer.go:29: 2021-01-29T19:34:18.783Z [INFO] TestAgent_Token.server: Handled event for server in area: event=member-update server=Node-3af00ac6-d252-b4ef-34d9-ff9205043064.dc1 area=wan writer.go:29: 2021-01-29T19:34:18.789Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:18.808Z [WARN] TestAgent_Token: Node info update blocked by ACLs: node=3af00ac6-d252-b4ef-34d9-ff9205043064 accessorID= writer.go:29: 2021-01-29T19:34:18.811Z [INFO] TestAgent_Token.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:18.813Z [INFO] TestAgent_Token.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:18.814Z [DEBUG] TestAgent_Token.server: Skipping self join check for node since the cluster is too small: node=Node-3af00ac6-d252-b4ef-34d9-ff9205043064 writer.go:29: 2021-01-29T19:34:18.816Z [INFO] TestAgent_Token.server: member joined, marking health alive: member=Node-3af00ac6-d252-b4ef-34d9-ff9205043064 writer.go:29: 2021-01-29T19:34:18.811Z [DEBUG] TestAgent_Token: Node info in sync writer.go:29: 2021-01-29T19:34:18.822Z [DEBUG] TestAgent_Token.server: Skipping self join check for node since the cluster is too small: node=Node-3af00ac6-d252-b4ef-34d9-ff9205043064 === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:18.875Z [WARN] TestAgent_TokenTriggersFullSync/acl_agent_token: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:18.876Z [WARN] TestAgent_TokenTriggersFullSync/acl_agent_token: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:18.878Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:18.881Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:18.930Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0de8a861-2a80-78e1-6df1-ea72e0f5a450 Address:127.0.0.1:30251}]" writer.go:29: 2021-01-29T19:34:18.938Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.raft: entering follower state: follower="Node at 127.0.0.1:30251 [Follower]" leader= === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:18.945Z [DEBUG] TestAgent_Token.acl: dropping node from result due to ACLs: node=Node-3af00ac6-d252-b4ef-34d9-ff9205043064 writer.go:29: 2021-01-29T19:34:18.957Z [DEBUG] TestAgent_Token.acl: dropping node from result due to ACLs: node=Node-3af00ac6-d252-b4ef-34d9-ff9205043064 === RUN TestAgent_Token/bad_token_name === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:18.961Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.serf.wan: serf: EventMemberJoin: Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.973Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.serf.lan: serf: EventMemberJoin: Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.977Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Started DNS server: address=127.0.0.1:30246 network=udp writer.go:29: 2021-01-29T19:34:18.980Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Handled event for server in area: event=member-join server=Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450.dc1 area=wan === RUN TestAgent_Token/bad_JSON === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:18.981Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Adding LAN server: server="Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 (Addr: tcp/127.0.0.1:30251) (DC: dc1)" writer.go:29: 2021-01-29T19:34:18.985Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Started DNS server: address=127.0.0.1:30246 network=tcp writer.go:29: 2021-01-29T19:34:18.989Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Started HTTP server: address=127.0.0.1:30247 network=tcp writer.go:29: 2021-01-29T19:34:18.992Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: started state syncer === RUN TestAgent_Token/set_user_legacy === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.001Z [INFO] TestAgent_Token: Updated agent's ACL token: token=acl_token === RUN TestAgent_Token/set_default === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.010Z [INFO] TestAgent_Token: Updated agent's ACL token: token=default === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.014Z [WARN] TestAgent_TokenTriggersFullSync/acl_agent_token.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:19.015Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.raft: entering candidate state: node="Node at 127.0.0.1:30251 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:19.018Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:19.019Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.server.raft: vote granted: from=0de8a861-2a80-78e1-6df1-ea72e0f5a450 term=2 tally=1 writer.go:29: 2021-01-29T19:34:19.019Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:19.020Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.raft: entering leader state: leader="Node at 127.0.0.1:30251 [Leader]" === RUN TestAgent_Token/set_agent_legacy === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.022Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: cluster leadership acquired === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.023Z [INFO] TestAgent_Token: Updated agent's ACL token: token=acl_agent_token === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.023Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: New leader elected: payload=Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 writer.go:29: 2021-01-29T19:34:19.030Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: initializing acls === RUN TestAgent_Token/set_agent === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.035Z [INFO] TestAgent_Token: Updated agent's ACL token: token=agent === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.034Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: initializing acls writer.go:29: 2021-01-29T19:34:19.038Z [WARN] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:19.035Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:19.041Z [WARN] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:19.041Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:19.044Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:19.046Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:19.049Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:19.051Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:34:19.051Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.serf.lan: serf: EventMemberUpdate: Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 === RUN TestAgent_Token/set_master_legacy === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.054Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Updating LAN server: server="Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 (Addr: tcp/127.0.0.1:30251) (DC: dc1)" === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.058Z [INFO] TestAgent_Token: Updated agent's ACL token: token=acl_agent_master_token === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.054Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.serf.lan: serf: EventMemberUpdate: Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 writer.go:29: 2021-01-29T19:34:19.054Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.serf.wan: serf: EventMemberUpdate: Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450.dc1 writer.go:29: 2021-01-29T19:34:19.059Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Updating LAN server: server="Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 (Addr: tcp/127.0.0.1:30251) (DC: dc1)" writer.go:29: 2021-01-29T19:34:19.062Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Handled event for server in area: event=member-update server=Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450.dc1 area=wan === RUN TestAgent_Token/set_master_ === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.062Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.serf.wan: serf: EventMemberUpdate: Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450.dc1 writer.go:29: 2021-01-29T19:34:19.067Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:19.069Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Handled event for server in area: event=member-update server=Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450.dc1 area=wan === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.071Z [INFO] TestAgent_Token: Updated agent's ACL token: token=agent_master === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.075Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:19.077Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:19.079Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Skipping self join check for node since the cluster is too small: node=Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 === RUN TestAgent_Token/set_repl_legacy === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.081Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: member joined, marking health alive: member=Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.084Z [INFO] TestAgent_Token: Updated agent's ACL token: token=acl_replication_token === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.085Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Skipping self join check for node since the cluster is too small: node=Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 writer.go:29: 2021-01-29T19:34:19.087Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.server: Skipping self join check for node since the cluster is too small: node=Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 === RUN TestAgent_Token/set_repl === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.097Z [INFO] TestAgent_Token: Updated agent's ACL token: token=replication === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.102Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.acl: dropping check from result due to ACLs: check=serfHealth writer.go:29: 2021-01-29T19:34:19.104Z [WARN] TestAgent_TokenTriggersFullSync/acl_agent_token: Node info update blocked by ACLs: node=0de8a861-2a80-78e1-6df1-ea72e0f5a450 accessorID= === RUN TestAgent_Token/clear_user_legacy === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.110Z [INFO] TestAgent_Token: Updated agent's ACL token: token=acl_token === RUN TestAgent_Token/clear_default === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.123Z [INFO] TestAgent_Token: Updated agent's ACL token: token=default === RUN TestAgent_Token/clear_agent_legacy === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.136Z [INFO] TestAgent_Token: Updated agent's ACL token: token=acl_agent_token === RUN TestAgent_Token/clear_agent === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.149Z [INFO] TestAgent_Token: Updated agent's ACL token: token=agent === RUN TestAgent_Token/clear_master_legacy === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.163Z [INFO] TestAgent_Token: Updated agent's ACL token: token=acl_agent_master_token === RUN TestAgent_Token/clear_master === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.183Z [INFO] TestAgent_Token: Updated agent's ACL token: token=agent_master === RUN TestAgent_Token/clear_repl_legacy === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.196Z [INFO] TestAgent_Token: Updated agent's ACL token: token=acl_replication_token === RUN TestAgent_Token/clear_repl === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.212Z [INFO] TestAgent_Token: Updated agent's ACL token: token=replication === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:19.215Z [ERROR] TestAgent_StartStop.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" === RUN TestAgent_Token/permission_denied === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.227Z [INFO] TestAgent_Token: Requesting shutdown writer.go:29: 2021-01-29T19:34:19.229Z [INFO] TestAgent_Token.server: shutting down server writer.go:29: 2021-01-29T19:34:19.230Z [DEBUG] TestAgent_Token.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:19.232Z [DEBUG] TestAgent_Token.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:19.234Z [DEBUG] TestAgent_Token.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:19.233Z [DEBUG] TestAgent_Token.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:19.234Z [DEBUG] TestAgent_Token.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:19.238Z [DEBUG] TestAgent_Token.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:19.238Z [WARN] TestAgent_Token.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:19.241Z [DEBUG] TestAgent_Token.acl: dropping check from result due to ACLs: check=serfHealth writer.go:29: 2021-01-29T19:34:19.250Z [WARN] TestAgent_Token: Node info update blocked by ACLs: node=3af00ac6-d252-b4ef-34d9-ff9205043064 accessorID= writer.go:29: 2021-01-29T19:34:19.252Z [WARN] TestAgent_Token.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:19.255Z [INFO] TestAgent_Token.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:19.255Z [INFO] TestAgent_Token: consul server down writer.go:29: 2021-01-29T19:34:19.258Z [INFO] TestAgent_Token: shutdown complete writer.go:29: 2021-01-29T19:34:19.259Z [INFO] TestAgent_Token: Stopping server: protocol=DNS address=127.0.0.1:30282 network=tcp writer.go:29: 2021-01-29T19:34:19.261Z [INFO] TestAgent_Token: Stopping server: protocol=DNS address=127.0.0.1:30282 network=udp writer.go:29: 2021-01-29T19:34:19.261Z [INFO] TestAgent_Token: Stopping server: protocol=HTTP address=127.0.0.1:30283 network=tcp === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:19.302Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.acl: dropping node from result due to ACLs: node=Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 writer.go:29: 2021-01-29T19:34:19.305Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.acl: dropping node from result due to ACLs: node=Node-0de8a861-2a80-78e1-6df1-ea72e0f5a450 writer.go:29: 2021-01-29T19:34:19.318Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Updated agent's ACL token: token=acl_agent_token === CONT TestAgent_Token writer.go:29: 2021-01-29T19:34:19.763Z [INFO] TestAgent_Token: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:19.766Z [INFO] TestAgent_Token: Endpoints down --- PASS: TestAgent_Token (1.28s) --- PASS: TestAgent_Token/bad_token_name (0.02s) --- PASS: TestAgent_Token/bad_JSON (0.01s) --- PASS: TestAgent_Token/set_user_legacy (0.01s) --- PASS: TestAgent_Token/set_default (0.01s) --- PASS: TestAgent_Token/set_agent_legacy (0.01s) --- PASS: TestAgent_Token/set_agent (0.02s) --- PASS: TestAgent_Token/set_master_legacy (0.01s) --- PASS: TestAgent_Token/set_master_ (0.01s) --- PASS: TestAgent_Token/set_repl_legacy (0.01s) --- PASS: TestAgent_Token/set_repl (0.01s) --- PASS: TestAgent_Token/clear_user_legacy (0.01s) --- PASS: TestAgent_Token/clear_default (0.01s) --- PASS: TestAgent_Token/clear_agent_legacy (0.01s) --- PASS: TestAgent_Token/clear_agent (0.01s) --- PASS: TestAgent_Token/clear_master_legacy (0.02s) --- PASS: TestAgent_Token/clear_master (0.01s) --- PASS: TestAgent_Token/clear_repl_legacy (0.02s) --- PASS: TestAgent_Token/clear_repl (0.01s) --- PASS: TestAgent_Token/permission_denied (0.00s) === CONT TestAgent_Monitor_ACLDeny writer.go:29: 2021-01-29T19:34:19.776Z [WARN] TestAgent_Monitor_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:19.777Z [WARN] TestAgent_Monitor_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:19.778Z [DEBUG] TestAgent_Monitor_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:19.779Z [DEBUG] TestAgent_Monitor_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:19.784Z [INFO] TestAgent_Monitor_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f37f30bf-880b-37f0-5382-095e3eaafe01 Address:127.0.0.1:30281}]" writer.go:29: 2021-01-29T19:34:19.788Z [INFO] TestAgent_Monitor_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30281 [Follower]" leader= writer.go:29: 2021-01-29T19:34:19.789Z [INFO] TestAgent_Monitor_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-f37f30bf-880b-37f0-5382-095e3eaafe01.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:19.793Z [INFO] TestAgent_Monitor_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-f37f30bf-880b-37f0-5382-095e3eaafe01 127.0.0.1 writer.go:29: 2021-01-29T19:34:19.797Z [INFO] TestAgent_Monitor_ACLDeny.server: Adding LAN server: server="Node-f37f30bf-880b-37f0-5382-095e3eaafe01 (Addr: tcp/127.0.0.1:30281) (DC: dc1)" writer.go:29: 2021-01-29T19:34:19.798Z [INFO] TestAgent_Monitor_ACLDeny.server: Handled event for server in area: event=member-join server=Node-f37f30bf-880b-37f0-5382-095e3eaafe01.dc1 area=wan writer.go:29: 2021-01-29T19:34:19.799Z [INFO] TestAgent_Monitor_ACLDeny: Started DNS server: address=127.0.0.1:30276 network=udp writer.go:29: 2021-01-29T19:34:19.801Z [INFO] TestAgent_Monitor_ACLDeny: Started DNS server: address=127.0.0.1:30276 network=tcp writer.go:29: 2021-01-29T19:34:19.805Z [INFO] TestAgent_Monitor_ACLDeny: Started HTTP server: address=127.0.0.1:30277 network=tcp writer.go:29: 2021-01-29T19:34:19.809Z [INFO] TestAgent_Monitor_ACLDeny: started state syncer writer.go:29: 2021-01-29T19:34:19.840Z [WARN] TestAgent_Monitor_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:19.841Z [INFO] TestAgent_Monitor_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30281 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:19.843Z [DEBUG] TestAgent_Monitor_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:19.845Z [DEBUG] TestAgent_Monitor_ACLDeny.server.raft: vote granted: from=f37f30bf-880b-37f0-5382-095e3eaafe01 term=2 tally=1 writer.go:29: 2021-01-29T19:34:19.846Z [INFO] TestAgent_Monitor_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:19.847Z [INFO] TestAgent_Monitor_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30281 [Leader]" writer.go:29: 2021-01-29T19:34:19.848Z [INFO] TestAgent_Monitor_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:19.850Z [INFO] TestAgent_Monitor_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:19.848Z [INFO] TestAgent_Monitor_ACLDeny.server: New leader elected: payload=Node-f37f30bf-880b-37f0-5382-095e3eaafe01 writer.go:29: 2021-01-29T19:34:19.849Z [INFO] TestAgent_Monitor_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:19.853Z [INFO] TestAgent_Monitor_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:19.857Z [WARN] TestAgent_Monitor_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:19.856Z [WARN] TestAgent_Monitor_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:19.860Z [INFO] TestAgent_Monitor_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:19.866Z [INFO] TestAgent_Monitor_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:19.868Z [INFO] TestAgent_Monitor_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:19.869Z [INFO] TestAgent_Monitor_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:19.871Z [INFO] TestAgent_Monitor_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-f37f30bf-880b-37f0-5382-095e3eaafe01 writer.go:29: 2021-01-29T19:34:19.872Z [INFO] TestAgent_Monitor_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:19.875Z [DEBUG] TestAgent_Monitor_ACLDeny.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:34:19.874Z [INFO] TestAgent_Monitor_ACLDeny.server: Updating LAN server: server="Node-f37f30bf-880b-37f0-5382-095e3eaafe01 (Addr: tcp/127.0.0.1:30281) (DC: dc1)" writer.go:29: 2021-01-29T19:34:19.875Z [INFO] TestAgent_Monitor_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-f37f30bf-880b-37f0-5382-095e3eaafe01.dc1 writer.go:29: 2021-01-29T19:34:19.877Z [INFO] TestAgent_Monitor_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-f37f30bf-880b-37f0-5382-095e3eaafe01 writer.go:29: 2021-01-29T19:34:19.880Z [INFO] TestAgent_Monitor_ACLDeny.server: Handled event for server in area: event=member-update server=Node-f37f30bf-880b-37f0-5382-095e3eaafe01.dc1 area=wan writer.go:29: 2021-01-29T19:34:19.882Z [INFO] TestAgent_Monitor_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-f37f30bf-880b-37f0-5382-095e3eaafe01.dc1 writer.go:29: 2021-01-29T19:34:19.884Z [INFO] TestAgent_Monitor_ACLDeny.server: Handled event for server in area: event=member-update server=Node-f37f30bf-880b-37f0-5382-095e3eaafe01.dc1 area=wan writer.go:29: 2021-01-29T19:34:19.882Z [INFO] TestAgent_Monitor_ACLDeny.server: Updating LAN server: server="Node-f37f30bf-880b-37f0-5382-095e3eaafe01 (Addr: tcp/127.0.0.1:30281) (DC: dc1)" writer.go:29: 2021-01-29T19:34:19.883Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:19.889Z [INFO] TestAgent_Monitor_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:19.890Z [INFO] TestAgent_Monitor_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:19.891Z [DEBUG] TestAgent_Monitor_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-f37f30bf-880b-37f0-5382-095e3eaafe01 writer.go:29: 2021-01-29T19:34:19.892Z [INFO] TestAgent_Monitor_ACLDeny.server: member joined, marking health alive: member=Node-f37f30bf-880b-37f0-5382-095e3eaafe01 writer.go:29: 2021-01-29T19:34:19.899Z [DEBUG] TestAgent_Monitor_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-f37f30bf-880b-37f0-5382-095e3eaafe01 writer.go:29: 2021-01-29T19:34:19.906Z [DEBUG] TestAgent_Monitor_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-f37f30bf-880b-37f0-5382-095e3eaafe01 writer.go:29: 2021-01-29T19:34:19.977Z [DEBUG] TestAgent_Monitor_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:19.979Z [INFO] TestAgent_Monitor_ACLDeny: Synced node info writer.go:29: 2021-01-29T19:34:19.979Z [DEBUG] TestAgent_Monitor_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:20.199Z [DEBUG] TestAgent_Monitor_ACLDeny.acl: dropping node from result due to ACLs: node=Node-f37f30bf-880b-37f0-5382-095e3eaafe01 writer.go:29: 2021-01-29T19:34:20.204Z [DEBUG] TestAgent_Monitor_ACLDeny.acl: dropping node from result due to ACLs: node=Node-f37f30bf-880b-37f0-5382-095e3eaafe01 writer.go:29: 2021-01-29T19:34:20.206Z [INFO] TestAgent_Monitor_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:20.209Z [INFO] TestAgent_Monitor_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:20.213Z [DEBUG] TestAgent_Monitor_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:20.216Z [DEBUG] TestAgent_Monitor_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:20.219Z [DEBUG] TestAgent_Monitor_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:20.216Z [DEBUG] TestAgent_Monitor_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:20.219Z [DEBUG] TestAgent_Monitor_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:20.223Z [WARN] TestAgent_Monitor_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:20.223Z [DEBUG] TestAgent_Monitor_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:20.226Z [DEBUG] TestAgent_Monitor_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:20.248Z [WARN] TestAgent_Monitor_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:20.258Z [DEBUG] TestAgent_Monitor_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:20.263Z [INFO] TestAgent_Monitor_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:20.263Z [INFO] TestAgent_Monitor_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:20.265Z [INFO] TestAgent_Monitor_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:20.268Z [INFO] TestAgent_Monitor_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30276 network=tcp writer.go:29: 2021-01-29T19:34:20.271Z [INFO] TestAgent_Monitor_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30276 network=udp writer.go:29: 2021-01-29T19:34:20.273Z [INFO] TestAgent_Monitor_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30277 network=tcp === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:20.381Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.acl: dropping service from result due to ACLs: service="{consul {}}" writer.go:29: 2021-01-29T19:34:20.383Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:20.385Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Synced node info writer.go:29: 2021-01-29T19:34:20.386Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token: Node info in sync writer.go:29: 2021-01-29T19:34:20.400Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Requesting shutdown writer.go:29: 2021-01-29T19:34:20.402Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server: shutting down server writer.go:29: 2021-01-29T19:34:20.403Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:20.404Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:20.406Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:20.405Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:20.406Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:20.409Z [WARN] TestAgent_TokenTriggersFullSync/acl_agent_token.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:20.409Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_agent_token.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:20.413Z [WARN] TestAgent_TokenTriggersFullSync/acl_agent_token.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:20.416Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:20.418Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: consul server down writer.go:29: 2021-01-29T19:34:20.419Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: shutdown complete writer.go:29: 2021-01-29T19:34:20.420Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Stopping server: protocol=DNS address=127.0.0.1:30246 network=tcp writer.go:29: 2021-01-29T19:34:20.421Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Stopping server: protocol=DNS address=127.0.0.1:30246 network=udp writer.go:29: 2021-01-29T19:34:20.423Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Stopping server: protocol=HTTP address=127.0.0.1:30247 network=tcp === CONT TestAgent_Monitor_ACLDeny writer.go:29: 2021-01-29T19:34:20.775Z [INFO] TestAgent_Monitor_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:20.777Z [INFO] TestAgent_Monitor_ACLDeny: Endpoints down --- PASS: TestAgent_Monitor_ACLDeny (1.01s) === CONT TestAgent_RegisterCheck_Service writer.go:29: 2021-01-29T19:34:20.798Z [WARN] TestAgent_RegisterCheck_Service: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:20.800Z [DEBUG] TestAgent_RegisterCheck_Service.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:20.804Z [DEBUG] TestAgent_RegisterCheck_Service.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:20.809Z [INFO] TestAgent_RegisterCheck_Service.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ef46d7e4-6b55-65d9-2724-f7afbf79a9c2 Address:127.0.0.1:30299}]" writer.go:29: 2021-01-29T19:34:20.812Z [INFO] TestAgent_RegisterCheck_Service.server.raft: entering follower state: follower="Node at 127.0.0.1:30299 [Follower]" leader= writer.go:29: 2021-01-29T19:34:20.816Z [INFO] TestAgent_RegisterCheck_Service.server.serf.wan: serf: EventMemberJoin: Node-ef46d7e4-6b55-65d9-2724-f7afbf79a9c2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:20.823Z [INFO] TestAgent_RegisterCheck_Service.server.serf.lan: serf: EventMemberJoin: Node-ef46d7e4-6b55-65d9-2724-f7afbf79a9c2 127.0.0.1 writer.go:29: 2021-01-29T19:34:20.831Z [INFO] TestAgent_RegisterCheck_Service.server: Adding LAN server: server="Node-ef46d7e4-6b55-65d9-2724-f7afbf79a9c2 (Addr: tcp/127.0.0.1:30299) (DC: dc1)" writer.go:29: 2021-01-29T19:34:20.832Z [INFO] TestAgent_RegisterCheck_Service.server: Handled event for server in area: event=member-join server=Node-ef46d7e4-6b55-65d9-2724-f7afbf79a9c2.dc1 area=wan writer.go:29: 2021-01-29T19:34:20.834Z [INFO] TestAgent_RegisterCheck_Service: Started DNS server: address=127.0.0.1:30294 network=udp writer.go:29: 2021-01-29T19:34:20.835Z [INFO] TestAgent_RegisterCheck_Service: Started DNS server: address=127.0.0.1:30294 network=tcp writer.go:29: 2021-01-29T19:34:20.836Z [INFO] TestAgent_RegisterCheck_Service: Started HTTP server: address=127.0.0.1:30295 network=tcp writer.go:29: 2021-01-29T19:34:20.838Z [INFO] TestAgent_RegisterCheck_Service: started state syncer writer.go:29: 2021-01-29T19:34:20.859Z [WARN] TestAgent_RegisterCheck_Service.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:20.860Z [INFO] TestAgent_RegisterCheck_Service.server.raft: entering candidate state: node="Node at 127.0.0.1:30299 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:20.861Z [DEBUG] TestAgent_RegisterCheck_Service.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:20.862Z [DEBUG] TestAgent_RegisterCheck_Service.server.raft: vote granted: from=ef46d7e4-6b55-65d9-2724-f7afbf79a9c2 term=2 tally=1 writer.go:29: 2021-01-29T19:34:20.863Z [INFO] TestAgent_RegisterCheck_Service.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:20.864Z [INFO] TestAgent_RegisterCheck_Service.server.raft: entering leader state: leader="Node at 127.0.0.1:30299 [Leader]" writer.go:29: 2021-01-29T19:34:20.865Z [INFO] TestAgent_RegisterCheck_Service.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:20.865Z [INFO] TestAgent_RegisterCheck_Service.server: New leader elected: payload=Node-ef46d7e4-6b55-65d9-2724-f7afbf79a9c2 writer.go:29: 2021-01-29T19:34:20.866Z [DEBUG] TestAgent_RegisterCheck_Service.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30299 writer.go:29: 2021-01-29T19:34:20.873Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:20.881Z [INFO] TestAgent_RegisterCheck_Service.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:20.882Z [INFO] TestAgent_RegisterCheck_Service.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:20.883Z [DEBUG] TestAgent_RegisterCheck_Service.server: Skipping self join check for node since the cluster is too small: node=Node-ef46d7e4-6b55-65d9-2724-f7afbf79a9c2 writer.go:29: 2021-01-29T19:34:20.883Z [INFO] TestAgent_RegisterCheck_Service.server: member joined, marking health alive: member=Node-ef46d7e4-6b55-65d9-2724-f7afbf79a9c2 === CONT TestAgent_TokenTriggersFullSync/acl_agent_token writer.go:29: 2021-01-29T19:34:20.925Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:20.927Z [INFO] TestAgent_TokenTriggersFullSync/acl_agent_token: Endpoints down === RUN TestAgent_TokenTriggersFullSync/agent writer.go:29: 2021-01-29T19:34:20.939Z [WARN] TestAgent_TokenTriggersFullSync/agent: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:20.940Z [WARN] TestAgent_TokenTriggersFullSync/agent: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:20.941Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:20.942Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:20.947Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cb1a7a18-41cd-11b6-a106-5e63123e4c5a Address:127.0.0.1:30305}]" writer.go:29: 2021-01-29T19:34:20.948Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.raft: entering follower state: follower="Node at 127.0.0.1:30305 [Follower]" leader= writer.go:29: 2021-01-29T19:34:20.948Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.serf.wan: serf: EventMemberJoin: Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:20.956Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.serf.lan: serf: EventMemberJoin: Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a 127.0.0.1 writer.go:29: 2021-01-29T19:34:20.960Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: Adding LAN server: server="Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a (Addr: tcp/127.0.0.1:30305) (DC: dc1)" writer.go:29: 2021-01-29T19:34:20.961Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: Handled event for server in area: event=member-join server=Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a.dc1 area=wan writer.go:29: 2021-01-29T19:34:20.961Z [INFO] TestAgent_TokenTriggersFullSync/agent: Started DNS server: address=127.0.0.1:30300 network=udp writer.go:29: 2021-01-29T19:34:20.963Z [INFO] TestAgent_TokenTriggersFullSync/agent: Started DNS server: address=127.0.0.1:30300 network=tcp writer.go:29: 2021-01-29T19:34:20.965Z [INFO] TestAgent_TokenTriggersFullSync/agent: Started HTTP server: address=127.0.0.1:30301 network=tcp writer.go:29: 2021-01-29T19:34:20.967Z [INFO] TestAgent_TokenTriggersFullSync/agent: started state syncer writer.go:29: 2021-01-29T19:34:20.992Z [WARN] TestAgent_TokenTriggersFullSync/agent.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:20.994Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.raft: entering candidate state: node="Node at 127.0.0.1:30305 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:20.996Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:20.998Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.server.raft: vote granted: from=cb1a7a18-41cd-11b6-a106-5e63123e4c5a term=2 tally=1 writer.go:29: 2021-01-29T19:34:20.999Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:21.000Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.raft: entering leader state: leader="Node at 127.0.0.1:30305 [Leader]" writer.go:29: 2021-01-29T19:34:21.001Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:21.002Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: New leader elected: payload=Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a writer.go:29: 2021-01-29T19:34:21.003Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: initializing acls writer.go:29: 2021-01-29T19:34:21.005Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:21.006Z [WARN] TestAgent_TokenTriggersFullSync/agent.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:21.007Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:21.009Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:21.010Z [INFO] TestAgent_TokenTriggersFullSync/agent.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:21.011Z [INFO] TestAgent_TokenTriggersFullSync/agent.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:21.012Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.serf.lan: serf: EventMemberUpdate: Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a writer.go:29: 2021-01-29T19:34:21.022Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: Updating LAN server: server="Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a (Addr: tcp/127.0.0.1:30305) (DC: dc1)" writer.go:29: 2021-01-29T19:34:21.022Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.serf.wan: serf: EventMemberUpdate: Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a.dc1 writer.go:29: 2021-01-29T19:34:21.026Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: Handled event for server in area: event=member-update server=Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a.dc1 area=wan writer.go:29: 2021-01-29T19:34:21.030Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:21.036Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:21.037Z [INFO] TestAgent_TokenTriggersFullSync/agent.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:21.040Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.server: Skipping self join check for node since the cluster is too small: node=Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a writer.go:29: 2021-01-29T19:34:21.041Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: member joined, marking health alive: member=Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a writer.go:29: 2021-01-29T19:34:21.048Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.server: Skipping self join check for node since the cluster is too small: node=Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a writer.go:29: 2021-01-29T19:34:21.098Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.acl: dropping check from result due to ACLs: check=serfHealth writer.go:29: 2021-01-29T19:34:21.101Z [WARN] TestAgent_TokenTriggersFullSync/agent: Node info update blocked by ACLs: node=cb1a7a18-41cd-11b6-a106-5e63123e4c5a accessorID= === CONT TestAgent_RegisterCheck_Service writer.go:29: 2021-01-29T19:34:21.208Z [DEBUG] TestAgent_RegisterCheck_Service: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:21.211Z [INFO] TestAgent_RegisterCheck_Service: Synced node info === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:21.215Z [ERROR] TestAgent_StartStop.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" writer.go:29: 2021-01-29T19:34:21.215Z [ERROR] TestAgent_StartStop.server.autopilot: Error promoting servers: error="error getting server raft protocol versions: No servers found" === CONT TestAgent_TokenTriggersFullSync/agent writer.go:29: 2021-01-29T19:34:21.222Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.acl: dropping node from result due to ACLs: node=Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a writer.go:29: 2021-01-29T19:34:21.226Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.acl: dropping node from result due to ACLs: node=Node-cb1a7a18-41cd-11b6-a106-5e63123e4c5a writer.go:29: 2021-01-29T19:34:21.242Z [INFO] TestAgent_TokenTriggersFullSync/agent: Updated agent's ACL token: token=agent === CONT TestAgent_RegisterCheck_Service writer.go:29: 2021-01-29T19:34:21.274Z [DEBUG] TestAgent_RegisterCheck_Service: Node info in sync writer.go:29: 2021-01-29T19:34:21.275Z [INFO] TestAgent_RegisterCheck_Service: Synced service: service=memcache writer.go:29: 2021-01-29T19:34:21.276Z [DEBUG] TestAgent_RegisterCheck_Service: Check in sync: check=service:memcache writer.go:29: 2021-01-29T19:34:21.278Z [DEBUG] TestAgent_RegisterCheck_Service: Node info in sync writer.go:29: 2021-01-29T19:34:21.279Z [DEBUG] TestAgent_RegisterCheck_Service: Service in sync: service=memcache writer.go:29: 2021-01-29T19:34:21.283Z [INFO] TestAgent_RegisterCheck_Service: Synced check: check=memcache_check2 writer.go:29: 2021-01-29T19:34:21.283Z [DEBUG] TestAgent_RegisterCheck_Service: Check in sync: check=service:memcache writer.go:29: 2021-01-29T19:34:21.284Z [INFO] TestAgent_RegisterCheck_Service: Requesting shutdown writer.go:29: 2021-01-29T19:34:21.285Z [INFO] TestAgent_RegisterCheck_Service.server: shutting down server writer.go:29: 2021-01-29T19:34:21.286Z [DEBUG] TestAgent_RegisterCheck_Service.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:21.287Z [WARN] TestAgent_RegisterCheck_Service.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:21.287Z [DEBUG] TestAgent_RegisterCheck_Service.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:21.289Z [WARN] TestAgent_RegisterCheck_Service.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:21.294Z [INFO] TestAgent_RegisterCheck_Service.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:21.294Z [INFO] TestAgent_RegisterCheck_Service: consul server down writer.go:29: 2021-01-29T19:34:21.296Z [INFO] TestAgent_RegisterCheck_Service: shutdown complete writer.go:29: 2021-01-29T19:34:21.296Z [INFO] TestAgent_RegisterCheck_Service: Stopping server: protocol=DNS address=127.0.0.1:30294 network=tcp writer.go:29: 2021-01-29T19:34:21.298Z [INFO] TestAgent_RegisterCheck_Service: Stopping server: protocol=DNS address=127.0.0.1:30294 network=udp writer.go:29: 2021-01-29T19:34:21.298Z [INFO] TestAgent_RegisterCheck_Service: Stopping server: protocol=HTTP address=127.0.0.1:30295 network=tcp 2021-01-29T19:34:21.732Z [ERROR] watch.watch: Watch errored: type=key error="Get "https://127.0.0.1:29774/v1/kv/asdf": dial tcp 127.0.0.1:29774: connect: connection refused" retry=45s writer.go:29: 2021-01-29T19:34:21.805Z [INFO] TestAgent_RegisterCheck_Service: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:21.806Z [INFO] TestAgent_RegisterCheck_Service: Endpoints down --- PASS: TestAgent_RegisterCheck_Service (1.03s) === CONT TestAgent_NodeMaintenance_ACLDeny writer.go:29: 2021-01-29T19:34:21.820Z [WARN] TestAgent_NodeMaintenance_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:21.821Z [WARN] TestAgent_NodeMaintenance_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:21.823Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:21.825Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:21.830Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bfca0fe9-7db4-cd4e-744b-bdcc59b853db Address:127.0.0.1:30221}]" writer.go:29: 2021-01-29T19:34:21.833Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30221 [Follower]" leader= writer.go:29: 2021-01-29T19:34:21.833Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:21.839Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db 127.0.0.1 writer.go:29: 2021-01-29T19:34:21.843Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: Adding LAN server: server="Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db (Addr: tcp/127.0.0.1:30221) (DC: dc1)" writer.go:29: 2021-01-29T19:34:21.843Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: Handled event for server in area: event=member-join server=Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db.dc1 area=wan writer.go:29: 2021-01-29T19:34:21.845Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Started DNS server: address=127.0.0.1:30216 network=tcp writer.go:29: 2021-01-29T19:34:21.847Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Started DNS server: address=127.0.0.1:30216 network=udp writer.go:29: 2021-01-29T19:34:21.849Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Started HTTP server: address=127.0.0.1:30217 network=tcp writer.go:29: 2021-01-29T19:34:21.849Z [INFO] TestAgent_NodeMaintenance_ACLDeny: started state syncer writer.go:29: 2021-01-29T19:34:21.895Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:21.904Z [WARN] TestAgent_NodeMaintenance_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:21.906Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30221 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:21.908Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:21.910Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.server.raft: vote granted: from=bfca0fe9-7db4-cd4e-744b-bdcc59b853db term=2 tally=1 writer.go:29: 2021-01-29T19:34:21.911Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:21.912Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30221 [Leader]" writer.go:29: 2021-01-29T19:34:21.914Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:21.916Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: New leader elected: payload=Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db writer.go:29: 2021-01-29T19:34:21.929Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:21.933Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:21.938Z [WARN] TestAgent_NodeMaintenance_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:21.941Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:21.955Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:21.962Z [INFO] TestAgent_NodeMaintenance_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:21.963Z [INFO] TestAgent_NodeMaintenance_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:21.964Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db writer.go:29: 2021-01-29T19:34:21.982Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db.dc1 writer.go:29: 2021-01-29T19:34:21.984Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: Updating LAN server: server="Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db (Addr: tcp/127.0.0.1:30221) (DC: dc1)" writer.go:29: 2021-01-29T19:34:21.986Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: Handled event for server in area: event=member-update server=Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db.dc1 area=wan writer.go:29: 2021-01-29T19:34:21.989Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:22.003Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:22.005Z [INFO] TestAgent_NodeMaintenance_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.007Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db writer.go:29: 2021-01-29T19:34:22.008Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: member joined, marking health alive: member=Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db writer.go:29: 2021-01-29T19:34:22.011Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db writer.go:29: 2021-01-29T19:34:22.087Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:22.090Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Synced node info writer.go:29: 2021-01-29T19:34:22.093Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:22.215Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.acl: dropping node from result due to ACLs: node=Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db writer.go:29: 2021-01-29T19:34:22.217Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.acl: dropping node from result due to ACLs: node=Node-bfca0fe9-7db4-cd4e-744b-bdcc59b853db === RUN TestAgent_NodeMaintenance_ACLDeny/no_token === RUN TestAgent_NodeMaintenance_ACLDeny/root_token === CONT TestAgent_NodeMaintenance_ACLDeny writer.go:29: 2021-01-29T19:34:22.222Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Node entered maintenance mode writer.go:29: 2021-01-29T19:34:22.223Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:22.225Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Synced check: check=_node_maintenance writer.go:29: 2021-01-29T19:34:22.226Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:22.228Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:22.229Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:22.230Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:22.230Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.231Z [WARN] TestAgent_NodeMaintenance_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:22.230Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:22.230Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:22.231Z [DEBUG] TestAgent_NodeMaintenance_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.234Z [WARN] TestAgent_NodeMaintenance_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:22.241Z [INFO] TestAgent_NodeMaintenance_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:22.242Z [INFO] TestAgent_NodeMaintenance_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:22.243Z [INFO] TestAgent_NodeMaintenance_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:22.244Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30216 network=tcp writer.go:29: 2021-01-29T19:34:22.245Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30216 network=udp writer.go:29: 2021-01-29T19:34:22.246Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30217 network=tcp === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:22.324Z [INFO] TestAgent_StartStop: Requesting shutdown writer.go:29: 2021-01-29T19:34:22.327Z [INFO] TestAgent_StartStop.server: shutting down server writer.go:29: 2021-01-29T19:34:22.328Z [DEBUG] TestAgent_StartStop.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.330Z [DEBUG] TestAgent_StartStop.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.335Z [INFO] TestAgent_StartStop: consul server down writer.go:29: 2021-01-29T19:34:22.336Z [INFO] TestAgent_StartStop: shutdown complete writer.go:29: 2021-01-29T19:34:22.337Z [INFO] TestAgent_StartStop: Stopping server: protocol=DNS address=127.0.0.1:30132 network=tcp writer.go:29: 2021-01-29T19:34:22.339Z [INFO] TestAgent_StartStop: Stopping server: protocol=DNS address=127.0.0.1:30132 network=udp writer.go:29: 2021-01-29T19:34:22.340Z [INFO] TestAgent_StartStop: Stopping server: protocol=HTTP address=127.0.0.1:30133 network=tcp === CONT TestAgent_NodeMaintenance_ACLDeny writer.go:29: 2021-01-29T19:34:22.747Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:22.749Z [INFO] TestAgent_NodeMaintenance_ACLDeny: Endpoints down --- PASS: TestAgent_NodeMaintenance_ACLDeny (0.94s) --- PASS: TestAgent_NodeMaintenance_ACLDeny/no_token (0.00s) --- PASS: TestAgent_NodeMaintenance_ACLDeny/root_token (0.01s) === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:22.760Z [WARN] TestAgent_NodeMaintenance_Disable: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:22.762Z [DEBUG] TestAgent_NodeMaintenance_Disable.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:22.764Z [DEBUG] TestAgent_NodeMaintenance_Disable.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:22.772Z [INFO] TestAgent_NodeMaintenance_Disable.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ac8cf621-bad7-8d39-3278-90e4363091a9 Address:127.0.0.1:30293}]" writer.go:29: 2021-01-29T19:34:22.774Z [INFO] TestAgent_NodeMaintenance_Disable.server.serf.wan: serf: EventMemberJoin: Node-ac8cf621-bad7-8d39-3278-90e4363091a9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.774Z [INFO] TestAgent_NodeMaintenance_Disable.server.raft: entering follower state: follower="Node at 127.0.0.1:30293 [Follower]" leader= writer.go:29: 2021-01-29T19:34:22.777Z [INFO] TestAgent_NodeMaintenance_Disable.server.serf.lan: serf: EventMemberJoin: Node-ac8cf621-bad7-8d39-3278-90e4363091a9 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.788Z [INFO] TestAgent_NodeMaintenance_Disable.server: Adding LAN server: server="Node-ac8cf621-bad7-8d39-3278-90e4363091a9 (Addr: tcp/127.0.0.1:30293) (DC: dc1)" writer.go:29: 2021-01-29T19:34:22.791Z [INFO] TestAgent_NodeMaintenance_Disable.server: Handled event for server in area: event=member-join server=Node-ac8cf621-bad7-8d39-3278-90e4363091a9.dc1 area=wan writer.go:29: 2021-01-29T19:34:22.799Z [INFO] TestAgent_NodeMaintenance_Disable: Started DNS server: address=127.0.0.1:30288 network=udp writer.go:29: 2021-01-29T19:34:22.802Z [INFO] TestAgent_NodeMaintenance_Disable: Started DNS server: address=127.0.0.1:30288 network=tcp writer.go:29: 2021-01-29T19:34:22.805Z [INFO] TestAgent_NodeMaintenance_Disable: Started HTTP server: address=127.0.0.1:30289 network=tcp writer.go:29: 2021-01-29T19:34:22.806Z [INFO] TestAgent_NodeMaintenance_Disable: started state syncer writer.go:29: 2021-01-29T19:34:22.826Z [WARN] TestAgent_NodeMaintenance_Disable.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:22.828Z [INFO] TestAgent_NodeMaintenance_Disable.server.raft: entering candidate state: node="Node at 127.0.0.1:30293 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:22.830Z [DEBUG] TestAgent_NodeMaintenance_Disable.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:22.832Z [DEBUG] TestAgent_NodeMaintenance_Disable.server.raft: vote granted: from=ac8cf621-bad7-8d39-3278-90e4363091a9 term=2 tally=1 writer.go:29: 2021-01-29T19:34:22.834Z [INFO] TestAgent_NodeMaintenance_Disable.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:22.835Z [INFO] TestAgent_NodeMaintenance_Disable.server.raft: entering leader state: leader="Node at 127.0.0.1:30293 [Leader]" writer.go:29: 2021-01-29T19:34:22.837Z [INFO] TestAgent_NodeMaintenance_Disable.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:22.838Z [INFO] TestAgent_NodeMaintenance_Disable.server: New leader elected: payload=Node-ac8cf621-bad7-8d39-3278-90e4363091a9 === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:22.842Z [INFO] TestAgent_StartStop: Waiting for endpoints to shut down === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:22.842Z [DEBUG] TestAgent_NodeMaintenance_Disable.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30293 === CONT TestAgent_StartStop writer.go:29: 2021-01-29T19:34:22.844Z [INFO] TestAgent_StartStop: Endpoints down --- PASS: TestAgent_StartStop (11.81s) === CONT TestAgent_NodeMaintenance_Enable === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:22.849Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:22.854Z [INFO] TestAgent_NodeMaintenance_Disable.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:22.855Z [INFO] TestAgent_NodeMaintenance_Disable.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.856Z [DEBUG] TestAgent_NodeMaintenance_Disable.server: Skipping self join check for node since the cluster is too small: node=Node-ac8cf621-bad7-8d39-3278-90e4363091a9 writer.go:29: 2021-01-29T19:34:22.857Z [INFO] TestAgent_NodeMaintenance_Disable.server: member joined, marking health alive: member=Node-ac8cf621-bad7-8d39-3278-90e4363091a9 === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:22.858Z [WARN] TestAgent_NodeMaintenance_Enable: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:22.860Z [DEBUG] TestAgent_NodeMaintenance_Enable.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:22.862Z [DEBUG] TestAgent_NodeMaintenance_Enable.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:22.866Z [INFO] TestAgent_NodeMaintenance_Enable.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:68cd5408-2e57-7830-8e3f-6fc140e97a06 Address:127.0.0.1:30317}]" writer.go:29: 2021-01-29T19:34:22.869Z [INFO] TestAgent_NodeMaintenance_Enable.server.raft: entering follower state: follower="Node at 127.0.0.1:30317 [Follower]" leader= writer.go:29: 2021-01-29T19:34:22.870Z [INFO] TestAgent_NodeMaintenance_Enable.server.serf.wan: serf: EventMemberJoin: Node-68cd5408-2e57-7830-8e3f-6fc140e97a06.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.876Z [INFO] TestAgent_NodeMaintenance_Enable.server.serf.lan: serf: EventMemberJoin: Node-68cd5408-2e57-7830-8e3f-6fc140e97a06 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.883Z [INFO] TestAgent_NodeMaintenance_Enable.server: Adding LAN server: server="Node-68cd5408-2e57-7830-8e3f-6fc140e97a06 (Addr: tcp/127.0.0.1:30317) (DC: dc1)" writer.go:29: 2021-01-29T19:34:22.884Z [INFO] TestAgent_NodeMaintenance_Enable.server: Handled event for server in area: event=member-join server=Node-68cd5408-2e57-7830-8e3f-6fc140e97a06.dc1 area=wan writer.go:29: 2021-01-29T19:34:22.888Z [INFO] TestAgent_NodeMaintenance_Enable: Started DNS server: address=127.0.0.1:30312 network=tcp writer.go:29: 2021-01-29T19:34:22.890Z [INFO] TestAgent_NodeMaintenance_Enable: Started DNS server: address=127.0.0.1:30312 network=udp writer.go:29: 2021-01-29T19:34:22.893Z [INFO] TestAgent_NodeMaintenance_Enable: Started HTTP server: address=127.0.0.1:30313 network=tcp writer.go:29: 2021-01-29T19:34:22.902Z [INFO] TestAgent_NodeMaintenance_Enable: started state syncer writer.go:29: 2021-01-29T19:34:22.931Z [WARN] TestAgent_NodeMaintenance_Enable.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:22.937Z [INFO] TestAgent_NodeMaintenance_Enable.server.raft: entering candidate state: node="Node at 127.0.0.1:30317 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:22.946Z [DEBUG] TestAgent_NodeMaintenance_Enable.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:22.947Z [DEBUG] TestAgent_NodeMaintenance_Enable.server.raft: vote granted: from=68cd5408-2e57-7830-8e3f-6fc140e97a06 term=2 tally=1 writer.go:29: 2021-01-29T19:34:22.948Z [INFO] TestAgent_NodeMaintenance_Enable.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:22.950Z [INFO] TestAgent_NodeMaintenance_Enable.server.raft: entering leader state: leader="Node at 127.0.0.1:30317 [Leader]" writer.go:29: 2021-01-29T19:34:22.956Z [INFO] TestAgent_NodeMaintenance_Enable.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:22.959Z [INFO] TestAgent_NodeMaintenance_Enable.server: New leader elected: payload=Node-68cd5408-2e57-7830-8e3f-6fc140e97a06 writer.go:29: 2021-01-29T19:34:22.964Z [DEBUG] TestAgent_NodeMaintenance_Enable.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30317 writer.go:29: 2021-01-29T19:34:22.986Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:22.993Z [INFO] TestAgent_NodeMaintenance_Enable.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:22.994Z [INFO] TestAgent_NodeMaintenance_Enable.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.995Z [DEBUG] TestAgent_NodeMaintenance_Enable.server: Skipping self join check for node since the cluster is too small: node=Node-68cd5408-2e57-7830-8e3f-6fc140e97a06 writer.go:29: 2021-01-29T19:34:22.996Z [INFO] TestAgent_NodeMaintenance_Enable.server: member joined, marking health alive: member=Node-68cd5408-2e57-7830-8e3f-6fc140e97a06 === CONT TestAgent_TokenTriggersFullSync/agent writer.go:29: 2021-01-29T19:34:23.029Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:23.101Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.acl: dropping service from result due to ACLs: service="{consul {}}" writer.go:29: 2021-01-29T19:34:23.103Z [DEBUG] TestAgent_TokenTriggersFullSync/agent: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:23.104Z [INFO] TestAgent_TokenTriggersFullSync/agent: Synced node info writer.go:29: 2021-01-29T19:34:23.127Z [INFO] TestAgent_TokenTriggersFullSync/agent: Requesting shutdown writer.go:29: 2021-01-29T19:34:23.129Z [INFO] TestAgent_TokenTriggersFullSync/agent.server: shutting down server writer.go:29: 2021-01-29T19:34:23.130Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:23.132Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:23.134Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.132Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:23.134Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:23.135Z [WARN] TestAgent_TokenTriggersFullSync/agent.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.136Z [DEBUG] TestAgent_TokenTriggersFullSync/agent.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.151Z [INFO] TestAgent_NodeMaintenance_Disable: Node entered maintenance mode === CONT TestAgent_TokenTriggersFullSync/agent writer.go:29: 2021-01-29T19:34:23.149Z [WARN] TestAgent_TokenTriggersFullSync/agent.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.153Z [DEBUG] TestAgent_NodeMaintenance_Disable: removed check: check=_node_maintenance writer.go:29: 2021-01-29T19:34:23.154Z [INFO] TestAgent_NodeMaintenance_Disable: Node left maintenance mode === CONT TestAgent_TokenTriggersFullSync/agent writer.go:29: 2021-01-29T19:34:23.155Z [INFO] TestAgent_TokenTriggersFullSync/agent.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:23.156Z [INFO] TestAgent_TokenTriggersFullSync/agent: consul server down writer.go:29: 2021-01-29T19:34:23.159Z [INFO] TestAgent_TokenTriggersFullSync/agent: shutdown complete writer.go:29: 2021-01-29T19:34:23.160Z [INFO] TestAgent_TokenTriggersFullSync/agent: Stopping server: protocol=DNS address=127.0.0.1:30300 network=tcp writer.go:29: 2021-01-29T19:34:23.166Z [INFO] TestAgent_TokenTriggersFullSync/agent: Stopping server: protocol=DNS address=127.0.0.1:30300 network=udp === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.167Z [INFO] TestAgent_NodeMaintenance_Enable: Node entered maintenance mode === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.167Z [INFO] TestAgent_NodeMaintenance_Disable: Synced node info === CONT TestAgent_TokenTriggersFullSync/agent writer.go:29: 2021-01-29T19:34:23.168Z [INFO] TestAgent_TokenTriggersFullSync/agent: Stopping server: protocol=HTTP address=127.0.0.1:30301 network=tcp === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.170Z [INFO] TestAgent_NodeMaintenance_Enable: Synced node info === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.172Z [INFO] TestAgent_NodeMaintenance_Disable: Deregistered check: check=_node_maintenance writer.go:29: 2021-01-29T19:34:23.173Z [INFO] TestAgent_NodeMaintenance_Disable: Requesting shutdown === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.173Z [INFO] TestAgent_NodeMaintenance_Enable: Synced check: check=_node_maintenance === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.175Z [INFO] TestAgent_NodeMaintenance_Disable.server: shutting down server === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.175Z [INFO] TestAgent_NodeMaintenance_Enable: Requesting shutdown === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.176Z [DEBUG] TestAgent_NodeMaintenance_Disable.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.176Z [INFO] TestAgent_NodeMaintenance_Enable.server: shutting down server === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.177Z [WARN] TestAgent_NodeMaintenance_Disable.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.178Z [DEBUG] TestAgent_NodeMaintenance_Enable.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.180Z [ERROR] TestAgent_NodeMaintenance_Disable.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.179Z [WARN] TestAgent_NodeMaintenance_Enable.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.183Z [DEBUG] connect.ca.consul: consul CA provider configured: id=d9:68:ba:57:ab:40:51:fd:5d:5e:ad:5a:a6:ab:5c:30:03:ec:59:e5:a3:ac:fa:a5:02:5d:74:05:e6:34:a3:6f is_primary=true === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.182Z [ERROR] TestAgent_NodeMaintenance_Enable.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:23.185Z [WARN] TestAgent_NodeMaintenance_Enable.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.187Z [DEBUG] TestAgent_NodeMaintenance_Enable.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.191Z [INFO] TestAgent_NodeMaintenance_Enable.server.router.manager: shutting down === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.194Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.connect: CA rotated to new root under provider: provider=consul === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.193Z [INFO] TestAgent_NodeMaintenance_Enable: consul server down === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.201Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: received new intermediate certificate from primary datacenter writer.go:29: 2021-01-29T19:34:23.207Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: updated root certificates from primary datacenter === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.197Z [INFO] TestAgent_NodeMaintenance_Enable: shutdown complete === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.181Z [DEBUG] TestAgent_NodeMaintenance_Disable.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.213Z [INFO] TestAgent_NodeMaintenance_Enable: Stopping server: protocol=DNS address=127.0.0.1:30312 network=tcp === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.187Z [WARN] TestAgent_NodeMaintenance_Disable.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.216Z [INFO] TestAgent_NodeMaintenance_Enable: Stopping server: protocol=DNS address=127.0.0.1:30312 network=udp writer.go:29: 2021-01-29T19:34:23.219Z [INFO] TestAgent_NodeMaintenance_Enable: Stopping server: protocol=HTTP address=127.0.0.1:30313 network=tcp === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.219Z [INFO] TestAgent_NodeMaintenance_Disable.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:23.219Z [INFO] TestAgent_NodeMaintenance_Disable: consul server down writer.go:29: 2021-01-29T19:34:23.225Z [INFO] TestAgent_NodeMaintenance_Disable: shutdown complete writer.go:29: 2021-01-29T19:34:23.226Z [INFO] TestAgent_NodeMaintenance_Disable: Stopping server: protocol=DNS address=127.0.0.1:30288 network=tcp writer.go:29: 2021-01-29T19:34:23.228Z [INFO] TestAgent_NodeMaintenance_Disable: Stopping server: protocol=DNS address=127.0.0.1:30288 network=udp writer.go:29: 2021-01-29T19:34:23.229Z [INFO] TestAgent_NodeMaintenance_Disable: Stopping server: protocol=HTTP address=127.0.0.1:30289 network=tcp === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.245Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Requesting shutdown writer.go:29: 2021-01-29T19:34:23.246Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server: shutting down server writer.go:29: 2021-01-29T19:34:23.247Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: stopping routine: routine="config entry replication" writer.go:29: 2021-01-29T19:34:23.249Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: stopping routine: routine="secondary CA roots watch" writer.go:29: 2021-01-29T19:34:23.251Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: stopping routine: routine="intention replication" writer.go:29: 2021-01-29T19:34:23.253Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: stopping routine: routine="secondary cert renew watch" writer.go:29: 2021-01-29T19:34:23.254Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.255Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.255Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: stopped routine: routine="secondary cert renew watch" writer.go:29: 2021-01-29T19:34:23.255Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.264Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.271Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:23.275Z [ERROR] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.rpc: RPC failed to server in DC: server=127.0.0.1:30215 datacenter=dc1 method=ConnectCA.Roots error="rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:23.275Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:23.275Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: consul server down writer.go:29: 2021-01-29T19:34:23.280Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: shutdown complete writer.go:29: 2021-01-29T19:34:23.281Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Stopping server: protocol=DNS address=127.0.0.1:30234 network=tcp writer.go:29: 2021-01-29T19:34:23.280Z [ERROR] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: CA root replication failed, will retry: routine="secondary CA roots watch" error="Error retrieving the primary datacenter's roots: rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:23.279Z [ERROR] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.rpc: RPC failed to server in DC: server=127.0.0.1:30215 datacenter=dc1 method=ConfigEntry.ListAll error="rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:23.279Z [ERROR] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.rpc: RPC failed to server in DC: server=127.0.0.1:30215 datacenter=dc1 method=Intention.List error="rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:23.283Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Stopping server: protocol=DNS address=127.0.0.1:30234 network=udp writer.go:29: 2021-01-29T19:34:23.287Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.replication.config_entry: stopped replication writer.go:29: 2021-01-29T19:34:23.297Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc2.leader: stopped routine: routine="config entry replication" writer.go:29: 2021-01-29T19:34:23.290Z [ERROR] TestAgentConnectCALeafCert_secondaryDC_good-dc2.server.connect: error replicating intentions: routine="intention replication" error="rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:23.290Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Stopping server: protocol=HTTP address=127.0.0.1:30235 network=tcp === CONT TestAgent_TokenTriggersFullSync/agent writer.go:29: 2021-01-29T19:34:23.672Z [INFO] TestAgent_TokenTriggersFullSync/agent: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:23.674Z [INFO] TestAgent_TokenTriggersFullSync/agent: Endpoints down === RUN TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.694Z [WARN] TestAgent_TokenTriggersFullSync/acl_token: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:23.698Z [WARN] TestAgent_TokenTriggersFullSync/acl_token: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:23.700Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:23.703Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:23.717Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c573e839-baf4-a2f6-dfe5-11751c77bdfd Address:127.0.0.1:30329}]" writer.go:29: 2021-01-29T19:34:23.719Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.raft: entering follower state: follower="Node at 127.0.0.1:30329 [Follower]" leader= writer.go:29: 2021-01-29T19:34:23.720Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.serf.wan: serf: EventMemberJoin: Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd.dc1 127.0.0.1 === CONT TestAgent_NodeMaintenance_Enable writer.go:29: 2021-01-29T19:34:23.722Z [INFO] TestAgent_NodeMaintenance_Enable: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:23.723Z [INFO] TestAgent_NodeMaintenance_Enable: Endpoints down --- PASS: TestAgent_NodeMaintenance_Enable (0.88s) === CONT TestAgent_NodeMaintenance_BadRequest === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.723Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.serf.lan: serf: EventMemberJoin: Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd 127.0.0.1 writer.go:29: 2021-01-29T19:34:23.726Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Started DNS server: address=127.0.0.1:30324 network=udp === CONT TestAgent_NodeMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.731Z [INFO] TestAgent_NodeMaintenance_Disable: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:23.732Z [INFO] TestAgent_NodeMaintenance_Disable: Endpoints down --- PASS: TestAgent_NodeMaintenance_Disable (0.98s) === CONT TestAgent_ServiceMaintenance_Disable === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.734Z [WARN] TestAgent_NodeMaintenance_BadRequest: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:23.735Z [DEBUG] TestAgent_NodeMaintenance_BadRequest.tlsutil: Update: version=1 === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.730Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: Adding LAN server: server="Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd (Addr: tcp/127.0.0.1:30329) (DC: dc1)" === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.737Z [DEBUG] TestAgent_NodeMaintenance_BadRequest.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.730Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: Handled event for server in area: event=member-join server=Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd.dc1 area=wan writer.go:29: 2021-01-29T19:34:23.732Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Started DNS server: address=127.0.0.1:30324 network=tcp writer.go:29: 2021-01-29T19:34:23.742Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Started HTTP server: address=127.0.0.1:30325 network=tcp writer.go:29: 2021-01-29T19:34:23.744Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: started state syncer === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.747Z [INFO] TestAgent_NodeMaintenance_BadRequest.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1cd07684-75ae-f183-5c18-c7cdcd8c2cc0 Address:127.0.0.1:30323}]" === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.748Z [WARN] TestAgent_ServiceMaintenance_Disable: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:23.749Z [DEBUG] TestAgent_ServiceMaintenance_Disable.tlsutil: Update: version=1 === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.749Z [INFO] TestAgent_NodeMaintenance_BadRequest.server.raft: entering follower state: follower="Node at 127.0.0.1:30323 [Follower]" leader= === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.751Z [DEBUG] TestAgent_ServiceMaintenance_Disable.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.751Z [INFO] TestAgent_NodeMaintenance_BadRequest.server.serf.wan: serf: EventMemberJoin: Node-1cd07684-75ae-f183-5c18-c7cdcd8c2cc0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:23.754Z [INFO] TestAgent_NodeMaintenance_BadRequest.server.serf.lan: serf: EventMemberJoin: Node-1cd07684-75ae-f183-5c18-c7cdcd8c2cc0 127.0.0.1 writer.go:29: 2021-01-29T19:34:23.756Z [INFO] TestAgent_NodeMaintenance_BadRequest.server: Handled event for server in area: event=member-join server=Node-1cd07684-75ae-f183-5c18-c7cdcd8c2cc0.dc1 area=wan writer.go:29: 2021-01-29T19:34:23.756Z [INFO] TestAgent_NodeMaintenance_BadRequest.server: Adding LAN server: server="Node-1cd07684-75ae-f183-5c18-c7cdcd8c2cc0 (Addr: tcp/127.0.0.1:30323) (DC: dc1)" writer.go:29: 2021-01-29T19:34:23.756Z [INFO] TestAgent_NodeMaintenance_BadRequest: Started DNS server: address=127.0.0.1:30318 network=udp === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.758Z [INFO] TestAgent_ServiceMaintenance_Disable.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:73b50da3-27f3-ace1-636d-209c1bb1bbcc Address:127.0.0.1:30341}]" === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.759Z [INFO] TestAgent_NodeMaintenance_BadRequest: Started DNS server: address=127.0.0.1:30318 network=tcp writer.go:29: 2021-01-29T19:34:23.761Z [INFO] TestAgent_NodeMaintenance_BadRequest: Started HTTP server: address=127.0.0.1:30319 network=tcp === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.761Z [INFO] TestAgent_ServiceMaintenance_Disable.server.serf.wan: serf: EventMemberJoin: Node-73b50da3-27f3-ace1-636d-209c1bb1bbcc.dc1 127.0.0.1 === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.762Z [INFO] TestAgent_NodeMaintenance_BadRequest: started state syncer === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.762Z [INFO] TestAgent_ServiceMaintenance_Disable.server.raft: entering follower state: follower="Node at 127.0.0.1:30341 [Follower]" leader= writer.go:29: 2021-01-29T19:34:23.763Z [INFO] TestAgent_ServiceMaintenance_Disable.server.serf.lan: serf: EventMemberJoin: Node-73b50da3-27f3-ace1-636d-209c1bb1bbcc 127.0.0.1 writer.go:29: 2021-01-29T19:34:23.769Z [INFO] TestAgent_ServiceMaintenance_Disable.server: Handled event for server in area: event=member-join server=Node-73b50da3-27f3-ace1-636d-209c1bb1bbcc.dc1 area=wan writer.go:29: 2021-01-29T19:34:23.769Z [INFO] TestAgent_ServiceMaintenance_Disable.server: Adding LAN server: server="Node-73b50da3-27f3-ace1-636d-209c1bb1bbcc (Addr: tcp/127.0.0.1:30341) (DC: dc1)" writer.go:29: 2021-01-29T19:34:23.771Z [INFO] TestAgent_ServiceMaintenance_Disable: Started DNS server: address=127.0.0.1:30336 network=udp === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.772Z [WARN] TestAgent_TokenTriggersFullSync/acl_token.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.772Z [INFO] TestAgent_ServiceMaintenance_Disable: Started DNS server: address=127.0.0.1:30336 network=tcp === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.772Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.raft: entering candidate state: node="Node at 127.0.0.1:30329 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:23.774Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.server.raft: votes: needed=1 === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.774Z [INFO] TestAgent_ServiceMaintenance_Disable: Started HTTP server: address=127.0.0.1:30337 network=tcp === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.776Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.server.raft: vote granted: from=c573e839-baf4-a2f6-dfe5-11751c77bdfd term=2 tally=1 === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.776Z [INFO] TestAgent_ServiceMaintenance_Disable: started state syncer === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.777Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:23.778Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.raft: entering leader state: leader="Node at 127.0.0.1:30329 [Leader]" writer.go:29: 2021-01-29T19:34:23.779Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:23.781Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: initializing acls writer.go:29: 2021-01-29T19:34:23.780Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: New leader elected: payload=Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd writer.go:29: 2021-01-29T19:34:23.780Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: initializing acls writer.go:29: 2021-01-29T19:34:23.784Z [WARN] TestAgent_TokenTriggersFullSync/acl_token.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:23.783Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:23.786Z [WARN] TestAgent_TokenTriggersFullSync/acl_token.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:23.786Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:23.788Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:23.789Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:23.788Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:23.791Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:34:23.790Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.serf.lan: serf: EventMemberUpdate: Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd writer.go:29: 2021-01-29T19:34:23.793Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.serf.wan: serf: EventMemberUpdate: Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd.dc1 writer.go:29: 2021-01-29T19:34:23.794Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: Updating LAN server: server="Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd (Addr: tcp/127.0.0.1:30329) (DC: dc1)" writer.go:29: 2021-01-29T19:34:23.794Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.serf.lan: serf: EventMemberUpdate: Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd writer.go:29: 2021-01-29T19:34:23.796Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: Handled event for server in area: event=member-update server=Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd.dc1 area=wan writer.go:29: 2021-01-29T19:34:23.796Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:23.798Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.serf.wan: serf: EventMemberUpdate: Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd.dc1 writer.go:29: 2021-01-29T19:34:23.801Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: Handled event for server in area: event=member-update server=Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd.dc1 area=wan writer.go:29: 2021-01-29T19:34:23.798Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: Updating LAN server: server="Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd (Addr: tcp/127.0.0.1:30329) (DC: dc1)" === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.803Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:23.805Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc2: Endpoints down === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.803Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.806Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Requesting shutdown === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.807Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.leader: started routine: routine="CA root pruning" === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.808Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server: shutting down server === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.809Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.server: Skipping self join check for node since the cluster is too small: node=Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.809Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.810Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: member joined, marking health alive: member=Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.811Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.811Z [DEBUG] TestAgentConnectCALeafCert_secondaryDC_good-dc1.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.813Z [WARN] TestAgent_ServiceMaintenance_Disable.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:23.814Z [INFO] TestAgent_ServiceMaintenance_Disable.server.raft: entering candidate state: node="Node at 127.0.0.1:30341 [Candidate]" term=2 === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.812Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.server: Skipping self join check for node since the cluster is too small: node=Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.813Z [WARN] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.819Z [WARN] TestAgent_NodeMaintenance_BadRequest.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.819Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.server: Skipping self join check for node since the cluster is too small: node=Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.821Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.router.manager: shutting down === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.822Z [DEBUG] TestAgent_ServiceMaintenance_Disable.server.raft: votes: needed=1 === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.822Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: consul server down === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.820Z [INFO] TestAgent_NodeMaintenance_BadRequest.server.raft: entering candidate state: node="Node at 127.0.0.1:30323 [Candidate]" term=2 === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.823Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: shutdown complete === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.823Z [DEBUG] TestAgent_ServiceMaintenance_Disable.server.raft: vote granted: from=73b50da3-27f3-ace1-636d-209c1bb1bbcc term=2 tally=1 === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.824Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Stopping server: protocol=DNS address=127.0.0.1:30210 network=tcp === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.825Z [INFO] TestAgent_ServiceMaintenance_Disable.server.raft: election won: tally=1 === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.823Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1.server.router.manager: shutting down === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.825Z [DEBUG] TestAgent_NodeMaintenance_BadRequest.server.raft: votes: needed=1 === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.825Z [INFO] TestAgent_ServiceMaintenance_Disable.server.raft: entering leader state: leader="Node at 127.0.0.1:30341 [Leader]" === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:23.825Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Stopping server: protocol=DNS address=127.0.0.1:30210 network=udp writer.go:29: 2021-01-29T19:34:23.827Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Stopping server: protocol=HTTP address=127.0.0.1:30211 network=tcp === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.826Z [DEBUG] TestAgent_NodeMaintenance_BadRequest.server.raft: vote granted: from=1cd07684-75ae-f183-5c18-c7cdcd8c2cc0 term=2 tally=1 === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.826Z [INFO] TestAgent_ServiceMaintenance_Disable.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:23.827Z [INFO] TestAgent_ServiceMaintenance_Disable.server: New leader elected: payload=Node-73b50da3-27f3-ace1-636d-209c1bb1bbcc === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.828Z [INFO] TestAgent_NodeMaintenance_BadRequest.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:23.829Z [INFO] TestAgent_NodeMaintenance_BadRequest.server.raft: entering leader state: leader="Node at 127.0.0.1:30323 [Leader]" === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.828Z [DEBUG] TestAgent_ServiceMaintenance_Disable.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30341 === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.830Z [INFO] TestAgent_NodeMaintenance_BadRequest.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:23.831Z [INFO] TestAgent_NodeMaintenance_BadRequest.server: New leader elected: payload=Node-1cd07684-75ae-f183-5c18-c7cdcd8c2cc0 === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.834Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.832Z [DEBUG] TestAgent_NodeMaintenance_BadRequest.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30323 === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.834Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.acl: dropping node from result due to ACLs: node=Node-c573e839-baf4-a2f6-dfe5-11751c77bdfd === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.838Z [INFO] TestAgent_ServiceMaintenance_Disable.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:23.839Z [INFO] TestAgent_ServiceMaintenance_Disable.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.840Z [DEBUG] TestAgent_ServiceMaintenance_Disable.server: Skipping self join check for node since the cluster is too small: node=Node-73b50da3-27f3-ace1-636d-209c1bb1bbcc writer.go:29: 2021-01-29T19:34:23.840Z [INFO] TestAgent_ServiceMaintenance_Disable.server: member joined, marking health alive: member=Node-73b50da3-27f3-ace1-636d-209c1bb1bbcc === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.841Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:23.848Z [INFO] TestAgent_NodeMaintenance_BadRequest.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:23.850Z [INFO] TestAgent_NodeMaintenance_BadRequest.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.851Z [DEBUG] TestAgent_NodeMaintenance_BadRequest.server: Skipping self join check for node since the cluster is too small: node=Node-1cd07684-75ae-f183-5c18-c7cdcd8c2cc0 === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:23.852Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Updated agent's ACL token: token=acl_token === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.853Z [INFO] TestAgent_NodeMaintenance_BadRequest.server: member joined, marking health alive: member=Node-1cd07684-75ae-f183-5c18-c7cdcd8c2cc0 === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:23.867Z [INFO] TestAgent_ServiceMaintenance_Disable: Service entered maintenance mode: service=test writer.go:29: 2021-01-29T19:34:23.869Z [DEBUG] TestAgent_ServiceMaintenance_Disable: removed check: check=_service_maintenance:test writer.go:29: 2021-01-29T19:34:23.869Z [INFO] TestAgent_ServiceMaintenance_Disable: Service left maintenance mode: service=test writer.go:29: 2021-01-29T19:34:23.871Z [INFO] TestAgent_ServiceMaintenance_Disable: Synced node info writer.go:29: 2021-01-29T19:34:23.873Z [INFO] TestAgent_ServiceMaintenance_Disable: Synced service: service=test writer.go:29: 2021-01-29T19:34:23.874Z [INFO] TestAgent_ServiceMaintenance_Disable: Deregistered check: check=_service_maintenance:test writer.go:29: 2021-01-29T19:34:23.875Z [INFO] TestAgent_ServiceMaintenance_Disable: Requesting shutdown writer.go:29: 2021-01-29T19:34:23.876Z [INFO] TestAgent_ServiceMaintenance_Disable.server: shutting down server writer.go:29: 2021-01-29T19:34:23.877Z [DEBUG] TestAgent_ServiceMaintenance_Disable.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.878Z [WARN] TestAgent_ServiceMaintenance_Disable.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.877Z [ERROR] TestAgent_ServiceMaintenance_Disable.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:23.878Z [DEBUG] TestAgent_ServiceMaintenance_Disable.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.879Z [WARN] TestAgent_ServiceMaintenance_Disable.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.886Z [INFO] TestAgent_ServiceMaintenance_Disable.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:23.888Z [INFO] TestAgent_ServiceMaintenance_Disable: consul server down writer.go:29: 2021-01-29T19:34:23.888Z [INFO] TestAgent_ServiceMaintenance_Disable: shutdown complete writer.go:29: 2021-01-29T19:34:23.889Z [INFO] TestAgent_ServiceMaintenance_Disable: Stopping server: protocol=DNS address=127.0.0.1:30336 network=tcp writer.go:29: 2021-01-29T19:34:23.890Z [INFO] TestAgent_ServiceMaintenance_Disable: Stopping server: protocol=DNS address=127.0.0.1:30336 network=udp writer.go:29: 2021-01-29T19:34:23.891Z [INFO] TestAgent_ServiceMaintenance_Disable: Stopping server: protocol=HTTP address=127.0.0.1:30337 network=tcp === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:23.893Z [INFO] TestAgent_NodeMaintenance_BadRequest: Requesting shutdown writer.go:29: 2021-01-29T19:34:23.894Z [INFO] TestAgent_NodeMaintenance_BadRequest.server: shutting down server writer.go:29: 2021-01-29T19:34:23.895Z [DEBUG] TestAgent_NodeMaintenance_BadRequest.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.896Z [WARN] TestAgent_NodeMaintenance_BadRequest.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.895Z [ERROR] TestAgent_NodeMaintenance_BadRequest.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:23.896Z [DEBUG] TestAgent_NodeMaintenance_BadRequest.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.898Z [WARN] TestAgent_NodeMaintenance_BadRequest.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.901Z [INFO] TestAgent_NodeMaintenance_BadRequest.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:23.901Z [INFO] TestAgent_NodeMaintenance_BadRequest: consul server down writer.go:29: 2021-01-29T19:34:23.904Z [INFO] TestAgent_NodeMaintenance_BadRequest: shutdown complete writer.go:29: 2021-01-29T19:34:23.905Z [INFO] TestAgent_NodeMaintenance_BadRequest: Stopping server: protocol=DNS address=127.0.0.1:30318 network=tcp writer.go:29: 2021-01-29T19:34:23.906Z [INFO] TestAgent_NodeMaintenance_BadRequest: Stopping server: protocol=DNS address=127.0.0.1:30318 network=udp writer.go:29: 2021-01-29T19:34:23.907Z [INFO] TestAgent_NodeMaintenance_BadRequest: Stopping server: protocol=HTTP address=127.0.0.1:30319 network=tcp === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:24.014Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.acl: dropping check from result due to ACLs: check=serfHealth writer.go:29: 2021-01-29T19:34:24.018Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Synced node info writer.go:29: 2021-01-29T19:34:24.020Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token: Node info in sync writer.go:29: 2021-01-29T19:34:24.021Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Requesting shutdown writer.go:29: 2021-01-29T19:34:24.022Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server: shutting down server writer.go:29: 2021-01-29T19:34:24.023Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:24.024Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:24.024Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:24.025Z [WARN] TestAgent_TokenTriggersFullSync/acl_token.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:24.024Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:24.024Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:24.025Z [DEBUG] TestAgent_TokenTriggersFullSync/acl_token.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:24.028Z [WARN] TestAgent_TokenTriggersFullSync/acl_token.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:24.030Z [INFO] TestAgent_TokenTriggersFullSync/acl_token.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:24.031Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: consul server down writer.go:29: 2021-01-29T19:34:24.032Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: shutdown complete writer.go:29: 2021-01-29T19:34:24.033Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Stopping server: protocol=DNS address=127.0.0.1:30324 network=tcp writer.go:29: 2021-01-29T19:34:24.033Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Stopping server: protocol=DNS address=127.0.0.1:30324 network=udp writer.go:29: 2021-01-29T19:34:24.034Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Stopping server: protocol=HTTP address=127.0.0.1:30325 network=tcp === CONT TestAgentConnectCALeafCert_secondaryDC_good writer.go:29: 2021-01-29T19:34:24.327Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:24.329Z [INFO] TestAgentConnectCALeafCert_secondaryDC_good-dc1: Endpoints down === CONT TestAgent_ServiceMaintenance_BadRequest --- PASS: TestAgentConnectCALeafCert_secondaryDC_good (8.52s) === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.352Z [WARN] TestAgent_ServiceMaintenance_BadRequest: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:24.354Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:24.357Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:24.362Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:176df84f-8b6a-4121-7e4a-bdc9a71974dc Address:127.0.0.1:30311}]" writer.go:29: 2021-01-29T19:34:24.364Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server.raft: entering follower state: follower="Node at 127.0.0.1:30311 [Follower]" leader= writer.go:29: 2021-01-29T19:34:24.364Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server.serf.wan: serf: EventMemberJoin: Node-176df84f-8b6a-4121-7e4a-bdc9a71974dc.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:24.373Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server.serf.lan: serf: EventMemberJoin: Node-176df84f-8b6a-4121-7e4a-bdc9a71974dc 127.0.0.1 writer.go:29: 2021-01-29T19:34:24.378Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server: Adding LAN server: server="Node-176df84f-8b6a-4121-7e4a-bdc9a71974dc (Addr: tcp/127.0.0.1:30311) (DC: dc1)" writer.go:29: 2021-01-29T19:34:24.380Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server: Handled event for server in area: event=member-join server=Node-176df84f-8b6a-4121-7e4a-bdc9a71974dc.dc1 area=wan writer.go:29: 2021-01-29T19:34:24.382Z [INFO] TestAgent_ServiceMaintenance_BadRequest: Started DNS server: address=127.0.0.1:30306 network=tcp writer.go:29: 2021-01-29T19:34:24.385Z [INFO] TestAgent_ServiceMaintenance_BadRequest: Started DNS server: address=127.0.0.1:30306 network=udp writer.go:29: 2021-01-29T19:34:24.387Z [INFO] TestAgent_ServiceMaintenance_BadRequest: Started HTTP server: address=127.0.0.1:30307 network=tcp writer.go:29: 2021-01-29T19:34:24.389Z [INFO] TestAgent_ServiceMaintenance_BadRequest: started state syncer === CONT TestAgent_ServiceMaintenance_Disable writer.go:29: 2021-01-29T19:34:24.393Z [INFO] TestAgent_ServiceMaintenance_Disable: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:24.394Z [INFO] TestAgent_ServiceMaintenance_Disable: Endpoints down --- PASS: TestAgent_ServiceMaintenance_Disable (0.66s) === CONT TestAgent_DeregisterService_ACLDeny === CONT TestAgent_NodeMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.409Z [INFO] TestAgent_NodeMaintenance_BadRequest: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:24.411Z [INFO] TestAgent_NodeMaintenance_BadRequest: Endpoints down --- PASS: TestAgent_NodeMaintenance_BadRequest (0.69s) === CONT TestAgent_DeregisterService === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.415Z [WARN] TestAgent_DeregisterService_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:24.417Z [WARN] TestAgent_DeregisterService_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:24.418Z [DEBUG] TestAgent_DeregisterService_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:24.421Z [DEBUG] TestAgent_DeregisterService_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.428Z [WARN] TestAgent_ServiceMaintenance_BadRequest.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:24.430Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server.raft: entering candidate state: node="Node at 127.0.0.1:30311 [Candidate]" term=2 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.442Z [INFO] TestAgent_DeregisterService_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:476e92b3-5c91-a76e-6cb3-e3c26093834f Address:127.0.0.1:30335}]" === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.445Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest.server.raft: votes: needed=1 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.444Z [INFO] TestAgent_DeregisterService_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30335 [Follower]" leader= === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.447Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest.server.raft: vote granted: from=176df84f-8b6a-4121-7e4a-bdc9a71974dc term=2 tally=1 writer.go:29: 2021-01-29T19:34:24.448Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server.raft: election won: tally=1 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.445Z [INFO] TestAgent_DeregisterService_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-476e92b3-5c91-a76e-6cb3-e3c26093834f.dc1 127.0.0.1 === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.449Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server.raft: entering leader state: leader="Node at 127.0.0.1:30311 [Leader]" writer.go:29: 2021-01-29T19:34:24.451Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:24.454Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server: New leader elected: payload=Node-176df84f-8b6a-4121-7e4a-bdc9a71974dc === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.455Z [INFO] TestAgent_DeregisterService_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-476e92b3-5c91-a76e-6cb3-e3c26093834f 127.0.0.1 writer.go:29: 2021-01-29T19:34:24.460Z [INFO] TestAgent_DeregisterService_ACLDeny: Started DNS server: address=127.0.0.1:30330 network=udp writer.go:29: 2021-01-29T19:34:24.462Z [INFO] TestAgent_DeregisterService_ACLDeny.server: Adding LAN server: server="Node-476e92b3-5c91-a76e-6cb3-e3c26093834f (Addr: tcp/127.0.0.1:30335) (DC: dc1)" === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.463Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30311 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.466Z [INFO] TestAgent_DeregisterService_ACLDeny.server: Handled event for server in area: event=member-join server=Node-476e92b3-5c91-a76e-6cb3-e3c26093834f.dc1 area=wan writer.go:29: 2021-01-29T19:34:24.468Z [INFO] TestAgent_DeregisterService_ACLDeny: Started DNS server: address=127.0.0.1:30330 network=tcp writer.go:29: 2021-01-29T19:34:24.476Z [INFO] TestAgent_DeregisterService_ACLDeny: Started HTTP server: address=127.0.0.1:30331 network=tcp writer.go:29: 2021-01-29T19:34:24.477Z [INFO] TestAgent_DeregisterService_ACLDeny: started state syncer === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.483Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.511Z [WARN] TestAgent_DeregisterService_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:24.514Z [INFO] TestAgent_DeregisterService_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30335 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:24.522Z [DEBUG] TestAgent_DeregisterService_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:24.524Z [DEBUG] TestAgent_DeregisterService_ACLDeny.server.raft: votes: needed=1 === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.525Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.525Z [DEBUG] TestAgent_DeregisterService_ACLDeny.server.raft: vote granted: from=476e92b3-5c91-a76e-6cb3-e3c26093834f term=2 tally=1 === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.526Z [INFO] TestAgent_ServiceMaintenance_BadRequest.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:24.527Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest.server: Skipping self join check for node since the cluster is too small: node=Node-176df84f-8b6a-4121-7e4a-bdc9a71974dc === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.527Z [INFO] TestAgent_DeregisterService_ACLDeny.server.raft: election won: tally=1 === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.528Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server: member joined, marking health alive: member=Node-176df84f-8b6a-4121-7e4a-bdc9a71974dc === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.528Z [INFO] TestAgent_DeregisterService_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30335 [Leader]" writer.go:29: 2021-01-29T19:34:24.531Z [INFO] TestAgent_DeregisterService_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:24.532Z [INFO] TestAgent_DeregisterService_ACLDeny.server: New leader elected: payload=Node-476e92b3-5c91-a76e-6cb3-e3c26093834f === CONT TestAgent_TokenTriggersFullSync/acl_token writer.go:29: 2021-01-29T19:34:24.536Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:24.537Z [INFO] TestAgent_TokenTriggersFullSync/acl_token: Endpoints down === RUN TestAgent_TokenTriggersFullSync/default === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.538Z [INFO] TestAgent_DeregisterService_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:24.548Z [INFO] TestAgent_DeregisterService_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:24.549Z [WARN] TestAgent_DeregisterService_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:24.573Z [INFO] TestAgent_DeregisterService_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:24.578Z [INFO] TestAgent_DeregisterService_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:24.579Z [INFO] TestAgent_DeregisterService_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:24.580Z [INFO] TestAgent_DeregisterService_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:24.581Z [INFO] TestAgent_DeregisterService_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-476e92b3-5c91-a76e-6cb3-e3c26093834f writer.go:29: 2021-01-29T19:34:24.582Z [INFO] TestAgent_DeregisterService_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-476e92b3-5c91-a76e-6cb3-e3c26093834f.dc1 writer.go:29: 2021-01-29T19:34:24.586Z [INFO] TestAgent_DeregisterService_ACLDeny.server: Updating LAN server: server="Node-476e92b3-5c91-a76e-6cb3-e3c26093834f (Addr: tcp/127.0.0.1:30335) (DC: dc1)" writer.go:29: 2021-01-29T19:34:24.588Z [INFO] TestAgent_DeregisterService_ACLDeny.server: Handled event for server in area: event=member-update server=Node-476e92b3-5c91-a76e-6cb3-e3c26093834f.dc1 area=wan writer.go:29: 2021-01-29T19:34:24.637Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:24.655Z [INFO] TestAgent_DeregisterService_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:24.656Z [INFO] TestAgent_DeregisterService_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:24.658Z [DEBUG] TestAgent_DeregisterService_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-476e92b3-5c91-a76e-6cb3-e3c26093834f writer.go:29: 2021-01-29T19:34:24.661Z [INFO] TestAgent_DeregisterService_ACLDeny.server: member joined, marking health alive: member=Node-476e92b3-5c91-a76e-6cb3-e3c26093834f === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.664Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:24.667Z [INFO] TestAgent_ServiceMaintenance_BadRequest: Synced node info === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.685Z [WARN] TestAgent_TokenTriggersFullSync/default: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:24.686Z [WARN] TestAgent_TokenTriggersFullSync/default: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:24.688Z [DEBUG] TestAgent_TokenTriggersFullSync/default.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:24.690Z [DEBUG] TestAgent_TokenTriggersFullSync/default.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.689Z [DEBUG] TestAgent_DeregisterService_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-476e92b3-5c91-a76e-6cb3-e3c26093834f === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.716Z [INFO] TestAgent_TokenTriggersFullSync/default.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6ae2a6f1-0459-1513-c53b-d07212569993 Address:127.0.0.1:30353}]" writer.go:29: 2021-01-29T19:34:24.722Z [INFO] TestAgent_TokenTriggersFullSync/default.server.serf.wan: serf: EventMemberJoin: Node-6ae2a6f1-0459-1513-c53b-d07212569993.dc1 127.0.0.1 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.732Z [DEBUG] TestAgent_DeregisterService_ACLDeny.acl: dropping node from result due to ACLs: node=Node-476e92b3-5c91-a76e-6cb3-e3c26093834f writer.go:29: 2021-01-29T19:34:24.734Z [DEBUG] TestAgent_DeregisterService_ACLDeny.acl: dropping node from result due to ACLs: node=Node-476e92b3-5c91-a76e-6cb3-e3c26093834f === RUN TestAgent_DeregisterService_ACLDeny/no_token === RUN TestAgent_DeregisterService_ACLDeny/root_token === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.746Z [DEBUG] TestAgent_DeregisterService_ACLDeny: removed service: service=test === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.731Z [INFO] TestAgent_TokenTriggersFullSync/default.server.raft: entering follower state: follower="Node at 127.0.0.1:30353 [Follower]" leader= writer.go:29: 2021-01-29T19:34:24.732Z [INFO] TestAgent_TokenTriggersFullSync/default.server.serf.lan: serf: EventMemberJoin: Node-6ae2a6f1-0459-1513-c53b-d07212569993 127.0.0.1 writer.go:29: 2021-01-29T19:34:24.752Z [INFO] TestAgent_TokenTriggersFullSync/default.server: Handled event for server in area: event=member-join server=Node-6ae2a6f1-0459-1513-c53b-d07212569993.dc1 area=wan writer.go:29: 2021-01-29T19:34:24.762Z [INFO] TestAgent_TokenTriggersFullSync/default: Started DNS server: address=127.0.0.1:30348 network=udp writer.go:29: 2021-01-29T19:34:24.767Z [INFO] TestAgent_TokenTriggersFullSync/default: Started DNS server: address=127.0.0.1:30348 network=tcp writer.go:29: 2021-01-29T19:34:24.768Z [INFO] TestAgent_TokenTriggersFullSync/default.server: Adding LAN server: server="Node-6ae2a6f1-0459-1513-c53b-d07212569993 (Addr: tcp/127.0.0.1:30353) (DC: dc1)" writer.go:29: 2021-01-29T19:34:24.770Z [INFO] TestAgent_TokenTriggersFullSync/default: Started HTTP server: address=127.0.0.1:30349 network=tcp writer.go:29: 2021-01-29T19:34:24.772Z [INFO] TestAgent_TokenTriggersFullSync/default: started state syncer === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.777Z [INFO] TestAgent_DeregisterService_ACLDeny: Synced node info === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.779Z [WARN] TestAgent_DeregisterService: bootstrap = true: do not enable unless necessary === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.780Z [INFO] TestAgent_DeregisterService_ACLDeny: Deregistered service: service=test === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.781Z [DEBUG] TestAgent_DeregisterService.tlsutil: Update: version=1 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.781Z [INFO] TestAgent_DeregisterService_ACLDeny: Requesting shutdown === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.783Z [DEBUG] TestAgent_DeregisterService.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.784Z [INFO] TestAgent_DeregisterService_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:24.786Z [DEBUG] TestAgent_DeregisterService_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:24.788Z [DEBUG] TestAgent_DeregisterService_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:24.786Z [ERROR] TestAgent_DeregisterService_ACLDeny.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:24.788Z [DEBUG] TestAgent_DeregisterService_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:24.789Z [DEBUG] TestAgent_DeregisterService_ACLDeny.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.793Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:24.794Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest: Node info in sync === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.789Z [DEBUG] TestAgent_DeregisterService_ACLDeny.leader: stopped routine: routine="acl token reaping" === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.795Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest: Node info in sync === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.794Z [WARN] TestAgent_DeregisterService_ACLDeny.server.serf.lan: serf: Shutdown without a Leave === RUN TestAgent_ServiceMaintenance_BadRequest/not_enabled === RUN TestAgent_ServiceMaintenance_BadRequest/no_service_id === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.794Z [DEBUG] TestAgent_DeregisterService_ACLDeny.leader: stopped routine: routine="CA root pruning" === RUN TestAgent_ServiceMaintenance_BadRequest/bad_service_id === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.800Z [INFO] TestAgent_ServiceMaintenance_BadRequest: Requesting shutdown === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.791Z [WARN] TestAgent_TokenTriggersFullSync/default.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.800Z [INFO] TestAgent_DeregisterService.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:682b3e94-ce9c-90cb-9ee3-1d9f931add79 Address:127.0.0.1:30347}]" === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.801Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server: shutting down server writer.go:29: 2021-01-29T19:34:24.804Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.804Z [INFO] TestAgent_DeregisterService.server.raft: entering follower state: follower="Node at 127.0.0.1:30347 [Follower]" leader= === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.805Z [WARN] TestAgent_DeregisterService_ACLDeny.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.805Z [INFO] TestAgent_DeregisterService.server.serf.wan: serf: EventMemberJoin: Node-682b3e94-ce9c-90cb-9ee3-1d9f931add79.dc1 127.0.0.1 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.808Z [INFO] TestAgent_DeregisterService_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:24.809Z [INFO] TestAgent_DeregisterService_ACLDeny: consul server down === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.810Z [INFO] TestAgent_DeregisterService.server.serf.lan: serf: EventMemberJoin: Node-682b3e94-ce9c-90cb-9ee3-1d9f931add79 127.0.0.1 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.810Z [INFO] TestAgent_DeregisterService_ACLDeny: shutdown complete === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.802Z [INFO] TestAgent_TokenTriggersFullSync/default.server.raft: entering candidate state: node="Node at 127.0.0.1:30353 [Candidate]" term=2 === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.811Z [INFO] TestAgent_DeregisterService_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30330 network=tcp writer.go:29: 2021-01-29T19:34:24.812Z [INFO] TestAgent_DeregisterService_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30330 network=udp === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.813Z [INFO] TestAgent_DeregisterService.server: Handled event for server in area: event=member-join server=Node-682b3e94-ce9c-90cb-9ee3-1d9f931add79.dc1 area=wan writer.go:29: 2021-01-29T19:34:24.813Z [INFO] TestAgent_DeregisterService.server: Adding LAN server: server="Node-682b3e94-ce9c-90cb-9ee3-1d9f931add79 (Addr: tcp/127.0.0.1:30347) (DC: dc1)" === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:24.814Z [INFO] TestAgent_DeregisterService_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30331 network=tcp === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.806Z [WARN] TestAgent_ServiceMaintenance_BadRequest.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.815Z [INFO] TestAgent_DeregisterService: Started DNS server: address=127.0.0.1:30342 network=tcp === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.817Z [DEBUG] TestAgent_TokenTriggersFullSync/default.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.818Z [INFO] TestAgent_DeregisterService: Started DNS server: address=127.0.0.1:30342 network=udp === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.817Z [DEBUG] TestAgent_ServiceMaintenance_BadRequest.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.820Z [INFO] TestAgent_DeregisterService: Started HTTP server: address=127.0.0.1:30343 network=tcp === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.819Z [WARN] TestAgent_ServiceMaintenance_BadRequest.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.821Z [INFO] TestAgent_DeregisterService: started state syncer === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.819Z [DEBUG] TestAgent_TokenTriggersFullSync/default.server.raft: votes: needed=1 === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:24.824Z [INFO] TestAgent_ServiceMaintenance_BadRequest.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:24.824Z [INFO] TestAgent_ServiceMaintenance_BadRequest: consul server down writer.go:29: 2021-01-29T19:34:24.825Z [INFO] TestAgent_ServiceMaintenance_BadRequest: shutdown complete writer.go:29: 2021-01-29T19:34:24.826Z [INFO] TestAgent_ServiceMaintenance_BadRequest: Stopping server: protocol=DNS address=127.0.0.1:30306 network=tcp writer.go:29: 2021-01-29T19:34:24.828Z [INFO] TestAgent_ServiceMaintenance_BadRequest: Stopping server: protocol=DNS address=127.0.0.1:30306 network=udp writer.go:29: 2021-01-29T19:34:24.829Z [INFO] TestAgent_ServiceMaintenance_BadRequest: Stopping server: protocol=HTTP address=127.0.0.1:30307 network=tcp === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.824Z [DEBUG] TestAgent_TokenTriggersFullSync/default.server.raft: vote granted: from=6ae2a6f1-0459-1513-c53b-d07212569993 term=2 tally=1 writer.go:29: 2021-01-29T19:34:24.831Z [INFO] TestAgent_TokenTriggersFullSync/default.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:24.832Z [INFO] TestAgent_TokenTriggersFullSync/default.server.raft: entering leader state: leader="Node at 127.0.0.1:30353 [Leader]" writer.go:29: 2021-01-29T19:34:24.834Z [INFO] TestAgent_TokenTriggersFullSync/default.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:24.835Z [INFO] TestAgent_TokenTriggersFullSync/default.server: New leader elected: payload=Node-6ae2a6f1-0459-1513-c53b-d07212569993 writer.go:29: 2021-01-29T19:34:24.838Z [INFO] TestAgent_TokenTriggersFullSync/default.server: initializing acls writer.go:29: 2021-01-29T19:34:24.842Z [INFO] TestAgent_TokenTriggersFullSync/default.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:24.843Z [WARN] TestAgent_TokenTriggersFullSync/default.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:24.845Z [INFO] TestAgent_TokenTriggersFullSync/default.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:24.848Z [INFO] TestAgent_TokenTriggersFullSync/default.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:24.849Z [INFO] TestAgent_TokenTriggersFullSync/default.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:24.852Z [INFO] TestAgent_TokenTriggersFullSync/default.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:24.854Z [INFO] TestAgent_TokenTriggersFullSync/default.server.serf.lan: serf: EventMemberUpdate: Node-6ae2a6f1-0459-1513-c53b-d07212569993 writer.go:29: 2021-01-29T19:34:24.857Z [INFO] TestAgent_TokenTriggersFullSync/default.server: Updating LAN server: server="Node-6ae2a6f1-0459-1513-c53b-d07212569993 (Addr: tcp/127.0.0.1:30353) (DC: dc1)" writer.go:29: 2021-01-29T19:34:24.858Z [INFO] TestAgent_TokenTriggersFullSync/default.server.serf.wan: serf: EventMemberUpdate: Node-6ae2a6f1-0459-1513-c53b-d07212569993.dc1 writer.go:29: 2021-01-29T19:34:24.860Z [INFO] TestAgent_TokenTriggersFullSync/default.server: Handled event for server in area: event=member-update server=Node-6ae2a6f1-0459-1513-c53b-d07212569993.dc1 area=wan === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.865Z [WARN] TestAgent_DeregisterService.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:24.866Z [INFO] TestAgent_DeregisterService.server.raft: entering candidate state: node="Node at 127.0.0.1:30347 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:24.868Z [DEBUG] TestAgent_DeregisterService.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:24.870Z [DEBUG] TestAgent_DeregisterService.server.raft: vote granted: from=682b3e94-ce9c-90cb-9ee3-1d9f931add79 term=2 tally=1 writer.go:29: 2021-01-29T19:34:24.871Z [INFO] TestAgent_DeregisterService.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:24.873Z [INFO] TestAgent_DeregisterService.server.raft: entering leader state: leader="Node at 127.0.0.1:30347 [Leader]" writer.go:29: 2021-01-29T19:34:24.874Z [INFO] TestAgent_DeregisterService.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:24.875Z [INFO] TestAgent_DeregisterService.server: New leader elected: payload=Node-682b3e94-ce9c-90cb-9ee3-1d9f931add79 writer.go:29: 2021-01-29T19:34:24.880Z [DEBUG] TestAgent_DeregisterService.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30347 === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.884Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.892Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.896Z [INFO] TestAgent_TokenTriggersFullSync/default.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:24.897Z [INFO] TestAgent_TokenTriggersFullSync/default.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:24.899Z [DEBUG] TestAgent_TokenTriggersFullSync/default.server: Skipping self join check for node since the cluster is too small: node=Node-6ae2a6f1-0459-1513-c53b-d07212569993 writer.go:29: 2021-01-29T19:34:24.901Z [INFO] TestAgent_TokenTriggersFullSync/default.server: member joined, marking health alive: member=Node-6ae2a6f1-0459-1513-c53b-d07212569993 === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.903Z [INFO] TestAgent_DeregisterService.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:24.904Z [INFO] TestAgent_DeregisterService.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:24.906Z [DEBUG] TestAgent_DeregisterService.server: Skipping self join check for node since the cluster is too small: node=Node-682b3e94-ce9c-90cb-9ee3-1d9f931add79 writer.go:29: 2021-01-29T19:34:24.907Z [INFO] TestAgent_DeregisterService.server: member joined, marking health alive: member=Node-682b3e94-ce9c-90cb-9ee3-1d9f931add79 === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:24.909Z [DEBUG] TestAgent_TokenTriggersFullSync/default.server: Skipping self join check for node since the cluster is too small: node=Node-6ae2a6f1-0459-1513-c53b-d07212569993 === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:24.919Z [DEBUG] TestAgent_DeregisterService: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:24.921Z [INFO] TestAgent_DeregisterService: Synced node info writer.go:29: 2021-01-29T19:34:24.922Z [DEBUG] TestAgent_DeregisterService: Node info in sync writer.go:29: 2021-01-29T19:34:24.930Z [DEBUG] TestAgent_DeregisterService: removed service: service=test writer.go:29: 2021-01-29T19:34:24.931Z [DEBUG] TestAgent_DeregisterService: Node info in sync writer.go:29: 2021-01-29T19:34:24.932Z [INFO] TestAgent_DeregisterService: Deregistered service: service=test writer.go:29: 2021-01-29T19:34:24.933Z [INFO] TestAgent_DeregisterService: Requesting shutdown writer.go:29: 2021-01-29T19:34:24.934Z [INFO] TestAgent_DeregisterService.server: shutting down server writer.go:29: 2021-01-29T19:34:24.935Z [DEBUG] TestAgent_DeregisterService.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:24.936Z [WARN] TestAgent_DeregisterService.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:24.936Z [DEBUG] TestAgent_DeregisterService.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:24.939Z [WARN] TestAgent_DeregisterService.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:24.941Z [INFO] TestAgent_DeregisterService.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:24.941Z [INFO] TestAgent_DeregisterService: consul server down writer.go:29: 2021-01-29T19:34:24.943Z [INFO] TestAgent_DeregisterService: shutdown complete writer.go:29: 2021-01-29T19:34:24.944Z [INFO] TestAgent_DeregisterService: Stopping server: protocol=DNS address=127.0.0.1:30342 network=tcp writer.go:29: 2021-01-29T19:34:24.945Z [INFO] TestAgent_DeregisterService: Stopping server: protocol=DNS address=127.0.0.1:30342 network=udp writer.go:29: 2021-01-29T19:34:24.946Z [INFO] TestAgent_DeregisterService: Stopping server: protocol=HTTP address=127.0.0.1:30343 network=tcp === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:25.028Z [DEBUG] TestAgent_TokenTriggersFullSync/default.acl: dropping check from result due to ACLs: check=serfHealth writer.go:29: 2021-01-29T19:34:25.031Z [WARN] TestAgent_TokenTriggersFullSync/default: Node info update blocked by ACLs: node=6ae2a6f1-0459-1513-c53b-d07212569993 accessorID= writer.go:29: 2021-01-29T19:34:25.032Z [DEBUG] TestAgent_TokenTriggersFullSync/default: Node info in sync writer.go:29: 2021-01-29T19:34:25.208Z [DEBUG] TestAgent_TokenTriggersFullSync/default.acl: dropping node from result due to ACLs: node=Node-6ae2a6f1-0459-1513-c53b-d07212569993 writer.go:29: 2021-01-29T19:34:25.211Z [DEBUG] TestAgent_TokenTriggersFullSync/default.acl: dropping node from result due to ACLs: node=Node-6ae2a6f1-0459-1513-c53b-d07212569993 writer.go:29: 2021-01-29T19:34:25.227Z [INFO] TestAgent_TokenTriggersFullSync/default: Updated agent's ACL token: token=default === CONT TestAgent_DeregisterService_ACLDeny writer.go:29: 2021-01-29T19:34:25.317Z [INFO] TestAgent_DeregisterService_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:25.318Z [INFO] TestAgent_DeregisterService_ACLDeny: Endpoints down --- PASS: TestAgent_DeregisterService_ACLDeny (0.92s) --- PASS: TestAgent_DeregisterService_ACLDeny/no_token (0.00s) --- PASS: TestAgent_DeregisterService_ACLDeny/root_token (0.04s) === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.326Z [WARN] TestAgent_UpdateCheck_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:25.327Z [WARN] TestAgent_UpdateCheck_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:25.328Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:25.330Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_ServiceMaintenance_BadRequest writer.go:29: 2021-01-29T19:34:25.331Z [INFO] TestAgent_ServiceMaintenance_BadRequest: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:25.331Z [INFO] TestAgent_ServiceMaintenance_BadRequest: Endpoints down --- PASS: TestAgent_ServiceMaintenance_BadRequest (1.00s) --- PASS: TestAgent_ServiceMaintenance_BadRequest/not_enabled (0.00s) --- PASS: TestAgent_ServiceMaintenance_BadRequest/no_service_id (0.00s) --- PASS: TestAgent_ServiceMaintenance_BadRequest/bad_service_id (0.00s) === CONT TestAgent_FailCheck_ACLDeny === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.339Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c01d5595-b535-429f-f6a0-3bc9d111b24a Address:127.0.0.1:30365}]" === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.341Z [WARN] TestAgent_FailCheck_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:25.342Z [WARN] TestAgent_FailCheck_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:25.343Z [DEBUG] TestAgent_FailCheck_ACLDeny.tlsutil: Update: version=1 === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.342Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30365 [Follower]" leader= === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.344Z [DEBUG] TestAgent_FailCheck_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.346Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-c01d5595-b535-429f-f6a0-3bc9d111b24a.dc1 127.0.0.1 === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.348Z [INFO] TestAgent_FailCheck_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1a6cf4b3-373a-4cea-4593-338730b4e8ae Address:127.0.0.1:30377}]" writer.go:29: 2021-01-29T19:34:25.352Z [INFO] TestAgent_FailCheck_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30377 [Follower]" leader= writer.go:29: 2021-01-29T19:34:25.354Z [INFO] TestAgent_FailCheck_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:25.360Z [INFO] TestAgent_FailCheck_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae 127.0.0.1 === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.364Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-c01d5595-b535-429f-f6a0-3bc9d111b24a 127.0.0.1 === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.368Z [INFO] TestAgent_FailCheck_ACLDeny.server: Adding LAN server: server="Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae (Addr: tcp/127.0.0.1:30377) (DC: dc1)" writer.go:29: 2021-01-29T19:34:25.369Z [INFO] TestAgent_FailCheck_ACLDeny.server: Handled event for server in area: event=member-join server=Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae.dc1 area=wan === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.373Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: Adding LAN server: server="Node-c01d5595-b535-429f-f6a0-3bc9d111b24a (Addr: tcp/127.0.0.1:30365) (DC: dc1)" writer.go:29: 2021-01-29T19:34:25.374Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: Handled event for server in area: event=member-join server=Node-c01d5595-b535-429f-f6a0-3bc9d111b24a.dc1 area=wan === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.376Z [INFO] TestAgent_FailCheck_ACLDeny: Started DNS server: address=127.0.0.1:30372 network=tcp === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.380Z [INFO] TestAgent_UpdateCheck_ACLDeny: Started DNS server: address=127.0.0.1:30360 network=tcp writer.go:29: 2021-01-29T19:34:25.382Z [INFO] TestAgent_UpdateCheck_ACLDeny: Started DNS server: address=127.0.0.1:30360 network=udp writer.go:29: 2021-01-29T19:34:25.385Z [INFO] TestAgent_UpdateCheck_ACLDeny: Started HTTP server: address=127.0.0.1:30361 network=tcp writer.go:29: 2021-01-29T19:34:25.386Z [INFO] TestAgent_UpdateCheck_ACLDeny: started state syncer === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.378Z [INFO] TestAgent_FailCheck_ACLDeny: Started DNS server: address=127.0.0.1:30372 network=udp writer.go:29: 2021-01-29T19:34:25.393Z [INFO] TestAgent_FailCheck_ACLDeny: Started HTTP server: address=127.0.0.1:30373 network=tcp === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.395Z [WARN] TestAgent_UpdateCheck_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.396Z [INFO] TestAgent_FailCheck_ACLDeny: started state syncer === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.396Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30365 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:25.401Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:25.402Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.server.raft: vote granted: from=c01d5595-b535-429f-f6a0-3bc9d111b24a term=2 tally=1 writer.go:29: 2021-01-29T19:34:25.403Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:25.405Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30365 [Leader]" writer.go:29: 2021-01-29T19:34:25.406Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:25.407Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: New leader elected: payload=Node-c01d5595-b535-429f-f6a0-3bc9d111b24a writer.go:29: 2021-01-29T19:34:25.408Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: initializing acls === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.415Z [WARN] TestAgent_FailCheck_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:25.417Z [INFO] TestAgent_FailCheck_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30377 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:25.422Z [DEBUG] TestAgent_FailCheck_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.420Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:25.424Z [WARN] TestAgent_UpdateCheck_ACLDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.425Z [DEBUG] TestAgent_FailCheck_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:25.427Z [DEBUG] TestAgent_FailCheck_ACLDeny.server.raft: vote granted: from=1a6cf4b3-373a-4cea-4593-338730b4e8ae term=2 tally=1 === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.425Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: initializing acls === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.428Z [INFO] TestAgent_FailCheck_ACLDeny.server.raft: election won: tally=1 === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.428Z [WARN] TestAgent_UpdateCheck_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:25.427Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: Bootstrapped ACL master token from configuration === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.429Z [INFO] TestAgent_FailCheck_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30377 [Leader]" === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.431Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: Created ACL anonymous token from configuration === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.431Z [INFO] TestAgent_FailCheck_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:25.432Z [INFO] TestAgent_FailCheck_ACLDeny.server: New leader elected: payload=Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.432Z [INFO] TestAgent_UpdateCheck_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:25.431Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: Created ACL anonymous token from configuration === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.433Z [INFO] TestAgent_FailCheck_ACLDeny.server: initializing acls === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.434Z [INFO] TestAgent_UpdateCheck_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:25.437Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.server: transitioning out of legacy ACL mode === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.437Z [INFO] TestAgent_FailCheck_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:25.439Z [WARN] TestAgent_FailCheck_ACLDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.437Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-c01d5595-b535-429f-f6a0-3bc9d111b24a writer.go:29: 2021-01-29T19:34:25.440Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: Updating LAN server: server="Node-c01d5595-b535-429f-f6a0-3bc9d111b24a (Addr: tcp/127.0.0.1:30365) (DC: dc1)" === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.441Z [INFO] TestAgent_FailCheck_ACLDeny.server: Bootstrapped ACL master token from configuration === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.440Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-c01d5595-b535-429f-f6a0-3bc9d111b24a.dc1 === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.443Z [INFO] TestAgent_FailCheck_ACLDeny.server: Created ACL anonymous token from configuration === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.440Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-c01d5595-b535-429f-f6a0-3bc9d111b24a === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.444Z [INFO] TestAgent_FailCheck_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:25.446Z [INFO] TestAgent_FailCheck_ACLDeny.leader: started routine: routine="acl token reaping" === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.445Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-c01d5595-b535-429f-f6a0-3bc9d111b24a.dc1 === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:25.447Z [INFO] TestAgent_DeregisterService: Waiting for endpoints to shut down === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.443Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: Handled event for server in area: event=member-update server=Node-c01d5595-b535-429f-f6a0-3bc9d111b24a.dc1 area=wan === CONT TestAgent_DeregisterService writer.go:29: 2021-01-29T19:34:25.448Z [INFO] TestAgent_DeregisterService: Endpoints down --- PASS: TestAgent_DeregisterService (1.04s) === CONT TestAgent_HealthServiceByName === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.445Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: Updating LAN server: server="Node-c01d5595-b535-429f-f6a0-3bc9d111b24a (Addr: tcp/127.0.0.1:30365) (DC: dc1)" writer.go:29: 2021-01-29T19:34:25.450Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: Handled event for server in area: event=member-update server=Node-c01d5595-b535-429f-f6a0-3bc9d111b24a.dc1 area=wan writer.go:29: 2021-01-29T19:34:25.452Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.456Z [WARN] TestAgent_HealthServiceByName: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:25.457Z [DEBUG] TestAgent_HealthServiceByName.tlsutil: Update: version=1 === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.457Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.458Z [DEBUG] TestAgent_HealthServiceByName.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.459Z [INFO] TestAgent_UpdateCheck_ACLDeny.leader: started routine: routine="CA root pruning" === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.459Z [INFO] TestAgent_FailCheck_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.460Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-c01d5595-b535-429f-f6a0-3bc9d111b24a === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.461Z [INFO] TestAgent_FailCheck_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae.dc1 writer.go:29: 2021-01-29T19:34:25.461Z [INFO] TestAgent_FailCheck_ACLDeny.server: Updating LAN server: server="Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae (Addr: tcp/127.0.0.1:30377) (DC: dc1)" writer.go:29: 2021-01-29T19:34:25.462Z [INFO] TestAgent_FailCheck_ACLDeny.server: Handled event for server in area: event=member-update server=Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae.dc1 area=wan writer.go:29: 2021-01-29T19:34:25.465Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.465Z [INFO] TestAgent_HealthServiceByName.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:856af2a9-1ea8-799a-2ba3-401506183bf9 Address:127.0.0.1:30395}]" === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.461Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: member joined, marking health alive: member=Node-c01d5595-b535-429f-f6a0-3bc9d111b24a === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.467Z [INFO] TestAgent_HealthServiceByName.server.serf.wan: serf: EventMemberJoin: Node-856af2a9-1ea8-799a-2ba3-401506183bf9.dc1 127.0.0.1 === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.468Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-c01d5595-b535-429f-f6a0-3bc9d111b24a === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.467Z [INFO] TestAgent_HealthServiceByName.server.raft: entering follower state: follower="Node at 127.0.0.1:30395 [Follower]" leader= === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.469Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-c01d5595-b535-429f-f6a0-3bc9d111b24a === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.471Z [INFO] TestAgent_FailCheck_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.471Z [INFO] TestAgent_HealthServiceByName.server.serf.lan: serf: EventMemberJoin: Node-856af2a9-1ea8-799a-2ba3-401506183bf9 127.0.0.1 === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.473Z [INFO] TestAgent_FailCheck_ACLDeny.leader: started routine: routine="CA root pruning" === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.474Z [INFO] TestAgent_HealthServiceByName.server: Adding LAN server: server="Node-856af2a9-1ea8-799a-2ba3-401506183bf9 (Addr: tcp/127.0.0.1:30395) (DC: dc1)" === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.474Z [DEBUG] TestAgent_FailCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.476Z [INFO] TestAgent_HealthServiceByName.server: Handled event for server in area: event=member-join server=Node-856af2a9-1ea8-799a-2ba3-401506183bf9.dc1 area=wan === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.476Z [INFO] TestAgent_FailCheck_ACLDeny.server: member joined, marking health alive: member=Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.478Z [INFO] TestAgent_HealthServiceByName: Started DNS server: address=127.0.0.1:30390 network=udp writer.go:29: 2021-01-29T19:34:25.479Z [INFO] TestAgent_HealthServiceByName: Started DNS server: address=127.0.0.1:30390 network=tcp writer.go:29: 2021-01-29T19:34:25.480Z [INFO] TestAgent_HealthServiceByName: Started HTTP server: address=127.0.0.1:30391 network=tcp writer.go:29: 2021-01-29T19:34:25.483Z [INFO] TestAgent_HealthServiceByName: started state syncer === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.479Z [DEBUG] TestAgent_FailCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae writer.go:29: 2021-01-29T19:34:25.482Z [DEBUG] TestAgent_FailCheck_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:25.488Z [INFO] TestAgent_FailCheck_ACLDeny: Synced node info === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.526Z [WARN] TestAgent_HealthServiceByName.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:25.527Z [INFO] TestAgent_HealthServiceByName.server.raft: entering candidate state: node="Node at 127.0.0.1:30395 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:25.530Z [DEBUG] TestAgent_HealthServiceByName.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:25.532Z [DEBUG] TestAgent_HealthServiceByName.server.raft: vote granted: from=856af2a9-1ea8-799a-2ba3-401506183bf9 term=2 tally=1 writer.go:29: 2021-01-29T19:34:25.533Z [INFO] TestAgent_HealthServiceByName.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:25.535Z [INFO] TestAgent_HealthServiceByName.server.raft: entering leader state: leader="Node at 127.0.0.1:30395 [Leader]" writer.go:29: 2021-01-29T19:34:25.536Z [INFO] TestAgent_HealthServiceByName.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:25.537Z [INFO] TestAgent_HealthServiceByName.server: New leader elected: payload=Node-856af2a9-1ea8-799a-2ba3-401506183bf9 writer.go:29: 2021-01-29T19:34:25.540Z [DEBUG] TestAgent_HealthServiceByName.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30395 writer.go:29: 2021-01-29T19:34:25.545Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:25.551Z [INFO] TestAgent_HealthServiceByName.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:25.553Z [INFO] TestAgent_HealthServiceByName.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:25.554Z [DEBUG] TestAgent_HealthServiceByName.server: Skipping self join check for node since the cluster is too small: node=Node-856af2a9-1ea8-799a-2ba3-401506183bf9 writer.go:29: 2021-01-29T19:34:25.557Z [INFO] TestAgent_HealthServiceByName.server: member joined, marking health alive: member=Node-856af2a9-1ea8-799a-2ba3-401506183bf9 === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.603Z [DEBUG] TestAgent_UpdateCheck_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:25.606Z [INFO] TestAgent_UpdateCheck_ACLDeny: Synced node info writer.go:29: 2021-01-29T19:34:25.728Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.acl: dropping node from result due to ACLs: node=Node-c01d5595-b535-429f-f6a0-3bc9d111b24a writer.go:29: 2021-01-29T19:34:25.733Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.acl: dropping node from result due to ACLs: node=Node-c01d5595-b535-429f-f6a0-3bc9d111b24a === RUN TestAgent_UpdateCheck_ACLDeny/no_token === RUN TestAgent_UpdateCheck_ACLDeny/root_token === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.743Z [DEBUG] TestAgent_UpdateCheck_ACLDeny: Check status updated: check=test status=passing writer.go:29: 2021-01-29T19:34:25.745Z [DEBUG] TestAgent_UpdateCheck_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:25.746Z [WARN] TestAgent_UpdateCheck_ACLDeny: Check registration blocked by ACLs: check=test accessorID= writer.go:29: 2021-01-29T19:34:25.747Z [INFO] TestAgent_UpdateCheck_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:25.748Z [INFO] TestAgent_UpdateCheck_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:25.749Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:25.751Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:25.752Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:25.754Z [WARN] TestAgent_UpdateCheck_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:25.754Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:25.754Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:25.754Z [DEBUG] TestAgent_UpdateCheck_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:25.759Z [WARN] TestAgent_UpdateCheck_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:25.765Z [INFO] TestAgent_UpdateCheck_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:25.765Z [INFO] TestAgent_UpdateCheck_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:25.766Z [INFO] TestAgent_UpdateCheck_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:25.767Z [INFO] TestAgent_UpdateCheck_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30360 network=tcp writer.go:29: 2021-01-29T19:34:25.768Z [INFO] TestAgent_UpdateCheck_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30360 network=udp writer.go:29: 2021-01-29T19:34:25.770Z [INFO] TestAgent_UpdateCheck_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30361 network=tcp === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.776Z [DEBUG] TestAgent_FailCheck_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:25.778Z [DEBUG] TestAgent_FailCheck_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:25.779Z [DEBUG] TestAgent_FailCheck_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:25.816Z [DEBUG] TestAgent_FailCheck_ACLDeny.acl: dropping node from result due to ACLs: node=Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae writer.go:29: 2021-01-29T19:34:25.824Z [DEBUG] TestAgent_FailCheck_ACLDeny.acl: dropping node from result due to ACLs: node=Node-1a6cf4b3-373a-4cea-4593-338730b4e8ae === RUN TestAgent_FailCheck_ACLDeny/no_token === RUN TestAgent_FailCheck_ACLDeny/root_token === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:25.827Z [DEBUG] TestAgent_FailCheck_ACLDeny: Check status updated: check=test status=critical writer.go:29: 2021-01-29T19:34:25.828Z [DEBUG] TestAgent_FailCheck_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:25.829Z [WARN] TestAgent_FailCheck_ACLDeny: Check registration blocked by ACLs: check=test accessorID= writer.go:29: 2021-01-29T19:34:25.830Z [DEBUG] TestAgent_FailCheck_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:25.831Z [WARN] TestAgent_FailCheck_ACLDeny: Check registration blocked by ACLs: check=test accessorID= writer.go:29: 2021-01-29T19:34:25.832Z [DEBUG] TestAgent_FailCheck_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:25.833Z [DEBUG] TestAgent_FailCheck_ACLDeny: Check in sync: check=test writer.go:29: 2021-01-29T19:34:25.832Z [INFO] TestAgent_FailCheck_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:25.835Z [INFO] TestAgent_FailCheck_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:25.837Z [DEBUG] TestAgent_FailCheck_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:25.838Z [DEBUG] TestAgent_FailCheck_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:25.839Z [DEBUG] TestAgent_FailCheck_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:25.838Z [DEBUG] TestAgent_FailCheck_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:25.839Z [DEBUG] TestAgent_FailCheck_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:25.840Z [WARN] TestAgent_FailCheck_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:25.840Z [DEBUG] TestAgent_FailCheck_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:25.852Z [WARN] TestAgent_FailCheck_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:25.856Z [INFO] TestAgent_FailCheck_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:25.856Z [INFO] TestAgent_FailCheck_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:25.859Z [INFO] TestAgent_FailCheck_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:25.860Z [INFO] TestAgent_FailCheck_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30372 network=tcp writer.go:29: 2021-01-29T19:34:25.862Z [INFO] TestAgent_FailCheck_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30372 network=udp writer.go:29: 2021-01-29T19:34:25.864Z [INFO] TestAgent_FailCheck_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30373 network=tcp === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:25.953Z [DEBUG] TestAgent_HealthServiceByName: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:25.956Z [INFO] TestAgent_HealthServiceByName: Synced node info === RUN TestAgent_HealthServiceByName/passing_checks === RUN TestAgent_HealthServiceByName/passing_checks/format=text === RUN TestAgent_HealthServiceByName/passing_checks/format=json === RUN TestAgent_HealthServiceByName/warning_checks === RUN TestAgent_HealthServiceByName/warning_checks/format=text === RUN TestAgent_HealthServiceByName/warning_checks/format=json === RUN TestAgent_HealthServiceByName/critical_checks === RUN TestAgent_HealthServiceByName/critical_checks/format=text === RUN TestAgent_HealthServiceByName/critical_checks/format=json === RUN TestAgent_HealthServiceByName/unknown_serviceName === RUN TestAgent_HealthServiceByName/unknown_serviceName/format=text === RUN TestAgent_HealthServiceByName/unknown_serviceName/format=json === RUN TestAgent_HealthServiceByName/critical_check_on_node === RUN TestAgent_HealthServiceByName/critical_check_on_node/format=text === RUN TestAgent_HealthServiceByName/critical_check_on_node/format=json === RUN TestAgent_HealthServiceByName/maintenance_check_on_node === RUN TestAgent_HealthServiceByName/maintenance_check_on_node/format=text === RUN TestAgent_HealthServiceByName/maintenance_check_on_node/format=json === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:26.009Z [INFO] TestAgent_HealthServiceByName: Requesting shutdown writer.go:29: 2021-01-29T19:34:26.011Z [INFO] TestAgent_HealthServiceByName.server: shutting down server writer.go:29: 2021-01-29T19:34:26.013Z [DEBUG] TestAgent_HealthServiceByName.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.015Z [WARN] TestAgent_HealthServiceByName.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.016Z [DEBUG] TestAgent_HealthServiceByName.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.019Z [WARN] TestAgent_HealthServiceByName.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.022Z [INFO] TestAgent_HealthServiceByName.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:26.022Z [INFO] TestAgent_HealthServiceByName: consul server down writer.go:29: 2021-01-29T19:34:26.026Z [INFO] TestAgent_HealthServiceByName: shutdown complete writer.go:29: 2021-01-29T19:34:26.027Z [INFO] TestAgent_HealthServiceByName: Stopping server: protocol=DNS address=127.0.0.1:30390 network=tcp writer.go:29: 2021-01-29T19:34:26.029Z [INFO] TestAgent_HealthServiceByName: Stopping server: protocol=DNS address=127.0.0.1:30390 network=udp writer.go:29: 2021-01-29T19:34:26.031Z [INFO] TestAgent_HealthServiceByName: Stopping server: protocol=HTTP address=127.0.0.1:30391 network=tcp === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:26.040Z [DEBUG] TestAgent_TokenTriggersFullSync/default.acl: dropping service from result due to ACLs: service="{consul {}}" writer.go:29: 2021-01-29T19:34:26.043Z [DEBUG] TestAgent_TokenTriggersFullSync/default: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:26.045Z [INFO] TestAgent_TokenTriggersFullSync/default: Synced node info writer.go:29: 2021-01-29T19:34:26.063Z [INFO] TestAgent_TokenTriggersFullSync/default: Requesting shutdown writer.go:29: 2021-01-29T19:34:26.066Z [INFO] TestAgent_TokenTriggersFullSync/default.server: shutting down server writer.go:29: 2021-01-29T19:34:26.068Z [DEBUG] TestAgent_TokenTriggersFullSync/default.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:26.069Z [DEBUG] TestAgent_TokenTriggersFullSync/default.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:26.070Z [DEBUG] TestAgent_TokenTriggersFullSync/default.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.069Z [DEBUG] TestAgent_TokenTriggersFullSync/default.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:26.071Z [DEBUG] TestAgent_TokenTriggersFullSync/default.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:26.072Z [WARN] TestAgent_TokenTriggersFullSync/default.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.072Z [DEBUG] TestAgent_TokenTriggersFullSync/default.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.077Z [WARN] TestAgent_TokenTriggersFullSync/default.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.081Z [INFO] TestAgent_TokenTriggersFullSync/default.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:26.081Z [INFO] TestAgent_TokenTriggersFullSync/default: consul server down writer.go:29: 2021-01-29T19:34:26.084Z [INFO] TestAgent_TokenTriggersFullSync/default: shutdown complete writer.go:29: 2021-01-29T19:34:26.085Z [INFO] TestAgent_TokenTriggersFullSync/default: Stopping server: protocol=DNS address=127.0.0.1:30348 network=tcp writer.go:29: 2021-01-29T19:34:26.087Z [INFO] TestAgent_TokenTriggersFullSync/default: Stopping server: protocol=DNS address=127.0.0.1:30348 network=udp writer.go:29: 2021-01-29T19:34:26.088Z [INFO] TestAgent_TokenTriggersFullSync/default: Stopping server: protocol=HTTP address=127.0.0.1:30349 network=tcp === CONT TestAgent_UpdateCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.271Z [INFO] TestAgent_UpdateCheck_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:26.275Z [INFO] TestAgent_UpdateCheck_ACLDeny: Endpoints down --- PASS: TestAgent_UpdateCheck_ACLDeny (0.96s) --- PASS: TestAgent_UpdateCheck_ACLDeny/no_token (0.00s) --- PASS: TestAgent_UpdateCheck_ACLDeny/root_token (0.01s) === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.305Z [WARN] TestAgent_WarnCheck_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:26.308Z [WARN] TestAgent_WarnCheck_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:26.311Z [DEBUG] TestAgent_WarnCheck_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:26.316Z [DEBUG] TestAgent_WarnCheck_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:26.327Z [INFO] TestAgent_WarnCheck_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:47283c4c-9193-05dd-abcd-55c865e1a985 Address:127.0.0.1:30389}]" writer.go:29: 2021-01-29T19:34:26.333Z [INFO] TestAgent_WarnCheck_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30389 [Follower]" leader= writer.go:29: 2021-01-29T19:34:26.336Z [INFO] TestAgent_WarnCheck_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-47283c4c-9193-05dd-abcd-55c865e1a985.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:26.343Z [INFO] TestAgent_WarnCheck_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-47283c4c-9193-05dd-abcd-55c865e1a985 127.0.0.1 writer.go:29: 2021-01-29T19:34:26.348Z [INFO] TestAgent_WarnCheck_ACLDeny.server: Handled event for server in area: event=member-join server=Node-47283c4c-9193-05dd-abcd-55c865e1a985.dc1 area=wan writer.go:29: 2021-01-29T19:34:26.348Z [INFO] TestAgent_WarnCheck_ACLDeny.server: Adding LAN server: server="Node-47283c4c-9193-05dd-abcd-55c865e1a985 (Addr: tcp/127.0.0.1:30389) (DC: dc1)" writer.go:29: 2021-01-29T19:34:26.349Z [INFO] TestAgent_WarnCheck_ACLDeny: Started DNS server: address=127.0.0.1:30384 network=udp writer.go:29: 2021-01-29T19:34:26.352Z [INFO] TestAgent_WarnCheck_ACLDeny: Started DNS server: address=127.0.0.1:30384 network=tcp writer.go:29: 2021-01-29T19:34:26.354Z [INFO] TestAgent_WarnCheck_ACLDeny: Started HTTP server: address=127.0.0.1:30385 network=tcp writer.go:29: 2021-01-29T19:34:26.355Z [INFO] TestAgent_WarnCheck_ACLDeny: started state syncer === CONT TestAgent_FailCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.366Z [INFO] TestAgent_FailCheck_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:26.367Z [INFO] TestAgent_FailCheck_ACLDeny: Endpoints down --- PASS: TestAgent_FailCheck_ACLDeny (1.04s) --- PASS: TestAgent_FailCheck_ACLDeny/no_token (0.00s) --- PASS: TestAgent_FailCheck_ACLDeny/root_token (0.01s) === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.378Z [WARN] TestAgent_WarnCheck: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:26.395Z [DEBUG] TestAgent_WarnCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:26.397Z [DEBUG] TestAgent_WarnCheck.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.399Z [DEBUG] TestAgent_WarnCheck_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:26.403Z [WARN] TestAgent_WarnCheck_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:26.405Z [INFO] TestAgent_WarnCheck_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30389 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:26.410Z [DEBUG] TestAgent_WarnCheck_ACLDeny.server.raft: votes: needed=1 === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.411Z [INFO] TestAgent_WarnCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:15130d0f-504f-2ce8-91df-0a9977dc6298 Address:127.0.0.1:30383}]" === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.412Z [DEBUG] TestAgent_WarnCheck_ACLDeny.server.raft: vote granted: from=47283c4c-9193-05dd-abcd-55c865e1a985 term=2 tally=1 === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.413Z [INFO] TestAgent_WarnCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:30383 [Follower]" leader= === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.414Z [INFO] TestAgent_WarnCheck_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:26.416Z [INFO] TestAgent_WarnCheck_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30389 [Leader]" === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.417Z [INFO] TestAgent_WarnCheck.server.serf.wan: serf: EventMemberJoin: Node-15130d0f-504f-2ce8-91df-0a9977dc6298.dc1 127.0.0.1 === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.419Z [INFO] TestAgent_WarnCheck_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:26.420Z [INFO] TestAgent_WarnCheck_ACLDeny.server: New leader elected: payload=Node-47283c4c-9193-05dd-abcd-55c865e1a985 writer.go:29: 2021-01-29T19:34:26.424Z [INFO] TestAgent_WarnCheck_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:26.429Z [INFO] TestAgent_WarnCheck_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:26.430Z [WARN] TestAgent_WarnCheck_ACLDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.420Z [INFO] TestAgent_WarnCheck.server.serf.lan: serf: EventMemberJoin: Node-15130d0f-504f-2ce8-91df-0a9977dc6298 127.0.0.1 writer.go:29: 2021-01-29T19:34:26.443Z [INFO] TestAgent_WarnCheck.server: Adding LAN server: server="Node-15130d0f-504f-2ce8-91df-0a9977dc6298 (Addr: tcp/127.0.0.1:30383) (DC: dc1)" writer.go:29: 2021-01-29T19:34:26.445Z [INFO] TestAgent_WarnCheck.server: Handled event for server in area: event=member-join server=Node-15130d0f-504f-2ce8-91df-0a9977dc6298.dc1 area=wan === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.448Z [ERROR] TestAgent_WarnCheck_ACLDeny.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.449Z [INFO] TestAgent_WarnCheck: Started DNS server: address=127.0.0.1:30378 network=tcp === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.448Z [INFO] TestAgent_WarnCheck_ACLDeny.server: Bootstrapped ACL master token from configuration === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.451Z [INFO] TestAgent_WarnCheck: Started DNS server: address=127.0.0.1:30378 network=udp writer.go:29: 2021-01-29T19:34:26.452Z [INFO] TestAgent_WarnCheck: Started HTTP server: address=127.0.0.1:30379 network=tcp === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.453Z [INFO] TestAgent_WarnCheck_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:26.455Z [INFO] TestAgent_WarnCheck_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.454Z [INFO] TestAgent_WarnCheck: started state syncer === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.456Z [INFO] TestAgent_WarnCheck_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:26.458Z [INFO] TestAgent_WarnCheck_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-47283c4c-9193-05dd-abcd-55c865e1a985 writer.go:29: 2021-01-29T19:34:26.460Z [INFO] TestAgent_WarnCheck_ACLDeny.server: Updating LAN server: server="Node-47283c4c-9193-05dd-abcd-55c865e1a985 (Addr: tcp/127.0.0.1:30389) (DC: dc1)" writer.go:29: 2021-01-29T19:34:26.462Z [INFO] TestAgent_WarnCheck_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-47283c4c-9193-05dd-abcd-55c865e1a985.dc1 writer.go:29: 2021-01-29T19:34:26.464Z [INFO] TestAgent_WarnCheck_ACLDeny.server: Handled event for server in area: event=member-update server=Node-47283c4c-9193-05dd-abcd-55c865e1a985.dc1 area=wan writer.go:29: 2021-01-29T19:34:26.470Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.473Z [WARN] TestAgent_WarnCheck.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.474Z [INFO] TestAgent_WarnCheck_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:26.475Z [INFO] TestAgent_WarnCheck_ACLDeny.leader: started routine: routine="CA root pruning" === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.474Z [INFO] TestAgent_WarnCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:30383 [Candidate]" term=2 === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.476Z [DEBUG] TestAgent_WarnCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-47283c4c-9193-05dd-abcd-55c865e1a985 writer.go:29: 2021-01-29T19:34:26.477Z [INFO] TestAgent_WarnCheck_ACLDeny.server: member joined, marking health alive: member=Node-47283c4c-9193-05dd-abcd-55c865e1a985 === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.477Z [DEBUG] TestAgent_WarnCheck.server.raft: votes: needed=1 === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.479Z [DEBUG] TestAgent_WarnCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-47283c4c-9193-05dd-abcd-55c865e1a985 === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.479Z [DEBUG] TestAgent_WarnCheck.server.raft: vote granted: from=15130d0f-504f-2ce8-91df-0a9977dc6298 term=2 tally=1 writer.go:29: 2021-01-29T19:34:26.480Z [INFO] TestAgent_WarnCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:26.482Z [INFO] TestAgent_WarnCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:30383 [Leader]" writer.go:29: 2021-01-29T19:34:26.483Z [INFO] TestAgent_WarnCheck.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:26.484Z [INFO] TestAgent_WarnCheck.server: New leader elected: payload=Node-15130d0f-504f-2ce8-91df-0a9977dc6298 writer.go:29: 2021-01-29T19:34:26.485Z [DEBUG] TestAgent_WarnCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30383 writer.go:29: 2021-01-29T19:34:26.489Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:26.505Z [INFO] TestAgent_WarnCheck.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:26.506Z [INFO] TestAgent_WarnCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.507Z [DEBUG] TestAgent_WarnCheck.server: Skipping self join check for node since the cluster is too small: node=Node-15130d0f-504f-2ce8-91df-0a9977dc6298 writer.go:29: 2021-01-29T19:34:26.508Z [INFO] TestAgent_WarnCheck.server: member joined, marking health alive: member=Node-15130d0f-504f-2ce8-91df-0a9977dc6298 === CONT TestAgent_HealthServiceByName writer.go:29: 2021-01-29T19:34:26.533Z [INFO] TestAgent_HealthServiceByName: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:26.534Z [INFO] TestAgent_HealthServiceByName: Endpoints down --- PASS: TestAgent_HealthServiceByName (1.09s) --- PASS: TestAgent_HealthServiceByName/passing_checks (0.01s) --- PASS: TestAgent_HealthServiceByName/passing_checks/format=text (0.00s) --- PASS: TestAgent_HealthServiceByName/passing_checks/format=json (0.00s) --- PASS: TestAgent_HealthServiceByName/warning_checks (0.01s) --- PASS: TestAgent_HealthServiceByName/warning_checks/format=text (0.00s) --- PASS: TestAgent_HealthServiceByName/warning_checks/format=json (0.00s) --- PASS: TestAgent_HealthServiceByName/critical_checks (0.01s) --- PASS: TestAgent_HealthServiceByName/critical_checks/format=text (0.00s) --- PASS: TestAgent_HealthServiceByName/critical_checks/format=json (0.00s) --- PASS: TestAgent_HealthServiceByName/unknown_serviceName (0.00s) --- PASS: TestAgent_HealthServiceByName/unknown_serviceName/format=text (0.00s) --- PASS: TestAgent_HealthServiceByName/unknown_serviceName/format=json (0.00s) --- PASS: TestAgent_HealthServiceByName/critical_check_on_node (0.01s) --- PASS: TestAgent_HealthServiceByName/critical_check_on_node/format=text (0.00s) --- PASS: TestAgent_HealthServiceByName/critical_check_on_node/format=json (0.00s) --- PASS: TestAgent_HealthServiceByName/maintenance_check_on_node (0.01s) --- PASS: TestAgent_HealthServiceByName/maintenance_check_on_node/format=text (0.00s) --- PASS: TestAgent_HealthServiceByName/maintenance_check_on_node/format=json (0.00s) === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.553Z [WARN] TestAgent_PassCheck_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:26.554Z [WARN] TestAgent_PassCheck_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:26.556Z [DEBUG] TestAgent_PassCheck_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:26.558Z [DEBUG] TestAgent_PassCheck_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:26.562Z [INFO] TestAgent_PassCheck_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec Address:127.0.0.1:30401}]" writer.go:29: 2021-01-29T19:34:26.564Z [INFO] TestAgent_PassCheck_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30401 [Follower]" leader= writer.go:29: 2021-01-29T19:34:26.564Z [INFO] TestAgent_PassCheck_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:26.568Z [INFO] TestAgent_PassCheck_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec 127.0.0.1 writer.go:29: 2021-01-29T19:34:26.570Z [INFO] TestAgent_PassCheck_ACLDeny.server: Adding LAN server: server="Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec (Addr: tcp/127.0.0.1:30401) (DC: dc1)" writer.go:29: 2021-01-29T19:34:26.571Z [INFO] TestAgent_PassCheck_ACLDeny.server: Handled event for server in area: event=member-join server=Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec.dc1 area=wan writer.go:29: 2021-01-29T19:34:26.573Z [INFO] TestAgent_PassCheck_ACLDeny: Started DNS server: address=127.0.0.1:30396 network=udp writer.go:29: 2021-01-29T19:34:26.575Z [INFO] TestAgent_PassCheck_ACLDeny: Started DNS server: address=127.0.0.1:30396 network=tcp writer.go:29: 2021-01-29T19:34:26.576Z [INFO] TestAgent_PassCheck_ACLDeny: Started HTTP server: address=127.0.0.1:30397 network=tcp writer.go:29: 2021-01-29T19:34:26.578Z [INFO] TestAgent_PassCheck_ACLDeny: started state syncer === CONT TestAgent_TokenTriggersFullSync/default writer.go:29: 2021-01-29T19:34:26.590Z [INFO] TestAgent_TokenTriggersFullSync/default: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:26.592Z [INFO] TestAgent_TokenTriggersFullSync/default: Endpoints down --- PASS: TestAgent_TokenTriggersFullSync (7.87s) --- PASS: TestAgent_TokenTriggersFullSync/acl_agent_token (2.21s) --- PASS: TestAgent_TokenTriggersFullSync/agent (2.75s) --- PASS: TestAgent_TokenTriggersFullSync/acl_token (0.86s) --- PASS: TestAgent_TokenTriggersFullSync/default (2.05s) === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.614Z [WARN] TestAgent_PassCheck: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:26.615Z [DEBUG] TestAgent_PassCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:26.618Z [DEBUG] TestAgent_PassCheck.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.622Z [DEBUG] TestAgent_PassCheck_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.624Z [INFO] TestAgent_PassCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a5253276-18a1-2ca8-777d-ba435b2c12ad Address:127.0.0.1:30413}]" writer.go:29: 2021-01-29T19:34:26.627Z [INFO] TestAgent_PassCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:30413 [Follower]" leader= === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.629Z [WARN] TestAgent_PassCheck_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:26.630Z [INFO] TestAgent_PassCheck_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30401 [Candidate]" term=2 === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.627Z [INFO] TestAgent_PassCheck.server.serf.wan: serf: EventMemberJoin: Node-a5253276-18a1-2ca8-777d-ba435b2c12ad.dc1 127.0.0.1 === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.633Z [DEBUG] TestAgent_PassCheck_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:26.634Z [DEBUG] TestAgent_PassCheck_ACLDeny.server.raft: vote granted: from=24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec term=2 tally=1 writer.go:29: 2021-01-29T19:34:26.636Z [INFO] TestAgent_PassCheck_ACLDeny.server.raft: election won: tally=1 === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.636Z [INFO] TestAgent_PassCheck.server.serf.lan: serf: EventMemberJoin: Node-a5253276-18a1-2ca8-777d-ba435b2c12ad 127.0.0.1 === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.638Z [INFO] TestAgent_PassCheck_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30401 [Leader]" writer.go:29: 2021-01-29T19:34:26.639Z [INFO] TestAgent_PassCheck_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:26.641Z [INFO] TestAgent_PassCheck_ACLDeny.server: New leader elected: payload=Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec writer.go:29: 2021-01-29T19:34:26.642Z [INFO] TestAgent_PassCheck_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:26.647Z [INFO] TestAgent_PassCheck_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:26.649Z [WARN] TestAgent_PassCheck_ACLDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.651Z [INFO] TestAgent_PassCheck.server: Adding LAN server: server="Node-a5253276-18a1-2ca8-777d-ba435b2c12ad (Addr: tcp/127.0.0.1:30413) (DC: dc1)" writer.go:29: 2021-01-29T19:34:26.651Z [INFO] TestAgent_PassCheck.server: Handled event for server in area: event=member-join server=Node-a5253276-18a1-2ca8-777d-ba435b2c12ad.dc1 area=wan writer.go:29: 2021-01-29T19:34:26.653Z [INFO] TestAgent_PassCheck: Started DNS server: address=127.0.0.1:30408 network=udp writer.go:29: 2021-01-29T19:34:26.655Z [INFO] TestAgent_PassCheck: Started DNS server: address=127.0.0.1:30408 network=tcp writer.go:29: 2021-01-29T19:34:26.656Z [INFO] TestAgent_PassCheck: Started HTTP server: address=127.0.0.1:30409 network=tcp writer.go:29: 2021-01-29T19:34:26.658Z [INFO] TestAgent_PassCheck: started state syncer === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.654Z [INFO] TestAgent_PassCheck_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:26.662Z [INFO] TestAgent_PassCheck_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:26.664Z [INFO] TestAgent_PassCheck_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:26.667Z [INFO] TestAgent_PassCheck_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:26.669Z [INFO] TestAgent_PassCheck_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec writer.go:29: 2021-01-29T19:34:26.672Z [INFO] TestAgent_PassCheck_ACLDeny.server: Updating LAN server: server="Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec (Addr: tcp/127.0.0.1:30401) (DC: dc1)" writer.go:29: 2021-01-29T19:34:26.674Z [INFO] TestAgent_PassCheck_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec.dc1 writer.go:29: 2021-01-29T19:34:26.676Z [INFO] TestAgent_PassCheck_ACLDeny.server: Handled event for server in area: event=member-update server=Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec.dc1 area=wan writer.go:29: 2021-01-29T19:34:26.686Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.692Z [WARN] TestAgent_PassCheck.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:26.695Z [INFO] TestAgent_PassCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:30413 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:26.698Z [DEBUG] TestAgent_PassCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:26.702Z [DEBUG] TestAgent_PassCheck.server.raft: vote granted: from=a5253276-18a1-2ca8-777d-ba435b2c12ad term=2 tally=1 === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.692Z [INFO] TestAgent_PassCheck_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.703Z [INFO] TestAgent_PassCheck.server.raft: election won: tally=1 === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.706Z [INFO] TestAgent_PassCheck_ACLDeny.leader: started routine: routine="CA root pruning" === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.706Z [INFO] TestAgent_PassCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:30413 [Leader]" === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.709Z [DEBUG] TestAgent_PassCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.709Z [INFO] TestAgent_PassCheck.server: cluster leadership acquired === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.711Z [INFO] TestAgent_PassCheck_ACLDeny.server: member joined, marking health alive: member=Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.710Z [INFO] TestAgent_PassCheck.server: New leader elected: payload=Node-a5253276-18a1-2ca8-777d-ba435b2c12ad writer.go:29: 2021-01-29T19:34:26.714Z [DEBUG] TestAgent_PassCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30413 === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.717Z [DEBUG] TestAgent_PassCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.721Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.727Z [DEBUG] TestAgent_WarnCheck_ACLDeny.acl: dropping node from result due to ACLs: node=Node-47283c4c-9193-05dd-abcd-55c865e1a985 === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.730Z [INFO] TestAgent_PassCheck.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.732Z [DEBUG] TestAgent_WarnCheck_ACLDeny.acl: dropping node from result due to ACLs: node=Node-47283c4c-9193-05dd-abcd-55c865e1a985 === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.732Z [INFO] TestAgent_PassCheck.leader: started routine: routine="CA root pruning" === RUN TestAgent_WarnCheck_ACLDeny/no_token === RUN TestAgent_WarnCheck_ACLDeny/root_token === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.740Z [DEBUG] TestAgent_WarnCheck_ACLDeny: Check status updated: check=test status=warning === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.734Z [DEBUG] TestAgent_PassCheck.server: Skipping self join check for node since the cluster is too small: node=Node-a5253276-18a1-2ca8-777d-ba435b2c12ad === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.740Z [DEBUG] TestAgent_WarnCheck: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.742Z [INFO] TestAgent_PassCheck.server: member joined, marking health alive: member=Node-a5253276-18a1-2ca8-777d-ba435b2c12ad === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.744Z [INFO] TestAgent_WarnCheck_ACLDeny: Synced node info === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.745Z [INFO] TestAgent_WarnCheck: Synced node info === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.746Z [WARN] TestAgent_WarnCheck_ACLDeny: Check registration blocked by ACLs: check=test accessorID= writer.go:29: 2021-01-29T19:34:26.761Z [INFO] TestAgent_WarnCheck_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:26.764Z [INFO] TestAgent_WarnCheck_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:26.765Z [DEBUG] TestAgent_WarnCheck_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.767Z [DEBUG] TestAgent_WarnCheck_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:26.769Z [DEBUG] TestAgent_WarnCheck_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:26.767Z [DEBUG] TestAgent_WarnCheck_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.769Z [DEBUG] TestAgent_WarnCheck_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:26.770Z [DEBUG] TestAgent_WarnCheck_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:26.770Z [WARN] TestAgent_WarnCheck_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.780Z [WARN] TestAgent_WarnCheck_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.783Z [INFO] TestAgent_WarnCheck_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:26.783Z [INFO] TestAgent_WarnCheck_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:26.785Z [INFO] TestAgent_WarnCheck_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:26.786Z [INFO] TestAgent_WarnCheck_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30384 network=tcp writer.go:29: 2021-01-29T19:34:26.787Z [INFO] TestAgent_WarnCheck_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30384 network=udp writer.go:29: 2021-01-29T19:34:26.788Z [INFO] TestAgent_WarnCheck_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30385 network=tcp === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:26.806Z [DEBUG] TestAgent_WarnCheck: Check status updated: check=test status=warning writer.go:29: 2021-01-29T19:34:26.807Z [DEBUG] TestAgent_WarnCheck: Node info in sync writer.go:29: 2021-01-29T19:34:26.809Z [INFO] TestAgent_WarnCheck: Synced check: check=test writer.go:29: 2021-01-29T19:34:26.810Z [INFO] TestAgent_WarnCheck: Requesting shutdown writer.go:29: 2021-01-29T19:34:26.810Z [INFO] TestAgent_WarnCheck.server: shutting down server writer.go:29: 2021-01-29T19:34:26.811Z [DEBUG] TestAgent_WarnCheck.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.812Z [WARN] TestAgent_WarnCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.812Z [DEBUG] TestAgent_WarnCheck.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.815Z [WARN] TestAgent_WarnCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.819Z [INFO] TestAgent_WarnCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:26.819Z [INFO] TestAgent_WarnCheck: consul server down writer.go:29: 2021-01-29T19:34:26.822Z [INFO] TestAgent_WarnCheck: shutdown complete writer.go:29: 2021-01-29T19:34:26.823Z [INFO] TestAgent_WarnCheck: Stopping server: protocol=DNS address=127.0.0.1:30378 network=tcp writer.go:29: 2021-01-29T19:34:26.825Z [INFO] TestAgent_WarnCheck: Stopping server: protocol=DNS address=127.0.0.1:30378 network=udp writer.go:29: 2021-01-29T19:34:26.826Z [INFO] TestAgent_WarnCheck: Stopping server: protocol=HTTP address=127.0.0.1:30379 network=tcp === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:26.871Z [DEBUG] TestAgent_PassCheck: Check status updated: check=test status=passing writer.go:29: 2021-01-29T19:34:26.874Z [INFO] TestAgent_PassCheck: Synced node info writer.go:29: 2021-01-29T19:34:26.875Z [INFO] TestAgent_PassCheck: Synced check: check=test writer.go:29: 2021-01-29T19:34:26.876Z [INFO] TestAgent_PassCheck: Requesting shutdown writer.go:29: 2021-01-29T19:34:26.876Z [INFO] TestAgent_PassCheck.server: shutting down server writer.go:29: 2021-01-29T19:34:26.877Z [DEBUG] TestAgent_PassCheck.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.878Z [WARN] TestAgent_PassCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.878Z [DEBUG] TestAgent_PassCheck.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.877Z [ERROR] TestAgent_PassCheck.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:26.881Z [WARN] TestAgent_PassCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.883Z [INFO] TestAgent_PassCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:26.885Z [INFO] TestAgent_PassCheck: consul server down writer.go:29: 2021-01-29T19:34:26.886Z [INFO] TestAgent_PassCheck: shutdown complete writer.go:29: 2021-01-29T19:34:26.887Z [INFO] TestAgent_PassCheck: Stopping server: protocol=DNS address=127.0.0.1:30408 network=tcp writer.go:29: 2021-01-29T19:34:26.890Z [INFO] TestAgent_PassCheck: Stopping server: protocol=DNS address=127.0.0.1:30408 network=udp writer.go:29: 2021-01-29T19:34:26.892Z [INFO] TestAgent_PassCheck: Stopping server: protocol=HTTP address=127.0.0.1:30409 network=tcp === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.913Z [DEBUG] TestAgent_PassCheck_ACLDeny.acl: dropping node from result due to ACLs: node=Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec writer.go:29: 2021-01-29T19:34:26.917Z [DEBUG] TestAgent_PassCheck_ACLDeny.acl: dropping node from result due to ACLs: node=Node-24eb47e3-9c77-a845-0d5c-d6ad8dfb02ec === RUN TestAgent_PassCheck_ACLDeny/no_token === RUN TestAgent_PassCheck_ACLDeny/root_token === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:26.926Z [DEBUG] TestAgent_PassCheck_ACLDeny: Check status updated: check=test status=passing writer.go:29: 2021-01-29T19:34:26.930Z [INFO] TestAgent_PassCheck_ACLDeny: Synced node info writer.go:29: 2021-01-29T19:34:26.931Z [WARN] TestAgent_PassCheck_ACLDeny: Check registration blocked by ACLs: check=test accessorID= writer.go:29: 2021-01-29T19:34:26.933Z [INFO] TestAgent_PassCheck_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:26.934Z [INFO] TestAgent_PassCheck_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:26.935Z [DEBUG] TestAgent_PassCheck_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:26.936Z [DEBUG] TestAgent_PassCheck_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.936Z [DEBUG] TestAgent_PassCheck_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:26.935Z [ERROR] TestAgent_PassCheck_ACLDeny.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:26.936Z [DEBUG] TestAgent_PassCheck_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:26.937Z [DEBUG] TestAgent_PassCheck_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:26.937Z [DEBUG] TestAgent_PassCheck_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:26.937Z [WARN] TestAgent_PassCheck_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.952Z [WARN] TestAgent_PassCheck_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:26.954Z [INFO] TestAgent_PassCheck_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:26.954Z [INFO] TestAgent_PassCheck_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:26.955Z [INFO] TestAgent_PassCheck_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:26.956Z [INFO] TestAgent_PassCheck_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30396 network=tcp writer.go:29: 2021-01-29T19:34:26.957Z [INFO] TestAgent_PassCheck_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30396 network=udp writer.go:29: 2021-01-29T19:34:26.958Z [INFO] TestAgent_PassCheck_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30397 network=tcp === CONT TestAgent_WarnCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.289Z [INFO] TestAgent_WarnCheck_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:27.290Z [INFO] TestAgent_WarnCheck_ACLDeny: Endpoints down --- PASS: TestAgent_WarnCheck_ACLDeny (1.01s) --- PASS: TestAgent_WarnCheck_ACLDeny/no_token (0.00s) --- PASS: TestAgent_WarnCheck_ACLDeny/root_token (0.03s) === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.300Z [WARN] TestAgent_DeregisterCheckACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:27.301Z [WARN] TestAgent_DeregisterCheckACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:27.302Z [DEBUG] TestAgent_DeregisterCheckACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:27.304Z [DEBUG] TestAgent_DeregisterCheckACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:27.318Z [INFO] TestAgent_DeregisterCheckACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:45caaacd-9b91-5cee-1eeb-f0630763e082 Address:127.0.0.1:30407}]" writer.go:29: 2021-01-29T19:34:27.320Z [INFO] TestAgent_DeregisterCheckACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30407 [Follower]" leader= === CONT TestAgent_WarnCheck writer.go:29: 2021-01-29T19:34:27.327Z [INFO] TestAgent_WarnCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:27.328Z [INFO] TestAgent_WarnCheck: Endpoints down --- PASS: TestAgent_WarnCheck (0.96s) === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.338Z [WARN] TestAgent_DeregisterCheck: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:27.339Z [DEBUG] TestAgent_DeregisterCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:27.340Z [DEBUG] TestAgent_DeregisterCheck.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.328Z [INFO] TestAgent_DeregisterCheckACLDeny.server.serf.wan: serf: EventMemberJoin: Node-45caaacd-9b91-5cee-1eeb-f0630763e082.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:27.344Z [INFO] TestAgent_DeregisterCheckACLDeny.server.serf.lan: serf: EventMemberJoin: Node-45caaacd-9b91-5cee-1eeb-f0630763e082 127.0.0.1 writer.go:29: 2021-01-29T19:34:27.350Z [INFO] TestAgent_DeregisterCheckACLDeny.server: Adding LAN server: server="Node-45caaacd-9b91-5cee-1eeb-f0630763e082 (Addr: tcp/127.0.0.1:30407) (DC: dc1)" writer.go:29: 2021-01-29T19:34:27.351Z [INFO] TestAgent_DeregisterCheckACLDeny.server: Handled event for server in area: event=member-join server=Node-45caaacd-9b91-5cee-1eeb-f0630763e082.dc1 area=wan writer.go:29: 2021-01-29T19:34:27.353Z [INFO] TestAgent_DeregisterCheckACLDeny: Started DNS server: address=127.0.0.1:30402 network=tcp === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.354Z [INFO] TestAgent_DeregisterCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:44bcc387-cada-fec5-1de0-f7b8596d460e Address:127.0.0.1:30419}]" writer.go:29: 2021-01-29T19:34:27.356Z [INFO] TestAgent_DeregisterCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:30419 [Follower]" leader= === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.358Z [INFO] TestAgent_DeregisterCheckACLDeny: Started DNS server: address=127.0.0.1:30402 network=udp === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.359Z [INFO] TestAgent_DeregisterCheck.server.serf.wan: serf: EventMemberJoin: Node-44bcc387-cada-fec5-1de0-f7b8596d460e.dc1 127.0.0.1 === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.362Z [INFO] TestAgent_DeregisterCheckACLDeny: Started HTTP server: address=127.0.0.1:30403 network=tcp writer.go:29: 2021-01-29T19:34:27.363Z [INFO] TestAgent_DeregisterCheckACLDeny: started state syncer === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.364Z [INFO] TestAgent_DeregisterCheck.server.serf.lan: serf: EventMemberJoin: Node-44bcc387-cada-fec5-1de0-f7b8596d460e 127.0.0.1 writer.go:29: 2021-01-29T19:34:27.366Z [INFO] TestAgent_DeregisterCheck.server: Handled event for server in area: event=member-join server=Node-44bcc387-cada-fec5-1de0-f7b8596d460e.dc1 area=wan writer.go:29: 2021-01-29T19:34:27.367Z [INFO] TestAgent_DeregisterCheck.server: Adding LAN server: server="Node-44bcc387-cada-fec5-1de0-f7b8596d460e (Addr: tcp/127.0.0.1:30419) (DC: dc1)" writer.go:29: 2021-01-29T19:34:27.370Z [INFO] TestAgent_DeregisterCheck: Started DNS server: address=127.0.0.1:30414 network=tcp writer.go:29: 2021-01-29T19:34:27.371Z [INFO] TestAgent_DeregisterCheck: Started DNS server: address=127.0.0.1:30414 network=udp writer.go:29: 2021-01-29T19:34:27.373Z [INFO] TestAgent_DeregisterCheck: Started HTTP server: address=127.0.0.1:30415 network=tcp writer.go:29: 2021-01-29T19:34:27.374Z [INFO] TestAgent_DeregisterCheck: started state syncer === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.382Z [WARN] TestAgent_DeregisterCheckACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:27.383Z [INFO] TestAgent_DeregisterCheckACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30407 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:27.385Z [DEBUG] TestAgent_DeregisterCheckACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:27.386Z [DEBUG] TestAgent_DeregisterCheckACLDeny.server.raft: vote granted: from=45caaacd-9b91-5cee-1eeb-f0630763e082 term=2 tally=1 writer.go:29: 2021-01-29T19:34:27.387Z [INFO] TestAgent_DeregisterCheckACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:27.388Z [INFO] TestAgent_DeregisterCheckACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30407 [Leader]" writer.go:29: 2021-01-29T19:34:27.389Z [INFO] TestAgent_DeregisterCheckACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:27.391Z [INFO] TestAgent_DeregisterCheckACLDeny.server: New leader elected: payload=Node-45caaacd-9b91-5cee-1eeb-f0630763e082 writer.go:29: 2021-01-29T19:34:27.392Z [INFO] TestAgent_DeregisterCheckACLDeny.server: initializing acls === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:27.394Z [INFO] TestAgent_PassCheck: Waiting for endpoints to shut down === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.394Z [INFO] TestAgent_DeregisterCheckACLDeny.server: Created ACL 'global-management' policy === CONT TestAgent_PassCheck writer.go:29: 2021-01-29T19:34:27.395Z [INFO] TestAgent_PassCheck: Endpoints down --- PASS: TestAgent_PassCheck (0.80s) === CONT TestAgent_RegisterCheck_ACLDeny === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.396Z [WARN] TestAgent_DeregisterCheckACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:27.398Z [INFO] TestAgent_DeregisterCheckACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:27.401Z [INFO] TestAgent_DeregisterCheckACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:27.403Z [WARN] TestAgent_DeregisterCheckACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:27.405Z [INFO] TestAgent_DeregisterCheckACLDeny.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.408Z [WARN] TestAgent_RegisterCheck_ACLDeny: bootstrap = true: do not enable unless necessary === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.403Z [INFO] TestAgent_DeregisterCheckACLDeny.server: Created ACL anonymous token from configuration === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.410Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.tlsutil: Update: version=1 === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.408Z [INFO] TestAgent_DeregisterCheckACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:27.411Z [DEBUG] TestAgent_DeregisterCheckACLDeny.server: transitioning out of legacy ACL mode === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.412Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.412Z [INFO] TestAgent_DeregisterCheckACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-45caaacd-9b91-5cee-1eeb-f0630763e082 writer.go:29: 2021-01-29T19:34:27.415Z [INFO] TestAgent_DeregisterCheckACLDeny.server: Updating LAN server: server="Node-45caaacd-9b91-5cee-1eeb-f0630763e082 (Addr: tcp/127.0.0.1:30407) (DC: dc1)" writer.go:29: 2021-01-29T19:34:27.415Z [INFO] TestAgent_DeregisterCheckACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-45caaacd-9b91-5cee-1eeb-f0630763e082 === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.418Z [WARN] TestAgent_DeregisterCheck.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.418Z [INFO] TestAgent_DeregisterCheckACLDeny.server: Updating LAN server: server="Node-45caaacd-9b91-5cee-1eeb-f0630763e082 (Addr: tcp/127.0.0.1:30407) (DC: dc1)" === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.419Z [INFO] TestAgent_DeregisterCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:30419 [Candidate]" term=2 === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.415Z [INFO] TestAgent_DeregisterCheckACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-45caaacd-9b91-5cee-1eeb-f0630763e082.dc1 === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.420Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4daf0826-65c1-49c9-d53a-ea8dde8e2df2 Address:127.0.0.1:29411}]" === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.421Z [INFO] TestAgent_DeregisterCheckACLDeny.server: Handled event for server in area: event=member-update server=Node-45caaacd-9b91-5cee-1eeb-f0630763e082.dc1 area=wan writer.go:29: 2021-01-29T19:34:27.421Z [INFO] TestAgent_DeregisterCheckACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-45caaacd-9b91-5cee-1eeb-f0630763e082.dc1 === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.422Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:27.422Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:29411 [Follower]" leader= === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.424Z [INFO] TestAgent_DeregisterCheckACLDeny.server: Handled event for server in area: event=member-update server=Node-45caaacd-9b91-5cee-1eeb-f0630763e082.dc1 area=wan === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.421Z [DEBUG] TestAgent_DeregisterCheck.server.raft: votes: needed=1 === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.427Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.428Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 127.0.0.1 === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.429Z [DEBUG] TestAgent_DeregisterCheck.server.raft: vote granted: from=44bcc387-cada-fec5-1de0-f7b8596d460e term=2 tally=1 === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.432Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: Adding LAN server: server="Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 (Addr: tcp/127.0.0.1:29411) (DC: dc1)" === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.431Z [INFO] TestAgent_DeregisterCheck.server.raft: election won: tally=1 === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.434Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: Handled event for server in area: event=member-join server=Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2.dc1 area=wan === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.434Z [INFO] TestAgent_DeregisterCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:30419 [Leader]" === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.437Z [INFO] TestAgent_RegisterCheck_ACLDeny: Started DNS server: address=127.0.0.1:29406 network=udp === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.438Z [INFO] TestAgent_DeregisterCheckACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.436Z [INFO] TestAgent_DeregisterCheck.server: cluster leadership acquired === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.439Z [INFO] TestAgent_DeregisterCheckACLDeny.leader: started routine: routine="CA root pruning" === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.437Z [INFO] TestAgent_DeregisterCheck.server: New leader elected: payload=Node-44bcc387-cada-fec5-1de0-f7b8596d460e === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.440Z [DEBUG] TestAgent_DeregisterCheckACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-45caaacd-9b91-5cee-1eeb-f0630763e082 === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.438Z [INFO] TestAgent_RegisterCheck_ACLDeny: Started DNS server: address=127.0.0.1:29406 network=tcp === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.442Z [INFO] TestAgent_DeregisterCheckACLDeny.server: member joined, marking health alive: member=Node-45caaacd-9b91-5cee-1eeb-f0630763e082 writer.go:29: 2021-01-29T19:34:27.444Z [DEBUG] TestAgent_DeregisterCheckACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-45caaacd-9b91-5cee-1eeb-f0630763e082 === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.444Z [INFO] TestAgent_RegisterCheck_ACLDeny: Started HTTP server: address=127.0.0.1:29407 network=tcp === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.446Z [DEBUG] TestAgent_DeregisterCheckACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-45caaacd-9b91-5cee-1eeb-f0630763e082 === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.446Z [INFO] TestAgent_RegisterCheck_ACLDeny: started state syncer === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.442Z [DEBUG] TestAgent_DeregisterCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30419 === CONT TestAgent_PassCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.461Z [INFO] TestAgent_PassCheck_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:27.463Z [INFO] TestAgent_PassCheck_ACLDeny: Endpoints down --- PASS: TestAgent_PassCheck_ACLDeny (0.93s) --- PASS: TestAgent_PassCheck_ACLDeny/no_token (0.00s) --- PASS: TestAgent_PassCheck_ACLDeny/root_token (0.01s) === CONT TestAgent_RegisterCheck_BadStatus === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.463Z [WARN] TestAgent_RegisterCheck_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:27.465Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:29411 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:27.467Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.server.raft: votes: needed=1 === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.467Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.468Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.server.raft: vote granted: from=4daf0826-65c1-49c9-d53a-ea8dde8e2df2 term=2 tally=1 writer.go:29: 2021-01-29T19:34:27.470Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:27.471Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:29411 [Leader]" writer.go:29: 2021-01-29T19:34:27.473Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:27.473Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: New leader elected: payload=Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.473Z [INFO] TestAgent_DeregisterCheck.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:27.476Z [INFO] TestAgent_DeregisterCheck.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.479Z [WARN] TestAgent_RegisterCheck_BadStatus: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:27.481Z [DEBUG] TestAgent_RegisterCheck_BadStatus.tlsutil: Update: version=1 === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.480Z [DEBUG] TestAgent_DeregisterCheck.server: Skipping self join check for node since the cluster is too small: node=Node-44bcc387-cada-fec5-1de0-f7b8596d460e === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.483Z [DEBUG] TestAgent_RegisterCheck_BadStatus.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.483Z [INFO] TestAgent_DeregisterCheck.server: member joined, marking health alive: member=Node-44bcc387-cada-fec5-1de0-f7b8596d460e === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.485Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:27.486Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:27.489Z [WARN] TestAgent_RegisterCheck_ACLDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.487Z [INFO] TestAgent_RegisterCheck_BadStatus.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:50b0c6e6-12c5-c231-596d-020582e180aa Address:127.0.0.1:29405}]" === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.489Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:27.492Z [WARN] TestAgent_RegisterCheck_ACLDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.493Z [INFO] TestAgent_RegisterCheck_BadStatus.server.raft: entering follower state: follower="Node at 127.0.0.1:29405 [Follower]" leader= === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.494Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: Bootstrapped ACL master token from configuration === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.493Z [INFO] TestAgent_RegisterCheck_BadStatus.server.serf.wan: serf: EventMemberJoin: Node-50b0c6e6-12c5-c231-596d-020582e180aa.dc1 127.0.0.1 === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.495Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:27.497Z [INFO] TestAgent_RegisterCheck_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:27.497Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: Created ACL anonymous token from configuration === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.499Z [INFO] TestAgent_RegisterCheck_BadStatus.server.serf.lan: serf: EventMemberJoin: Node-50b0c6e6-12c5-c231-596d-020582e180aa 127.0.0.1 writer.go:29: 2021-01-29T19:34:27.503Z [INFO] TestAgent_RegisterCheck_BadStatus.server: Handled event for server in area: event=member-join server=Node-50b0c6e6-12c5-c231-596d-020582e180aa.dc1 area=wan writer.go:29: 2021-01-29T19:34:27.504Z [INFO] TestAgent_RegisterCheck_BadStatus.server: Adding LAN server: server="Node-50b0c6e6-12c5-c231-596d-020582e180aa (Addr: tcp/127.0.0.1:29405) (DC: dc1)" === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.499Z [INFO] TestAgent_RegisterCheck_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:27.507Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.server: transitioning out of legacy ACL mode === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.504Z [INFO] TestAgent_RegisterCheck_BadStatus: Started DNS server: address=127.0.0.1:29400 network=udp === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.507Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.509Z [INFO] TestAgent_RegisterCheck_BadStatus: Started DNS server: address=127.0.0.1:29400 network=tcp writer.go:29: 2021-01-29T19:34:27.513Z [INFO] TestAgent_RegisterCheck_BadStatus: Started HTTP server: address=127.0.0.1:29401 network=tcp writer.go:29: 2021-01-29T19:34:27.514Z [INFO] TestAgent_RegisterCheck_BadStatus: started state syncer === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.510Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2.dc1 writer.go:29: 2021-01-29T19:34:27.510Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 writer.go:29: 2021-01-29T19:34:27.510Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: Updating LAN server: server="Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 (Addr: tcp/127.0.0.1:29411) (DC: dc1)" writer.go:29: 2021-01-29T19:34:27.523Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: Updating LAN server: server="Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 (Addr: tcp/127.0.0.1:29411) (DC: dc1)" writer.go:29: 2021-01-29T19:34:27.518Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: Handled event for server in area: event=member-update server=Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2.dc1 area=wan writer.go:29: 2021-01-29T19:34:27.522Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2.dc1 writer.go:29: 2021-01-29T19:34:27.525Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:27.529Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: Handled event for server in area: event=member-update server=Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2.dc1 area=wan === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.534Z [WARN] TestAgent_RegisterCheck_BadStatus.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:27.536Z [INFO] TestAgent_RegisterCheck_BadStatus.server.raft: entering candidate state: node="Node at 127.0.0.1:29405 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:27.539Z [DEBUG] TestAgent_RegisterCheck_BadStatus.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:27.540Z [DEBUG] TestAgent_RegisterCheck_BadStatus.server.raft: vote granted: from=50b0c6e6-12c5-c231-596d-020582e180aa term=2 tally=1 === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.541Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.542Z [INFO] TestAgent_RegisterCheck_BadStatus.server.raft: election won: tally=1 === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.543Z [INFO] TestAgent_RegisterCheck_ACLDeny.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.545Z [INFO] TestAgent_RegisterCheck_BadStatus.server.raft: entering leader state: leader="Node at 127.0.0.1:29405 [Leader]" === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.545Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.547Z [INFO] TestAgent_RegisterCheck_BadStatus.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:27.548Z [INFO] TestAgent_RegisterCheck_BadStatus.server: New leader elected: payload=Node-50b0c6e6-12c5-c231-596d-020582e180aa === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.547Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: member joined, marking health alive: member=Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 writer.go:29: 2021-01-29T19:34:27.554Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 writer.go:29: 2021-01-29T19:34:27.556Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.552Z [DEBUG] TestAgent_RegisterCheck_BadStatus.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29405 writer.go:29: 2021-01-29T19:34:27.561Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:27.567Z [INFO] TestAgent_RegisterCheck_BadStatus.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:27.569Z [INFO] TestAgent_RegisterCheck_BadStatus.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:27.571Z [DEBUG] TestAgent_RegisterCheck_BadStatus.server: Skipping self join check for node since the cluster is too small: node=Node-50b0c6e6-12c5-c231-596d-020582e180aa writer.go:29: 2021-01-29T19:34:27.572Z [INFO] TestAgent_RegisterCheck_BadStatus.server: member joined, marking health alive: member=Node-50b0c6e6-12c5-c231-596d-020582e180aa === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.583Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.acl: dropping node from result due to ACLs: node=Node-4daf0826-65c1-49c9-d53a-ea8dde8e2df2 === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.587Z [DEBUG] TestAgent_DeregisterCheck: removed check: check=test writer.go:29: 2021-01-29T19:34:27.590Z [INFO] TestAgent_DeregisterCheck: Synced node info === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.589Z [INFO] TestAgent_RegisterCheck_ACLDeny: Synced node info === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.592Z [INFO] TestAgent_DeregisterCheck: Deregistered check: check=test === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.592Z [INFO] TestAgent_RegisterCheck_BadStatus: Requesting shutdown writer.go:29: 2021-01-29T19:34:27.594Z [INFO] TestAgent_RegisterCheck_BadStatus.server: shutting down server === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.594Z [INFO] TestAgent_DeregisterCheck: Requesting shutdown === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.595Z [DEBUG] TestAgent_RegisterCheck_BadStatus.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.596Z [INFO] TestAgent_DeregisterCheck.server: shutting down server === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.596Z [WARN] TestAgent_RegisterCheck_BadStatus.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.597Z [DEBUG] TestAgent_DeregisterCheck.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.595Z [ERROR] TestAgent_RegisterCheck_BadStatus.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.598Z [WARN] TestAgent_DeregisterCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:27.597Z [ERROR] TestAgent_DeregisterCheck.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.596Z [DEBUG] TestAgent_RegisterCheck_BadStatus.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:27.599Z [WARN] TestAgent_RegisterCheck_BadStatus.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.599Z [DEBUG] TestAgent_DeregisterCheck.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:27.602Z [WARN] TestAgent_DeregisterCheck.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.606Z [INFO] TestAgent_RegisterCheck_BadStatus.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:27.609Z [INFO] TestAgent_RegisterCheck_BadStatus: consul server down === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.609Z [INFO] TestAgent_DeregisterCheck.server.router.manager: shutting down === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.610Z [INFO] TestAgent_RegisterCheck_BadStatus: shutdown complete === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.611Z [INFO] TestAgent_DeregisterCheck: consul server down === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.612Z [INFO] TestAgent_RegisterCheck_BadStatus: Stopping server: protocol=DNS address=127.0.0.1:29400 network=tcp === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.613Z [INFO] TestAgent_DeregisterCheck: shutdown complete === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.614Z [INFO] TestAgent_RegisterCheck_BadStatus: Stopping server: protocol=DNS address=127.0.0.1:29400 network=udp === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.614Z [INFO] TestAgent_DeregisterCheck: Stopping server: protocol=DNS address=127.0.0.1:30414 network=tcp === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.593Z [INFO] TestAgent_RegisterCheck_ACLDeny: Synced service: service=foo:1234 === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.616Z [INFO] TestAgent_DeregisterCheck: Stopping server: protocol=DNS address=127.0.0.1:30414 network=udp === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:27.615Z [INFO] TestAgent_RegisterCheck_BadStatus: Stopping server: protocol=HTTP address=127.0.0.1:29401 network=tcp === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:27.618Z [INFO] TestAgent_DeregisterCheck: Stopping server: protocol=HTTP address=127.0.0.1:30415 network=tcp === RUN TestAgent_RegisterCheck_ACLDeny/no_token_-_node_check === RUN TestAgent_RegisterCheck_ACLDeny/svc_token_-_node_check === RUN TestAgent_RegisterCheck_ACLDeny/node_token_-_node_check === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.650Z [DEBUG] TestAgent_RegisterCheck_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:27.652Z [DEBUG] TestAgent_RegisterCheck_ACLDeny: Service in sync: service=foo:1234 writer.go:29: 2021-01-29T19:34:27.657Z [INFO] TestAgent_RegisterCheck_ACLDeny: Synced check: check=test === RUN TestAgent_RegisterCheck_ACLDeny/no_token_-_svc_check === RUN TestAgent_RegisterCheck_ACLDeny/node_token_-_svc_check === RUN TestAgent_RegisterCheck_ACLDeny/svc_token_-_svc_check === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.675Z [DEBUG] TestAgent_RegisterCheck_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:27.676Z [DEBUG] TestAgent_RegisterCheck_ACLDeny: Service in sync: service=foo:1234 writer.go:29: 2021-01-29T19:34:27.677Z [DEBUG] TestAgent_RegisterCheck_ACLDeny: Check in sync: check=test writer.go:29: 2021-01-29T19:34:27.682Z [INFO] TestAgent_RegisterCheck_ACLDeny: Synced check: check=test2 === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.684Z [DEBUG] TestAgent_DeregisterCheckACLDeny: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.685Z [INFO] TestAgent_RegisterCheck_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:27.686Z [INFO] TestAgent_RegisterCheck_ACLDeny.server: shutting down server === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.687Z [INFO] TestAgent_DeregisterCheckACLDeny: Synced node info === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.688Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:27.689Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.689Z [DEBUG] TestAgent_DeregisterCheckACLDeny.acl: dropping node from result due to ACLs: node=Node-45caaacd-9b91-5cee-1eeb-f0630763e082 === RUN TestAgent_DeregisterCheckACLDeny/no_token === RUN TestAgent_DeregisterCheckACLDeny/root_token === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.707Z [DEBUG] TestAgent_DeregisterCheckACLDeny: removed check: check=test === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.688Z [ERROR] TestAgent_RegisterCheck_ACLDeny.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.710Z [DEBUG] TestAgent_DeregisterCheckACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:27.722Z [INFO] TestAgent_DeregisterCheckACLDeny: Deregistered check: check=test === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.689Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:27.691Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.723Z [INFO] TestAgent_DeregisterCheckACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:27.726Z [INFO] TestAgent_DeregisterCheckACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:27.728Z [DEBUG] TestAgent_DeregisterCheckACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:27.730Z [DEBUG] TestAgent_DeregisterCheckACLDeny.leader: stopping routine: routine="acl token reaping" === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.691Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.733Z [DEBUG] TestAgent_DeregisterCheckACLDeny.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.726Z [WARN] TestAgent_RegisterCheck_ACLDeny.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.730Z [DEBUG] TestAgent_DeregisterCheckACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.726Z [DEBUG] TestAgent_RegisterCheck_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.734Z [DEBUG] TestAgent_DeregisterCheckACLDeny.leader: stopped routine: routine="acl token reaping" === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.739Z [WARN] TestAgent_RegisterCheck_ACLDeny.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.737Z [WARN] TestAgent_DeregisterCheckACLDeny.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.742Z [INFO] TestAgent_RegisterCheck_ACLDeny.server.router.manager: shutting down === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.737Z [DEBUG] TestAgent_DeregisterCheckACLDeny.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.743Z [INFO] TestAgent_RegisterCheck_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:27.746Z [INFO] TestAgent_RegisterCheck_ACLDeny: shutdown complete === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.746Z [WARN] TestAgent_DeregisterCheckACLDeny.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.747Z [INFO] TestAgent_RegisterCheck_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29406 network=tcp writer.go:29: 2021-01-29T19:34:27.749Z [INFO] TestAgent_RegisterCheck_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29406 network=udp === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.748Z [INFO] TestAgent_DeregisterCheckACLDeny.server.router.manager: shutting down === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:27.750Z [INFO] TestAgent_RegisterCheck_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:29407 network=tcp === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:27.749Z [INFO] TestAgent_DeregisterCheckACLDeny: consul server down writer.go:29: 2021-01-29T19:34:27.752Z [INFO] TestAgent_DeregisterCheckACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:27.757Z [INFO] TestAgent_DeregisterCheckACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30402 network=tcp writer.go:29: 2021-01-29T19:34:27.763Z [INFO] TestAgent_DeregisterCheckACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30402 network=udp writer.go:29: 2021-01-29T19:34:27.766Z [INFO] TestAgent_DeregisterCheckACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30403 network=tcp === CONT TestAgent_RegisterCheck_BadStatus writer.go:29: 2021-01-29T19:34:28.120Z [INFO] TestAgent_RegisterCheck_BadStatus: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:28.121Z [INFO] TestAgent_RegisterCheck_BadStatus: Endpoints down --- PASS: TestAgent_RegisterCheck_BadStatus (0.66s) === CONT TestAgent_RegisterCheck_Passing === CONT TestAgent_DeregisterCheck writer.go:29: 2021-01-29T19:34:28.134Z [INFO] TestAgent_DeregisterCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:28.136Z [INFO] TestAgent_DeregisterCheck: Endpoints down --- PASS: TestAgent_DeregisterCheck (0.81s) === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable === CONT TestAgent_RegisterCheck_ACLDeny writer.go:29: 2021-01-29T19:34:28.269Z [INFO] TestAgent_RegisterCheck_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:28.271Z [INFO] TestAgent_RegisterCheck_ACLDeny: Endpoints down --- PASS: TestAgent_RegisterCheck_ACLDeny (0.88s) --- PASS: TestAgent_RegisterCheck_ACLDeny/no_token_-_node_check (0.00s) --- PASS: TestAgent_RegisterCheck_ACLDeny/svc_token_-_node_check (0.00s) --- PASS: TestAgent_RegisterCheck_ACLDeny/node_token_-_node_check (0.01s) --- PASS: TestAgent_RegisterCheck_ACLDeny/no_token_-_svc_check (0.00s) --- PASS: TestAgent_RegisterCheck_ACLDeny/node_token_-_svc_check (0.00s) --- PASS: TestAgent_RegisterCheck_ACLDeny/svc_token_-_svc_check (0.01s) === CONT TestAgent_RegisterCheckScriptsExecDisable === CONT TestAgent_DeregisterCheckACLDeny writer.go:29: 2021-01-29T19:34:28.283Z [INFO] TestAgent_DeregisterCheckACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:28.289Z [INFO] TestAgent_DeregisterCheckACLDeny: Endpoints down --- PASS: TestAgent_DeregisterCheckACLDeny (1.00s) --- PASS: TestAgent_DeregisterCheckACLDeny/no_token (0.00s) --- PASS: TestAgent_DeregisterCheckACLDeny/root_token (0.02s) === CONT TestAgent_RegisterCheck === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.295Z [WARN] TestAgent_RegisterCheckScriptsExecRemoteDisable: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:28.317Z [DEBUG] TestAgent_RegisterCheckScriptsExecRemoteDisable.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:28.325Z [DEBUG] TestAgent_RegisterCheckScriptsExecRemoteDisable.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.327Z [WARN] TestAgent_RegisterCheckScriptsExecDisable: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:28.331Z [DEBUG] TestAgent_RegisterCheckScriptsExecDisable.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:28.333Z [DEBUG] TestAgent_RegisterCheckScriptsExecDisable.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.339Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:37d9a224-3f96-6d4f-4068-c4d858add53f Address:127.0.0.1:29423}]" writer.go:29: 2021-01-29T19:34:28.343Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.raft: entering follower state: follower="Node at 127.0.0.1:29423 [Follower]" leader= === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.349Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9693af06-d69d-d49e-2c51-f1a27f9890a0 Address:127.0.0.1:29417}]" writer.go:29: 2021-01-29T19:34:28.352Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server.serf.wan: serf: EventMemberJoin: Node-9693af06-d69d-d49e-2c51-f1a27f9890a0.dc1 127.0.0.1 === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.354Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.serf.wan: serf: EventMemberJoin: Node-37d9a224-3f96-6d4f-4068-c4d858add53f.dc1 127.0.0.1 === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.356Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server.serf.lan: serf: EventMemberJoin: Node-9693af06-d69d-d49e-2c51-f1a27f9890a0 127.0.0.1 writer.go:29: 2021-01-29T19:34:28.361Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server.raft: entering follower state: follower="Node at 127.0.0.1:29417 [Follower]" leader= writer.go:29: 2021-01-29T19:34:28.402Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: Started DNS server: address=127.0.0.1:29412 network=udp === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.415Z [WARN] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:28.440Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.raft: entering candidate state: node="Node at 127.0.0.1:29423 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:28.440Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.serf.lan: serf: EventMemberJoin: Node-37d9a224-3f96-6d4f-4068-c4d858add53f 127.0.0.1 writer.go:29: 2021-01-29T19:34:28.447Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: Started DNS server: address=127.0.0.1:29418 network=udp === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.448Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server: Adding LAN server: server="Node-9693af06-d69d-d49e-2c51-f1a27f9890a0 (Addr: tcp/127.0.0.1:29417) (DC: dc1)" === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.457Z [WARN] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.raft: unable to get address for sever, using fallback address: id=37d9a224-3f96-6d4f-4068-c4d858add53f fallback=127.0.0.1:29423 error="Could not find address for server id 37d9a224-3f96-6d4f-4068-c4d858add53f" writer.go:29: 2021-01-29T19:34:28.472Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server: Adding LAN server: server="Node-37d9a224-3f96-6d4f-4068-c4d858add53f (Addr: tcp/127.0.0.1:29423) (DC: dc1)" === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.482Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server: Handled event for server in area: event=member-join server=Node-9693af06-d69d-d49e-2c51-f1a27f9890a0.dc1 area=wan writer.go:29: 2021-01-29T19:34:28.486Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: Started DNS server: address=127.0.0.1:29412 network=tcp writer.go:29: 2021-01-29T19:34:28.489Z [WARN] TestAgent_RegisterCheckScriptsExecDisable.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:28.490Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server.raft: entering candidate state: node="Node at 127.0.0.1:29417 [Candidate]" term=2 === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.494Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server: Handled event for server in area: event=member-join server=Node-37d9a224-3f96-6d4f-4068-c4d858add53f.dc1 area=wan === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.501Z [WARN] TestAgent_RegisterCheck: bootstrap = true: do not enable unless necessary === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.503Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: Started HTTP server: address=127.0.0.1:29413 network=tcp === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.503Z [DEBUG] TestAgent_RegisterCheck.tlsutil: Update: version=1 === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.501Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: Started DNS server: address=127.0.0.1:29418 network=tcp writer.go:29: 2021-01-29T19:34:28.509Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: Started HTTP server: address=127.0.0.1:29419 network=tcp writer.go:29: 2021-01-29T19:34:28.510Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: started state syncer === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.508Z [DEBUG] TestAgent_RegisterCheck.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.513Z [WARN] TestAgent_RegisterCheck_Passing: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:28.515Z [DEBUG] TestAgent_RegisterCheck_Passing.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:28.516Z [DEBUG] TestAgent_RegisterCheck_Passing.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.506Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: started state syncer === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.506Z [DEBUG] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:28.519Z [DEBUG] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.raft: vote granted: from=37d9a224-3f96-6d4f-4068-c4d858add53f term=2 tally=1 writer.go:29: 2021-01-29T19:34:28.520Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:28.521Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.raft: entering leader state: leader="Node at 127.0.0.1:29423 [Leader]" === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.507Z [DEBUG] TestAgent_RegisterCheckScriptsExecDisable.server.raft: votes: needed=1 === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.522Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server: cluster leadership acquired === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.523Z [DEBUG] TestAgent_RegisterCheckScriptsExecDisable.server.raft: vote granted: from=9693af06-d69d-d49e-2c51-f1a27f9890a0 term=2 tally=1 writer.go:29: 2021-01-29T19:34:28.528Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:28.528Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server.raft: entering leader state: leader="Node at 127.0.0.1:29417 [Leader]" writer.go:29: 2021-01-29T19:34:28.529Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server: cluster leadership acquired === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.530Z [INFO] TestAgent_RegisterCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a24a4dcf-57d1-230e-10b0-7243a2be7b88 Address:127.0.0.1:29429}]" === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.531Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server: New leader elected: payload=Node-9693af06-d69d-d49e-2c51-f1a27f9890a0 === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.533Z [INFO] TestAgent_RegisterCheck.server.serf.wan: serf: EventMemberJoin: Node-a24a4dcf-57d1-230e-10b0-7243a2be7b88.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:28.535Z [INFO] TestAgent_RegisterCheck.server.serf.lan: serf: EventMemberJoin: Node-a24a4dcf-57d1-230e-10b0-7243a2be7b88 127.0.0.1 === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.525Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server: New leader elected: payload=Node-37d9a224-3f96-6d4f-4068-c4d858add53f === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.535Z [DEBUG] TestAgent_RegisterCheckScriptsExecDisable.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29417 === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.537Z [INFO] TestAgent_RegisterCheck: Started DNS server: address=127.0.0.1:29424 network=udp writer.go:29: 2021-01-29T19:34:28.539Z [INFO] TestAgent_RegisterCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:29429 [Follower]" leader= writer.go:29: 2021-01-29T19:34:28.543Z [INFO] TestAgent_RegisterCheck.server: Adding LAN server: server="Node-a24a4dcf-57d1-230e-10b0-7243a2be7b88 (Addr: tcp/127.0.0.1:29429) (DC: dc1)" writer.go:29: 2021-01-29T19:34:28.545Z [INFO] TestAgent_RegisterCheck.server: Handled event for server in area: event=member-join server=Node-a24a4dcf-57d1-230e-10b0-7243a2be7b88.dc1 area=wan === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.526Z [DEBUG] TestAgent_RegisterCheckScriptsExecRemoteDisable.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29423 === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.547Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.547Z [INFO] TestAgent_RegisterCheck: Started DNS server: address=127.0.0.1:29424 network=tcp writer.go:29: 2021-01-29T19:34:28.552Z [INFO] TestAgent_RegisterCheck: Started HTTP server: address=127.0.0.1:29425 network=tcp writer.go:29: 2021-01-29T19:34:28.553Z [INFO] TestAgent_RegisterCheck: started state syncer === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.549Z [INFO] TestAgent_RegisterCheck_Passing.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:554ba9f0-dba8-ac35-249d-ae9369f448e7 Address:127.0.0.1:30371}]" === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.559Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:28.561Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:28.562Z [DEBUG] TestAgent_RegisterCheckScriptsExecDisable.server: Skipping self join check for node since the cluster is too small: node=Node-9693af06-d69d-d49e-2c51-f1a27f9890a0 === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.557Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.564Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server: member joined, marking health alive: member=Node-9693af06-d69d-d49e-2c51-f1a27f9890a0 === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.562Z [INFO] TestAgent_RegisterCheck_Passing.server.raft: entering follower state: follower="Node at 127.0.0.1:30371 [Follower]" leader= === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.570Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:28.573Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: Synced node info writer.go:29: 2021-01-29T19:34:28.583Z [DEBUG] TestAgent_RegisterCheckScriptsExecRemoteDisable: Node info in sync writer.go:29: 2021-01-29T19:34:28.582Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:28.586Z [DEBUG] TestAgent_RegisterCheckScriptsExecRemoteDisable.server: Skipping self join check for node since the cluster is too small: node=Node-37d9a224-3f96-6d4f-4068-c4d858add53f writer.go:29: 2021-01-29T19:34:28.588Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server: member joined, marking health alive: member=Node-37d9a224-3f96-6d4f-4068-c4d858add53f === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.572Z [INFO] TestAgent_RegisterCheck_Passing.server.serf.wan: serf: EventMemberJoin: Node-554ba9f0-dba8-ac35-249d-ae9369f448e7.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:28.591Z [INFO] TestAgent_RegisterCheck_Passing.server.serf.lan: serf: EventMemberJoin: Node-554ba9f0-dba8-ac35-249d-ae9369f448e7 127.0.0.1 === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.594Z [WARN] TestAgent_RegisterCheck.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:28.605Z [INFO] TestAgent_RegisterCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:29429 [Candidate]" term=2 === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.606Z [INFO] TestAgent_RegisterCheck_Passing.server: Handled event for server in area: event=member-join server=Node-554ba9f0-dba8-ac35-249d-ae9369f448e7.dc1 area=wan writer.go:29: 2021-01-29T19:34:28.606Z [INFO] TestAgent_RegisterCheck_Passing.server: Adding LAN server: server="Node-554ba9f0-dba8-ac35-249d-ae9369f448e7 (Addr: tcp/127.0.0.1:30371) (DC: dc1)" writer.go:29: 2021-01-29T19:34:28.609Z [INFO] TestAgent_RegisterCheck_Passing: Started DNS server: address=127.0.0.1:30366 network=tcp writer.go:29: 2021-01-29T19:34:28.618Z [INFO] TestAgent_RegisterCheck_Passing: Started DNS server: address=127.0.0.1:30366 network=udp writer.go:29: 2021-01-29T19:34:28.621Z [INFO] TestAgent_RegisterCheck_Passing: Started HTTP server: address=127.0.0.1:30367 network=tcp writer.go:29: 2021-01-29T19:34:28.624Z [INFO] TestAgent_RegisterCheck_Passing: started state syncer writer.go:29: 2021-01-29T19:34:28.626Z [WARN] TestAgent_RegisterCheck_Passing.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.644Z [DEBUG] TestAgent_RegisterCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:28.646Z [DEBUG] TestAgent_RegisterCheck.server.raft: vote granted: from=a24a4dcf-57d1-230e-10b0-7243a2be7b88 term=2 tally=1 writer.go:29: 2021-01-29T19:34:28.647Z [INFO] TestAgent_RegisterCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:28.648Z [INFO] TestAgent_RegisterCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:29429 [Leader]" writer.go:29: 2021-01-29T19:34:28.650Z [INFO] TestAgent_RegisterCheck.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:28.652Z [INFO] TestAgent_RegisterCheck.server: New leader elected: payload=Node-a24a4dcf-57d1-230e-10b0-7243a2be7b88 === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.632Z [INFO] TestAgent_RegisterCheck_Passing.server.raft: entering candidate state: node="Node at 127.0.0.1:30371 [Candidate]" term=2 === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.656Z [DEBUG] TestAgent_RegisterCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29429 === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.685Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: Requesting shutdown writer.go:29: 2021-01-29T19:34:28.687Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server: shutting down server writer.go:29: 2021-01-29T19:34:28.688Z [DEBUG] TestAgent_RegisterCheckScriptsExecRemoteDisable.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.689Z [DEBUG] TestAgent_RegisterCheck_Passing.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:28.690Z [DEBUG] TestAgent_RegisterCheck_Passing.server.raft: vote granted: from=554ba9f0-dba8-ac35-249d-ae9369f448e7 term=2 tally=1 === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.689Z [WARN] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.690Z [INFO] TestAgent_RegisterCheck_Passing.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:28.691Z [INFO] TestAgent_RegisterCheck_Passing.server.raft: entering leader state: leader="Node at 127.0.0.1:30371 [Leader]" === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.692Z [DEBUG] TestAgent_RegisterCheckScriptsExecRemoteDisable.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.693Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.692Z [INFO] TestAgent_RegisterCheck_Passing.server: cluster leadership acquired === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.694Z [WARN] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.694Z [INFO] TestAgent_RegisterCheck_Passing.server: New leader elected: payload=Node-554ba9f0-dba8-ac35-249d-ae9369f448e7 writer.go:29: 2021-01-29T19:34:28.696Z [DEBUG] TestAgent_RegisterCheck_Passing.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30371 === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.698Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:28.698Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: consul server down writer.go:29: 2021-01-29T19:34:28.699Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: shutdown complete writer.go:29: 2021-01-29T19:34:28.700Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: Stopping server: protocol=DNS address=127.0.0.1:29418 network=tcp writer.go:29: 2021-01-29T19:34:28.701Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: Stopping server: protocol=DNS address=127.0.0.1:29418 network=udp === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.703Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:28.703Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: Stopping server: protocol=HTTP address=127.0.0.1:29419 network=tcp === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.708Z [INFO] TestAgent_RegisterCheck.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.709Z [INFO] TestAgent_RegisterCheck_Passing.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.709Z [INFO] TestAgent_RegisterCheck.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.711Z [INFO] TestAgent_RegisterCheck_Passing.leader: started routine: routine="CA root pruning" === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.711Z [DEBUG] TestAgent_RegisterCheck.server: Skipping self join check for node since the cluster is too small: node=Node-a24a4dcf-57d1-230e-10b0-7243a2be7b88 === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.713Z [DEBUG] TestAgent_RegisterCheck_Passing.server: Skipping self join check for node since the cluster is too small: node=Node-554ba9f0-dba8-ac35-249d-ae9369f448e7 === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.713Z [INFO] TestAgent_RegisterCheck.server: member joined, marking health alive: member=Node-a24a4dcf-57d1-230e-10b0-7243a2be7b88 === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:28.715Z [INFO] TestAgent_RegisterCheck_Passing.server: member joined, marking health alive: member=Node-554ba9f0-dba8-ac35-249d-ae9369f448e7 === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:28.743Z [INFO] TestAgent_RegisterCheck: Synced node info writer.go:29: 2021-01-29T19:34:28.745Z [INFO] TestAgent_RegisterCheck: Synced check: check=test writer.go:29: 2021-01-29T19:34:28.746Z [INFO] TestAgent_RegisterCheck: Requesting shutdown writer.go:29: 2021-01-29T19:34:28.746Z [INFO] TestAgent_RegisterCheck.server: shutting down server writer.go:29: 2021-01-29T19:34:28.747Z [DEBUG] TestAgent_RegisterCheck.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:28.747Z [ERROR] TestAgent_RegisterCheck.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:28.748Z [DEBUG] TestAgent_RegisterCheck.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:28.750Z [WARN] TestAgent_RegisterCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:28.752Z [WARN] TestAgent_RegisterCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:28.754Z [INFO] TestAgent_RegisterCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:28.754Z [INFO] TestAgent_RegisterCheck: consul server down writer.go:29: 2021-01-29T19:34:28.756Z [INFO] TestAgent_RegisterCheck: shutdown complete writer.go:29: 2021-01-29T19:34:28.756Z [INFO] TestAgent_RegisterCheck: Stopping server: protocol=DNS address=127.0.0.1:29424 network=tcp writer.go:29: 2021-01-29T19:34:28.757Z [INFO] TestAgent_RegisterCheck: Stopping server: protocol=DNS address=127.0.0.1:29424 network=udp writer.go:29: 2021-01-29T19:34:28.758Z [INFO] TestAgent_RegisterCheck: Stopping server: protocol=HTTP address=127.0.0.1:29425 network=tcp === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:28.786Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: Requesting shutdown writer.go:29: 2021-01-29T19:34:28.787Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server: shutting down server writer.go:29: 2021-01-29T19:34:28.788Z [DEBUG] TestAgent_RegisterCheckScriptsExecDisable.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:28.789Z [WARN] TestAgent_RegisterCheckScriptsExecDisable.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:28.788Z [ERROR] TestAgent_RegisterCheckScriptsExecDisable.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:28.789Z [DEBUG] TestAgent_RegisterCheckScriptsExecDisable.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:28.791Z [WARN] TestAgent_RegisterCheckScriptsExecDisable.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:28.794Z [INFO] TestAgent_RegisterCheckScriptsExecDisable.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:28.794Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: consul server down writer.go:29: 2021-01-29T19:34:28.797Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: shutdown complete writer.go:29: 2021-01-29T19:34:28.799Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: Stopping server: protocol=DNS address=127.0.0.1:29412 network=tcp writer.go:29: 2021-01-29T19:34:28.800Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: Stopping server: protocol=DNS address=127.0.0.1:29412 network=udp writer.go:29: 2021-01-29T19:34:28.802Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: Stopping server: protocol=HTTP address=127.0.0.1:29413 network=tcp === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:29.020Z [DEBUG] TestAgent_RegisterCheck_Passing: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:29.023Z [INFO] TestAgent_RegisterCheck_Passing: Synced node info writer.go:29: 2021-01-29T19:34:29.037Z [DEBUG] TestAgent_RegisterCheck_Passing: Node info in sync writer.go:29: 2021-01-29T19:34:29.045Z [INFO] TestAgent_RegisterCheck_Passing: Synced check: check=test writer.go:29: 2021-01-29T19:34:29.046Z [INFO] TestAgent_RegisterCheck_Passing: Requesting shutdown writer.go:29: 2021-01-29T19:34:29.048Z [INFO] TestAgent_RegisterCheck_Passing.server: shutting down server writer.go:29: 2021-01-29T19:34:29.049Z [DEBUG] TestAgent_RegisterCheck_Passing.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:29.050Z [WARN] TestAgent_RegisterCheck_Passing.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:29.052Z [DEBUG] TestAgent_RegisterCheck_Passing.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:29.056Z [WARN] TestAgent_RegisterCheck_Passing.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:29.069Z [INFO] TestAgent_RegisterCheck_Passing.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:29.069Z [INFO] TestAgent_RegisterCheck_Passing: consul server down writer.go:29: 2021-01-29T19:34:29.072Z [INFO] TestAgent_RegisterCheck_Passing: shutdown complete writer.go:29: 2021-01-29T19:34:29.073Z [INFO] TestAgent_RegisterCheck_Passing: Stopping server: protocol=DNS address=127.0.0.1:30366 network=tcp writer.go:29: 2021-01-29T19:34:29.075Z [INFO] TestAgent_RegisterCheck_Passing: Stopping server: protocol=DNS address=127.0.0.1:30366 network=udp writer.go:29: 2021-01-29T19:34:29.076Z [INFO] TestAgent_RegisterCheck_Passing: Stopping server: protocol=HTTP address=127.0.0.1:30367 network=tcp === CONT TestAgent_RegisterCheckScriptsExecRemoteDisable writer.go:29: 2021-01-29T19:34:29.205Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:29.207Z [INFO] TestAgent_RegisterCheckScriptsExecRemoteDisable: Endpoints down --- PASS: TestAgent_RegisterCheckScriptsExecRemoteDisable (1.07s) === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.220Z [WARN] TestAgent_ForceLeavePrune-a1: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:29.222Z [DEBUG] TestAgent_ForceLeavePrune-a1.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:29.224Z [DEBUG] TestAgent_ForceLeavePrune-a1.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:29.229Z [INFO] TestAgent_ForceLeavePrune-a1.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 Address:127.0.0.1:29441}]" writer.go:29: 2021-01-29T19:34:29.233Z [INFO] TestAgent_ForceLeavePrune-a1.server.raft: entering follower state: follower="Node at 127.0.0.1:29441 [Follower]" leader= writer.go:29: 2021-01-29T19:34:29.235Z [INFO] TestAgent_ForceLeavePrune-a1.server.serf.wan: serf: EventMemberJoin: Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:29.245Z [INFO] TestAgent_ForceLeavePrune-a1.server.serf.lan: serf: EventMemberJoin: Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 127.0.0.1 writer.go:29: 2021-01-29T19:34:29.248Z [INFO] TestAgent_ForceLeavePrune-a1.server: Handled event for server in area: event=member-join server=Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524.dc1 area=wan writer.go:29: 2021-01-29T19:34:29.249Z [INFO] TestAgent_ForceLeavePrune-a1.server: Adding LAN server: server="Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 (Addr: tcp/127.0.0.1:29441) (DC: dc1)" writer.go:29: 2021-01-29T19:34:29.254Z [INFO] TestAgent_ForceLeavePrune-a1: Started DNS server: address=127.0.0.1:29436 network=tcp writer.go:29: 2021-01-29T19:34:29.258Z [INFO] TestAgent_ForceLeavePrune-a1: Started DNS server: address=127.0.0.1:29436 network=udp === CONT TestAgent_RegisterCheck writer.go:29: 2021-01-29T19:34:29.260Z [INFO] TestAgent_RegisterCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:29.261Z [INFO] TestAgent_RegisterCheck: Endpoints down --- PASS: TestAgent_RegisterCheck (0.97s) === CONT TestAgent_ForceLeave_ACLDeny === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.269Z [INFO] TestAgent_ForceLeavePrune-a1: Started HTTP server: address=127.0.0.1:29437 network=tcp writer.go:29: 2021-01-29T19:34:29.271Z [INFO] TestAgent_ForceLeavePrune-a1: started state syncer === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.282Z [WARN] TestAgent_ForceLeave_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:29.284Z [WARN] TestAgent_ForceLeave_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:29.286Z [DEBUG] TestAgent_ForceLeave_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:29.288Z [DEBUG] TestAgent_ForceLeave_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:29.294Z [INFO] TestAgent_ForceLeave_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0d91ab8a-b239-5dc6-a71d-a16b0f712a67 Address:127.0.0.1:30359}]" === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.294Z [WARN] TestAgent_ForceLeavePrune-a1.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:29.296Z [INFO] TestAgent_ForceLeavePrune-a1.server.raft: entering candidate state: node="Node at 127.0.0.1:29441 [Candidate]" term=2 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.297Z [INFO] TestAgent_ForceLeave_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:30359 [Follower]" leader= === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.299Z [DEBUG] TestAgent_ForceLeavePrune-a1.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:29.301Z [DEBUG] TestAgent_ForceLeavePrune-a1.server.raft: vote granted: from=a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 term=2 tally=1 === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:29.304Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: Waiting for endpoints to shut down === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.303Z [INFO] TestAgent_ForceLeavePrune-a1.server.raft: election won: tally=1 === CONT TestAgent_RegisterCheckScriptsExecDisable writer.go:29: 2021-01-29T19:34:29.305Z [INFO] TestAgent_RegisterCheckScriptsExecDisable: Endpoints down --- PASS: TestAgent_RegisterCheckScriptsExecDisable (1.03s) === CONT TestOpenMetricsMimeTypeHeaders === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.305Z [INFO] TestAgent_ForceLeavePrune-a1.server.raft: entering leader state: leader="Node at 127.0.0.1:29441 [Leader]" writer.go:29: 2021-01-29T19:34:29.307Z [INFO] TestAgent_ForceLeavePrune-a1.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:29.308Z [INFO] TestAgent_ForceLeavePrune-a1.server: New leader elected: payload=Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 === CONT TestAgent_Leave_ACLDeny === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.311Z [DEBUG] TestAgent_ForceLeavePrune-a1.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29441 --- PASS: TestOpenMetricsMimeTypeHeaders (0.01s) === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.302Z [INFO] TestAgent_ForceLeave_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67.dc1 127.0.0.1 === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.317Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:29.322Z [INFO] TestAgent_ForceLeavePrune-a1.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:29.324Z [INFO] TestAgent_ForceLeavePrune-a1.leader: started routine: routine="CA root pruning" === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.324Z [WARN] TestAgent_Leave_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.325Z [DEBUG] TestAgent_ForceLeavePrune-a1.server: Skipping self join check for node since the cluster is too small: node=Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.326Z [WARN] TestAgent_Leave_ACLDeny: bootstrap = true: do not enable unless necessary === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.327Z [INFO] TestAgent_ForceLeavePrune-a1.server: member joined, marking health alive: member=Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.327Z [DEBUG] TestAgent_Leave_ACLDeny.tlsutil: Update: version=1 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.329Z [INFO] TestAgent_ForceLeave_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 127.0.0.1 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.330Z [DEBUG] TestAgent_Leave_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.334Z [INFO] TestAgent_ForceLeave_ACLDeny.server: Adding LAN server: server="Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 (Addr: tcp/127.0.0.1:30359) (DC: dc1)" writer.go:29: 2021-01-29T19:34:29.336Z [INFO] TestAgent_ForceLeave_ACLDeny.server: Handled event for server in area: event=member-join server=Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67.dc1 area=wan === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.336Z [INFO] TestAgent_Leave_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7a21a0d0-aa4d-63bc-2166-734ab8477f8a Address:127.0.0.1:29435}]" writer.go:29: 2021-01-29T19:34:29.339Z [INFO] TestAgent_Leave_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a.dc1 127.0.0.1 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.340Z [WARN] TestAgent_ForceLeave_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.340Z [INFO] TestAgent_Leave_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:29435 [Follower]" leader= writer.go:29: 2021-01-29T19:34:29.343Z [INFO] TestAgent_Leave_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a 127.0.0.1 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.344Z [INFO] TestAgent_ForceLeave_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:30359 [Candidate]" term=2 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.348Z [INFO] TestAgent_Leave_ACLDeny.server: Handled event for server in area: event=member-join server=Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a.dc1 area=wan writer.go:29: 2021-01-29T19:34:29.350Z [INFO] TestAgent_Leave_ACLDeny: Started DNS server: address=127.0.0.1:29430 network=udp === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.355Z [INFO] TestAgent_ForceLeave_ACLDeny: Started DNS server: address=127.0.0.1:30354 network=udp === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.353Z [INFO] TestAgent_Leave_ACLDeny: Started DNS server: address=127.0.0.1:29430 network=tcp === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.356Z [INFO] TestAgent_ForceLeave_ACLDeny: Started DNS server: address=127.0.0.1:30354 network=tcp writer.go:29: 2021-01-29T19:34:29.358Z [INFO] TestAgent_ForceLeave_ACLDeny: Started HTTP server: address=127.0.0.1:30355 network=tcp === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.351Z [INFO] TestAgent_Leave_ACLDeny.server: Adding LAN server: server="Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a (Addr: tcp/127.0.0.1:29435) (DC: dc1)" === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.359Z [INFO] TestAgent_ForceLeave_ACLDeny: started state syncer writer.go:29: 2021-01-29T19:34:29.357Z [DEBUG] TestAgent_ForceLeave_ACLDeny.server.raft: votes: needed=1 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.360Z [INFO] TestAgent_Leave_ACLDeny: Started HTTP server: address=127.0.0.1:29431 network=tcp === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.362Z [DEBUG] TestAgent_ForceLeave_ACLDeny.server.raft: vote granted: from=0d91ab8a-b239-5dc6-a71d-a16b0f712a67 term=2 tally=1 writer.go:29: 2021-01-29T19:34:29.363Z [INFO] TestAgent_ForceLeave_ACLDeny.server.raft: election won: tally=1 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.363Z [INFO] TestAgent_Leave_ACLDeny: started state syncer === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.364Z [INFO] TestAgent_ForceLeave_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:30359 [Leader]" writer.go:29: 2021-01-29T19:34:29.367Z [INFO] TestAgent_ForceLeave_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:29.368Z [INFO] TestAgent_ForceLeave_ACLDeny.server: New leader elected: payload=Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 writer.go:29: 2021-01-29T19:34:29.370Z [INFO] TestAgent_ForceLeave_ACLDeny.server: initializing acls === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.374Z [WARN] TestAgent_ForceLeavePrune-a2: bootstrap = true: do not enable unless necessary === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.375Z [INFO] TestAgent_ForceLeave_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:29.377Z [WARN] TestAgent_ForceLeave_ACLDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.377Z [DEBUG] TestAgent_ForceLeavePrune-a2.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:29.380Z [DEBUG] TestAgent_ForceLeavePrune-a2.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.383Z [INFO] TestAgent_ForceLeave_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:29.385Z [INFO] TestAgent_ForceLeave_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:29.386Z [INFO] TestAgent_ForceLeave_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:29.387Z [INFO] TestAgent_ForceLeave_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:29.387Z [INFO] TestAgent_ForceLeave_ACLDeny.server: initializing acls === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.390Z [INFO] TestAgent_ForceLeavePrune-a2.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:63a671bf-78e4-baef-0dd1-26ae5f6ea044 Address:127.0.0.1:29471}]" === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.390Z [WARN] TestAgent_ForceLeave_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:29.389Z [INFO] TestAgent_ForceLeave_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 writer.go:29: 2021-01-29T19:34:29.394Z [INFO] TestAgent_ForceLeave_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67.dc1 writer.go:29: 2021-01-29T19:34:29.394Z [INFO] TestAgent_ForceLeave_ACLDeny.server: Updating LAN server: server="Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 (Addr: tcp/127.0.0.1:30359) (DC: dc1)" writer.go:29: 2021-01-29T19:34:29.392Z [DEBUG] TestAgent_ForceLeave_ACLDeny.server: transitioning out of legacy ACL mode === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.397Z [WARN] TestAgent_Leave_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.398Z [INFO] TestAgent_ForceLeave_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.392Z [INFO] TestAgent_ForceLeavePrune-a2.server.raft: entering follower state: follower="Node at 127.0.0.1:29471 [Follower]" leader= === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.397Z [INFO] TestAgent_ForceLeave_ACLDeny.server: Handled event for server in area: event=member-update server=Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67.dc1 area=wan === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.399Z [INFO] TestAgent_Leave_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:29435 [Candidate]" term=2 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.399Z [INFO] TestAgent_ForceLeave_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67.dc1 writer.go:29: 2021-01-29T19:34:29.401Z [INFO] TestAgent_ForceLeave_ACLDeny.server: Handled event for server in area: event=member-update server=Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67.dc1 area=wan === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.401Z [DEBUG] TestAgent_Leave_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.399Z [INFO] TestAgent_ForceLeave_ACLDeny.server: Updating LAN server: server="Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 (Addr: tcp/127.0.0.1:30359) (DC: dc1)" === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.398Z [INFO] TestAgent_ForceLeavePrune-a2.server.serf.wan: serf: EventMemberJoin: Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044.dc1 127.0.0.1 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.400Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.406Z [DEBUG] TestAgent_Leave_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:29.407Z [DEBUG] TestAgent_Leave_ACLDeny.server.raft: vote granted: from=7a21a0d0-aa4d-63bc-2166-734ab8477f8a term=2 tally=1 writer.go:29: 2021-01-29T19:34:29.408Z [INFO] TestAgent_Leave_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:29.410Z [INFO] TestAgent_Leave_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:29435 [Leader]" writer.go:29: 2021-01-29T19:34:29.412Z [INFO] TestAgent_Leave_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:29.416Z [INFO] TestAgent_Leave_ACLDeny.server: New leader elected: payload=Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.416Z [INFO] TestAgent_ForceLeave_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:29.417Z [INFO] TestAgent_ForceLeave_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:29.419Z [DEBUG] TestAgent_ForceLeave_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 writer.go:29: 2021-01-29T19:34:29.420Z [INFO] TestAgent_ForceLeave_ACLDeny.server: member joined, marking health alive: member=Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 writer.go:29: 2021-01-29T19:34:29.425Z [DEBUG] TestAgent_ForceLeave_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 writer.go:29: 2021-01-29T19:34:29.426Z [DEBUG] TestAgent_ForceLeave_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.416Z [INFO] TestAgent_ForceLeavePrune-a2.server.serf.lan: serf: EventMemberJoin: Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 127.0.0.1 writer.go:29: 2021-01-29T19:34:29.435Z [INFO] TestAgent_ForceLeavePrune-a2.server: Adding LAN server: server="Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 (Addr: tcp/127.0.0.1:29471) (DC: dc1)" === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.429Z [INFO] TestAgent_Leave_ACLDeny.server: initializing acls === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.437Z [INFO] TestAgent_ForceLeavePrune-a2.server: Handled event for server in area: event=member-join server=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044.dc1 area=wan writer.go:29: 2021-01-29T19:34:29.439Z [WARN] TestAgent_ForceLeavePrune-a2.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:29.440Z [INFO] TestAgent_ForceLeavePrune-a2.server.raft: entering candidate state: node="Node at 127.0.0.1:29471 [Candidate]" term=2 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.443Z [INFO] TestAgent_Leave_ACLDeny.server: Created ACL 'global-management' policy === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.443Z [DEBUG] TestAgent_ForceLeavePrune-a2.server.raft: votes: needed=1 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.445Z [WARN] TestAgent_Leave_ACLDeny.server: Configuring a non-UUID master token is deprecated === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.446Z [DEBUG] TestAgent_ForceLeavePrune-a2.server.raft: vote granted: from=63a671bf-78e4-baef-0dd1-26ae5f6ea044 term=2 tally=1 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.448Z [INFO] TestAgent_Leave_ACLDeny.server: Bootstrapped ACL master token from configuration === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.444Z [INFO] TestAgent_ForceLeavePrune-a2: Started DNS server: address=127.0.0.1:29466 network=tcp writer.go:29: 2021-01-29T19:34:29.450Z [INFO] TestAgent_ForceLeavePrune-a2: Started DNS server: address=127.0.0.1:29466 network=udp === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.451Z [INFO] TestAgent_Leave_ACLDeny.server: Created ACL anonymous token from configuration === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.454Z [DEBUG] TestAgent_ForceLeave_ACLDeny.acl: dropping node from result due to ACLs: node=Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 === RUN TestAgent_ForceLeave_ACLDeny/no_token === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.453Z [INFO] TestAgent_Leave_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" === RUN TestAgent_ForceLeave_ACLDeny/agent_master_token === RUN TestAgent_ForceLeave_ACLDeny/read-only_token === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.457Z [INFO] TestAgent_Leave_ACLDeny.leader: started routine: routine="acl token reaping" === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.449Z [INFO] TestAgent_ForceLeavePrune-a2.server.raft: election won: tally=1 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.465Z [INFO] TestAgent_Leave_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.467Z [INFO] TestAgent_ForceLeavePrune-a2.server.raft: entering leader state: leader="Node at 127.0.0.1:29471 [Leader]" === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.472Z [INFO] TestAgent_Leave_ACLDeny.server: Updating LAN server: server="Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a (Addr: tcp/127.0.0.1:29435) (DC: dc1)" writer.go:29: 2021-01-29T19:34:29.472Z [INFO] TestAgent_Leave_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a.dc1 writer.go:29: 2021-01-29T19:34:29.476Z [INFO] TestAgent_Leave_ACLDeny.server: Handled event for server in area: event=member-update server=Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a.dc1 area=wan === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.463Z [INFO] TestAgent_ForceLeavePrune-a2: Started HTTP server: address=127.0.0.1:29467 network=tcp writer.go:29: 2021-01-29T19:34:29.472Z [INFO] TestAgent_ForceLeavePrune-a2.server: cluster leadership acquired === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.477Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.474Z [INFO] TestAgent_ForceLeavePrune-a2.server: New leader elected: payload=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 writer.go:29: 2021-01-29T19:34:29.475Z [DEBUG] TestAgent_ForceLeavePrune-a2.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29471 writer.go:29: 2021-01-29T19:34:29.482Z [INFO] TestAgent_ForceLeavePrune-a2: started state syncer === RUN TestAgent_ForceLeave_ACLDeny/operator_write_token === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.490Z [INFO] TestAgent_ForceLeavePrune-a2: Synced node info === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.495Z [INFO] TestAgent_Leave_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.495Z [DEBUG] TestAgent_ForceLeavePrune-a2: Node info in sync writer.go:29: 2021-01-29T19:34:29.495Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.497Z [INFO] TestAgent_Leave_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:29.500Z [DEBUG] TestAgent_Leave_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a writer.go:29: 2021-01-29T19:34:29.502Z [INFO] TestAgent_Leave_ACLDeny.server: member joined, marking health alive: member=Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.500Z [INFO] TestAgent_ForceLeavePrune-a1: (LAN) joining: lan_addresses=[127.0.0.1:29469] writer.go:29: 2021-01-29T19:34:29.506Z [INFO] TestAgent_ForceLeavePrune-a2.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:29.507Z [INFO] TestAgent_ForceLeavePrune-a2.leader: started routine: routine="CA root pruning" === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.508Z [INFO] TestAgent_ForceLeave_ACLDeny: Force leaving node: node=Node-0d91ab8a-b239-5dc6-a71d-a16b0f712a67 === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.509Z [DEBUG] TestAgent_ForceLeavePrune-a2.server: Skipping self join check for node since the cluster is too small: node=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.509Z [DEBUG] TestAgent_ForceLeave_ACLDeny.server.serf.lan: serf: Refuting an older leave intent === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.510Z [INFO] TestAgent_ForceLeavePrune-a2.server: member joined, marking health alive: member=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.511Z [DEBUG] TestAgent_ForceLeave_ACLDeny.server.serf.wan: serf: Refuting an older leave intent === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.513Z [DEBUG] TestAgent_Leave_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.512Z [DEBUG] TestAgent_ForceLeavePrune-a2.server.memberlist.lan: memberlist: Stream connection from=127.0.0.1:38208 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.513Z [INFO] TestAgent_ForceLeave_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:29.517Z [INFO] TestAgent_ForceLeave_ACLDeny.server: shutting down server === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.505Z [DEBUG] TestAgent_ForceLeavePrune-a1.server.memberlist.lan: memberlist: Initiating push/pull sync with: 127.0.0.1:29469 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.519Z [DEBUG] TestAgent_ForceLeave_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:29.522Z [DEBUG] TestAgent_ForceLeave_ACLDeny.leader: stopping routine: routine="acl token reaping" === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.524Z [INFO] TestAgent_ForceLeavePrune-a2.server.serf.lan: serf: EventMemberJoin: Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 127.0.0.1 writer.go:29: 2021-01-29T19:34:29.524Z [INFO] TestAgent_ForceLeavePrune-a1.server.serf.lan: serf: EventMemberJoin: Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 127.0.0.1 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.520Z [ERROR] TestAgent_ForceLeave_ACLDeny.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:29.522Z [DEBUG] TestAgent_ForceLeave_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.526Z [INFO] TestAgent_ForceLeavePrune-a2.server: Adding LAN server: server="Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 (Addr: tcp/127.0.0.1:29441) (DC: dc1)" writer.go:29: 2021-01-29T19:34:29.528Z [INFO] TestAgent_ForceLeavePrune-a1: (LAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:34:29.529Z [INFO] TestAgent_ForceLeavePrune-a2.server: New leader elected: payload=Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 writer.go:29: 2021-01-29T19:34:29.531Z [DEBUG] TestAgent_ForceLeavePrune-a1: systemd notify failed: error="No socket" writer.go:29: 2021-01-29T19:34:29.530Z [ERROR] TestAgent_ForceLeavePrune-a2.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 other=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.525Z [DEBUG] TestAgent_ForceLeave_ACLDeny.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.528Z [INFO] TestAgent_ForceLeavePrune-a1.server: Adding LAN server: server="Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 (Addr: tcp/127.0.0.1:29471) (DC: dc1)" writer.go:29: 2021-01-29T19:34:29.535Z [INFO] TestAgent_ForceLeavePrune-a2.server: member joined, marking health alive: member=Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.525Z [DEBUG] TestAgent_ForceLeave_ACLDeny.leader: stopped routine: routine="acl token reaping" === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.536Z [ERROR] TestAgent_ForceLeavePrune-a1.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 other=Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524 === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.535Z [WARN] TestAgent_ForceLeave_ACLDeny.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.534Z [INFO] TestAgent_ForceLeavePrune-a2: Requesting shutdown writer.go:29: 2021-01-29T19:34:29.537Z [WARN] TestAgent_ForceLeavePrune-a2.server.memberlist.wan: memberlist: Failed to resolve Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524.dc1/127.0.0.1:29440: lookup Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524.dc1/127.0.0.1: no such host === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.535Z [DEBUG] TestAgent_ForceLeave_ACLDeny.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.540Z [INFO] TestAgent_ForceLeavePrune-a1.server: member joined, marking health alive: member=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 writer.go:29: 2021-01-29T19:34:29.544Z [DEBUG] TestAgent_ForceLeavePrune-a2.server: Failed to flood-join server at address: server=Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524.dc1 address=127.0.0.1:29440 error="1 error occurred: * Failed to resolve Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524.dc1/127.0.0.1:29440: lookup Node-a9ea64ce-3aa8-4ff4-cdd5-31af4f965524.dc1/127.0.0.1: no such host " writer.go:29: 2021-01-29T19:34:29.544Z [WARN] TestAgent_ForceLeavePrune-a1.server.memberlist.wan: memberlist: Failed to resolve Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044.dc1/127.0.0.1:29470: lookup Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044.dc1/127.0.0.1: no such host writer.go:29: 2021-01-29T19:34:29.542Z [INFO] TestAgent_ForceLeavePrune-a2.server: shutting down server writer.go:29: 2021-01-29T19:34:29.548Z [DEBUG] TestAgent_ForceLeavePrune-a2.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:29.547Z [DEBUG] TestAgent_ForceLeavePrune-a1.server: Failed to flood-join server at address: server=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044.dc1 address=127.0.0.1:29470 error="1 error occurred: * Failed to resolve Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044.dc1/127.0.0.1:29470: lookup Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044.dc1/127.0.0.1: no such host " === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.542Z [WARN] TestAgent_ForceLeave_ACLDeny.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.550Z [WARN] TestAgent_ForceLeavePrune-a2.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:29.550Z [DEBUG] TestAgent_ForceLeavePrune-a2.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.555Z [INFO] TestAgent_ForceLeave_ACLDeny.server.router.manager: shutting down === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.559Z [WARN] TestAgent_ForceLeavePrune-a2.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.555Z [INFO] TestAgent_ForceLeave_ACLDeny: consul server down === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.562Z [INFO] TestAgent_ForceLeavePrune-a2.server.router.manager: shutting down === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.563Z [INFO] TestAgent_ForceLeave_ACLDeny: shutdown complete === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.562Z [INFO] TestAgent_ForceLeavePrune-a2: consul server down === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.565Z [INFO] TestAgent_ForceLeave_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30354 network=tcp === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.566Z [INFO] TestAgent_ForceLeavePrune-a2: shutdown complete writer.go:29: 2021-01-29T19:34:29.567Z [INFO] TestAgent_ForceLeavePrune-a2: Stopping server: protocol=DNS address=127.0.0.1:29466 network=tcp writer.go:29: 2021-01-29T19:34:29.568Z [INFO] TestAgent_ForceLeavePrune-a2: Stopping server: protocol=DNS address=127.0.0.1:29466 network=udp === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.567Z [INFO] TestAgent_ForceLeave_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:30354 network=udp === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.570Z [INFO] TestAgent_ForceLeavePrune-a2: Stopping server: protocol=HTTP address=127.0.0.1:29467 network=tcp === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:29.570Z [INFO] TestAgent_ForceLeave_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:30355 network=tcp === CONT TestAgent_RegisterCheck_Passing writer.go:29: 2021-01-29T19:34:29.578Z [INFO] TestAgent_RegisterCheck_Passing: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:29.579Z [INFO] TestAgent_RegisterCheck_Passing: Endpoints down --- PASS: TestAgent_RegisterCheck_Passing (1.46s) === CONT TestAgent_JoinLANNotify === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.600Z [DEBUG] TestAgent_Leave_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.601Z [DEBUG] TestAgent_ForceLeavePrune-a1: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_JoinLANNotify writer.go:29: 2021-01-29T19:34:29.600Z [WARN] TestAgent_JoinLANNotify: bootstrap = true: do not enable unless necessary === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:29.603Z [INFO] TestAgent_ForceLeavePrune-a1: Synced node info writer.go:29: 2021-01-29T19:34:29.604Z [DEBUG] TestAgent_ForceLeavePrune-a1: Node info in sync === CONT TestAgent_JoinLANNotify writer.go:29: 2021-01-29T19:34:29.603Z [DEBUG] TestAgent_JoinLANNotify.tlsutil: Update: version=1 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.605Z [INFO] TestAgent_Leave_ACLDeny: Synced node info === CONT TestAgent_JoinLANNotify writer.go:29: 2021-01-29T19:34:29.605Z [DEBUG] TestAgent_JoinLANNotify.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:29.613Z [INFO] TestAgent_JoinLANNotify.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f76a97a3-d600-b7fe-1185-d1e572616b8e Address:127.0.0.1:29447}]" writer.go:29: 2021-01-29T19:34:29.616Z [INFO] TestAgent_JoinLANNotify.server.raft: entering follower state: follower="Node at 127.0.0.1:29447 [Follower]" leader= writer.go:29: 2021-01-29T19:34:29.618Z [INFO] TestAgent_JoinLANNotify.server.serf.wan: serf: EventMemberJoin: Node-f76a97a3-d600-b7fe-1185-d1e572616b8e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:29.624Z [INFO] TestAgent_JoinLANNotify.server.serf.lan: serf: EventMemberJoin: Node-f76a97a3-d600-b7fe-1185-d1e572616b8e 127.0.0.1 writer.go:29: 2021-01-29T19:34:29.632Z [INFO] TestAgent_JoinLANNotify: Started DNS server: address=127.0.0.1:29442 network=udp writer.go:29: 2021-01-29T19:34:29.634Z [INFO] TestAgent_JoinLANNotify.server: Adding LAN server: server="Node-f76a97a3-d600-b7fe-1185-d1e572616b8e (Addr: tcp/127.0.0.1:29447) (DC: dc1)" writer.go:29: 2021-01-29T19:34:29.634Z [INFO] TestAgent_JoinLANNotify.server: Handled event for server in area: event=member-join server=Node-f76a97a3-d600-b7fe-1185-d1e572616b8e.dc1 area=wan writer.go:29: 2021-01-29T19:34:29.635Z [INFO] TestAgent_JoinLANNotify: Started DNS server: address=127.0.0.1:29442 network=tcp writer.go:29: 2021-01-29T19:34:29.637Z [INFO] TestAgent_JoinLANNotify: Started HTTP server: address=127.0.0.1:29443 network=tcp writer.go:29: 2021-01-29T19:34:29.638Z [INFO] TestAgent_JoinLANNotify: started state syncer writer.go:29: 2021-01-29T19:34:29.660Z [WARN] TestAgent_JoinLANNotify.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:29.662Z [INFO] TestAgent_JoinLANNotify.server.raft: entering candidate state: node="Node at 127.0.0.1:29447 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:29.664Z [DEBUG] TestAgent_JoinLANNotify.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:29.665Z [DEBUG] TestAgent_JoinLANNotify.server.raft: vote granted: from=f76a97a3-d600-b7fe-1185-d1e572616b8e term=2 tally=1 writer.go:29: 2021-01-29T19:34:29.665Z [INFO] TestAgent_JoinLANNotify.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:29.666Z [INFO] TestAgent_JoinLANNotify.server.raft: entering leader state: leader="Node at 127.0.0.1:29447 [Leader]" writer.go:29: 2021-01-29T19:34:29.667Z [INFO] TestAgent_JoinLANNotify.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:29.668Z [INFO] TestAgent_JoinLANNotify.server: New leader elected: payload=Node-f76a97a3-d600-b7fe-1185-d1e572616b8e writer.go:29: 2021-01-29T19:34:29.669Z [DEBUG] TestAgent_JoinLANNotify.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29447 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.672Z [DEBUG] TestAgent_Leave_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:29.673Z [DEBUG] TestAgent_Leave_ACLDeny: Node info in sync === CONT TestAgent_JoinLANNotify writer.go:29: 2021-01-29T19:34:29.674Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.675Z [DEBUG] TestAgent_Leave_ACLDeny: Node info in sync === CONT TestAgent_JoinLANNotify writer.go:29: 2021-01-29T19:34:29.679Z [INFO] TestAgent_JoinLANNotify.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:29.680Z [INFO] TestAgent_JoinLANNotify.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:29.681Z [DEBUG] TestAgent_JoinLANNotify.server: Skipping self join check for node since the cluster is too small: node=Node-f76a97a3-d600-b7fe-1185-d1e572616b8e writer.go:29: 2021-01-29T19:34:29.681Z [INFO] TestAgent_JoinLANNotify.server: member joined, marking health alive: member=Node-f76a97a3-d600-b7fe-1185-d1e572616b8e === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.685Z [DEBUG] TestAgent_Leave_ACLDeny.acl: dropping node from result due to ACLs: node=Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a writer.go:29: 2021-01-29T19:34:29.690Z [DEBUG] TestAgent_Leave_ACLDeny.acl: dropping node from result due to ACLs: node=Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a === RUN TestAgent_Leave_ACLDeny/no_token === RUN TestAgent_Leave_ACLDeny/read-only_token === RUN TestAgent_Leave_ACLDeny/agent_master_token === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:29.702Z [INFO] TestAgent_Leave_ACLDeny.server: server starting leave writer.go:29: 2021-01-29T19:34:29.705Z [INFO] TestAgent_Leave_ACLDeny.server.serf.wan: serf: EventMemberLeave: Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:29.707Z [INFO] TestAgent_Leave_ACLDeny.server: Handled event for server in area: event=member-leave server=Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a.dc1 area=wan writer.go:29: 2021-01-29T19:34:29.707Z [INFO] TestAgent_Leave_ACLDeny.server.router.manager: shutting down === CONT TestAgent_JoinLANNotify writer.go:29: 2021-01-29T19:34:29.912Z [DEBUG] TestAgent_JoinLANNotify.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:29.914Z [INFO] TestAgent_JoinLANNotify.client.serf.lan: serf: EventMemberJoin: Node-b6b1624b-de9f-0366-6e78-3ba6acc786a0 127.0.0.1 writer.go:29: 2021-01-29T19:34:29.917Z [INFO] TestAgent_JoinLANNotify: Started DNS server: address=127.0.0.1:29448 network=tcp writer.go:29: 2021-01-29T19:34:29.918Z [INFO] TestAgent_JoinLANNotify: Started DNS server: address=127.0.0.1:29448 network=udp writer.go:29: 2021-01-29T19:34:29.920Z [INFO] TestAgent_JoinLANNotify: Started HTTP server: address=127.0.0.1:29449 network=tcp writer.go:29: 2021-01-29T19:34:29.921Z [INFO] TestAgent_JoinLANNotify: started state syncer writer.go:29: 2021-01-29T19:34:29.921Z [WARN] TestAgent_JoinLANNotify.client.manager: No servers available writer.go:29: 2021-01-29T19:34:29.923Z [ERROR] TestAgent_JoinLANNotify.anti_entropy: failed to sync remote state: error="No known Consul servers" writer.go:29: 2021-01-29T19:34:29.922Z [INFO] TestAgent_JoinLANNotify: (LAN) joining: lan_addresses=[127.0.0.1:29451] writer.go:29: 2021-01-29T19:34:29.926Z [DEBUG] TestAgent_JoinLANNotify.server.memberlist.lan: memberlist: Initiating push/pull sync with: 127.0.0.1:29451 writer.go:29: 2021-01-29T19:34:29.926Z [DEBUG] TestAgent_JoinLANNotify.client.memberlist.lan: memberlist: Stream connection from=127.0.0.1:44118 writer.go:29: 2021-01-29T19:34:29.929Z [INFO] TestAgent_JoinLANNotify.client.serf.lan: serf: EventMemberJoin: Node-f76a97a3-d600-b7fe-1185-d1e572616b8e 127.0.0.1 writer.go:29: 2021-01-29T19:34:29.930Z [INFO] TestAgent_JoinLANNotify.client: adding server: server="Node-f76a97a3-d600-b7fe-1185-d1e572616b8e (Addr: tcp/127.0.0.1:29447) (DC: dc1)" writer.go:29: 2021-01-29T19:34:29.929Z [INFO] TestAgent_JoinLANNotify.server.serf.lan: serf: EventMemberJoin: Node-b6b1624b-de9f-0366-6e78-3ba6acc786a0 127.0.0.1 writer.go:29: 2021-01-29T19:34:29.932Z [INFO] TestAgent_JoinLANNotify.client: New leader elected: payload=Node-f76a97a3-d600-b7fe-1185-d1e572616b8e writer.go:29: 2021-01-29T19:34:29.933Z [INFO] TestAgent_JoinLANNotify: (LAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:34:29.935Z [INFO] TestAgent_JoinLANNotify: Requesting shutdown writer.go:29: 2021-01-29T19:34:29.933Z [INFO] TestAgent_JoinLANNotify.server: member joined, marking health alive: member=Node-b6b1624b-de9f-0366-6e78-3ba6acc786a0 writer.go:29: 2021-01-29T19:34:29.936Z [INFO] TestAgent_JoinLANNotify.client: shutting down client writer.go:29: 2021-01-29T19:34:29.938Z [WARN] TestAgent_JoinLANNotify.client.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:29.938Z [INFO] TestAgent_JoinLANNotify.client.manager: shutting down writer.go:29: 2021-01-29T19:34:29.940Z [INFO] TestAgent_JoinLANNotify: consul client down writer.go:29: 2021-01-29T19:34:29.942Z [INFO] TestAgent_JoinLANNotify: shutdown complete writer.go:29: 2021-01-29T19:34:29.943Z [INFO] TestAgent_JoinLANNotify: Stopping server: protocol=DNS address=127.0.0.1:29448 network=tcp writer.go:29: 2021-01-29T19:34:29.944Z [INFO] TestAgent_JoinLANNotify: Stopping server: protocol=DNS address=127.0.0.1:29448 network=udp writer.go:29: 2021-01-29T19:34:29.962Z [INFO] TestAgent_JoinLANNotify: Stopping server: protocol=HTTP address=127.0.0.1:29449 network=tcp writer.go:29: 2021-01-29T19:34:30.058Z [DEBUG] TestAgent_JoinLANNotify: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:30.062Z [INFO] TestAgent_JoinLANNotify: Synced node info === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:30.072Z [INFO] TestAgent_ForceLeavePrune-a2: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:30.073Z [INFO] TestAgent_ForceLeavePrune-a2: Endpoints down === CONT TestAgent_ForceLeave_ACLDeny writer.go:29: 2021-01-29T19:34:30.076Z [INFO] TestAgent_ForceLeave_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:30.078Z [INFO] TestAgent_ForceLeave_ACLDeny: Endpoints down --- PASS: TestAgent_ForceLeave_ACLDeny (0.82s) --- PASS: TestAgent_ForceLeave_ACLDeny/no_token (0.00s) --- PASS: TestAgent_ForceLeave_ACLDeny/agent_master_token (0.00s) --- PASS: TestAgent_ForceLeave_ACLDeny/read-only_token (0.03s) --- PASS: TestAgent_ForceLeave_ACLDeny/operator_write_token (0.02s) === CONT TestAgent_Join_ACLDeny writer.go:29: 2021-01-29T19:34:30.104Z [WARN] TestAgent_Join_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:30.106Z [WARN] TestAgent_Join_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:30.109Z [DEBUG] TestAgent_Join_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:30.113Z [DEBUG] TestAgent_Join_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:30.122Z [INFO] TestAgent_Join_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ca172914-74e0-b789-2f09-28fa74291107 Address:127.0.0.1:29477}]" writer.go:29: 2021-01-29T19:34:30.130Z [INFO] TestAgent_Join_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:29477 [Follower]" leader= writer.go:29: 2021-01-29T19:34:30.135Z [INFO] TestAgent_Join_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-ca172914-74e0-b789-2f09-28fa74291107.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:30.149Z [INFO] TestAgent_Join_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-ca172914-74e0-b789-2f09-28fa74291107 127.0.0.1 writer.go:29: 2021-01-29T19:34:30.157Z [INFO] TestAgent_Join_ACLDeny.server: Adding LAN server: server="Node-ca172914-74e0-b789-2f09-28fa74291107 (Addr: tcp/127.0.0.1:29477) (DC: dc1)" writer.go:29: 2021-01-29T19:34:30.162Z [INFO] TestAgent_Join_ACLDeny.server: Handled event for server in area: event=member-join server=Node-ca172914-74e0-b789-2f09-28fa74291107.dc1 area=wan writer.go:29: 2021-01-29T19:34:30.163Z [INFO] TestAgent_Join_ACLDeny: Started DNS server: address=127.0.0.1:29472 network=udp writer.go:29: 2021-01-29T19:34:30.169Z [INFO] TestAgent_Join_ACLDeny: Started DNS server: address=127.0.0.1:29472 network=tcp writer.go:29: 2021-01-29T19:34:30.171Z [WARN] TestAgent_Join_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:30.173Z [INFO] TestAgent_Join_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:29477 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:30.175Z [DEBUG] TestAgent_Join_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:30.176Z [DEBUG] TestAgent_Join_ACLDeny.server.raft: vote granted: from=ca172914-74e0-b789-2f09-28fa74291107 term=2 tally=1 writer.go:29: 2021-01-29T19:34:30.178Z [INFO] TestAgent_Join_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:30.179Z [INFO] TestAgent_Join_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:29477 [Leader]" writer.go:29: 2021-01-29T19:34:30.180Z [INFO] TestAgent_Join_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:30.181Z [INFO] TestAgent_Join_ACLDeny: Started HTTP server: address=127.0.0.1:29473 network=tcp writer.go:29: 2021-01-29T19:34:30.182Z [INFO] TestAgent_Join_ACLDeny.server: New leader elected: payload=Node-ca172914-74e0-b789-2f09-28fa74291107 writer.go:29: 2021-01-29T19:34:30.183Z [INFO] TestAgent_Join_ACLDeny: started state syncer writer.go:29: 2021-01-29T19:34:30.183Z [INFO] TestAgent_Join_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:30.184Z [ERROR] TestAgent_Join_ACLDeny.anti_entropy: failed to sync remote state: error="ACL not found" writer.go:29: 2021-01-29T19:34:30.194Z [INFO] TestAgent_Join_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:30.196Z [WARN] TestAgent_Join_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:30.199Z [INFO] TestAgent_Join_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:30.202Z [INFO] TestAgent_Join_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:30.204Z [INFO] TestAgent_Join_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:30.205Z [INFO] TestAgent_Join_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:30.207Z [INFO] TestAgent_Join_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-ca172914-74e0-b789-2f09-28fa74291107 writer.go:29: 2021-01-29T19:34:30.209Z [INFO] TestAgent_Join_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-ca172914-74e0-b789-2f09-28fa74291107.dc1 writer.go:29: 2021-01-29T19:34:30.209Z [INFO] TestAgent_Join_ACLDeny.server: Updating LAN server: server="Node-ca172914-74e0-b789-2f09-28fa74291107 (Addr: tcp/127.0.0.1:29477) (DC: dc1)" writer.go:29: 2021-01-29T19:34:30.212Z [INFO] TestAgent_Join_ACLDeny.server: Handled event for server in area: event=member-update server=Node-ca172914-74e0-b789-2f09-28fa74291107.dc1 area=wan writer.go:29: 2021-01-29T19:34:30.214Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:30.222Z [WARN] TestAgent_Join_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:30.223Z [DEBUG] TestAgent_Join_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:30.226Z [INFO] TestAgent_Join_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:30.228Z [INFO] TestAgent_Join_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:30.229Z [DEBUG] TestAgent_Join_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-ca172914-74e0-b789-2f09-28fa74291107 writer.go:29: 2021-01-29T19:34:30.231Z [INFO] TestAgent_Join_ACLDeny.server: member joined, marking health alive: member=Node-ca172914-74e0-b789-2f09-28fa74291107 writer.go:29: 2021-01-29T19:34:30.231Z [DEBUG] TestAgent_Join_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:30.238Z [INFO] TestAgent_Join_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4a852d68-0c07-c29e-db52-5763c6794667 Address:127.0.0.1:29465}]" writer.go:29: 2021-01-29T19:34:30.241Z [INFO] TestAgent_Join_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:29465 [Follower]" leader= writer.go:29: 2021-01-29T19:34:30.241Z [INFO] TestAgent_Join_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-4a852d68-0c07-c29e-db52-5763c6794667.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:30.248Z [INFO] TestAgent_Join_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-4a852d68-0c07-c29e-db52-5763c6794667 127.0.0.1 writer.go:29: 2021-01-29T19:34:30.251Z [INFO] TestAgent_Join_ACLDeny.server: Handled event for server in area: event=member-join server=Node-4a852d68-0c07-c29e-db52-5763c6794667.dc1 area=wan writer.go:29: 2021-01-29T19:34:30.251Z [INFO] TestAgent_Join_ACLDeny.server: Adding LAN server: server="Node-4a852d68-0c07-c29e-db52-5763c6794667 (Addr: tcp/127.0.0.1:29465) (DC: dc1)" writer.go:29: 2021-01-29T19:34:30.252Z [DEBUG] TestAgent_Join_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-ca172914-74e0-b789-2f09-28fa74291107 writer.go:29: 2021-01-29T19:34:30.262Z [INFO] TestAgent_Join_ACLDeny: Started DNS server: address=127.0.0.1:29460 network=tcp writer.go:29: 2021-01-29T19:34:30.264Z [INFO] TestAgent_Join_ACLDeny: Started DNS server: address=127.0.0.1:29460 network=udp writer.go:29: 2021-01-29T19:34:30.266Z [INFO] TestAgent_Join_ACLDeny: Started HTTP server: address=127.0.0.1:29461 network=tcp writer.go:29: 2021-01-29T19:34:30.267Z [INFO] TestAgent_Join_ACLDeny: started state syncer writer.go:29: 2021-01-29T19:34:30.296Z [WARN] TestAgent_Join_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:30.298Z [INFO] TestAgent_Join_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:29465 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:30.301Z [DEBUG] TestAgent_Join_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:30.302Z [DEBUG] TestAgent_Join_ACLDeny.server.raft: vote granted: from=4a852d68-0c07-c29e-db52-5763c6794667 term=2 tally=1 writer.go:29: 2021-01-29T19:34:30.304Z [INFO] TestAgent_Join_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:30.305Z [INFO] TestAgent_Join_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:29465 [Leader]" writer.go:29: 2021-01-29T19:34:30.307Z [INFO] TestAgent_Join_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:30.307Z [INFO] TestAgent_Join_ACLDeny.server: New leader elected: payload=Node-4a852d68-0c07-c29e-db52-5763c6794667 writer.go:29: 2021-01-29T19:34:30.323Z [DEBUG] TestAgent_Join_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29465 writer.go:29: 2021-01-29T19:34:30.329Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:30.352Z [INFO] TestAgent_Join_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:30.354Z [INFO] TestAgent_Join_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:30.356Z [DEBUG] TestAgent_Join_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-4a852d68-0c07-c29e-db52-5763c6794667 writer.go:29: 2021-01-29T19:34:30.357Z [INFO] TestAgent_Join_ACLDeny.server: member joined, marking health alive: member=Node-4a852d68-0c07-c29e-db52-5763c6794667 writer.go:29: 2021-01-29T19:34:30.377Z [DEBUG] TestAgent_Join_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:30.388Z [INFO] TestAgent_Join_ACLDeny: Synced node info writer.go:29: 2021-01-29T19:34:30.390Z [DEBUG] TestAgent_Join_ACLDeny: Node info in sync === CONT TestAgent_JoinLANNotify writer.go:29: 2021-01-29T19:34:30.463Z [INFO] TestAgent_JoinLANNotify: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:30.464Z [INFO] TestAgent_JoinLANNotify: Endpoints down writer.go:29: 2021-01-29T19:34:30.465Z [INFO] TestAgent_JoinLANNotify: Requesting shutdown writer.go:29: 2021-01-29T19:34:30.467Z [INFO] TestAgent_JoinLANNotify.server: shutting down server writer.go:29: 2021-01-29T19:34:30.468Z [DEBUG] TestAgent_JoinLANNotify.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:30.469Z [WARN] TestAgent_JoinLANNotify.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:30.471Z [DEBUG] TestAgent_JoinLANNotify.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:30.483Z [WARN] TestAgent_JoinLANNotify.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:30.489Z [INFO] TestAgent_JoinLANNotify.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:30.492Z [INFO] TestAgent_JoinLANNotify: consul server down writer.go:29: 2021-01-29T19:34:30.493Z [INFO] TestAgent_JoinLANNotify: shutdown complete writer.go:29: 2021-01-29T19:34:30.495Z [INFO] TestAgent_JoinLANNotify: Stopping server: protocol=DNS address=127.0.0.1:29442 network=tcp writer.go:29: 2021-01-29T19:34:30.503Z [INFO] TestAgent_JoinLANNotify: Stopping server: protocol=DNS address=127.0.0.1:29442 network=udp writer.go:29: 2021-01-29T19:34:30.505Z [INFO] TestAgent_JoinLANNotify: Stopping server: protocol=HTTP address=127.0.0.1:29443 network=tcp === CONT TestAgent_Join_ACLDeny writer.go:29: 2021-01-29T19:34:30.590Z [DEBUG] TestAgent_Join_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:30.593Z [INFO] TestAgent_Join_ACLDeny: Synced node info writer.go:29: 2021-01-29T19:34:30.594Z [DEBUG] TestAgent_Join_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:30.676Z [DEBUG] TestAgent_Join_ACLDeny.acl: dropping node from result due to ACLs: node=Node-ca172914-74e0-b789-2f09-28fa74291107 === RUN TestAgent_Join_ACLDeny/no_token === RUN TestAgent_Join_ACLDeny/agent_master_token === CONT TestAgent_Join_ACLDeny writer.go:29: 2021-01-29T19:34:30.683Z [INFO] TestAgent_Join_ACLDeny: (LAN) joining: lan_addresses=[127.0.0.1:29463] writer.go:29: 2021-01-29T19:34:30.685Z [DEBUG] TestAgent_Join_ACLDeny.server.memberlist.lan: memberlist: Initiating push/pull sync with: 127.0.0.1:29463 writer.go:29: 2021-01-29T19:34:30.686Z [DEBUG] TestAgent_Join_ACLDeny.server.memberlist.lan: memberlist: Stream connection from=127.0.0.1:43054 writer.go:29: 2021-01-29T19:34:30.690Z [INFO] TestAgent_Join_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-ca172914-74e0-b789-2f09-28fa74291107 127.0.0.1 writer.go:29: 2021-01-29T19:34:30.693Z [INFO] TestAgent_Join_ACLDeny.server: Adding LAN server: server="Node-ca172914-74e0-b789-2f09-28fa74291107 (Addr: tcp/127.0.0.1:29477) (DC: dc1)" writer.go:29: 2021-01-29T19:34:30.694Z [INFO] TestAgent_Join_ACLDeny.server: New leader elected: payload=Node-ca172914-74e0-b789-2f09-28fa74291107 writer.go:29: 2021-01-29T19:34:30.694Z [ERROR] TestAgent_Join_ACLDeny.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-ca172914-74e0-b789-2f09-28fa74291107 other=Node-4a852d68-0c07-c29e-db52-5763c6794667 writer.go:29: 2021-01-29T19:34:30.695Z [INFO] TestAgent_Join_ACLDeny.server: member joined, marking health alive: member=Node-ca172914-74e0-b789-2f09-28fa74291107 writer.go:29: 2021-01-29T19:34:30.690Z [INFO] TestAgent_Join_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-4a852d68-0c07-c29e-db52-5763c6794667 127.0.0.1 writer.go:29: 2021-01-29T19:34:30.697Z [INFO] TestAgent_Join_ACLDeny: (LAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:34:30.698Z [DEBUG] TestAgent_Join_ACLDeny: systemd notify failed: error="No socket" === RUN TestAgent_Join_ACLDeny/read-only_token === CONT TestAgent_Join_ACLDeny writer.go:29: 2021-01-29T19:34:30.698Z [INFO] TestAgent_Join_ACLDeny.server: Adding LAN server: server="Node-4a852d68-0c07-c29e-db52-5763c6794667 (Addr: tcp/127.0.0.1:29465) (DC: dc1)" writer.go:29: 2021-01-29T19:34:30.701Z [ERROR] TestAgent_Join_ACLDeny.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-4a852d68-0c07-c29e-db52-5763c6794667 other=Node-ca172914-74e0-b789-2f09-28fa74291107 writer.go:29: 2021-01-29T19:34:30.703Z [INFO] TestAgent_Join_ACLDeny.server: member joined, marking health alive: member=Node-4a852d68-0c07-c29e-db52-5763c6794667 writer.go:29: 2021-01-29T19:34:30.704Z [INFO] TestAgent_Join_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:30.706Z [INFO] TestAgent_Join_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:30.707Z [DEBUG] TestAgent_Join_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:30.711Z [WARN] TestAgent_Join_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:30.711Z [WARN] TestAgent_Join_ACLDeny.server.memberlist.wan: memberlist: Failed to resolve Node-4a852d68-0c07-c29e-db52-5763c6794667.dc1/127.0.0.1:29464: lookup Node-4a852d68-0c07-c29e-db52-5763c6794667.dc1/127.0.0.1: no such host writer.go:29: 2021-01-29T19:34:30.714Z [DEBUG] TestAgent_Join_ACLDeny.server: Failed to flood-join server at address: server=Node-4a852d68-0c07-c29e-db52-5763c6794667.dc1 address=127.0.0.1:29464 error="1 error occurred: * Failed to resolve Node-4a852d68-0c07-c29e-db52-5763c6794667.dc1/127.0.0.1:29464: lookup Node-4a852d68-0c07-c29e-db52-5763c6794667.dc1/127.0.0.1: no such host " writer.go:29: 2021-01-29T19:34:30.707Z [WARN] TestAgent_Join_ACLDeny.server.memberlist.wan: memberlist: Failed to resolve Node-ca172914-74e0-b789-2f09-28fa74291107.dc1/127.0.0.1:29476: lookup Node-ca172914-74e0-b789-2f09-28fa74291107.dc1/127.0.0.1: no such host writer.go:29: 2021-01-29T19:34:30.711Z [DEBUG] TestAgent_Join_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:30.713Z [WARN] TestAgent_Join_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:30.717Z [DEBUG] TestAgent_Join_ACLDeny.server: Failed to flood-join server at address: server=Node-ca172914-74e0-b789-2f09-28fa74291107.dc1 address=127.0.0.1:29476 error="1 error occurred: * Failed to resolve Node-ca172914-74e0-b789-2f09-28fa74291107.dc1/127.0.0.1:29476: lookup Node-ca172914-74e0-b789-2f09-28fa74291107.dc1/127.0.0.1: no such host " writer.go:29: 2021-01-29T19:34:30.722Z [INFO] TestAgent_Join_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:30.722Z [INFO] TestAgent_Join_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:30.724Z [INFO] TestAgent_Join_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:30.725Z [INFO] TestAgent_Join_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29460 network=tcp writer.go:29: 2021-01-29T19:34:30.727Z [INFO] TestAgent_Join_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29460 network=udp writer.go:29: 2021-01-29T19:34:30.728Z [INFO] TestAgent_Join_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:29461 network=tcp === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:30.748Z [DEBUG] TestAgent_ForceLeavePrune-a1.server.memberlist.lan: memberlist: Failed ping: Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 (timeout reached) === CONT TestAgent_JoinLANNotify writer.go:29: 2021-01-29T19:34:31.006Z [INFO] TestAgent_JoinLANNotify: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:31.008Z [INFO] TestAgent_JoinLANNotify: Endpoints down --- PASS: TestAgent_JoinLANNotify (1.43s) === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:31.034Z [WARN] TestAgent_Join_WAN: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:31.041Z [DEBUG] TestAgent_Join_WAN.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:31.044Z [DEBUG] TestAgent_Join_WAN.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:31.085Z [INFO] TestAgent_Join_WAN.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a Address:127.0.0.1:29459}]" writer.go:29: 2021-01-29T19:34:31.087Z [INFO] TestAgent_Join_WAN.server.raft: entering follower state: follower="Node at 127.0.0.1:29459 [Follower]" leader= writer.go:29: 2021-01-29T19:34:31.088Z [INFO] TestAgent_Join_WAN.server.serf.wan: serf: EventMemberJoin: Node-aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:31.102Z [INFO] TestAgent_Join_WAN.server.serf.lan: serf: EventMemberJoin: Node-aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a 127.0.0.1 writer.go:29: 2021-01-29T19:34:31.105Z [INFO] TestAgent_Join_WAN: Started DNS server: address=127.0.0.1:29454 network=udp writer.go:29: 2021-01-29T19:34:31.106Z [INFO] TestAgent_Join_WAN.server: Adding LAN server: server="Node-aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a (Addr: tcp/127.0.0.1:29459) (DC: dc1)" writer.go:29: 2021-01-29T19:34:31.107Z [INFO] TestAgent_Join_WAN.server: Handled event for server in area: event=member-join server=Node-aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a.dc1 area=wan writer.go:29: 2021-01-29T19:34:31.108Z [INFO] TestAgent_Join_WAN: Started DNS server: address=127.0.0.1:29454 network=tcp writer.go:29: 2021-01-29T19:34:31.110Z [INFO] TestAgent_Join_WAN: Started HTTP server: address=127.0.0.1:29455 network=tcp writer.go:29: 2021-01-29T19:34:31.111Z [INFO] TestAgent_Join_WAN: started state syncer writer.go:29: 2021-01-29T19:34:31.127Z [WARN] TestAgent_Join_WAN.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:31.132Z [INFO] TestAgent_Join_WAN.server.raft: entering candidate state: node="Node at 127.0.0.1:29459 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:31.160Z [DEBUG] TestAgent_Join_WAN.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:31.162Z [DEBUG] TestAgent_Join_WAN.server.raft: vote granted: from=aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a term=2 tally=1 writer.go:29: 2021-01-29T19:34:31.164Z [INFO] TestAgent_Join_WAN.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:31.165Z [INFO] TestAgent_Join_WAN.server.raft: entering leader state: leader="Node at 127.0.0.1:29459 [Leader]" writer.go:29: 2021-01-29T19:34:31.166Z [INFO] TestAgent_Join_WAN.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:31.169Z [INFO] TestAgent_Join_WAN.server: New leader elected: payload=Node-aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a writer.go:29: 2021-01-29T19:34:31.171Z [DEBUG] TestAgent_Join_WAN.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29459 writer.go:29: 2021-01-29T19:34:31.177Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:31.194Z [INFO] TestAgent_Join_WAN.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:31.196Z [INFO] TestAgent_Join_WAN.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:31.198Z [DEBUG] TestAgent_Join_WAN.server: Skipping self join check for node since the cluster is too small: node=Node-aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a writer.go:29: 2021-01-29T19:34:31.211Z [INFO] TestAgent_Join_WAN.server: member joined, marking health alive: member=Node-aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a === CONT TestAgent_Join_ACLDeny writer.go:29: 2021-01-29T19:34:31.230Z [INFO] TestAgent_Join_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:31.232Z [INFO] TestAgent_Join_ACLDeny: Endpoints down writer.go:29: 2021-01-29T19:34:31.234Z [INFO] TestAgent_Join_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:31.235Z [INFO] TestAgent_Join_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:31.236Z [DEBUG] TestAgent_Join_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:31.245Z [DEBUG] TestAgent_Join_ACLDeny.leader: stopping routine: routine="acl token reaping" === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:31.245Z [DEBUG] TestAgent_Join_WAN: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:31.247Z [INFO] TestAgent_ForceLeavePrune-a1.server.memberlist.lan: memberlist: Suspect Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 has failed, no acks received === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:31.249Z [INFO] TestAgent_Join_WAN: Synced node info === CONT TestAgent_Join_ACLDeny writer.go:29: 2021-01-29T19:34:31.247Z [DEBUG] TestAgent_Join_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:31.245Z [DEBUG] TestAgent_Join_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:31.247Z [DEBUG] TestAgent_Join_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:31.251Z [WARN] TestAgent_Join_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:31.251Z [DEBUG] TestAgent_Join_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:31.262Z [WARN] TestAgent_Join_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:31.270Z [INFO] TestAgent_Join_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:31.272Z [INFO] TestAgent_Join_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:31.273Z [INFO] TestAgent_Join_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:31.275Z [INFO] TestAgent_Join_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29472 network=tcp writer.go:29: 2021-01-29T19:34:31.276Z [INFO] TestAgent_Join_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29472 network=udp writer.go:29: 2021-01-29T19:34:31.277Z [INFO] TestAgent_Join_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:29473 network=tcp === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:31.316Z [DEBUG] TestAgent_ForceLeavePrune-a1.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:31.318Z [WARN] TestAgent_ForceLeavePrune-a1: error getting server health from server: server=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 error="rpc error getting client: failed to get conn: dial tcp 127.0.0.1:0->127.0.0.1:29471: connect: connection refused" writer.go:29: 2021-01-29T19:34:31.318Z [DEBUG] TestAgent_ForceLeavePrune-a1.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:31.477Z [DEBUG] TestAgent_Leave_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:31.511Z [WARN] TestAgent_Join_WAN: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:31.513Z [DEBUG] TestAgent_Join_WAN.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:31.516Z [DEBUG] TestAgent_Join_WAN.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:31.524Z [INFO] TestAgent_Join_WAN.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5bbcdbae-605d-bf16-de2b-b2ed48bce341 Address:127.0.0.1:29489}]" writer.go:29: 2021-01-29T19:34:31.526Z [INFO] TestAgent_Join_WAN.server.raft: entering follower state: follower="Node at 127.0.0.1:29489 [Follower]" leader= writer.go:29: 2021-01-29T19:34:31.527Z [INFO] TestAgent_Join_WAN.server.serf.wan: serf: EventMemberJoin: Node-5bbcdbae-605d-bf16-de2b-b2ed48bce341.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:31.530Z [INFO] TestAgent_Join_WAN.server.serf.lan: serf: EventMemberJoin: Node-5bbcdbae-605d-bf16-de2b-b2ed48bce341 127.0.0.1 writer.go:29: 2021-01-29T19:34:31.531Z [INFO] TestAgent_Join_WAN.server: Handled event for server in area: event=member-join server=Node-5bbcdbae-605d-bf16-de2b-b2ed48bce341.dc1 area=wan writer.go:29: 2021-01-29T19:34:31.532Z [INFO] TestAgent_Join_WAN.server: Adding LAN server: server="Node-5bbcdbae-605d-bf16-de2b-b2ed48bce341 (Addr: tcp/127.0.0.1:29489) (DC: dc1)" writer.go:29: 2021-01-29T19:34:31.532Z [INFO] TestAgent_Join_WAN: Started DNS server: address=127.0.0.1:29484 network=udp writer.go:29: 2021-01-29T19:34:31.535Z [INFO] TestAgent_Join_WAN: Started DNS server: address=127.0.0.1:29484 network=tcp writer.go:29: 2021-01-29T19:34:31.537Z [INFO] TestAgent_Join_WAN: Started HTTP server: address=127.0.0.1:29485 network=tcp writer.go:29: 2021-01-29T19:34:31.538Z [INFO] TestAgent_Join_WAN: started state syncer writer.go:29: 2021-01-29T19:34:31.571Z [WARN] TestAgent_Join_WAN.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:31.574Z [INFO] TestAgent_Join_WAN.server.raft: entering candidate state: node="Node at 127.0.0.1:29489 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:31.579Z [DEBUG] TestAgent_Join_WAN.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:31.582Z [DEBUG] TestAgent_Join_WAN.server.raft: vote granted: from=5bbcdbae-605d-bf16-de2b-b2ed48bce341 term=2 tally=1 writer.go:29: 2021-01-29T19:34:31.586Z [INFO] TestAgent_Join_WAN.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:31.589Z [INFO] TestAgent_Join_WAN.server.raft: entering leader state: leader="Node at 127.0.0.1:29489 [Leader]" writer.go:29: 2021-01-29T19:34:31.592Z [INFO] TestAgent_Join_WAN.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:31.593Z [INFO] TestAgent_Join_WAN.server: New leader elected: payload=Node-5bbcdbae-605d-bf16-de2b-b2ed48bce341 writer.go:29: 2021-01-29T19:34:31.595Z [DEBUG] TestAgent_Join_WAN.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29489 writer.go:29: 2021-01-29T19:34:31.609Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:31.622Z [INFO] TestAgent_Join_WAN.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:31.624Z [INFO] TestAgent_Join_WAN.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:31.625Z [DEBUG] TestAgent_Join_WAN.server: Skipping self join check for node since the cluster is too small: node=Node-5bbcdbae-605d-bf16-de2b-b2ed48bce341 writer.go:29: 2021-01-29T19:34:31.626Z [INFO] TestAgent_Join_WAN.server: member joined, marking health alive: member=Node-5bbcdbae-605d-bf16-de2b-b2ed48bce341 === CONT TestAgent_Join_ACLDeny writer.go:29: 2021-01-29T19:34:31.662Z [DEBUG] TestAgent_Join_ACLDeny.server.memberlist.lan: memberlist: Failed ping: Node-4a852d68-0c07-c29e-db52-5763c6794667 (timeout reached) === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:31.729Z [DEBUG] TestAgent_Join_WAN: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:31.742Z [INFO] TestAgent_Join_WAN: Synced node info writer.go:29: 2021-01-29T19:34:31.744Z [DEBUG] TestAgent_Join_WAN: Node info in sync === CONT TestAgent_Join_ACLDeny writer.go:29: 2021-01-29T19:34:31.783Z [INFO] TestAgent_Join_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:31.796Z [INFO] TestAgent_Join_ACLDeny: Endpoints down --- PASS: TestAgent_Join_ACLDeny (1.72s) --- PASS: TestAgent_Join_ACLDeny/no_token (0.00s) --- PASS: TestAgent_Join_ACLDeny/agent_master_token (0.02s) --- PASS: TestAgent_Join_ACLDeny/read-only_token (0.00s) === CONT TestAgent_Join === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:31.809Z [INFO] TestAgent_Join_WAN: (WAN) joining: wan_addresses=[127.0.0.1:29488] writer.go:29: 2021-01-29T19:34:31.813Z [DEBUG] TestAgent_Join_WAN.server.memberlist.wan: memberlist: Initiating push/pull sync with: 127.0.0.1:29488 writer.go:29: 2021-01-29T19:34:31.818Z [DEBUG] TestAgent_Join_WAN.server.memberlist.wan: memberlist: Stream connection from=127.0.0.1:41638 writer.go:29: 2021-01-29T19:34:31.825Z [INFO] TestAgent_Join_WAN.server.serf.wan: serf: EventMemberJoin: Node-aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:31.830Z [INFO] TestAgent_Join_WAN.server: Handled event for server in area: event=member-join server=Node-aafbfa92-6fe2-5a74-4b4f-db948dbfdd8a.dc1 area=wan writer.go:29: 2021-01-29T19:34:31.826Z [INFO] TestAgent_Join_WAN.server.serf.wan: serf: EventMemberJoin: Node-5bbcdbae-605d-bf16-de2b-b2ed48bce341.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:31.832Z [INFO] TestAgent_Join_WAN: (WAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:34:31.833Z [INFO] TestAgent_Join_WAN: Requesting shutdown writer.go:29: 2021-01-29T19:34:31.834Z [INFO] TestAgent_Join_WAN.server: shutting down server writer.go:29: 2021-01-29T19:34:31.834Z [DEBUG] TestAgent_Join_WAN.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:31.835Z [WARN] TestAgent_Join_WAN.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:31.836Z [INFO] TestAgent_Join_WAN.server: Handled event for server in area: event=member-join server=Node-5bbcdbae-605d-bf16-de2b-b2ed48bce341.dc1 area=wan writer.go:29: 2021-01-29T19:34:31.837Z [DEBUG] TestAgent_Join_WAN.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:31.840Z [WARN] TestAgent_Join: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:31.842Z [DEBUG] TestAgent_Join.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:31.847Z [DEBUG] TestAgent_Join.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:31.840Z [WARN] TestAgent_Join_WAN.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:31.862Z [INFO] TestAgent_Join_WAN.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:31.864Z [INFO] TestAgent_Join_WAN: consul server down writer.go:29: 2021-01-29T19:34:31.865Z [INFO] TestAgent_Join_WAN: shutdown complete === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:31.866Z [INFO] TestAgent_Join.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:41bf4686-d226-02a5-42df-f786a4ff07f0 Address:127.0.0.1:29483}]" writer.go:29: 2021-01-29T19:34:31.868Z [INFO] TestAgent_Join.server.serf.wan: serf: EventMemberJoin: Node-41bf4686-d226-02a5-42df-f786a4ff07f0.dc1 127.0.0.1 === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:31.866Z [INFO] TestAgent_Join_WAN: Stopping server: protocol=DNS address=127.0.0.1:29484 network=tcp writer.go:29: 2021-01-29T19:34:31.875Z [INFO] TestAgent_Join_WAN: Stopping server: protocol=DNS address=127.0.0.1:29484 network=udp === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:31.871Z [INFO] TestAgent_Join.server.raft: entering follower state: follower="Node at 127.0.0.1:29483 [Follower]" leader= writer.go:29: 2021-01-29T19:34:31.871Z [INFO] TestAgent_Join.server.serf.lan: serf: EventMemberJoin: Node-41bf4686-d226-02a5-42df-f786a4ff07f0 127.0.0.1 === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:31.877Z [INFO] TestAgent_Join_WAN: Stopping server: protocol=HTTP address=127.0.0.1:29485 network=tcp === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:31.896Z [INFO] TestAgent_Join.server: Adding LAN server: server="Node-41bf4686-d226-02a5-42df-f786a4ff07f0 (Addr: tcp/127.0.0.1:29483) (DC: dc1)" writer.go:29: 2021-01-29T19:34:31.896Z [INFO] TestAgent_Join.server: Handled event for server in area: event=member-join server=Node-41bf4686-d226-02a5-42df-f786a4ff07f0.dc1 area=wan writer.go:29: 2021-01-29T19:34:31.896Z [INFO] TestAgent_Join: Started DNS server: address=127.0.0.1:29478 network=udp writer.go:29: 2021-01-29T19:34:31.904Z [INFO] TestAgent_Join: Started DNS server: address=127.0.0.1:29478 network=tcp writer.go:29: 2021-01-29T19:34:31.922Z [INFO] TestAgent_Join: Started HTTP server: address=127.0.0.1:29479 network=tcp writer.go:29: 2021-01-29T19:34:31.926Z [INFO] TestAgent_Join: started state syncer writer.go:29: 2021-01-29T19:34:31.947Z [WARN] TestAgent_Join.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:31.955Z [INFO] TestAgent_Join.server.raft: entering candidate state: node="Node at 127.0.0.1:29483 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:31.976Z [DEBUG] TestAgent_Join.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:31.988Z [DEBUG] TestAgent_Join.server.raft: vote granted: from=41bf4686-d226-02a5-42df-f786a4ff07f0 term=2 tally=1 writer.go:29: 2021-01-29T19:34:31.991Z [INFO] TestAgent_Join.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:31.995Z [INFO] TestAgent_Join.server.raft: entering leader state: leader="Node at 127.0.0.1:29483 [Leader]" writer.go:29: 2021-01-29T19:34:31.997Z [INFO] TestAgent_Join.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:31.998Z [INFO] TestAgent_Join.server: New leader elected: payload=Node-41bf4686-d226-02a5-42df-f786a4ff07f0 writer.go:29: 2021-01-29T19:34:32.002Z [DEBUG] TestAgent_Join.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29483 writer.go:29: 2021-01-29T19:34:32.007Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:32.022Z [DEBUG] TestAgent_ForceLeavePrune-a1: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:32.024Z [DEBUG] TestAgent_ForceLeavePrune-a1: Node info in sync === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:32.047Z [INFO] TestAgent_Join.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:32.048Z [INFO] TestAgent_Join.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.049Z [DEBUG] TestAgent_Join.server: Skipping self join check for node since the cluster is too small: node=Node-41bf4686-d226-02a5-42df-f786a4ff07f0 writer.go:29: 2021-01-29T19:34:32.050Z [INFO] TestAgent_Join.server: member joined, marking health alive: member=Node-41bf4686-d226-02a5-42df-f786a4ff07f0 === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:32.316Z [WARN] TestAgent_ForceLeavePrune-a1: error getting server health from server: server=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 error="context deadline exceeded" === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:32.344Z [DEBUG] TestAgent_Join: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:32.347Z [INFO] TestAgent_Join: Synced node info writer.go:29: 2021-01-29T19:34:32.348Z [DEBUG] TestAgent_Join: Node info in sync === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:32.393Z [INFO] TestAgent_Join_WAN: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:32.397Z [INFO] TestAgent_Join_WAN: Endpoints down writer.go:29: 2021-01-29T19:34:32.403Z [INFO] TestAgent_Join_WAN: Requesting shutdown writer.go:29: 2021-01-29T19:34:32.408Z [INFO] TestAgent_Join_WAN.server: shutting down server writer.go:29: 2021-01-29T19:34:32.415Z [DEBUG] TestAgent_Join_WAN.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:32.417Z [WARN] TestAgent_Join: bootstrap = true: do not enable unless necessary === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:32.420Z [WARN] TestAgent_Join_WAN.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:32.421Z [DEBUG] TestAgent_Join.tlsutil: Update: version=1 === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:32.420Z [DEBUG] TestAgent_Join_WAN.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:32.426Z [DEBUG] TestAgent_Join.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:32.433Z [WARN] TestAgent_Join_WAN.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.440Z [INFO] TestAgent_Join_WAN.server.router.manager: shutting down === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:32.439Z [INFO] TestAgent_Join.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 Address:127.0.0.1:29495}]" === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:32.445Z [INFO] TestAgent_Join_WAN: consul server down writer.go:29: 2021-01-29T19:34:32.452Z [INFO] TestAgent_Join_WAN: shutdown complete === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:32.450Z [INFO] TestAgent_Join.server.raft: entering follower state: follower="Node at 127.0.0.1:29495 [Follower]" leader= === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:32.457Z [INFO] TestAgent_Join_WAN: Stopping server: protocol=DNS address=127.0.0.1:29454 network=tcp writer.go:29: 2021-01-29T19:34:32.463Z [INFO] TestAgent_Join_WAN: Stopping server: protocol=DNS address=127.0.0.1:29454 network=udp writer.go:29: 2021-01-29T19:34:32.465Z [INFO] TestAgent_Join_WAN: Stopping server: protocol=HTTP address=127.0.0.1:29455 network=tcp === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:32.450Z [INFO] TestAgent_Join.server.serf.wan: serf: EventMemberJoin: Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:32.477Z [INFO] TestAgent_Join.server.serf.lan: serf: EventMemberJoin: Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 127.0.0.1 writer.go:29: 2021-01-29T19:34:32.485Z [INFO] TestAgent_Join.server: Adding LAN server: server="Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 (Addr: tcp/127.0.0.1:29495) (DC: dc1)" writer.go:29: 2021-01-29T19:34:32.489Z [INFO] TestAgent_Join.server: Handled event for server in area: event=member-join server=Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9.dc1 area=wan writer.go:29: 2021-01-29T19:34:32.493Z [INFO] TestAgent_Join: Started DNS server: address=127.0.0.1:29490 network=udp writer.go:29: 2021-01-29T19:34:32.500Z [INFO] TestAgent_Join: Started DNS server: address=127.0.0.1:29490 network=tcp writer.go:29: 2021-01-29T19:34:32.504Z [INFO] TestAgent_Join: Started HTTP server: address=127.0.0.1:29491 network=tcp writer.go:29: 2021-01-29T19:34:32.507Z [INFO] TestAgent_Join: started state syncer writer.go:29: 2021-01-29T19:34:32.530Z [WARN] TestAgent_Join.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:32.532Z [INFO] TestAgent_Join.server.raft: entering candidate state: node="Node at 127.0.0.1:29495 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:32.534Z [DEBUG] TestAgent_Join.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:32.537Z [DEBUG] TestAgent_Join.server.raft: vote granted: from=45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 term=2 tally=1 writer.go:29: 2021-01-29T19:34:32.543Z [INFO] TestAgent_Join.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:32.553Z [INFO] TestAgent_Join.server.raft: entering leader state: leader="Node at 127.0.0.1:29495 [Leader]" writer.go:29: 2021-01-29T19:34:32.557Z [INFO] TestAgent_Join.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:32.558Z [INFO] TestAgent_Join.server: New leader elected: payload=Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 writer.go:29: 2021-01-29T19:34:32.566Z [DEBUG] TestAgent_Join.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29495 writer.go:29: 2021-01-29T19:34:32.574Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:32.585Z [INFO] TestAgent_Join.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:32.587Z [INFO] TestAgent_Join.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.589Z [DEBUG] TestAgent_Join.server: Skipping self join check for node since the cluster is too small: node=Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 writer.go:29: 2021-01-29T19:34:32.591Z [INFO] TestAgent_Join.server: member joined, marking health alive: member=Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:32.707Z [INFO] TestAgent_Leave_ACLDeny.server.serf.lan: serf: EventMemberLeave: Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a 127.0.0.1 writer.go:29: 2021-01-29T19:34:32.711Z [INFO] TestAgent_Leave_ACLDeny.server: Removing LAN server: server="Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a (Addr: tcp/127.0.0.1:29435) (DC: dc1)" writer.go:29: 2021-01-29T19:34:32.713Z [WARN] TestAgent_Leave_ACLDeny.server: deregistering self should be done by follower: name=Node-7a21a0d0-aa4d-63bc-2166-734ab8477f8a === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:32.732Z [DEBUG] TestAgent_Join: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:32.735Z [INFO] TestAgent_Join: Synced node info writer.go:29: 2021-01-29T19:34:32.737Z [DEBUG] TestAgent_Join: Node info in sync === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:32.747Z [DEBUG] TestAgent_ForceLeavePrune-a1.server.memberlist.lan: memberlist: Failed ping: Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 (timeout reached) === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:32.785Z [INFO] TestAgent_Join: (LAN) joining: lan_addresses=[127.0.0.1:29493] writer.go:29: 2021-01-29T19:34:32.787Z [DEBUG] TestAgent_Join.server.memberlist.lan: memberlist: Initiating push/pull sync with: 127.0.0.1:29493 writer.go:29: 2021-01-29T19:34:32.788Z [DEBUG] TestAgent_Join.server.memberlist.lan: memberlist: Stream connection from=127.0.0.1:50194 writer.go:29: 2021-01-29T19:34:32.797Z [INFO] TestAgent_Join.server.serf.lan: serf: EventMemberJoin: Node-41bf4686-d226-02a5-42df-f786a4ff07f0 127.0.0.1 writer.go:29: 2021-01-29T19:34:32.798Z [INFO] TestAgent_Join.server: Adding LAN server: server="Node-41bf4686-d226-02a5-42df-f786a4ff07f0 (Addr: tcp/127.0.0.1:29483) (DC: dc1)" writer.go:29: 2021-01-29T19:34:32.800Z [INFO] TestAgent_Join.server: New leader elected: payload=Node-41bf4686-d226-02a5-42df-f786a4ff07f0 writer.go:29: 2021-01-29T19:34:32.801Z [ERROR] TestAgent_Join.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-41bf4686-d226-02a5-42df-f786a4ff07f0 other=Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 writer.go:29: 2021-01-29T19:34:32.803Z [INFO] TestAgent_Join.server: member joined, marking health alive: member=Node-41bf4686-d226-02a5-42df-f786a4ff07f0 writer.go:29: 2021-01-29T19:34:32.802Z [INFO] TestAgent_Join.server.serf.lan: serf: EventMemberJoin: Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 127.0.0.1 writer.go:29: 2021-01-29T19:34:32.806Z [INFO] TestAgent_Join: (LAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:34:32.807Z [DEBUG] TestAgent_Join: systemd notify failed: error="No socket" writer.go:29: 2021-01-29T19:34:32.806Z [WARN] TestAgent_Join.server.memberlist.wan: memberlist: Failed to resolve Node-41bf4686-d226-02a5-42df-f786a4ff07f0.dc1/127.0.0.1:29482: lookup Node-41bf4686-d226-02a5-42df-f786a4ff07f0.dc1/127.0.0.1: no such host writer.go:29: 2021-01-29T19:34:32.806Z [INFO] TestAgent_Join.server: Adding LAN server: server="Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 (Addr: tcp/127.0.0.1:29495) (DC: dc1)" writer.go:29: 2021-01-29T19:34:32.810Z [DEBUG] TestAgent_Join.server: Failed to flood-join server at address: server=Node-41bf4686-d226-02a5-42df-f786a4ff07f0.dc1 address=127.0.0.1:29482 error="1 error occurred: * Failed to resolve Node-41bf4686-d226-02a5-42df-f786a4ff07f0.dc1/127.0.0.1:29482: lookup Node-41bf4686-d226-02a5-42df-f786a4ff07f0.dc1/127.0.0.1: no such host " writer.go:29: 2021-01-29T19:34:32.809Z [INFO] TestAgent_Join: Requesting shutdown writer.go:29: 2021-01-29T19:34:32.814Z [ERROR] TestAgent_Join.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 other=Node-41bf4686-d226-02a5-42df-f786a4ff07f0 writer.go:29: 2021-01-29T19:34:32.821Z [INFO] TestAgent_Join.server: member joined, marking health alive: member=Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 writer.go:29: 2021-01-29T19:34:32.815Z [INFO] TestAgent_Join.server: shutting down server writer.go:29: 2021-01-29T19:34:32.816Z [WARN] TestAgent_Join.server.memberlist.wan: memberlist: Failed to resolve Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9.dc1/127.0.0.1:29494: lookup Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9.dc1/127.0.0.1: no such host writer.go:29: 2021-01-29T19:34:32.831Z [DEBUG] TestAgent_Join.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.836Z [DEBUG] TestAgent_Join.server: Failed to flood-join server at address: server=Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9.dc1 address=127.0.0.1:29494 error="1 error occurred: * Failed to resolve Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9.dc1/127.0.0.1:29494: lookup Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9.dc1/127.0.0.1: no such host " writer.go:29: 2021-01-29T19:34:32.840Z [WARN] TestAgent_Join.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.841Z [DEBUG] TestAgent_Join.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.845Z [WARN] TestAgent_Join.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.848Z [INFO] TestAgent_Join.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:32.848Z [INFO] TestAgent_Join: consul server down writer.go:29: 2021-01-29T19:34:32.850Z [INFO] TestAgent_Join: shutdown complete writer.go:29: 2021-01-29T19:34:32.851Z [INFO] TestAgent_Join: Stopping server: protocol=DNS address=127.0.0.1:29490 network=tcp writer.go:29: 2021-01-29T19:34:32.852Z [INFO] TestAgent_Join: Stopping server: protocol=DNS address=127.0.0.1:29490 network=udp writer.go:29: 2021-01-29T19:34:32.853Z [INFO] TestAgent_Join: Stopping server: protocol=HTTP address=127.0.0.1:29491 network=tcp === CONT TestAgent_Join_WAN writer.go:29: 2021-01-29T19:34:32.968Z [INFO] TestAgent_Join_WAN: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:32.969Z [INFO] TestAgent_Join_WAN: Endpoints down --- PASS: TestAgent_Join_WAN (1.96s) === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:32.988Z [WARN] TestAgent_Members_ACLFilter: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:32.989Z [WARN] TestAgent_Members_ACLFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:32.991Z [DEBUG] TestAgent_Members_ACLFilter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:32.993Z [DEBUG] TestAgent_Members_ACLFilter.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:32.999Z [INFO] TestAgent_Members_ACLFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9979b11f-4a73-4171-7202-7febe0fc3041 Address:127.0.0.1:29501}]" writer.go:29: 2021-01-29T19:34:33.001Z [INFO] TestAgent_Members_ACLFilter.server.serf.wan: serf: EventMemberJoin: Node-9979b11f-4a73-4171-7202-7febe0fc3041.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:33.001Z [INFO] TestAgent_Members_ACLFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:29501 [Follower]" leader= writer.go:29: 2021-01-29T19:34:33.003Z [INFO] TestAgent_Members_ACLFilter.server.serf.lan: serf: EventMemberJoin: Node-9979b11f-4a73-4171-7202-7febe0fc3041 127.0.0.1 writer.go:29: 2021-01-29T19:34:33.009Z [INFO] TestAgent_Members_ACLFilter.server: Adding LAN server: server="Node-9979b11f-4a73-4171-7202-7febe0fc3041 (Addr: tcp/127.0.0.1:29501) (DC: dc1)" writer.go:29: 2021-01-29T19:34:33.009Z [INFO] TestAgent_Members_ACLFilter.server: Handled event for server in area: event=member-join server=Node-9979b11f-4a73-4171-7202-7febe0fc3041.dc1 area=wan writer.go:29: 2021-01-29T19:34:33.012Z [INFO] TestAgent_Members_ACLFilter: Started DNS server: address=127.0.0.1:29496 network=tcp writer.go:29: 2021-01-29T19:34:33.013Z [INFO] TestAgent_Members_ACLFilter: Started DNS server: address=127.0.0.1:29496 network=udp writer.go:29: 2021-01-29T19:34:33.016Z [INFO] TestAgent_Members_ACLFilter: Started HTTP server: address=127.0.0.1:29497 network=tcp writer.go:29: 2021-01-29T19:34:33.017Z [INFO] TestAgent_Members_ACLFilter: started state syncer writer.go:29: 2021-01-29T19:34:33.060Z [DEBUG] TestAgent_Members_ACLFilter.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:33.073Z [WARN] TestAgent_Members_ACLFilter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:33.075Z [INFO] TestAgent_Members_ACLFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:29501 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:33.077Z [DEBUG] TestAgent_Members_ACLFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:33.079Z [DEBUG] TestAgent_Members_ACLFilter.server.raft: vote granted: from=9979b11f-4a73-4171-7202-7febe0fc3041 term=2 tally=1 writer.go:29: 2021-01-29T19:34:33.082Z [INFO] TestAgent_Members_ACLFilter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:33.086Z [INFO] TestAgent_Members_ACLFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:29501 [Leader]" writer.go:29: 2021-01-29T19:34:33.087Z [INFO] TestAgent_Members_ACLFilter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:33.088Z [INFO] TestAgent_Members_ACLFilter.server: New leader elected: payload=Node-9979b11f-4a73-4171-7202-7febe0fc3041 writer.go:29: 2021-01-29T19:34:33.090Z [INFO] TestAgent_Members_ACLFilter.server: initializing acls writer.go:29: 2021-01-29T19:34:33.097Z [INFO] TestAgent_Members_ACLFilter.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:33.099Z [WARN] TestAgent_Members_ACLFilter.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:33.102Z [INFO] TestAgent_Members_ACLFilter.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:33.107Z [INFO] TestAgent_Members_ACLFilter.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:33.109Z [INFO] TestAgent_Members_ACLFilter.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:33.112Z [INFO] TestAgent_Members_ACLFilter.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:33.117Z [INFO] TestAgent_Members_ACLFilter.server.serf.lan: serf: EventMemberUpdate: Node-9979b11f-4a73-4171-7202-7febe0fc3041 writer.go:29: 2021-01-29T19:34:33.121Z [INFO] TestAgent_Members_ACLFilter.server: Updating LAN server: server="Node-9979b11f-4a73-4171-7202-7febe0fc3041 (Addr: tcp/127.0.0.1:29501) (DC: dc1)" writer.go:29: 2021-01-29T19:34:33.121Z [INFO] TestAgent_Members_ACLFilter.server.serf.wan: serf: EventMemberUpdate: Node-9979b11f-4a73-4171-7202-7febe0fc3041.dc1 writer.go:29: 2021-01-29T19:34:33.124Z [INFO] TestAgent_Members_ACLFilter.server: Handled event for server in area: event=member-update server=Node-9979b11f-4a73-4171-7202-7febe0fc3041.dc1 area=wan === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.124Z [DEBUG] TestAgent_Join: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:33.126Z [DEBUG] TestAgent_Join: Node info in sync === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.127Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:33.135Z [INFO] TestAgent_Members_ACLFilter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:33.137Z [INFO] TestAgent_Members_ACLFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:33.140Z [DEBUG] TestAgent_Members_ACLFilter.server: Skipping self join check for node since the cluster is too small: node=Node-9979b11f-4a73-4171-7202-7febe0fc3041 writer.go:29: 2021-01-29T19:34:33.148Z [INFO] TestAgent_Members_ACLFilter.server: member joined, marking health alive: member=Node-9979b11f-4a73-4171-7202-7febe0fc3041 writer.go:29: 2021-01-29T19:34:33.155Z [DEBUG] TestAgent_Members_ACLFilter.server: Skipping self join check for node since the cluster is too small: node=Node-9979b11f-4a73-4171-7202-7febe0fc3041 === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:33.316Z [DEBUG] TestAgent_ForceLeavePrune-a1.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:33.318Z [WARN] TestAgent_ForceLeavePrune-a1: error getting server health from server: server=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 error="rpc error getting client: failed to get conn: dial tcp 127.0.0.1:0->127.0.0.1:29471: connect: connection refused" === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.343Z [DEBUG] TestAgent_Members_ACLFilter.acl: dropping node from result due to ACLs: node=Node-9979b11f-4a73-4171-7202-7febe0fc3041 writer.go:29: 2021-01-29T19:34:33.346Z [DEBUG] TestAgent_Members_ACLFilter.acl: dropping node from result due to ACLs: node=Node-9979b11f-4a73-4171-7202-7febe0fc3041 === RUN TestAgent_Members_ACLFilter/no_token === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.348Z [DEBUG] TestAgent_Members_ACLFilter: dropping node from result due to ACLs: node=Node-9979b11f-4a73-4171-7202-7febe0fc3041 accessorID= === RUN TestAgent_Members_ACLFilter/root_token === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.346Z [DEBUG] TestAgent_Members_ACLFilter: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:33.354Z [INFO] TestAgent_Members_ACLFilter: Requesting shutdown === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.355Z [INFO] TestAgent_Join: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:33.356Z [INFO] TestAgent_Join: Endpoints down writer.go:29: 2021-01-29T19:34:33.357Z [INFO] TestAgent_Join: Requesting shutdown === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.356Z [INFO] TestAgent_Members_ACLFilter.server: shutting down server === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.358Z [INFO] TestAgent_Join.server: shutting down server writer.go:29: 2021-01-29T19:34:33.359Z [DEBUG] TestAgent_Join.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.358Z [DEBUG] TestAgent_Members_ACLFilter.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.359Z [WARN] TestAgent_Join.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.359Z [DEBUG] TestAgent_Members_ACLFilter.leader: stopping routine: routine="acl token reaping" === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.361Z [DEBUG] TestAgent_Join.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.361Z [DEBUG] TestAgent_Members_ACLFilter.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.363Z [WARN] TestAgent_Join.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:33.367Z [INFO] TestAgent_Join.server.router.manager: shutting down === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.355Z [INFO] TestAgent_Members_ACLFilter: Synced node info === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.368Z [INFO] TestAgent_Join: consul server down writer.go:29: 2021-01-29T19:34:33.369Z [INFO] TestAgent_Join: shutdown complete === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.362Z [DEBUG] TestAgent_Members_ACLFilter.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.370Z [INFO] TestAgent_Join: Stopping server: protocol=DNS address=127.0.0.1:29478 network=tcp === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.362Z [DEBUG] TestAgent_Members_ACLFilter.leader: stopped routine: routine="acl token reaping" === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.371Z [INFO] TestAgent_Join: Stopping server: protocol=DNS address=127.0.0.1:29478 network=udp writer.go:29: 2021-01-29T19:34:33.372Z [INFO] TestAgent_Join: Stopping server: protocol=HTTP address=127.0.0.1:29479 network=tcp === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.365Z [WARN] TestAgent_Members_ACLFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:33.365Z [DEBUG] TestAgent_Members_ACLFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:33.370Z [DEBUG] TestAgent_Members_ACLFilter: Node info in sync writer.go:29: 2021-01-29T19:34:33.375Z [WARN] TestAgent_Members_ACLFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:33.381Z [INFO] TestAgent_Members_ACLFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:33.381Z [INFO] TestAgent_Members_ACLFilter: consul server down writer.go:29: 2021-01-29T19:34:33.384Z [INFO] TestAgent_Members_ACLFilter: shutdown complete writer.go:29: 2021-01-29T19:34:33.385Z [INFO] TestAgent_Members_ACLFilter: Stopping server: protocol=DNS address=127.0.0.1:29496 network=tcp writer.go:29: 2021-01-29T19:34:33.387Z [INFO] TestAgent_Members_ACLFilter: Stopping server: protocol=DNS address=127.0.0.1:29496 network=udp writer.go:29: 2021-01-29T19:34:33.388Z [INFO] TestAgent_Members_ACLFilter: Stopping server: protocol=HTTP address=127.0.0.1:29497 network=tcp === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.394Z [DEBUG] TestAgent_Join.server.memberlist.lan: memberlist: Failed ping: Node-45a75bd2-8fd3-edbd-c1b5-c741e67dfcb9 (timeout reached) === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:33.477Z [ERROR] TestAgent_Leave_ACLDeny.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" === CONT TestAgent_Join writer.go:29: 2021-01-29T19:34:33.873Z [INFO] TestAgent_Join: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:33.874Z [INFO] TestAgent_Join: Endpoints down --- PASS: TestAgent_Join (2.08s) === CONT TestAgent_Members_WAN === CONT TestAgent_Members_ACLFilter writer.go:29: 2021-01-29T19:34:33.890Z [INFO] TestAgent_Members_ACLFilter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:33.891Z [INFO] TestAgent_Members_ACLFilter: Endpoints down --- PASS: TestAgent_Members_ACLFilter (0.92s) --- PASS: TestAgent_Members_ACLFilter/no_token (0.00s) --- PASS: TestAgent_Members_ACLFilter/root_token (0.00s) === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:33.899Z [WARN] TestAgent_Members: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:33.900Z [DEBUG] TestAgent_Members.tlsutil: Update: version=1 === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:33.917Z [WARN] TestAgent_Members_WAN: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:33.918Z [DEBUG] TestAgent_Members_WAN.tlsutil: Update: version=1 === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:33.917Z [DEBUG] TestAgent_Members.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:33.922Z [DEBUG] TestAgent_Members_WAN.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:33.926Z [INFO] TestAgent_Members.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:46e07cf5-e4b9-179d-9e65-ca87094efdf6 Address:127.0.0.1:29531}]" === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:33.927Z [INFO] TestAgent_Members_WAN.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:97a0b0cb-b3df-8736-e0b8-9c444a9b925b Address:127.0.0.1:29519}]" === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:33.927Z [INFO] TestAgent_Members.server.raft: entering follower state: follower="Node at 127.0.0.1:29531 [Follower]" leader= === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:33.929Z [INFO] TestAgent_Members_WAN.server.serf.wan: serf: EventMemberJoin: Node-97a0b0cb-b3df-8736-e0b8-9c444a9b925b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:33.929Z [INFO] TestAgent_Members_WAN.server.raft: entering follower state: follower="Node at 127.0.0.1:29519 [Follower]" leader= writer.go:29: 2021-01-29T19:34:33.931Z [INFO] TestAgent_Members_WAN.server.serf.lan: serf: EventMemberJoin: Node-97a0b0cb-b3df-8736-e0b8-9c444a9b925b 127.0.0.1 writer.go:29: 2021-01-29T19:34:33.934Z [INFO] TestAgent_Members_WAN.server: Adding LAN server: server="Node-97a0b0cb-b3df-8736-e0b8-9c444a9b925b (Addr: tcp/127.0.0.1:29519) (DC: dc1)" === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:33.930Z [INFO] TestAgent_Members.server.serf.wan: serf: EventMemberJoin: Node-46e07cf5-e4b9-179d-9e65-ca87094efdf6.dc1 127.0.0.1 === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:33.934Z [INFO] TestAgent_Members_WAN.server: Handled event for server in area: event=member-join server=Node-97a0b0cb-b3df-8736-e0b8-9c444a9b925b.dc1 area=wan writer.go:29: 2021-01-29T19:34:33.936Z [INFO] TestAgent_Members_WAN: Started DNS server: address=127.0.0.1:29514 network=udp writer.go:29: 2021-01-29T19:34:33.938Z [INFO] TestAgent_Members_WAN: Started DNS server: address=127.0.0.1:29514 network=tcp === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:33.937Z [INFO] TestAgent_Members.server.serf.lan: serf: EventMemberJoin: Node-46e07cf5-e4b9-179d-9e65-ca87094efdf6 127.0.0.1 writer.go:29: 2021-01-29T19:34:33.941Z [INFO] TestAgent_Members.server: Adding LAN server: server="Node-46e07cf5-e4b9-179d-9e65-ca87094efdf6 (Addr: tcp/127.0.0.1:29531) (DC: dc1)" === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:33.941Z [INFO] TestAgent_Members_WAN: Started HTTP server: address=127.0.0.1:29515 network=tcp === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:33.942Z [INFO] TestAgent_Members.server: Handled event for server in area: event=member-join server=Node-46e07cf5-e4b9-179d-9e65-ca87094efdf6.dc1 area=wan writer.go:29: 2021-01-29T19:34:33.944Z [INFO] TestAgent_Members: Started DNS server: address=127.0.0.1:29526 network=udp writer.go:29: 2021-01-29T19:34:33.944Z [INFO] TestAgent_Members: Started DNS server: address=127.0.0.1:29526 network=tcp === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:33.943Z [INFO] TestAgent_Members_WAN: started state syncer === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:33.946Z [INFO] TestAgent_Members: Started HTTP server: address=127.0.0.1:29527 network=tcp writer.go:29: 2021-01-29T19:34:33.950Z [INFO] TestAgent_Members: started state syncer writer.go:29: 2021-01-29T19:34:33.991Z [WARN] TestAgent_Members.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:33.993Z [INFO] TestAgent_Members.server.raft: entering candidate state: node="Node at 127.0.0.1:29531 [Candidate]" term=2 === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:33.994Z [WARN] TestAgent_Members_WAN.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:33.996Z [DEBUG] TestAgent_Members.server.raft: votes: needed=1 === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:33.996Z [INFO] TestAgent_Members_WAN.server.raft: entering candidate state: node="Node at 127.0.0.1:29519 [Candidate]" term=2 === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:33.998Z [DEBUG] TestAgent_Members.server.raft: vote granted: from=46e07cf5-e4b9-179d-9e65-ca87094efdf6 term=2 tally=1 writer.go:29: 2021-01-29T19:34:34.000Z [INFO] TestAgent_Members.server.raft: election won: tally=1 === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:33.999Z [DEBUG] TestAgent_Members_WAN.server.raft: votes: needed=1 === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:34.002Z [INFO] TestAgent_Members.server.raft: entering leader state: leader="Node at 127.0.0.1:29531 [Leader]" writer.go:29: 2021-01-29T19:34:34.004Z [INFO] TestAgent_Members.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:34.004Z [INFO] TestAgent_Members.server: New leader elected: payload=Node-46e07cf5-e4b9-179d-9e65-ca87094efdf6 === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:34.002Z [DEBUG] TestAgent_Members_WAN.server.raft: vote granted: from=97a0b0cb-b3df-8736-e0b8-9c444a9b925b term=2 tally=1 === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:34.005Z [DEBUG] TestAgent_Members.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29531 === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:34.005Z [INFO] TestAgent_Members_WAN.server.raft: election won: tally=1 === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:34.009Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:34.010Z [INFO] TestAgent_Members_WAN.server.raft: entering leader state: leader="Node at 127.0.0.1:29519 [Leader]" === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:34.014Z [INFO] TestAgent_Members.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:34.015Z [INFO] TestAgent_Members.leader: started routine: routine="CA root pruning" === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:34.014Z [INFO] TestAgent_Members_WAN.server: cluster leadership acquired === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:34.016Z [DEBUG] TestAgent_Members.server: Skipping self join check for node since the cluster is too small: node=Node-46e07cf5-e4b9-179d-9e65-ca87094efdf6 === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:34.014Z [INFO] TestAgent_Members_WAN.server: New leader elected: payload=Node-97a0b0cb-b3df-8736-e0b8-9c444a9b925b === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:34.017Z [INFO] TestAgent_Members.server: member joined, marking health alive: member=Node-46e07cf5-e4b9-179d-9e65-ca87094efdf6 === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:34.016Z [DEBUG] TestAgent_Members_WAN.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29519 writer.go:29: 2021-01-29T19:34:34.025Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:34.035Z [INFO] TestAgent_Members_WAN.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:34.037Z [INFO] TestAgent_Members_WAN.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:34.038Z [DEBUG] TestAgent_Members_WAN.server: Skipping self join check for node since the cluster is too small: node=Node-97a0b0cb-b3df-8736-e0b8-9c444a9b925b writer.go:29: 2021-01-29T19:34:34.040Z [INFO] TestAgent_Members_WAN.server: member joined, marking health alive: member=Node-97a0b0cb-b3df-8736-e0b8-9c444a9b925b === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:34.247Z [INFO] TestAgent_ForceLeavePrune-a1.server.memberlist.lan: memberlist: Suspect Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 has failed, no acks received === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:34.273Z [INFO] TestAgent_Members_WAN: Requesting shutdown writer.go:29: 2021-01-29T19:34:34.275Z [INFO] TestAgent_Members_WAN.server: shutting down server writer.go:29: 2021-01-29T19:34:34.276Z [DEBUG] TestAgent_Members_WAN.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:34.276Z [WARN] TestAgent_Members_WAN.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:34.278Z [ERROR] TestAgent_Members_WAN.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:34.279Z [DEBUG] TestAgent_Members_WAN.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:34.281Z [WARN] TestAgent_Members_WAN.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:34.282Z [INFO] TestAgent_Members_WAN.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:34.283Z [INFO] TestAgent_Members_WAN: consul server down writer.go:29: 2021-01-29T19:34:34.284Z [INFO] TestAgent_Members_WAN: shutdown complete writer.go:29: 2021-01-29T19:34:34.285Z [INFO] TestAgent_Members_WAN: Stopping server: protocol=DNS address=127.0.0.1:29514 network=tcp writer.go:29: 2021-01-29T19:34:34.286Z [INFO] TestAgent_Members_WAN: Stopping server: protocol=DNS address=127.0.0.1:29514 network=udp writer.go:29: 2021-01-29T19:34:34.287Z [INFO] TestAgent_Members_WAN: Stopping server: protocol=HTTP address=127.0.0.1:29515 network=tcp === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:34.316Z [WARN] TestAgent_ForceLeavePrune-a1: error getting server health from server: server=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 error="context deadline exceeded" === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:34.350Z [DEBUG] TestAgent_Members: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:34.354Z [INFO] TestAgent_Members: Synced node info writer.go:29: 2021-01-29T19:34:34.355Z [DEBUG] TestAgent_Members: Node info in sync writer.go:29: 2021-01-29T19:34:34.372Z [INFO] TestAgent_Members: Requesting shutdown writer.go:29: 2021-01-29T19:34:34.374Z [INFO] TestAgent_Members.server: shutting down server writer.go:29: 2021-01-29T19:34:34.376Z [DEBUG] TestAgent_Members.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:34.377Z [WARN] TestAgent_Members.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:34.377Z [DEBUG] TestAgent_Members.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:34.382Z [WARN] TestAgent_Members.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:34.386Z [INFO] TestAgent_Members.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:34.386Z [INFO] TestAgent_Members: consul server down writer.go:29: 2021-01-29T19:34:34.389Z [INFO] TestAgent_Members: shutdown complete writer.go:29: 2021-01-29T19:34:34.390Z [INFO] TestAgent_Members: Stopping server: protocol=DNS address=127.0.0.1:29526 network=tcp writer.go:29: 2021-01-29T19:34:34.391Z [INFO] TestAgent_Members: Stopping server: protocol=DNS address=127.0.0.1:29526 network=udp writer.go:29: 2021-01-29T19:34:34.392Z [INFO] TestAgent_Members: Stopping server: protocol=HTTP address=127.0.0.1:29527 network=tcp === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:34.750Z [DEBUG] TestAgent_ForceLeavePrune-a1.server.memberlist.lan: memberlist: Failed ping: Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 (timeout reached) === CONT TestAgent_Members_WAN writer.go:29: 2021-01-29T19:34:34.789Z [INFO] TestAgent_Members_WAN: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:34.790Z [INFO] TestAgent_Members_WAN: Endpoints down --- PASS: TestAgent_Members_WAN (0.92s) === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.813Z [WARN] TestAgent_Checks_ACLFilter: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:34.816Z [WARN] TestAgent_Checks_ACLFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:34.819Z [DEBUG] TestAgent_Checks_ACLFilter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:34.824Z [DEBUG] TestAgent_Checks_ACLFilter.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:34.835Z [INFO] TestAgent_Checks_ACLFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 Address:127.0.0.1:29525}]" writer.go:29: 2021-01-29T19:34:34.838Z [INFO] TestAgent_Checks_ACLFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:29525 [Follower]" leader= writer.go:29: 2021-01-29T19:34:34.843Z [INFO] TestAgent_Checks_ACLFilter.server.serf.wan: serf: EventMemberJoin: Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:34.853Z [INFO] TestAgent_Checks_ACLFilter.server.serf.lan: serf: EventMemberJoin: Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 127.0.0.1 writer.go:29: 2021-01-29T19:34:34.857Z [INFO] TestAgent_Checks_ACLFilter.server: Adding LAN server: server="Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 (Addr: tcp/127.0.0.1:29525) (DC: dc1)" writer.go:29: 2021-01-29T19:34:34.858Z [INFO] TestAgent_Checks_ACLFilter.server: Handled event for server in area: event=member-join server=Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4.dc1 area=wan writer.go:29: 2021-01-29T19:34:34.861Z [INFO] TestAgent_Checks_ACLFilter: Started DNS server: address=127.0.0.1:29520 network=tcp writer.go:29: 2021-01-29T19:34:34.864Z [INFO] TestAgent_Checks_ACLFilter: Started DNS server: address=127.0.0.1:29520 network=udp writer.go:29: 2021-01-29T19:34:34.871Z [INFO] TestAgent_Checks_ACLFilter: Started HTTP server: address=127.0.0.1:29521 network=tcp writer.go:29: 2021-01-29T19:34:34.873Z [INFO] TestAgent_Checks_ACLFilter: started state syncer writer.go:29: 2021-01-29T19:34:34.888Z [WARN] TestAgent_Checks_ACLFilter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:34.889Z [INFO] TestAgent_Checks_ACLFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:29525 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:34.890Z [DEBUG] TestAgent_Checks_ACLFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:34.891Z [DEBUG] TestAgent_Checks_ACLFilter.server.raft: vote granted: from=08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 term=2 tally=1 writer.go:29: 2021-01-29T19:34:34.893Z [INFO] TestAgent_Checks_ACLFilter.server.raft: election won: tally=1 === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:34.895Z [INFO] TestAgent_Members: Waiting for endpoints to shut down === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.894Z [INFO] TestAgent_Checks_ACLFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:29525 [Leader]" === CONT TestAgent_Members writer.go:29: 2021-01-29T19:34:34.896Z [INFO] TestAgent_Members: Endpoints down --- PASS: TestAgent_Members (1.00s) === CONT TestAgent_Reload_ACLDeny === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.897Z [INFO] TestAgent_Checks_ACLFilter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:34.899Z [INFO] TestAgent_Checks_ACLFilter.server: New leader elected: payload=Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 writer.go:29: 2021-01-29T19:34:34.906Z [INFO] TestAgent_Checks_ACLFilter.server: initializing acls writer.go:29: 2021-01-29T19:34:34.909Z [INFO] TestAgent_Checks_ACLFilter.server: initializing acls writer.go:29: 2021-01-29T19:34:34.911Z [WARN] TestAgent_Checks_ACLFilter.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:34.910Z [INFO] TestAgent_Checks_ACLFilter.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:34.916Z [WARN] TestAgent_Checks_ACLFilter.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:34.915Z [INFO] TestAgent_Checks_ACLFilter.server: Bootstrapped ACL master token from configuration === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:34.917Z [WARN] TestAgent_Reload_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:34.919Z [WARN] TestAgent_Reload_ACLDeny: bootstrap = true: do not enable unless necessary === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.921Z [INFO] TestAgent_Checks_ACLFilter.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:34.922Z [INFO] TestAgent_Checks_ACLFilter.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:34.923Z [INFO] TestAgent_Checks_ACLFilter.leader: started routine: routine="acl token reaping" === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:34.922Z [DEBUG] TestAgent_Reload_ACLDeny.tlsutil: Update: version=1 === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.921Z [INFO] TestAgent_Checks_ACLFilter.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:34.924Z [DEBUG] TestAgent_Checks_ACLFilter.server: transitioning out of legacy ACL mode === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:34.925Z [DEBUG] TestAgent_Reload_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.924Z [INFO] TestAgent_Checks_ACLFilter.server.serf.lan: serf: EventMemberUpdate: Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 writer.go:29: 2021-01-29T19:34:34.927Z [INFO] TestAgent_Checks_ACLFilter.server.serf.wan: serf: EventMemberUpdate: Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4.dc1 writer.go:29: 2021-01-29T19:34:34.927Z [INFO] TestAgent_Checks_ACLFilter.server: Updating LAN server: server="Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 (Addr: tcp/127.0.0.1:29525) (DC: dc1)" === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:34.930Z [INFO] TestAgent_Reload_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:406da990-3833-5e99-6be1-a25d249cb989 Address:127.0.0.1:29507}]" === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.927Z [INFO] TestAgent_Checks_ACLFilter.server.serf.lan: serf: EventMemberUpdate: Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 writer.go:29: 2021-01-29T19:34:34.928Z [INFO] TestAgent_Checks_ACLFilter.server: Handled event for server in area: event=member-update server=Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4.dc1 area=wan writer.go:29: 2021-01-29T19:34:34.930Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:34.933Z [INFO] TestAgent_Checks_ACLFilter.server.serf.wan: serf: EventMemberUpdate: Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4.dc1 writer.go:29: 2021-01-29T19:34:34.933Z [INFO] TestAgent_Checks_ACLFilter.server: Updating LAN server: server="Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 (Addr: tcp/127.0.0.1:29525) (DC: dc1)" writer.go:29: 2021-01-29T19:34:34.937Z [INFO] TestAgent_Checks_ACLFilter.server: Handled event for server in area: event=member-update server=Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4.dc1 area=wan === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:34.933Z [INFO] TestAgent_Reload_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:29507 [Follower]" leader= === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.940Z [INFO] TestAgent_Checks_ACLFilter.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:34.941Z [INFO] TestAgent_Reload_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-406da990-3833-5e99-6be1-a25d249cb989.dc1 127.0.0.1 === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.942Z [INFO] TestAgent_Checks_ACLFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:34.944Z [DEBUG] TestAgent_Checks_ACLFilter.server: Skipping self join check for node since the cluster is too small: node=Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:34.943Z [INFO] TestAgent_Reload_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-406da990-3833-5e99-6be1-a25d249cb989 127.0.0.1 === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.945Z [INFO] TestAgent_Checks_ACLFilter.server: member joined, marking health alive: member=Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:34.947Z [INFO] TestAgent_Reload_ACLDeny.server: Adding LAN server: server="Node-406da990-3833-5e99-6be1-a25d249cb989 (Addr: tcp/127.0.0.1:29507) (DC: dc1)" writer.go:29: 2021-01-29T19:34:34.947Z [INFO] TestAgent_Reload_ACLDeny.server: Handled event for server in area: event=member-join server=Node-406da990-3833-5e99-6be1-a25d249cb989.dc1 area=wan === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.948Z [DEBUG] TestAgent_Checks_ACLFilter.server: Skipping self join check for node since the cluster is too small: node=Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 writer.go:29: 2021-01-29T19:34:34.949Z [DEBUG] TestAgent_Checks_ACLFilter.server: Skipping self join check for node since the cluster is too small: node=Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:34.951Z [INFO] TestAgent_Reload_ACLDeny: Started DNS server: address=127.0.0.1:29502 network=tcp writer.go:29: 2021-01-29T19:34:34.953Z [INFO] TestAgent_Reload_ACLDeny: Started DNS server: address=127.0.0.1:29502 network=udp writer.go:29: 2021-01-29T19:34:34.955Z [INFO] TestAgent_Reload_ACLDeny: Started HTTP server: address=127.0.0.1:29503 network=tcp writer.go:29: 2021-01-29T19:34:34.957Z [INFO] TestAgent_Reload_ACLDeny: started state syncer === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.962Z [DEBUG] TestAgent_Checks_ACLFilter.acl: dropping node from result due to ACLs: node=Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 writer.go:29: 2021-01-29T19:34:34.967Z [DEBUG] TestAgent_Checks_ACLFilter.acl: dropping node from result due to ACLs: node=Node-08cd2c7e-4c28-fd9c-4dbc-b5cfd96cb5c4 === RUN TestAgent_Checks_ACLFilter/no_token === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.971Z [DEBUG] TestAgent_Checks_ACLFilter: dropping check from result due to ACLs: check=mysql === RUN TestAgent_Checks_ACLFilter/root_token === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:34.979Z [INFO] TestAgent_Checks_ACLFilter: Requesting shutdown writer.go:29: 2021-01-29T19:34:34.981Z [INFO] TestAgent_Checks_ACLFilter.server: shutting down server writer.go:29: 2021-01-29T19:34:34.982Z [DEBUG] TestAgent_Checks_ACLFilter.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:34.983Z [DEBUG] TestAgent_Checks_ACLFilter.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:34.983Z [DEBUG] TestAgent_Checks_ACLFilter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:34.982Z [ERROR] TestAgent_Checks_ACLFilter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:34.983Z [DEBUG] TestAgent_Checks_ACLFilter.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:34.984Z [DEBUG] TestAgent_Checks_ACLFilter.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:34.984Z [WARN] TestAgent_Checks_ACLFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:34.984Z [DEBUG] TestAgent_Checks_ACLFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:34.988Z [WARN] TestAgent_Checks_ACLFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:34.990Z [INFO] TestAgent_Checks_ACLFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:34.990Z [INFO] TestAgent_Checks_ACLFilter: consul server down writer.go:29: 2021-01-29T19:34:34.992Z [INFO] TestAgent_Checks_ACLFilter: shutdown complete writer.go:29: 2021-01-29T19:34:34.993Z [INFO] TestAgent_Checks_ACLFilter: Stopping server: protocol=DNS address=127.0.0.1:29520 network=tcp writer.go:29: 2021-01-29T19:34:34.994Z [INFO] TestAgent_Checks_ACLFilter: Stopping server: protocol=DNS address=127.0.0.1:29520 network=udp writer.go:29: 2021-01-29T19:34:34.994Z [INFO] TestAgent_Checks_ACLFilter: Stopping server: protocol=HTTP address=127.0.0.1:29521 network=tcp === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:34.997Z [DEBUG] TestAgent_Reload_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:35.007Z [WARN] TestAgent_Reload_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:35.009Z [INFO] TestAgent_Reload_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:29507 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:35.011Z [DEBUG] TestAgent_Reload_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:35.013Z [DEBUG] TestAgent_Reload_ACLDeny.server.raft: vote granted: from=406da990-3833-5e99-6be1-a25d249cb989 term=2 tally=1 writer.go:29: 2021-01-29T19:34:35.015Z [INFO] TestAgent_Reload_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:35.017Z [INFO] TestAgent_Reload_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:29507 [Leader]" writer.go:29: 2021-01-29T19:34:35.020Z [INFO] TestAgent_Reload_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:35.021Z [INFO] TestAgent_Reload_ACLDeny.server: New leader elected: payload=Node-406da990-3833-5e99-6be1-a25d249cb989 writer.go:29: 2021-01-29T19:34:35.025Z [INFO] TestAgent_Reload_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:35.038Z [INFO] TestAgent_Reload_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:35.040Z [WARN] TestAgent_Reload_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:35.047Z [INFO] TestAgent_Reload_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:35.052Z [INFO] TestAgent_Reload_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:35.058Z [INFO] TestAgent_Reload_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:35.062Z [INFO] TestAgent_Reload_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:35.066Z [INFO] TestAgent_Reload_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-406da990-3833-5e99-6be1-a25d249cb989 writer.go:29: 2021-01-29T19:34:35.073Z [INFO] TestAgent_Reload_ACLDeny.server: Updating LAN server: server="Node-406da990-3833-5e99-6be1-a25d249cb989 (Addr: tcp/127.0.0.1:29507) (DC: dc1)" writer.go:29: 2021-01-29T19:34:35.075Z [INFO] TestAgent_Reload_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-406da990-3833-5e99-6be1-a25d249cb989.dc1 writer.go:29: 2021-01-29T19:34:35.080Z [INFO] TestAgent_Reload_ACLDeny.server: Handled event for server in area: event=member-update server=Node-406da990-3833-5e99-6be1-a25d249cb989.dc1 area=wan writer.go:29: 2021-01-29T19:34:35.094Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:35.100Z [INFO] TestAgent_Reload_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:35.101Z [INFO] TestAgent_Reload_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:35.102Z [DEBUG] TestAgent_Reload_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-406da990-3833-5e99-6be1-a25d249cb989 writer.go:29: 2021-01-29T19:34:35.103Z [INFO] TestAgent_Reload_ACLDeny.server: member joined, marking health alive: member=Node-406da990-3833-5e99-6be1-a25d249cb989 writer.go:29: 2021-01-29T19:34:35.105Z [DEBUG] TestAgent_Reload_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-406da990-3833-5e99-6be1-a25d249cb989 === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.255Z [INFO] TestAgent_ForceLeavePrune-a1.server.memberlist.lan: memberlist: Marking Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 as failed, suspect timeout reached (0 peer confirmations) writer.go:29: 2021-01-29T19:34:35.256Z [INFO] TestAgent_ForceLeavePrune-a1.server.serf.lan: serf: EventMemberFailed: Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 127.0.0.1 === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:35.259Z [DEBUG] TestAgent_Reload_ACLDeny.acl: dropping node from result due to ACLs: node=Node-406da990-3833-5e99-6be1-a25d249cb989 === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.260Z [INFO] TestAgent_ForceLeavePrune-a1.server: Removing LAN server: server="Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 (Addr: tcp/127.0.0.1:29471) (DC: dc1)" === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:35.262Z [DEBUG] TestAgent_Reload_ACLDeny.acl: dropping node from result due to ACLs: node=Node-406da990-3833-5e99-6be1-a25d249cb989 === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.261Z [INFO] TestAgent_ForceLeavePrune-a1.server: member failed, marking health critical: member=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 === RUN TestAgent_Reload_ACLDeny/no_token === RUN TestAgent_Reload_ACLDeny/read-only_token === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:35.269Z [INFO] TestAgent_Reload_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:35.270Z [INFO] TestAgent_Reload_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:35.271Z [DEBUG] TestAgent_Reload_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:35.272Z [DEBUG] TestAgent_Reload_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:35.273Z [DEBUG] TestAgent_Reload_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:35.271Z [ERROR] TestAgent_Reload_ACLDeny.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:35.272Z [DEBUG] TestAgent_Reload_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:35.273Z [DEBUG] TestAgent_Reload_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:35.273Z [DEBUG] TestAgent_Reload_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:35.273Z [WARN] TestAgent_Reload_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:35.279Z [WARN] TestAgent_Reload_ACLDeny.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.281Z [INFO] TestAgent_ForceLeavePrune-a1: Force leaving node: node=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:35.282Z [INFO] TestAgent_Reload_ACLDeny.server.router.manager: shutting down === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.284Z [INFO] TestAgent_ForceLeavePrune-a1.server.serf.lan: serf: EventMemberLeave (forced): Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 127.0.0.1 === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:35.283Z [INFO] TestAgent_Reload_ACLDeny: consul server down === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.285Z [INFO] TestAgent_ForceLeavePrune-a1.server.serf.lan: serf: EventMemberReap (forced): Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 127.0.0.1 === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:35.287Z [INFO] TestAgent_Reload_ACLDeny: shutdown complete === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.287Z [INFO] TestAgent_ForceLeavePrune-a1: Requesting shutdown writer.go:29: 2021-01-29T19:34:35.289Z [INFO] TestAgent_ForceLeavePrune-a1.server: shutting down server === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:35.288Z [INFO] TestAgent_Reload_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29502 network=tcp === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.288Z [INFO] TestAgent_ForceLeavePrune-a1.server: Removing LAN server: server="Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 (Addr: tcp/127.0.0.1:29471) (DC: dc1)" writer.go:29: 2021-01-29T19:34:35.291Z [INFO] TestAgent_ForceLeavePrune-a1.server: Removing LAN server: server="Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 (Addr: tcp/127.0.0.1:29471) (DC: dc1)" writer.go:29: 2021-01-29T19:34:35.291Z [DEBUG] TestAgent_ForceLeavePrune-a1.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:35.294Z [WARN] TestAgent_ForceLeavePrune-a1.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:35.292Z [INFO] TestAgent_ForceLeavePrune-a1.server: deregistering member: member=Node-63a671bf-78e4-baef-0dd1-26ae5f6ea044 reason=left writer.go:29: 2021-01-29T19:34:35.299Z [DEBUG] TestAgent_ForceLeavePrune-a1.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:35.302Z [WARN] TestAgent_ForceLeavePrune-a1.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:35.305Z [INFO] TestAgent_ForceLeavePrune-a1.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:35.306Z [INFO] TestAgent_ForceLeavePrune-a1: consul server down writer.go:29: 2021-01-29T19:34:35.309Z [INFO] TestAgent_ForceLeavePrune-a1: shutdown complete === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:35.291Z [INFO] TestAgent_Reload_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29502 network=udp === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.310Z [INFO] TestAgent_ForceLeavePrune-a1: Stopping server: protocol=DNS address=127.0.0.1:29436 network=tcp === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:35.314Z [INFO] TestAgent_Reload_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:29503 network=tcp === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.314Z [INFO] TestAgent_ForceLeavePrune-a1: Stopping server: protocol=DNS address=127.0.0.1:29436 network=udp writer.go:29: 2021-01-29T19:34:35.317Z [INFO] TestAgent_ForceLeavePrune-a1: Stopping server: protocol=HTTP address=127.0.0.1:29437 network=tcp === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:35.477Z [ERROR] TestAgent_Leave_ACLDeny.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" === CONT TestAgent_Checks_ACLFilter writer.go:29: 2021-01-29T19:34:35.495Z [INFO] TestAgent_Checks_ACLFilter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:35.497Z [INFO] TestAgent_Checks_ACLFilter: Endpoints down --- PASS: TestAgent_Checks_ACLFilter (0.71s) --- PASS: TestAgent_Checks_ACLFilter/no_token (0.00s) --- PASS: TestAgent_Checks_ACLFilter/root_token (0.00s) === CONT TestAgent_Reload writer.go:29: 2021-01-29T19:34:35.528Z [WARN] TestAgent_Reload: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:35.529Z [DEBUG] TestAgent_Reload.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:35.531Z [DEBUG] TestAgent_Reload.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:35.535Z [INFO] TestAgent_Reload.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6e126506-51cd-9523-b1ab-37b9779ae569 Address:127.0.0.1:29537}]" writer.go:29: 2021-01-29T19:34:35.537Z [INFO] TestAgent_Reload.server.raft: entering follower state: follower="Node at 127.0.0.1:29537 [Follower]" leader= writer.go:29: 2021-01-29T19:34:35.537Z [INFO] TestAgent_Reload.server.serf.wan: serf: EventMemberJoin: Node-6e126506-51cd-9523-b1ab-37b9779ae569.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:35.543Z [INFO] TestAgent_Reload.server.serf.lan: serf: EventMemberJoin: Node-6e126506-51cd-9523-b1ab-37b9779ae569 127.0.0.1 writer.go:29: 2021-01-29T19:34:35.548Z [INFO] TestAgent_Reload.server: Adding LAN server: server="Node-6e126506-51cd-9523-b1ab-37b9779ae569 (Addr: tcp/127.0.0.1:29537) (DC: dc1)" writer.go:29: 2021-01-29T19:34:35.549Z [INFO] TestAgent_Reload.server: Handled event for server in area: event=member-join server=Node-6e126506-51cd-9523-b1ab-37b9779ae569.dc1 area=wan writer.go:29: 2021-01-29T19:34:35.550Z [INFO] TestAgent_Reload: Started DNS server: address=127.0.0.1:29532 network=tcp writer.go:29: 2021-01-29T19:34:35.551Z [INFO] TestAgent_Reload: Started DNS server: address=127.0.0.1:29532 network=udp writer.go:29: 2021-01-29T19:34:35.553Z [INFO] TestAgent_Reload: Started HTTP server: address=127.0.0.1:29533 network=tcp writer.go:29: 2021-01-29T19:34:35.554Z [WARN] TestAgent_Reload: The 'handler' field in watches has been deprecated and replaced with the 'args' field. See https://www.consul.io/docs/agent/watches.html writer.go:29: 2021-01-29T19:34:35.555Z [INFO] TestAgent_Reload: started state syncer writer.go:29: 2021-01-29T19:34:35.578Z [WARN] TestAgent_Reload.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:35.579Z [INFO] TestAgent_Reload.server.raft: entering candidate state: node="Node at 127.0.0.1:29537 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:35.582Z [DEBUG] TestAgent_Reload.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:35.584Z [DEBUG] TestAgent_Reload.server.raft: vote granted: from=6e126506-51cd-9523-b1ab-37b9779ae569 term=2 tally=1 writer.go:29: 2021-01-29T19:34:35.586Z [INFO] TestAgent_Reload.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:35.587Z [INFO] TestAgent_Reload.server.raft: entering leader state: leader="Node at 127.0.0.1:29537 [Leader]" writer.go:29: 2021-01-29T19:34:35.590Z [INFO] TestAgent_Reload.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:35.591Z [INFO] TestAgent_Reload.server: New leader elected: payload=Node-6e126506-51cd-9523-b1ab-37b9779ae569 writer.go:29: 2021-01-29T19:34:35.596Z [DEBUG] TestAgent_Reload.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29537 writer.go:29: 2021-01-29T19:34:35.606Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:35.611Z [INFO] TestAgent_Reload.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:35.613Z [INFO] TestAgent_Reload.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:35.615Z [DEBUG] TestAgent_Reload.server: Skipping self join check for node since the cluster is too small: node=Node-6e126506-51cd-9523-b1ab-37b9779ae569 writer.go:29: 2021-01-29T19:34:35.616Z [INFO] TestAgent_Reload.server: member joined, marking health alive: member=Node-6e126506-51cd-9523-b1ab-37b9779ae569 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:35.711Z [INFO] TestAgent_Leave_ACLDeny.server: Waiting to drain RPC traffic: drain_time=5s === CONT TestAgent_Reload writer.go:29: 2021-01-29T19:34:35.804Z [DEBUG] TestAgent_Reload.http: Request finished: method=GET url=/v1/kv/test?dc=dc1 from=127.0.0.1:32896 latency=243.489304ms writer.go:29: 2021-01-29T19:34:35.807Z [DEBUG] TestAgent_Reload: watch handler output: watch_handler=true output= === CONT TestAgent_Reload_ACLDeny writer.go:29: 2021-01-29T19:34:35.815Z [INFO] TestAgent_Reload_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:35.816Z [INFO] TestAgent_Reload_ACLDeny: Endpoints down --- PASS: TestAgent_Reload_ACLDeny (0.92s) --- PASS: TestAgent_Reload_ACLDeny/no_token (0.00s) --- PASS: TestAgent_Reload_ACLDeny/read-only_token (0.00s) === CONT TestAgent_Metrics_ACLDeny === CONT TestAgent_ForceLeavePrune writer.go:29: 2021-01-29T19:34:35.820Z [INFO] TestAgent_ForceLeavePrune-a1: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:35.821Z [INFO] TestAgent_ForceLeavePrune-a1: Endpoints down --- PASS: TestAgent_ForceLeavePrune (6.61s) === CONT TestAgent_Self_ACLDeny === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.844Z [WARN] TestAgent_Metrics_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:35.846Z [WARN] TestAgent_Metrics_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:35.848Z [DEBUG] TestAgent_Metrics_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:35.850Z [DEBUG] TestAgent_Metrics_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.854Z [WARN] TestAgent_Self_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:35.855Z [WARN] TestAgent_Self_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:35.856Z [DEBUG] TestAgent_Self_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:35.862Z [DEBUG] TestAgent_Self_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.865Z [INFO] TestAgent_Metrics_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:89729710-0105-028a-5182-016ae9050f73 Address:127.0.0.1:29543}]" === CONT TestAgent_Reload writer.go:29: 2021-01-29T19:34:35.871Z [WARN] TestAgent_Reload: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:35.872Z [DEBUG] TestAgent_Reload: removed service: service=redis === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.867Z [INFO] TestAgent_Metrics_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:29543 [Follower]" leader= === CONT TestAgent_Reload writer.go:29: 2021-01-29T19:34:35.873Z [DEBUG] TestAgent_Reload.tlsutil: Update: version=2 writer.go:29: 2021-01-29T19:34:35.874Z [INFO] TestAgent_Reload: Requesting shutdown === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.870Z [INFO] TestAgent_Metrics_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-89729710-0105-028a-5182-016ae9050f73.dc1 127.0.0.1 === CONT TestAgent_Reload writer.go:29: 2021-01-29T19:34:35.882Z [INFO] TestAgent_Reload.server: shutting down server writer.go:29: 2021-01-29T19:34:35.883Z [DEBUG] TestAgent_Reload.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:35.884Z [WARN] TestAgent_Reload.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:35.883Z [ERROR] TestAgent_Reload.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:35.884Z [DEBUG] TestAgent_Reload.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:35.887Z [WARN] TestAgent_Reload.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.888Z [INFO] TestAgent_Self_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5c4a26f6-4839-bf70-3647-ad5a2926d9de Address:127.0.0.1:29513}]" === CONT TestAgent_Reload writer.go:29: 2021-01-29T19:34:35.889Z [INFO] TestAgent_Reload.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:35.891Z [INFO] TestAgent_Reload: consul server down writer.go:29: 2021-01-29T19:34:35.891Z [INFO] TestAgent_Reload: shutdown complete writer.go:29: 2021-01-29T19:34:35.892Z [INFO] TestAgent_Reload: Stopping server: protocol=DNS address=127.0.0.1:29532 network=tcp === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.891Z [INFO] TestAgent_Self_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de.dc1 127.0.0.1 === CONT TestAgent_Reload writer.go:29: 2021-01-29T19:34:35.893Z [INFO] TestAgent_Reload: Stopping server: protocol=DNS address=127.0.0.1:29532 network=udp writer.go:29: 2021-01-29T19:34:35.894Z [INFO] TestAgent_Reload: Stopping server: protocol=HTTP address=127.0.0.1:29533 network=tcp === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.884Z [INFO] TestAgent_Metrics_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-89729710-0105-028a-5182-016ae9050f73 127.0.0.1 writer.go:29: 2021-01-29T19:34:35.899Z [INFO] TestAgent_Metrics_ACLDeny.server: Adding LAN server: server="Node-89729710-0105-028a-5182-016ae9050f73 (Addr: tcp/127.0.0.1:29543) (DC: dc1)" writer.go:29: 2021-01-29T19:34:35.899Z [INFO] TestAgent_Metrics_ACLDeny.server: Handled event for server in area: event=member-join server=Node-89729710-0105-028a-5182-016ae9050f73.dc1 area=wan === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.895Z [INFO] TestAgent_Self_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de 127.0.0.1 writer.go:29: 2021-01-29T19:34:35.895Z [INFO] TestAgent_Self_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:29513 [Follower]" leader= writer.go:29: 2021-01-29T19:34:35.912Z [INFO] TestAgent_Self_ACLDeny.server: Adding LAN server: server="Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de (Addr: tcp/127.0.0.1:29513) (DC: dc1)" writer.go:29: 2021-01-29T19:34:35.912Z [INFO] TestAgent_Self_ACLDeny.server: Handled event for server in area: event=member-join server=Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de.dc1 area=wan === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.903Z [INFO] TestAgent_Metrics_ACLDeny: Started DNS server: address=127.0.0.1:29538 network=tcp === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.920Z [INFO] TestAgent_Self_ACLDeny: Started DNS server: address=127.0.0.1:29508 network=tcp === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.917Z [INFO] TestAgent_Metrics_ACLDeny: Started DNS server: address=127.0.0.1:29538 network=udp === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.921Z [INFO] TestAgent_Self_ACLDeny: Started DNS server: address=127.0.0.1:29508 network=udp writer.go:29: 2021-01-29T19:34:35.923Z [INFO] TestAgent_Self_ACLDeny: Started HTTP server: address=127.0.0.1:29509 network=tcp === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.913Z [WARN] TestAgent_Metrics_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.926Z [INFO] TestAgent_Self_ACLDeny: started state syncer === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.928Z [INFO] TestAgent_Metrics_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:29543 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:35.923Z [INFO] TestAgent_Metrics_ACLDeny: Started HTTP server: address=127.0.0.1:29539 network=tcp writer.go:29: 2021-01-29T19:34:35.932Z [DEBUG] TestAgent_Metrics_ACLDeny.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:35.935Z [DEBUG] TestAgent_Metrics_ACLDeny.server.raft: vote granted: from=89729710-0105-028a-5182-016ae9050f73 term=2 tally=1 writer.go:29: 2021-01-29T19:34:35.936Z [INFO] TestAgent_Metrics_ACLDeny: started state syncer writer.go:29: 2021-01-29T19:34:35.938Z [INFO] TestAgent_Metrics_ACLDeny.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:35.940Z [INFO] TestAgent_Metrics_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:29543 [Leader]" writer.go:29: 2021-01-29T19:34:35.941Z [INFO] TestAgent_Metrics_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:35.942Z [INFO] TestAgent_Metrics_ACLDeny.server: New leader elected: payload=Node-89729710-0105-028a-5182-016ae9050f73 writer.go:29: 2021-01-29T19:34:35.946Z [INFO] TestAgent_Metrics_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:35.949Z [INFO] TestAgent_Metrics_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:35.950Z [WARN] TestAgent_Metrics_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:35.951Z [INFO] TestAgent_Metrics_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:35.952Z [WARN] TestAgent_Metrics_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:35.954Z [INFO] TestAgent_Metrics_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:35.954Z [INFO] TestAgent_Metrics_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:35.956Z [INFO] TestAgent_Metrics_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:35.957Z [INFO] TestAgent_Metrics_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:35.958Z [INFO] TestAgent_Metrics_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:35.959Z [INFO] TestAgent_Metrics_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-89729710-0105-028a-5182-016ae9050f73 writer.go:29: 2021-01-29T19:34:35.960Z [INFO] TestAgent_Metrics_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-89729710-0105-028a-5182-016ae9050f73.dc1 writer.go:29: 2021-01-29T19:34:35.961Z [INFO] TestAgent_Metrics_ACLDeny.server: Updating LAN server: server="Node-89729710-0105-028a-5182-016ae9050f73 (Addr: tcp/127.0.0.1:29543) (DC: dc1)" === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.962Z [WARN] TestAgent_Self_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:35.963Z [INFO] TestAgent_Self_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:29513 [Candidate]" term=2 === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.962Z [INFO] TestAgent_Metrics_ACLDeny.server: Handled event for server in area: event=member-update server=Node-89729710-0105-028a-5182-016ae9050f73.dc1 area=wan === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.963Z [DEBUG] TestAgent_Self_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.962Z [DEBUG] TestAgent_Metrics_ACLDeny.server: transitioning out of legacy ACL mode === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.965Z [DEBUG] TestAgent_Self_ACLDeny.server.raft: votes: needed=1 === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.963Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.966Z [DEBUG] TestAgent_Self_ACLDeny.server.raft: vote granted: from=5c4a26f6-4839-bf70-3647-ad5a2926d9de term=2 tally=1 writer.go:29: 2021-01-29T19:34:35.967Z [INFO] TestAgent_Self_ACLDeny.server.raft: election won: tally=1 === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:35.965Z [INFO] TestAgent_Metrics_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-89729710-0105-028a-5182-016ae9050f73 writer.go:29: 2021-01-29T19:34:35.968Z [INFO] TestAgent_Metrics_ACLDeny.server: Updating LAN server: server="Node-89729710-0105-028a-5182-016ae9050f73 (Addr: tcp/127.0.0.1:29543) (DC: dc1)" writer.go:29: 2021-01-29T19:34:35.968Z [INFO] TestAgent_Metrics_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-89729710-0105-028a-5182-016ae9050f73.dc1 writer.go:29: 2021-01-29T19:34:35.970Z [INFO] TestAgent_Metrics_ACLDeny.server: Handled event for server in area: event=member-update server=Node-89729710-0105-028a-5182-016ae9050f73.dc1 area=wan writer.go:29: 2021-01-29T19:34:35.969Z [INFO] TestAgent_Metrics_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:35.972Z [INFO] TestAgent_Metrics_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:35.973Z [DEBUG] TestAgent_Metrics_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-89729710-0105-028a-5182-016ae9050f73 writer.go:29: 2021-01-29T19:34:35.973Z [INFO] TestAgent_Metrics_ACLDeny.server: member joined, marking health alive: member=Node-89729710-0105-028a-5182-016ae9050f73 writer.go:29: 2021-01-29T19:34:35.975Z [DEBUG] TestAgent_Metrics_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-89729710-0105-028a-5182-016ae9050f73 writer.go:29: 2021-01-29T19:34:35.976Z [DEBUG] TestAgent_Metrics_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-89729710-0105-028a-5182-016ae9050f73 === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:35.968Z [INFO] TestAgent_Self_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:29513 [Leader]" writer.go:29: 2021-01-29T19:34:35.977Z [INFO] TestAgent_Self_ACLDeny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:35.978Z [INFO] TestAgent_Self_ACLDeny.server: New leader elected: payload=Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de writer.go:29: 2021-01-29T19:34:35.979Z [INFO] TestAgent_Self_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:34:35.983Z [INFO] TestAgent_Self_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:35.984Z [WARN] TestAgent_Self_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:35.991Z [INFO] TestAgent_Self_ACLDeny.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:35.993Z [INFO] TestAgent_Self_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:35.993Z [INFO] TestAgent_Self_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:35.994Z [INFO] TestAgent_Self_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:35.996Z [INFO] TestAgent_Self_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de writer.go:29: 2021-01-29T19:34:35.997Z [INFO] TestAgent_Self_ACLDeny.server: Updating LAN server: server="Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de (Addr: tcp/127.0.0.1:29513) (DC: dc1)" writer.go:29: 2021-01-29T19:34:35.998Z [INFO] TestAgent_Self_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de.dc1 writer.go:29: 2021-01-29T19:34:35.999Z [INFO] TestAgent_Self_ACLDeny.server: Handled event for server in area: event=member-update server=Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de.dc1 area=wan writer.go:29: 2021-01-29T19:34:36.025Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:36.030Z [INFO] TestAgent_Self_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:36.032Z [INFO] TestAgent_Self_ACLDeny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:36.034Z [DEBUG] TestAgent_Self_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de writer.go:29: 2021-01-29T19:34:36.036Z [INFO] TestAgent_Self_ACLDeny.server: member joined, marking health alive: member=Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de writer.go:29: 2021-01-29T19:34:36.044Z [DEBUG] TestAgent_Self_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:36.064Z [DEBUG] TestAgent_Metrics_ACLDeny.acl: dropping node from result due to ACLs: node=Node-89729710-0105-028a-5182-016ae9050f73 writer.go:29: 2021-01-29T19:34:36.069Z [DEBUG] TestAgent_Metrics_ACLDeny.acl: dropping node from result due to ACLs: node=Node-89729710-0105-028a-5182-016ae9050f73 === RUN TestAgent_Metrics_ACLDeny/no_token === RUN TestAgent_Metrics_ACLDeny/agent_master_token === RUN TestAgent_Metrics_ACLDeny/read-only_token === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:36.079Z [INFO] TestAgent_Metrics_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:36.081Z [INFO] TestAgent_Metrics_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:36.081Z [DEBUG] TestAgent_Metrics_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:36.082Z [DEBUG] TestAgent_Metrics_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:36.083Z [DEBUG] TestAgent_Metrics_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:36.082Z [ERROR] TestAgent_Metrics_ACLDeny.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:36.082Z [DEBUG] TestAgent_Metrics_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:36.083Z [DEBUG] TestAgent_Metrics_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:36.084Z [DEBUG] TestAgent_Metrics_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:36.084Z [WARN] TestAgent_Metrics_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:36.104Z [WARN] TestAgent_Metrics_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:36.106Z [INFO] TestAgent_Metrics_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:36.106Z [INFO] TestAgent_Metrics_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:36.108Z [INFO] TestAgent_Metrics_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:36.109Z [INFO] TestAgent_Metrics_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29538 network=tcp writer.go:29: 2021-01-29T19:34:36.109Z [INFO] TestAgent_Metrics_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29538 network=udp writer.go:29: 2021-01-29T19:34:36.110Z [INFO] TestAgent_Metrics_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:29539 network=tcp === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:36.122Z [DEBUG] TestAgent_Self_ACLDeny: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:36.125Z [INFO] TestAgent_Self_ACLDeny: Synced node info writer.go:29: 2021-01-29T19:34:36.127Z [DEBUG] TestAgent_Self_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:34:36.364Z [DEBUG] TestAgent_Self_ACLDeny.acl: dropping node from result due to ACLs: node=Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de writer.go:29: 2021-01-29T19:34:36.368Z [DEBUG] TestAgent_Self_ACLDeny.acl: dropping node from result due to ACLs: node=Node-5c4a26f6-4839-bf70-3647-ad5a2926d9de === RUN TestAgent_Self_ACLDeny/no_token === RUN TestAgent_Self_ACLDeny/agent_master_token === RUN TestAgent_Self_ACLDeny/read-only_token === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:36.388Z [INFO] TestAgent_Self_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:34:36.391Z [INFO] TestAgent_Self_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:34:36.393Z [DEBUG] TestAgent_Self_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:36.396Z [DEBUG] TestAgent_Self_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:36.399Z [DEBUG] TestAgent_Self_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:36.397Z [DEBUG] TestAgent_Self_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:36.402Z [DEBUG] TestAgent_Self_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:36.405Z [WARN] TestAgent_Self_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:36.405Z [DEBUG] TestAgent_Self_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:36.410Z [WARN] TestAgent_Self_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:36.412Z [INFO] TestAgent_Self_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:36.412Z [INFO] TestAgent_Self_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:36.415Z [INFO] TestAgent_Self_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:34:36.416Z [INFO] TestAgent_Self_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29508 network=tcp writer.go:29: 2021-01-29T19:34:36.418Z [INFO] TestAgent_Self_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29508 network=udp writer.go:29: 2021-01-29T19:34:36.419Z [INFO] TestAgent_Self_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:29509 network=tcp === CONT TestAgent_Metrics_ACLDeny writer.go:29: 2021-01-29T19:34:36.612Z [INFO] TestAgent_Metrics_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:36.614Z [INFO] TestAgent_Metrics_ACLDeny: Endpoints down --- PASS: TestAgent_Metrics_ACLDeny (0.80s) --- PASS: TestAgent_Metrics_ACLDeny/no_token (0.00s) --- PASS: TestAgent_Metrics_ACLDeny/agent_master_token (0.00s) --- PASS: TestAgent_Metrics_ACLDeny/read-only_token (0.00s) === CONT TestAgent_Self writer.go:29: 2021-01-29T19:34:36.643Z [WARN] TestAgent_Self: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:36.645Z [DEBUG] TestAgent_Self.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:36.647Z [DEBUG] TestAgent_Self.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:36.656Z [INFO] TestAgent_Self.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:95fb5119-c52f-0086-8040-4e4553f1acec Address:127.0.0.1:29549}]" writer.go:29: 2021-01-29T19:34:36.659Z [INFO] TestAgent_Self.server.raft: entering follower state: follower="Node at 127.0.0.1:29549 [Follower]" leader= writer.go:29: 2021-01-29T19:34:36.661Z [INFO] TestAgent_Self.server.serf.wan: serf: EventMemberJoin: Node-95fb5119-c52f-0086-8040-4e4553f1acec.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:36.667Z [INFO] TestAgent_Self.server.serf.lan: serf: EventMemberJoin: Node-95fb5119-c52f-0086-8040-4e4553f1acec 127.0.0.1 writer.go:29: 2021-01-29T19:34:36.671Z [INFO] TestAgent_Self.server: Handled event for server in area: event=member-join server=Node-95fb5119-c52f-0086-8040-4e4553f1acec.dc1 area=wan writer.go:29: 2021-01-29T19:34:36.672Z [INFO] TestAgent_Self.server: Adding LAN server: server="Node-95fb5119-c52f-0086-8040-4e4553f1acec (Addr: tcp/127.0.0.1:29549) (DC: dc1)" writer.go:29: 2021-01-29T19:34:36.673Z [INFO] TestAgent_Self: Started DNS server: address=127.0.0.1:29544 network=tcp writer.go:29: 2021-01-29T19:34:36.680Z [INFO] TestAgent_Self: Started DNS server: address=127.0.0.1:29544 network=udp writer.go:29: 2021-01-29T19:34:36.688Z [INFO] TestAgent_Self: Started HTTP server: address=127.0.0.1:29545 network=tcp writer.go:29: 2021-01-29T19:34:36.690Z [INFO] TestAgent_Self: started state syncer writer.go:29: 2021-01-29T19:34:36.708Z [WARN] TestAgent_Self.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:36.709Z [INFO] TestAgent_Self.server.raft: entering candidate state: node="Node at 127.0.0.1:29549 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:36.711Z [DEBUG] TestAgent_Self.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:36.712Z [DEBUG] TestAgent_Self.server.raft: vote granted: from=95fb5119-c52f-0086-8040-4e4553f1acec term=2 tally=1 writer.go:29: 2021-01-29T19:34:36.713Z [INFO] TestAgent_Self.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:36.714Z [INFO] TestAgent_Self.server.raft: entering leader state: leader="Node at 127.0.0.1:29549 [Leader]" writer.go:29: 2021-01-29T19:34:36.715Z [INFO] TestAgent_Self.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:36.715Z [INFO] TestAgent_Self.server: New leader elected: payload=Node-95fb5119-c52f-0086-8040-4e4553f1acec writer.go:29: 2021-01-29T19:34:36.716Z [DEBUG] TestAgent_Self.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29549 writer.go:29: 2021-01-29T19:34:36.725Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:36.725Z [INFO] TestAgent_Self: Synced node info writer.go:29: 2021-01-29T19:34:36.728Z [DEBUG] TestAgent_Self: Node info in sync writer.go:29: 2021-01-29T19:34:36.728Z [INFO] TestAgent_Self.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:36.731Z [INFO] TestAgent_Self.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:36.733Z [DEBUG] TestAgent_Self.server: Skipping self join check for node since the cluster is too small: node=Node-95fb5119-c52f-0086-8040-4e4553f1acec writer.go:29: 2021-01-29T19:34:36.734Z [INFO] TestAgent_Self.server: member joined, marking health alive: member=Node-95fb5119-c52f-0086-8040-4e4553f1acec writer.go:29: 2021-01-29T19:34:36.895Z [INFO] TestAgent_Self: Requesting shutdown === CONT TestAgent_Reload writer.go:29: 2021-01-29T19:34:36.895Z [WARN] TestAgent_Reload: Timeout stopping server: protocol=HTTP address=127.0.0.1:29533 network=tcp === CONT TestAgent_Self writer.go:29: 2021-01-29T19:34:36.897Z [INFO] TestAgent_Self.server: shutting down server === CONT TestAgent_Reload writer.go:29: 2021-01-29T19:34:36.897Z [INFO] TestAgent_Reload: Waiting for endpoints to shut down === CONT TestAgent_Self writer.go:29: 2021-01-29T19:34:36.898Z [DEBUG] TestAgent_Self.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_Reload writer.go:29: 2021-01-29T19:34:36.898Z [INFO] TestAgent_Reload: Endpoints down --- PASS: TestAgent_Reload (1.40s) === CONT TestAgent_HealthServicesACLEnforcement === CONT TestAgent_Self writer.go:29: 2021-01-29T19:34:36.899Z [WARN] TestAgent_Self.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:36.899Z [DEBUG] TestAgent_Self.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:36.902Z [WARN] TestAgent_Self.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:36.906Z [INFO] TestAgent_Self.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:36.907Z [INFO] TestAgent_Self: consul server down writer.go:29: 2021-01-29T19:34:36.908Z [INFO] TestAgent_Self: shutdown complete writer.go:29: 2021-01-29T19:34:36.908Z [INFO] TestAgent_Self: Stopping server: protocol=DNS address=127.0.0.1:29544 network=tcp writer.go:29: 2021-01-29T19:34:36.909Z [INFO] TestAgent_Self: Stopping server: protocol=DNS address=127.0.0.1:29544 network=udp writer.go:29: 2021-01-29T19:34:36.910Z [INFO] TestAgent_Self: Stopping server: protocol=HTTP address=127.0.0.1:29545 network=tcp === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:36.918Z [WARN] TestAgent_HealthServicesACLEnforcement: bootstrap = true: do not enable unless necessary === CONT TestAgent_Self_ACLDeny writer.go:29: 2021-01-29T19:34:36.921Z [INFO] TestAgent_Self_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:36.922Z [INFO] TestAgent_Self_ACLDeny: Endpoints down --- PASS: TestAgent_Self_ACLDeny (1.10s) --- PASS: TestAgent_Self_ACLDeny/no_token (0.00s) --- PASS: TestAgent_Self_ACLDeny/agent_master_token (0.00s) --- PASS: TestAgent_Self_ACLDeny/read-only_token (0.01s) === CONT TestACL_vetServiceUpdate === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:36.922Z [DEBUG] TestAgent_HealthServicesACLEnforcement.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:36.945Z [DEBUG] TestAgent_HealthServicesACLEnforcement.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACL_vetServiceUpdate writer.go:29: 2021-01-29T19:34:36.984Z [WARN] TestACL_vetServiceUpdate: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:36.985Z [WARN] TestACL_vetServiceUpdate: bootstrap = true: do not enable unless necessary --- PASS: TestACL_vetServiceUpdate (0.07s) === CONT TestAgent_HealthServiceByID === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:36.991Z [INFO] TestAgent_HealthServicesACLEnforcement.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:34118ac6-f62d-e1b0-cc26-0a93d35b5497 Address:127.0.0.1:29567}]" writer.go:29: 2021-01-29T19:34:37.007Z [INFO] TestAgent_HealthServicesACLEnforcement.server.raft: entering follower state: follower="Node at 127.0.0.1:29567 [Follower]" leader= writer.go:29: 2021-01-29T19:34:37.013Z [INFO] TestAgent_HealthServicesACLEnforcement.server.serf.wan: serf: EventMemberJoin: Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:37.053Z [INFO] TestAgent_HealthServicesACLEnforcement.server.serf.lan: serf: EventMemberJoin: Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 127.0.0.1 === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.077Z [WARN] TestAgent_HealthServiceByID: bootstrap = true: do not enable unless necessary === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.078Z [INFO] TestAgent_HealthServicesACLEnforcement.server: Adding LAN server: server="Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 (Addr: tcp/127.0.0.1:29567) (DC: dc1)" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.100Z [DEBUG] TestAgent_HealthServiceByID.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:37.103Z [DEBUG] TestAgent_HealthServiceByID.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.104Z [INFO] TestAgent_HealthServicesACLEnforcement.server: Handled event for server in area: event=member-join server=Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497.dc1 area=wan writer.go:29: 2021-01-29T19:34:37.109Z [WARN] TestAgent_HealthServicesACLEnforcement.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:37.124Z [INFO] TestAgent_HealthServicesACLEnforcement.server.raft: entering candidate state: node="Node at 127.0.0.1:29567 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:37.112Z [INFO] TestAgent_HealthServicesACLEnforcement: Started DNS server: address=127.0.0.1:29562 network=tcp writer.go:29: 2021-01-29T19:34:37.133Z [INFO] TestAgent_HealthServicesACLEnforcement: Started DNS server: address=127.0.0.1:29562 network=udp writer.go:29: 2021-01-29T19:34:37.139Z [INFO] TestAgent_HealthServicesACLEnforcement: Started HTTP server: address=127.0.0.1:29563 network=tcp writer.go:29: 2021-01-29T19:34:37.143Z [INFO] TestAgent_HealthServicesACLEnforcement: started state syncer writer.go:29: 2021-01-29T19:34:37.140Z [DEBUG] TestAgent_HealthServicesACLEnforcement.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:37.155Z [DEBUG] TestAgent_HealthServicesACLEnforcement.server.raft: vote granted: from=34118ac6-f62d-e1b0-cc26-0a93d35b5497 term=2 tally=1 writer.go:29: 2021-01-29T19:34:37.158Z [INFO] TestAgent_HealthServicesACLEnforcement.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:37.154Z [DEBUG] TestAgent_HealthServicesACLEnforcement.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:37.165Z [INFO] TestAgent_HealthServicesACLEnforcement.server.raft: entering leader state: leader="Node at 127.0.0.1:29567 [Leader]" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.155Z [INFO] TestAgent_HealthServiceByID.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f628f31d-7f9b-cf25-39fe-eefa1e5c2a1e Address:127.0.0.1:29555}]" === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.169Z [INFO] TestAgent_HealthServicesACLEnforcement.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:37.171Z [INFO] TestAgent_HealthServicesACLEnforcement.server: New leader elected: payload=Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.173Z [INFO] TestAgent_HealthServiceByID.server.raft: entering follower state: follower="Node at 127.0.0.1:29555 [Follower]" leader= writer.go:29: 2021-01-29T19:34:37.180Z [INFO] TestAgent_HealthServiceByID.server.serf.wan: serf: EventMemberJoin: Node-f628f31d-7f9b-cf25-39fe-eefa1e5c2a1e.dc1 127.0.0.1 === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.193Z [INFO] TestAgent_HealthServicesACLEnforcement.server: initializing acls === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.201Z [INFO] TestAgent_HealthServiceByID.server.serf.lan: serf: EventMemberJoin: Node-f628f31d-7f9b-cf25-39fe-eefa1e5c2a1e 127.0.0.1 === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.208Z [INFO] TestAgent_HealthServicesACLEnforcement.server: Created ACL 'global-management' policy === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.214Z [INFO] TestAgent_HealthServiceByID.server: Adding LAN server: server="Node-f628f31d-7f9b-cf25-39fe-eefa1e5c2a1e (Addr: tcp/127.0.0.1:29555) (DC: dc1)" === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.224Z [INFO] TestAgent_HealthServicesACLEnforcement.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:37.229Z [INFO] TestAgent_HealthServicesACLEnforcement.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:37.230Z [INFO] TestAgent_HealthServicesACLEnforcement.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:37.232Z [INFO] TestAgent_HealthServicesACLEnforcement.leader: started routine: routine="acl token reaping" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.236Z [WARN] TestAgent_HealthServiceByID.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:37.237Z [INFO] TestAgent_HealthServiceByID.server.raft: entering candidate state: node="Node at 127.0.0.1:29555 [Candidate]" term=2 === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.239Z [INFO] TestAgent_HealthServicesACLEnforcement.server.serf.lan: serf: EventMemberUpdate: Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 writer.go:29: 2021-01-29T19:34:37.243Z [INFO] TestAgent_HealthServicesACLEnforcement.server.serf.wan: serf: EventMemberUpdate: Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497.dc1 writer.go:29: 2021-01-29T19:34:37.247Z [INFO] TestAgent_HealthServicesACLEnforcement.server: Updating LAN server: server="Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 (Addr: tcp/127.0.0.1:29567) (DC: dc1)" writer.go:29: 2021-01-29T19:34:37.247Z [INFO] TestAgent_HealthServicesACLEnforcement.server: Handled event for server in area: event=member-update server=Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497.dc1 area=wan === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.249Z [INFO] TestAgent_HealthServiceByID.server: Handled event for server in area: event=member-join server=Node-f628f31d-7f9b-cf25-39fe-eefa1e5c2a1e.dc1 area=wan writer.go:29: 2021-01-29T19:34:37.250Z [INFO] TestAgent_HealthServiceByID: Started DNS server: address=127.0.0.1:29550 network=udp === RUN TestAgent_HealthServicesACLEnforcement/no-token-health-by-id === RUN TestAgent_HealthServicesACLEnforcement/no-token-health-by-name === RUN TestAgent_HealthServicesACLEnforcement/root-token-health-by-id === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.321Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === RUN TestAgent_HealthServicesACLEnforcement/root-token-health-by-name === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.323Z [INFO] TestAgent_HealthServiceByID: Started DNS server: address=127.0.0.1:29550 network=tcp === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.335Z [INFO] TestAgent_HealthServicesACLEnforcement: Requesting shutdown writer.go:29: 2021-01-29T19:34:37.341Z [INFO] TestAgent_HealthServicesACLEnforcement.server: shutting down server === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.335Z [DEBUG] TestAgent_HealthServiceByID.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:37.347Z [DEBUG] TestAgent_HealthServiceByID.server.raft: vote granted: from=f628f31d-7f9b-cf25-39fe-eefa1e5c2a1e term=2 tally=1 === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.345Z [DEBUG] TestAgent_HealthServicesACLEnforcement.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.349Z [INFO] TestAgent_HealthServiceByID.server.raft: election won: tally=1 === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.354Z [DEBUG] TestAgent_HealthServicesACLEnforcement.leader: stopping routine: routine="acl token reaping" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.355Z [INFO] TestAgent_HealthServiceByID.server.raft: entering leader state: leader="Node at 127.0.0.1:29555 [Leader]" === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.346Z [ERROR] TestAgent_HealthServicesACLEnforcement.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.350Z [INFO] TestAgent_HealthServiceByID: Started HTTP server: address=127.0.0.1:29551 network=tcp === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.356Z [DEBUG] TestAgent_HealthServicesACLEnforcement.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.366Z [INFO] TestAgent_HealthServiceByID: started state syncer writer.go:29: 2021-01-29T19:34:37.370Z [INFO] TestAgent_HealthServiceByID.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:37.371Z [INFO] TestAgent_HealthServiceByID.server: New leader elected: payload=Node-f628f31d-7f9b-cf25-39fe-eefa1e5c2a1e === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.358Z [INFO] TestAgent_HealthServicesACLEnforcement.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.380Z [INFO] TestAgent_HealthServiceByID: Synced node info writer.go:29: 2021-01-29T19:34:37.382Z [DEBUG] TestAgent_HealthServiceByID: Node info in sync writer.go:29: 2021-01-29T19:34:37.384Z [DEBUG] TestAgent_HealthServiceByID.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29555 === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.358Z [WARN] TestAgent_HealthServicesACLEnforcement.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:37.358Z [DEBUG] TestAgent_HealthServicesACLEnforcement.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:37.387Z [INFO] TestAgent_HealthServicesACLEnforcement.leader: started routine: routine="CA root pruning" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.401Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.399Z [WARN] TestAgent_HealthServicesACLEnforcement.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_Self writer.go:29: 2021-01-29T19:34:37.411Z [INFO] TestAgent_Self: Waiting for endpoints to shut down === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.412Z [DEBUG] TestAgent_HealthServicesACLEnforcement.server: Skipping self join check for node since the cluster is too small: node=Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 === CONT TestAgent_Self writer.go:29: 2021-01-29T19:34:37.413Z [INFO] TestAgent_Self: Endpoints down --- PASS: TestAgent_Self (0.80s) === CONT TestAgent_ChecksWithFilter === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.414Z [INFO] TestAgent_HealthServicesACLEnforcement.server: member joined, marking health alive: member=Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 writer.go:29: 2021-01-29T19:34:37.419Z [INFO] TestAgent_HealthServicesACLEnforcement.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:37.421Z [INFO] TestAgent_HealthServicesACLEnforcement: consul server down === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.424Z [INFO] TestAgent_HealthServiceByID.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.423Z [INFO] TestAgent_HealthServicesACLEnforcement: shutdown complete === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.426Z [INFO] TestAgent_HealthServiceByID.leader: started routine: routine="CA root pruning" === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.422Z [ERROR] TestAgent_HealthServicesACLEnforcement.server: failed to reconcile member: member="{Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 127.0.0.1 29565 map[acls:1 bootstrap:1 build:1.7.4: dc:dc1 id:34118ac6-f62d-e1b0-cc26-0a93d35b5497 port:29567 raft_vsn:3 role:consul segment: vsn:2 vsn_max:3 vsn_min:2 wan_join_port:29566] alive 1 5 2 2 5 4}" error="leadership lost while committing log" writer.go:29: 2021-01-29T19:34:37.434Z [DEBUG] TestAgent_HealthServicesACLEnforcement.server: Skipping self join check for node since the cluster is too small: node=Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 writer.go:29: 2021-01-29T19:34:37.437Z [INFO] TestAgent_HealthServicesACLEnforcement.server: member joined, marking health alive: member=Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.430Z [DEBUG] TestAgent_HealthServiceByID.server: Skipping self join check for node since the cluster is too small: node=Node-f628f31d-7f9b-cf25-39fe-eefa1e5c2a1e writer.go:29: 2021-01-29T19:34:37.441Z [INFO] TestAgent_HealthServiceByID.server: member joined, marking health alive: member=Node-f628f31d-7f9b-cf25-39fe-eefa1e5c2a1e === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.440Z [ERROR] TestAgent_HealthServicesACLEnforcement.server: failed to reconcile member: member="{Node-34118ac6-f62d-e1b0-cc26-0a93d35b5497 127.0.0.1 29565 map[acls:1 bootstrap:1 build:1.7.4: dc:dc1 id:34118ac6-f62d-e1b0-cc26-0a93d35b5497 port:29567 raft_vsn:3 role:consul segment: vsn:2 vsn_max:3 vsn_min:2 wan_join_port:29566] alive 1 5 2 2 5 4}" error="raft is already shutdown" writer.go:29: 2021-01-29T19:34:37.443Z [DEBUG] TestAgent_HealthServicesACLEnforcement.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.442Z [INFO] TestAgent_HealthServicesACLEnforcement: Stopping server: protocol=DNS address=127.0.0.1:29562 network=tcp writer.go:29: 2021-01-29T19:34:37.446Z [INFO] TestAgent_HealthServicesACLEnforcement: Stopping server: protocol=DNS address=127.0.0.1:29562 network=udp writer.go:29: 2021-01-29T19:34:37.447Z [INFO] TestAgent_HealthServicesACLEnforcement: Stopping server: protocol=HTTP address=127.0.0.1:29563 network=tcp writer.go:29: 2021-01-29T19:34:37.449Z [DEBUG] TestAgent_HealthServicesACLEnforcement.leader: stopped routine: routine="CA root pruning" === RUN TestAgent_HealthServiceByID/passing_checks === RUN TestAgent_HealthServiceByID/passing_checks/format=text === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:37.492Z [ERROR] TestAgent_Leave_ACLDeny.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" === RUN TestAgent_HealthServiceByID/passing_checks/format=json === RUN TestAgent_HealthServiceByID/warning_checks === RUN TestAgent_HealthServiceByID/warning_checks/format=text === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.508Z [WARN] TestAgent_ChecksWithFilter: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:37.510Z [DEBUG] TestAgent_ChecksWithFilter.tlsutil: Update: version=1 === RUN TestAgent_HealthServiceByID/warning_checks/format=json === RUN TestAgent_HealthServiceByID/critical_checks === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.512Z [DEBUG] TestAgent_ChecksWithFilter.tlsutil: OutgoingRPCWrapper: version=1 === RUN TestAgent_HealthServiceByID/critical_checks/format=text === RUN TestAgent_HealthServiceByID/critical_checks/format=json === RUN TestAgent_HealthServiceByID/unknown_serviceid === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.519Z [INFO] TestAgent_ChecksWithFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c066db23-3914-6d0d-5847-96b8283f910d Address:127.0.0.1:29561}]" writer.go:29: 2021-01-29T19:34:37.523Z [INFO] TestAgent_ChecksWithFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:29561 [Follower]" leader= === RUN TestAgent_HealthServiceByID/unknown_serviceid/format=text === RUN TestAgent_HealthServiceByID/unknown_serviceid/format=json === RUN TestAgent_HealthServiceByID/critical_check_on_node === RUN TestAgent_HealthServiceByID/critical_check_on_node/format=text === RUN TestAgent_HealthServiceByID/critical_check_on_node/format=json === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.533Z [INFO] TestAgent_ChecksWithFilter.server.serf.wan: serf: EventMemberJoin: Node-c066db23-3914-6d0d-5847-96b8283f910d.dc1 127.0.0.1 === RUN TestAgent_HealthServiceByID/maintenance_check_on_node === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.537Z [INFO] TestAgent_ChecksWithFilter.server.serf.lan: serf: EventMemberJoin: Node-c066db23-3914-6d0d-5847-96b8283f910d 127.0.0.1 === RUN TestAgent_HealthServiceByID/maintenance_check_on_node/format=text === RUN TestAgent_HealthServiceByID/maintenance_check_on_node/format=json === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.544Z [INFO] TestAgent_ChecksWithFilter.server: Adding LAN server: server="Node-c066db23-3914-6d0d-5847-96b8283f910d (Addr: tcp/127.0.0.1:29561) (DC: dc1)" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.545Z [INFO] TestAgent_HealthServiceByID: Requesting shutdown === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.545Z [INFO] TestAgent_ChecksWithFilter.server: Handled event for server in area: event=member-join server=Node-c066db23-3914-6d0d-5847-96b8283f910d.dc1 area=wan === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.547Z [INFO] TestAgent_HealthServiceByID.server: shutting down server writer.go:29: 2021-01-29T19:34:37.547Z [DEBUG] TestAgent_HealthServiceByID.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.548Z [WARN] TestAgent_HealthServiceByID.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.549Z [INFO] TestAgent_ChecksWithFilter: Started DNS server: address=127.0.0.1:29556 network=udp === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.549Z [DEBUG] TestAgent_HealthServiceByID.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.550Z [INFO] TestAgent_ChecksWithFilter: Started DNS server: address=127.0.0.1:29556 network=tcp writer.go:29: 2021-01-29T19:34:37.552Z [INFO] TestAgent_ChecksWithFilter: Started HTTP server: address=127.0.0.1:29557 network=tcp writer.go:29: 2021-01-29T19:34:37.554Z [INFO] TestAgent_ChecksWithFilter: started state syncer === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.553Z [WARN] TestAgent_HealthServiceByID.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:37.557Z [INFO] TestAgent_HealthServiceByID.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:37.557Z [INFO] TestAgent_HealthServiceByID: consul server down === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.563Z [WARN] TestAgent_ChecksWithFilter.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.562Z [INFO] TestAgent_HealthServiceByID: shutdown complete === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.564Z [INFO] TestAgent_ChecksWithFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:29561 [Candidate]" term=2 === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.565Z [INFO] TestAgent_HealthServiceByID: Stopping server: protocol=DNS address=127.0.0.1:29550 network=tcp === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.566Z [DEBUG] TestAgent_ChecksWithFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:37.567Z [DEBUG] TestAgent_ChecksWithFilter.server.raft: vote granted: from=c066db23-3914-6d0d-5847-96b8283f910d term=2 tally=1 === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.567Z [INFO] TestAgent_HealthServiceByID: Stopping server: protocol=DNS address=127.0.0.1:29550 network=udp === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.568Z [INFO] TestAgent_ChecksWithFilter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:37.568Z [INFO] TestAgent_ChecksWithFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:29561 [Leader]" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:37.568Z [INFO] TestAgent_HealthServiceByID: Stopping server: protocol=HTTP address=127.0.0.1:29551 network=tcp === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:37.569Z [INFO] TestAgent_ChecksWithFilter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:37.571Z [INFO] TestAgent_ChecksWithFilter.server: New leader elected: payload=Node-c066db23-3914-6d0d-5847-96b8283f910d writer.go:29: 2021-01-29T19:34:37.572Z [DEBUG] TestAgent_ChecksWithFilter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29561 writer.go:29: 2021-01-29T19:34:37.580Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:37.591Z [INFO] TestAgent_ChecksWithFilter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:37.593Z [INFO] TestAgent_ChecksWithFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.594Z [DEBUG] TestAgent_ChecksWithFilter.server: Skipping self join check for node since the cluster is too small: node=Node-c066db23-3914-6d0d-5847-96b8283f910d writer.go:29: 2021-01-29T19:34:37.595Z [INFO] TestAgent_ChecksWithFilter.server: member joined, marking health alive: member=Node-c066db23-3914-6d0d-5847-96b8283f910d writer.go:29: 2021-01-29T19:34:37.856Z [INFO] TestAgent_ChecksWithFilter: Requesting shutdown writer.go:29: 2021-01-29T19:34:37.857Z [INFO] TestAgent_ChecksWithFilter.server: shutting down server writer.go:29: 2021-01-29T19:34:37.859Z [DEBUG] TestAgent_ChecksWithFilter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.861Z [WARN] TestAgent_ChecksWithFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:37.859Z [ERROR] TestAgent_ChecksWithFilter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:37.861Z [DEBUG] TestAgent_ChecksWithFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.864Z [WARN] TestAgent_ChecksWithFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:37.871Z [INFO] TestAgent_ChecksWithFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:37.872Z [INFO] TestAgent_ChecksWithFilter: consul server down writer.go:29: 2021-01-29T19:34:37.875Z [INFO] TestAgent_ChecksWithFilter: shutdown complete writer.go:29: 2021-01-29T19:34:37.877Z [INFO] TestAgent_ChecksWithFilter: Stopping server: protocol=DNS address=127.0.0.1:29556 network=tcp writer.go:29: 2021-01-29T19:34:37.879Z [INFO] TestAgent_ChecksWithFilter: Stopping server: protocol=DNS address=127.0.0.1:29556 network=udp writer.go:29: 2021-01-29T19:34:37.881Z [INFO] TestAgent_ChecksWithFilter: Stopping server: protocol=HTTP address=127.0.0.1:29557 network=tcp === CONT TestAgent_HealthServicesACLEnforcement writer.go:29: 2021-01-29T19:34:37.949Z [INFO] TestAgent_HealthServicesACLEnforcement: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:37.950Z [INFO] TestAgent_HealthServicesACLEnforcement: Endpoints down --- PASS: TestAgent_HealthServicesACLEnforcement (1.05s) --- PASS: TestAgent_HealthServicesACLEnforcement/no-token-health-by-id (0.00s) --- PASS: TestAgent_HealthServicesACLEnforcement/no-token-health-by-name (0.01s) --- PASS: TestAgent_HealthServicesACLEnforcement/root-token-health-by-id (0.02s) --- PASS: TestAgent_HealthServicesACLEnforcement/root-token-health-by-name (0.00s) === CONT TestAgent_Checks writer.go:29: 2021-01-29T19:34:37.963Z [WARN] TestAgent_Checks: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:37.966Z [DEBUG] TestAgent_Checks.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:37.967Z [DEBUG] TestAgent_Checks.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:37.972Z [INFO] TestAgent_Checks.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b31de50a-961e-0812-1960-fd6d3671a694 Address:127.0.0.1:29579}]" writer.go:29: 2021-01-29T19:34:37.973Z [INFO] TestAgent_Checks.server.raft: entering follower state: follower="Node at 127.0.0.1:29579 [Follower]" leader= writer.go:29: 2021-01-29T19:34:37.977Z [INFO] TestAgent_Checks.server.serf.wan: serf: EventMemberJoin: Node-b31de50a-961e-0812-1960-fd6d3671a694.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:37.980Z [INFO] TestAgent_Checks.server.serf.lan: serf: EventMemberJoin: Node-b31de50a-961e-0812-1960-fd6d3671a694 127.0.0.1 writer.go:29: 2021-01-29T19:34:37.983Z [INFO] TestAgent_Checks.server: Adding LAN server: server="Node-b31de50a-961e-0812-1960-fd6d3671a694 (Addr: tcp/127.0.0.1:29579) (DC: dc1)" writer.go:29: 2021-01-29T19:34:37.983Z [INFO] TestAgent_Checks.server: Handled event for server in area: event=member-join server=Node-b31de50a-961e-0812-1960-fd6d3671a694.dc1 area=wan writer.go:29: 2021-01-29T19:34:37.986Z [INFO] TestAgent_Checks: Started DNS server: address=127.0.0.1:29574 network=tcp writer.go:29: 2021-01-29T19:34:37.988Z [INFO] TestAgent_Checks: Started DNS server: address=127.0.0.1:29574 network=udp writer.go:29: 2021-01-29T19:34:37.989Z [INFO] TestAgent_Checks: Started HTTP server: address=127.0.0.1:29575 network=tcp writer.go:29: 2021-01-29T19:34:37.990Z [INFO] TestAgent_Checks: started state syncer writer.go:29: 2021-01-29T19:34:38.032Z [WARN] TestAgent_Checks.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:38.034Z [INFO] TestAgent_Checks.server.raft: entering candidate state: node="Node at 127.0.0.1:29579 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:38.037Z [DEBUG] TestAgent_Checks.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:38.038Z [DEBUG] TestAgent_Checks.server.raft: vote granted: from=b31de50a-961e-0812-1960-fd6d3671a694 term=2 tally=1 writer.go:29: 2021-01-29T19:34:38.039Z [INFO] TestAgent_Checks.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:38.040Z [INFO] TestAgent_Checks.server.raft: entering leader state: leader="Node at 127.0.0.1:29579 [Leader]" writer.go:29: 2021-01-29T19:34:38.041Z [INFO] TestAgent_Checks.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:38.042Z [INFO] TestAgent_Checks.server: New leader elected: payload=Node-b31de50a-961e-0812-1960-fd6d3671a694 writer.go:29: 2021-01-29T19:34:38.044Z [DEBUG] TestAgent_Checks.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29579 writer.go:29: 2021-01-29T19:34:38.052Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:38.059Z [INFO] TestAgent_Checks.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:38.064Z [INFO] TestAgent_Checks.leader: started routine: routine="CA root pruning" === CONT TestAgent_HealthServiceByID writer.go:29: 2021-01-29T19:34:38.070Z [INFO] TestAgent_HealthServiceByID: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:38.071Z [INFO] TestAgent_HealthServiceByID: Endpoints down --- PASS: TestAgent_HealthServiceByID (1.08s) --- PASS: TestAgent_HealthServiceByID/passing_checks (0.04s) --- PASS: TestAgent_HealthServiceByID/passing_checks/format=text (0.00s) --- PASS: TestAgent_HealthServiceByID/passing_checks/format=json (0.00s) --- PASS: TestAgent_HealthServiceByID/warning_checks (0.01s) --- PASS: TestAgent_HealthServiceByID/warning_checks/format=text (0.00s) --- PASS: TestAgent_HealthServiceByID/warning_checks/format=json (0.00s) --- PASS: TestAgent_HealthServiceByID/critical_checks (0.01s) --- PASS: TestAgent_HealthServiceByID/critical_checks/format=text (0.00s) --- PASS: TestAgent_HealthServiceByID/critical_checks/format=json (0.00s) --- PASS: TestAgent_HealthServiceByID/unknown_serviceid (0.01s) --- PASS: TestAgent_HealthServiceByID/unknown_serviceid/format=text (0.00s) --- PASS: TestAgent_HealthServiceByID/unknown_serviceid/format=json (0.00s) --- PASS: TestAgent_HealthServiceByID/critical_check_on_node (0.00s) --- PASS: TestAgent_HealthServiceByID/critical_check_on_node/format=text (0.00s) --- PASS: TestAgent_HealthServiceByID/critical_check_on_node/format=json (0.00s) --- PASS: TestAgent_HealthServiceByID/maintenance_check_on_node (0.01s) --- PASS: TestAgent_HealthServiceByID/maintenance_check_on_node/format=text (0.00s) --- PASS: TestAgent_HealthServiceByID/maintenance_check_on_node/format=json (0.00s) === CONT TestAgent_Services_ACLFilter === CONT TestAgent_Checks writer.go:29: 2021-01-29T19:34:38.070Z [DEBUG] TestAgent_Checks.server: Skipping self join check for node since the cluster is too small: node=Node-b31de50a-961e-0812-1960-fd6d3671a694 writer.go:29: 2021-01-29T19:34:38.083Z [INFO] TestAgent_Checks.server: member joined, marking health alive: member=Node-b31de50a-961e-0812-1960-fd6d3671a694 writer.go:29: 2021-01-29T19:34:38.098Z [INFO] TestAgent_Checks: Requesting shutdown writer.go:29: 2021-01-29T19:34:38.100Z [INFO] TestAgent_Checks.server: shutting down server writer.go:29: 2021-01-29T19:34:38.102Z [DEBUG] TestAgent_Checks.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:38.102Z [ERROR] TestAgent_Checks.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:38.106Z [DEBUG] TestAgent_Checks.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:38.106Z [WARN] TestAgent_Checks.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_Services_ACLFilter writer.go:29: 2021-01-29T19:34:38.107Z [WARN] TestAgent_Services_ACLFilter: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:38.109Z [WARN] TestAgent_Services_ACLFilter: bootstrap = true: do not enable unless necessary === CONT TestAgent_Checks writer.go:29: 2021-01-29T19:34:38.110Z [WARN] TestAgent_Checks.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:38.112Z [INFO] TestAgent_Checks.server.router.manager: shutting down === CONT TestAgent_Services_ACLFilter writer.go:29: 2021-01-29T19:34:38.111Z [DEBUG] TestAgent_Services_ACLFilter.tlsutil: Update: version=1 === CONT TestAgent_Checks writer.go:29: 2021-01-29T19:34:38.113Z [INFO] TestAgent_Checks: consul server down writer.go:29: 2021-01-29T19:34:38.115Z [INFO] TestAgent_Checks: shutdown complete writer.go:29: 2021-01-29T19:34:38.116Z [INFO] TestAgent_Checks: Stopping server: protocol=DNS address=127.0.0.1:29574 network=tcp writer.go:29: 2021-01-29T19:34:38.118Z [INFO] TestAgent_Checks: Stopping server: protocol=DNS address=127.0.0.1:29574 network=udp === CONT TestAgent_Services_ACLFilter writer.go:29: 2021-01-29T19:34:38.115Z [DEBUG] TestAgent_Services_ACLFilter.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Checks writer.go:29: 2021-01-29T19:34:38.120Z [INFO] TestAgent_Checks: Stopping server: protocol=HTTP address=127.0.0.1:29575 network=tcp === CONT TestAgent_Services_ACLFilter writer.go:29: 2021-01-29T19:34:38.127Z [INFO] TestAgent_Services_ACLFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d Address:127.0.0.1:29573}]" writer.go:29: 2021-01-29T19:34:38.131Z [INFO] TestAgent_Services_ACLFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:29573 [Follower]" leader= writer.go:29: 2021-01-29T19:34:38.138Z [INFO] TestAgent_Services_ACLFilter.server.serf.wan: serf: EventMemberJoin: Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:38.146Z [INFO] TestAgent_Services_ACLFilter.server.serf.lan: serf: EventMemberJoin: Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d 127.0.0.1 writer.go:29: 2021-01-29T19:34:38.153Z [INFO] TestAgent_Services_ACLFilter.server: Handled event for server in area: event=member-join server=Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d.dc1 area=wan writer.go:29: 2021-01-29T19:34:38.154Z [INFO] TestAgent_Services_ACLFilter.server: Adding LAN server: server="Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d (Addr: tcp/127.0.0.1:29573) (DC: dc1)" writer.go:29: 2021-01-29T19:34:38.158Z [INFO] TestAgent_Services_ACLFilter: Started DNS server: address=127.0.0.1:29568 network=tcp writer.go:29: 2021-01-29T19:34:38.161Z [INFO] TestAgent_Services_ACLFilter: Started DNS server: address=127.0.0.1:29568 network=udp writer.go:29: 2021-01-29T19:34:38.168Z [INFO] TestAgent_Services_ACLFilter: Started HTTP server: address=127.0.0.1:29569 network=tcp writer.go:29: 2021-01-29T19:34:38.169Z [INFO] TestAgent_Services_ACLFilter: started state syncer writer.go:29: 2021-01-29T19:34:38.186Z [WARN] TestAgent_Services_ACLFilter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:38.188Z [INFO] TestAgent_Services_ACLFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:29573 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:38.195Z [DEBUG] TestAgent_Services_ACLFilter.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:38.198Z [DEBUG] TestAgent_Services_ACLFilter.server.raft: vote granted: from=c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d term=2 tally=1 writer.go:29: 2021-01-29T19:34:38.202Z [INFO] TestAgent_Services_ACLFilter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:38.206Z [INFO] TestAgent_Services_ACLFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:29573 [Leader]" writer.go:29: 2021-01-29T19:34:38.204Z [DEBUG] TestAgent_Services_ACLFilter.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:38.213Z [INFO] TestAgent_Services_ACLFilter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:38.214Z [INFO] TestAgent_Services_ACLFilter.server: New leader elected: payload=Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d writer.go:29: 2021-01-29T19:34:38.216Z [INFO] TestAgent_Services_ACLFilter.server: initializing acls writer.go:29: 2021-01-29T19:34:38.224Z [INFO] TestAgent_Services_ACLFilter.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:38.226Z [WARN] TestAgent_Services_ACLFilter.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:38.228Z [INFO] TestAgent_Services_ACLFilter.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:38.232Z [INFO] TestAgent_Services_ACLFilter.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:38.233Z [INFO] TestAgent_Services_ACLFilter.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:38.234Z [INFO] TestAgent_Services_ACLFilter.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:38.235Z [INFO] TestAgent_Services_ACLFilter.server.serf.lan: serf: EventMemberUpdate: Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d writer.go:29: 2021-01-29T19:34:38.236Z [INFO] TestAgent_Services_ACLFilter.server.serf.wan: serf: EventMemberUpdate: Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d.dc1 writer.go:29: 2021-01-29T19:34:38.236Z [INFO] TestAgent_Services_ACLFilter.server: Updating LAN server: server="Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d (Addr: tcp/127.0.0.1:29573) (DC: dc1)" writer.go:29: 2021-01-29T19:34:38.238Z [INFO] TestAgent_Services_ACLFilter.server: Handled event for server in area: event=member-update server=Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d.dc1 area=wan writer.go:29: 2021-01-29T19:34:38.240Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:38.244Z [INFO] TestAgent_Services_ACLFilter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:38.245Z [INFO] TestAgent_Services_ACLFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:38.246Z [DEBUG] TestAgent_Services_ACLFilter.server: Skipping self join check for node since the cluster is too small: node=Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d writer.go:29: 2021-01-29T19:34:38.247Z [INFO] TestAgent_Services_ACLFilter.server: member joined, marking health alive: member=Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d writer.go:29: 2021-01-29T19:34:38.249Z [DEBUG] TestAgent_Services_ACLFilter.server: Skipping self join check for node since the cluster is too small: node=Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d writer.go:29: 2021-01-29T19:34:38.313Z [DEBUG] TestAgent_Services_ACLFilter: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:38.317Z [INFO] TestAgent_Services_ACLFilter: Synced node info === CONT TestAgent_ChecksWithFilter writer.go:29: 2021-01-29T19:34:38.382Z [INFO] TestAgent_ChecksWithFilter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:38.384Z [INFO] TestAgent_ChecksWithFilter: Endpoints down --- PASS: TestAgent_ChecksWithFilter (0.97s) === CONT TestAgent_Services_MeshGateway writer.go:29: 2021-01-29T19:34:38.401Z [WARN] TestAgent_Services_MeshGateway: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:38.404Z [DEBUG] TestAgent_Services_MeshGateway.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:38.406Z [DEBUG] TestAgent_Services_MeshGateway.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:38.413Z [INFO] TestAgent_Services_MeshGateway.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ee4db4c0-8ea3-6f4a-c478-2295784b0c9c Address:127.0.0.1:29591}]" writer.go:29: 2021-01-29T19:34:38.418Z [INFO] TestAgent_Services_MeshGateway.server.raft: entering follower state: follower="Node at 127.0.0.1:29591 [Follower]" leader= writer.go:29: 2021-01-29T19:34:38.425Z [INFO] TestAgent_Services_MeshGateway.server.serf.wan: serf: EventMemberJoin: Node-ee4db4c0-8ea3-6f4a-c478-2295784b0c9c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:38.432Z [INFO] TestAgent_Services_MeshGateway.server.serf.lan: serf: EventMemberJoin: Node-ee4db4c0-8ea3-6f4a-c478-2295784b0c9c 127.0.0.1 writer.go:29: 2021-01-29T19:34:38.437Z [INFO] TestAgent_Services_MeshGateway.server: Adding LAN server: server="Node-ee4db4c0-8ea3-6f4a-c478-2295784b0c9c (Addr: tcp/127.0.0.1:29591) (DC: dc1)" writer.go:29: 2021-01-29T19:34:38.437Z [INFO] TestAgent_Services_MeshGateway.server: Handled event for server in area: event=member-join server=Node-ee4db4c0-8ea3-6f4a-c478-2295784b0c9c.dc1 area=wan writer.go:29: 2021-01-29T19:34:38.442Z [INFO] TestAgent_Services_MeshGateway: Started DNS server: address=127.0.0.1:29586 network=tcp writer.go:29: 2021-01-29T19:34:38.445Z [INFO] TestAgent_Services_MeshGateway: Started DNS server: address=127.0.0.1:29586 network=udp writer.go:29: 2021-01-29T19:34:38.447Z [INFO] TestAgent_Services_MeshGateway: Started HTTP server: address=127.0.0.1:29587 network=tcp writer.go:29: 2021-01-29T19:34:38.449Z [INFO] TestAgent_Services_MeshGateway: started state syncer writer.go:29: 2021-01-29T19:34:38.462Z [WARN] TestAgent_Services_MeshGateway.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:38.464Z [INFO] TestAgent_Services_MeshGateway.server.raft: entering candidate state: node="Node at 127.0.0.1:29591 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:38.467Z [DEBUG] TestAgent_Services_MeshGateway.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:38.469Z [DEBUG] TestAgent_Services_MeshGateway.server.raft: vote granted: from=ee4db4c0-8ea3-6f4a-c478-2295784b0c9c term=2 tally=1 writer.go:29: 2021-01-29T19:34:38.471Z [INFO] TestAgent_Services_MeshGateway.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:38.473Z [INFO] TestAgent_Services_MeshGateway.server.raft: entering leader state: leader="Node at 127.0.0.1:29591 [Leader]" writer.go:29: 2021-01-29T19:34:38.475Z [INFO] TestAgent_Services_MeshGateway.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:38.475Z [INFO] TestAgent_Services_MeshGateway.server: New leader elected: payload=Node-ee4db4c0-8ea3-6f4a-c478-2295784b0c9c writer.go:29: 2021-01-29T19:34:38.476Z [DEBUG] TestAgent_Services_MeshGateway.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29591 writer.go:29: 2021-01-29T19:34:38.483Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:38.488Z [INFO] TestAgent_Services_MeshGateway.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:38.489Z [INFO] TestAgent_Services_MeshGateway.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:38.491Z [DEBUG] TestAgent_Services_MeshGateway.server: Skipping self join check for node since the cluster is too small: node=Node-ee4db4c0-8ea3-6f4a-c478-2295784b0c9c writer.go:29: 2021-01-29T19:34:38.492Z [INFO] TestAgent_Services_MeshGateway.server: member joined, marking health alive: member=Node-ee4db4c0-8ea3-6f4a-c478-2295784b0c9c === CONT TestAgent_Services_ACLFilter writer.go:29: 2021-01-29T19:34:38.564Z [DEBUG] TestAgent_Services_ACLFilter.acl: dropping node from result due to ACLs: node=Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d writer.go:29: 2021-01-29T19:34:38.569Z [DEBUG] TestAgent_Services_ACLFilter.acl: dropping node from result due to ACLs: node=Node-c3dbf9eb-3a25-0303-203e-2dc7b2c9be7d === RUN TestAgent_Services_ACLFilter/no_token === CONT TestAgent_Services_ACLFilter writer.go:29: 2021-01-29T19:34:38.572Z [DEBUG] TestAgent_Services_ACLFilter: dropping service from result due to ACLs: service=mysql === RUN TestAgent_Services_ACLFilter/root_token === CONT TestAgent_Services_ACLFilter writer.go:29: 2021-01-29T19:34:38.575Z [INFO] TestAgent_Services_ACLFilter: Requesting shutdown writer.go:29: 2021-01-29T19:34:38.577Z [INFO] TestAgent_Services_ACLFilter.server: shutting down server writer.go:29: 2021-01-29T19:34:38.578Z [DEBUG] TestAgent_Services_ACLFilter.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:38.579Z [DEBUG] TestAgent_Services_ACLFilter.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:38.579Z [DEBUG] TestAgent_Services_ACLFilter.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:38.580Z [WARN] TestAgent_Services_ACLFilter.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:38.579Z [DEBUG] TestAgent_Services_ACLFilter.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:38.580Z [DEBUG] TestAgent_Services_ACLFilter.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:38.581Z [DEBUG] TestAgent_Services_ACLFilter.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:38.582Z [WARN] TestAgent_Services_ACLFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:38.594Z [INFO] TestAgent_Services_ACLFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:38.594Z [INFO] TestAgent_Services_ACLFilter: consul server down writer.go:29: 2021-01-29T19:34:38.598Z [INFO] TestAgent_Services_ACLFilter: shutdown complete writer.go:29: 2021-01-29T19:34:38.601Z [INFO] TestAgent_Services_ACLFilter: Stopping server: protocol=DNS address=127.0.0.1:29568 network=tcp writer.go:29: 2021-01-29T19:34:38.604Z [INFO] TestAgent_Services_ACLFilter: Stopping server: protocol=DNS address=127.0.0.1:29568 network=udp writer.go:29: 2021-01-29T19:34:38.606Z [INFO] TestAgent_Services_ACLFilter: Stopping server: protocol=HTTP address=127.0.0.1:29569 network=tcp === CONT TestAgent_Checks writer.go:29: 2021-01-29T19:34:38.622Z [INFO] TestAgent_Checks: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:38.624Z [INFO] TestAgent_Checks: Endpoints down --- PASS: TestAgent_Checks (0.67s) === CONT TestAgent_Services_Sidecar === CONT TestAgent_Services_MeshGateway writer.go:29: 2021-01-29T19:34:38.636Z [INFO] TestAgent_Services_MeshGateway: Requesting shutdown === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:38.637Z [WARN] TestAgent_Services_Sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:38.639Z [DEBUG] TestAgent_Services_Sidecar.tlsutil: Update: version=1 === CONT TestAgent_Services_MeshGateway writer.go:29: 2021-01-29T19:34:38.639Z [INFO] TestAgent_Services_MeshGateway.server: shutting down server writer.go:29: 2021-01-29T19:34:38.641Z [DEBUG] TestAgent_Services_MeshGateway.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:38.641Z [DEBUG] TestAgent_Services_Sidecar.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Services_MeshGateway writer.go:29: 2021-01-29T19:34:38.643Z [WARN] TestAgent_Services_MeshGateway.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:38.642Z [ERROR] TestAgent_Services_MeshGateway.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:38.643Z [DEBUG] TestAgent_Services_MeshGateway.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:38.647Z [INFO] TestAgent_Services_Sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:95ef4675-05db-7cf2-d481-81c2248c9f81 Address:127.0.0.1:29597}]" === CONT TestAgent_Services_MeshGateway writer.go:29: 2021-01-29T19:34:38.651Z [WARN] TestAgent_Services_MeshGateway.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:38.653Z [INFO] TestAgent_Services_MeshGateway.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:38.654Z [INFO] TestAgent_Services_MeshGateway: consul server down writer.go:29: 2021-01-29T19:34:38.655Z [INFO] TestAgent_Services_MeshGateway: shutdown complete writer.go:29: 2021-01-29T19:34:38.656Z [INFO] TestAgent_Services_MeshGateway: Stopping server: protocol=DNS address=127.0.0.1:29586 network=tcp writer.go:29: 2021-01-29T19:34:38.657Z [INFO] TestAgent_Services_MeshGateway: Stopping server: protocol=DNS address=127.0.0.1:29586 network=udp writer.go:29: 2021-01-29T19:34:38.658Z [INFO] TestAgent_Services_MeshGateway: Stopping server: protocol=HTTP address=127.0.0.1:29587 network=tcp === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:38.651Z [INFO] TestAgent_Services_Sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:29597 [Follower]" leader= writer.go:29: 2021-01-29T19:34:38.652Z [INFO] TestAgent_Services_Sidecar.server.serf.wan: serf: EventMemberJoin: Node-95ef4675-05db-7cf2-d481-81c2248c9f81.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:38.665Z [INFO] TestAgent_Services_Sidecar.server.serf.lan: serf: EventMemberJoin: Node-95ef4675-05db-7cf2-d481-81c2248c9f81 127.0.0.1 writer.go:29: 2021-01-29T19:34:38.669Z [INFO] TestAgent_Services_Sidecar.server: Adding LAN server: server="Node-95ef4675-05db-7cf2-d481-81c2248c9f81 (Addr: tcp/127.0.0.1:29597) (DC: dc1)" writer.go:29: 2021-01-29T19:34:38.671Z [INFO] TestAgent_Services_Sidecar.server: Handled event for server in area: event=member-join server=Node-95ef4675-05db-7cf2-d481-81c2248c9f81.dc1 area=wan writer.go:29: 2021-01-29T19:34:38.675Z [INFO] TestAgent_Services_Sidecar: Started DNS server: address=127.0.0.1:29592 network=tcp writer.go:29: 2021-01-29T19:34:38.685Z [INFO] TestAgent_Services_Sidecar: Started DNS server: address=127.0.0.1:29592 network=udp writer.go:29: 2021-01-29T19:34:38.687Z [INFO] TestAgent_Services_Sidecar: Started HTTP server: address=127.0.0.1:29593 network=tcp writer.go:29: 2021-01-29T19:34:38.689Z [INFO] TestAgent_Services_Sidecar: started state syncer writer.go:29: 2021-01-29T19:34:38.715Z [WARN] TestAgent_Services_Sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:38.717Z [INFO] TestAgent_Services_Sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:29597 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:38.720Z [DEBUG] TestAgent_Services_Sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:38.722Z [DEBUG] TestAgent_Services_Sidecar.server.raft: vote granted: from=95ef4675-05db-7cf2-d481-81c2248c9f81 term=2 tally=1 writer.go:29: 2021-01-29T19:34:38.724Z [INFO] TestAgent_Services_Sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:38.726Z [INFO] TestAgent_Services_Sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:29597 [Leader]" writer.go:29: 2021-01-29T19:34:38.727Z [INFO] TestAgent_Services_Sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:38.728Z [INFO] TestAgent_Services_Sidecar.server: New leader elected: payload=Node-95ef4675-05db-7cf2-d481-81c2248c9f81 writer.go:29: 2021-01-29T19:34:38.743Z [DEBUG] TestAgent_Services_Sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29597 writer.go:29: 2021-01-29T19:34:38.757Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:38.764Z [INFO] TestAgent_Services_Sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:38.766Z [INFO] TestAgent_Services_Sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:38.768Z [DEBUG] TestAgent_Services_Sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-95ef4675-05db-7cf2-d481-81c2248c9f81 writer.go:29: 2021-01-29T19:34:38.769Z [INFO] TestAgent_Services_Sidecar.server: member joined, marking health alive: member=Node-95ef4675-05db-7cf2-d481-81c2248c9f81 writer.go:29: 2021-01-29T19:34:38.866Z [DEBUG] TestAgent_Services_Sidecar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:38.872Z [INFO] TestAgent_Services_Sidecar: Synced node info writer.go:29: 2021-01-29T19:34:38.873Z [DEBUG] TestAgent_Services_Sidecar: Node info in sync === CONT TestAgent_Services_ACLFilter writer.go:29: 2021-01-29T19:34:39.108Z [INFO] TestAgent_Services_ACLFilter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:39.110Z [INFO] TestAgent_Services_ACLFilter: Endpoints down --- PASS: TestAgent_Services_ACLFilter (1.04s) --- PASS: TestAgent_Services_ACLFilter/no_token (0.00s) --- PASS: TestAgent_Services_ACLFilter/root_token (0.00s) === CONT TestAgent_Services_ExternalConnectProxy === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:39.129Z [INFO] TestAgent_Services_Sidecar: Requesting shutdown === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.145Z [WARN] TestAgent_Services_ExternalConnectProxy: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:39.148Z [DEBUG] TestAgent_Services_ExternalConnectProxy.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:39.150Z [DEBUG] TestAgent_Services_ExternalConnectProxy.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:39.157Z [INFO] TestAgent_Services_ExternalConnectProxy.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b6a2890a-86b4-238c-8580-4213a124b89f Address:127.0.0.1:29585}]" === CONT TestAgent_Services_MeshGateway writer.go:29: 2021-01-29T19:34:39.160Z [INFO] TestAgent_Services_MeshGateway: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:39.161Z [INFO] TestAgent_Services_MeshGateway: Endpoints down --- PASS: TestAgent_Services_MeshGateway (0.78s) === CONT TestAgent_ServicesFiltered === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.160Z [INFO] TestAgent_Services_ExternalConnectProxy.server.raft: entering follower state: follower="Node at 127.0.0.1:29585 [Follower]" leader= === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:39.140Z [INFO] TestAgent_Services_Sidecar.server: shutting down server writer.go:29: 2021-01-29T19:34:39.163Z [DEBUG] TestAgent_Services_Sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:39.165Z [WARN] TestAgent_Services_Sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:39.167Z [DEBUG] TestAgent_Services_Sidecar.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.172Z [INFO] TestAgent_Services_ExternalConnectProxy.server.serf.wan: serf: EventMemberJoin: Node-b6a2890a-86b4-238c-8580-4213a124b89f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:39.177Z [INFO] TestAgent_Services_ExternalConnectProxy.server.serf.lan: serf: EventMemberJoin: Node-b6a2890a-86b4-238c-8580-4213a124b89f 127.0.0.1 writer.go:29: 2021-01-29T19:34:39.181Z [INFO] TestAgent_Services_ExternalConnectProxy.server: Adding LAN server: server="Node-b6a2890a-86b4-238c-8580-4213a124b89f (Addr: tcp/127.0.0.1:29585) (DC: dc1)" writer.go:29: 2021-01-29T19:34:39.182Z [INFO] TestAgent_Services_ExternalConnectProxy.server: Handled event for server in area: event=member-join server=Node-b6a2890a-86b4-238c-8580-4213a124b89f.dc1 area=wan === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:39.176Z [WARN] TestAgent_Services_Sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:39.185Z [INFO] TestAgent_Services_Sidecar.server.router.manager: shutting down === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.186Z [WARN] TestAgent_ServicesFiltered: bootstrap = true: do not enable unless necessary === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:39.187Z [INFO] TestAgent_Services_Sidecar: consul server down writer.go:29: 2021-01-29T19:34:39.192Z [INFO] TestAgent_Services_Sidecar: shutdown complete === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.197Z [INFO] TestAgent_Services_ExternalConnectProxy: Started DNS server: address=127.0.0.1:29580 network=udp writer.go:29: 2021-01-29T19:34:39.198Z [INFO] TestAgent_Services_ExternalConnectProxy: Started DNS server: address=127.0.0.1:29580 network=tcp === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.189Z [DEBUG] TestAgent_ServicesFiltered.tlsutil: Update: version=1 === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.200Z [INFO] TestAgent_Services_ExternalConnectProxy: Started HTTP server: address=127.0.0.1:29581 network=tcp === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:39.196Z [INFO] TestAgent_Services_Sidecar: Stopping server: protocol=DNS address=127.0.0.1:29592 network=tcp === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.201Z [INFO] TestAgent_Services_ExternalConnectProxy: started state syncer === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.200Z [DEBUG] TestAgent_ServicesFiltered.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:39.202Z [INFO] TestAgent_Services_Sidecar: Stopping server: protocol=DNS address=127.0.0.1:29592 network=udp writer.go:29: 2021-01-29T19:34:39.206Z [INFO] TestAgent_Services_Sidecar: Stopping server: protocol=HTTP address=127.0.0.1:29593 network=tcp === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.212Z [INFO] TestAgent_ServicesFiltered.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f24ccf9f-2f28-d13e-6571-4a85cc466510 Address:127.0.0.1:29603}]" writer.go:29: 2021-01-29T19:34:39.215Z [INFO] TestAgent_ServicesFiltered.server.raft: entering follower state: follower="Node at 127.0.0.1:29603 [Follower]" leader= writer.go:29: 2021-01-29T19:34:39.217Z [INFO] TestAgent_ServicesFiltered.server.serf.wan: serf: EventMemberJoin: Node-f24ccf9f-2f28-d13e-6571-4a85cc466510.dc1 127.0.0.1 === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.223Z [WARN] TestAgent_Services_ExternalConnectProxy.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:39.224Z [INFO] TestAgent_Services_ExternalConnectProxy.server.raft: entering candidate state: node="Node at 127.0.0.1:29585 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:39.227Z [DEBUG] TestAgent_Services_ExternalConnectProxy.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:39.229Z [DEBUG] TestAgent_Services_ExternalConnectProxy.server.raft: vote granted: from=b6a2890a-86b4-238c-8580-4213a124b89f term=2 tally=1 writer.go:29: 2021-01-29T19:34:39.232Z [INFO] TestAgent_Services_ExternalConnectProxy.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:39.236Z [INFO] TestAgent_Services_ExternalConnectProxy.server.raft: entering leader state: leader="Node at 127.0.0.1:29585 [Leader]" writer.go:29: 2021-01-29T19:34:39.239Z [INFO] TestAgent_Services_ExternalConnectProxy.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:39.240Z [INFO] TestAgent_Services_ExternalConnectProxy.server: New leader elected: payload=Node-b6a2890a-86b4-238c-8580-4213a124b89f writer.go:29: 2021-01-29T19:34:39.241Z [DEBUG] TestAgent_Services_ExternalConnectProxy.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29585 === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.229Z [INFO] TestAgent_ServicesFiltered.server.serf.lan: serf: EventMemberJoin: Node-f24ccf9f-2f28-d13e-6571-4a85cc466510 127.0.0.1 === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.253Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.256Z [INFO] TestAgent_ServicesFiltered.server: Handled event for server in area: event=member-join server=Node-f24ccf9f-2f28-d13e-6571-4a85cc466510.dc1 area=wan writer.go:29: 2021-01-29T19:34:39.256Z [INFO] TestAgent_ServicesFiltered.server: Adding LAN server: server="Node-f24ccf9f-2f28-d13e-6571-4a85cc466510 (Addr: tcp/127.0.0.1:29603) (DC: dc1)" writer.go:29: 2021-01-29T19:34:39.257Z [INFO] TestAgent_ServicesFiltered: Started DNS server: address=127.0.0.1:29598 network=udp writer.go:29: 2021-01-29T19:34:39.260Z [INFO] TestAgent_ServicesFiltered: Started DNS server: address=127.0.0.1:29598 network=tcp writer.go:29: 2021-01-29T19:34:39.259Z [WARN] TestAgent_ServicesFiltered.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.262Z [INFO] TestAgent_Services_ExternalConnectProxy: Synced node info === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.263Z [INFO] TestAgent_ServicesFiltered.server.raft: entering candidate state: node="Node at 127.0.0.1:29603 [Candidate]" term=2 === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.265Z [INFO] TestAgent_Services_ExternalConnectProxy.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.263Z [INFO] TestAgent_ServicesFiltered: Started HTTP server: address=127.0.0.1:29599 network=tcp === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.268Z [INFO] TestAgent_Services_ExternalConnectProxy.leader: started routine: routine="CA root pruning" === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.268Z [INFO] TestAgent_ServicesFiltered: started state syncer writer.go:29: 2021-01-29T19:34:39.267Z [DEBUG] TestAgent_ServicesFiltered.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:39.272Z [DEBUG] TestAgent_ServicesFiltered.server.raft: vote granted: from=f24ccf9f-2f28-d13e-6571-4a85cc466510 term=2 tally=1 writer.go:29: 2021-01-29T19:34:39.272Z [INFO] TestAgent_ServicesFiltered.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:39.273Z [INFO] TestAgent_ServicesFiltered.server.raft: entering leader state: leader="Node at 127.0.0.1:29603 [Leader]" writer.go:29: 2021-01-29T19:34:39.274Z [INFO] TestAgent_ServicesFiltered.server: cluster leadership acquired === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.270Z [DEBUG] TestAgent_Services_ExternalConnectProxy.server: Skipping self join check for node since the cluster is too small: node=Node-b6a2890a-86b4-238c-8580-4213a124b89f === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.278Z [INFO] TestAgent_ServicesFiltered.server: New leader elected: payload=Node-f24ccf9f-2f28-d13e-6571-4a85cc466510 === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.278Z [INFO] TestAgent_Services_ExternalConnectProxy.server: member joined, marking health alive: member=Node-b6a2890a-86b4-238c-8580-4213a124b89f === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.279Z [DEBUG] TestAgent_ServicesFiltered.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29603 writer.go:29: 2021-01-29T19:34:39.290Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:39.304Z [INFO] TestAgent_ServicesFiltered: Synced node info writer.go:29: 2021-01-29T19:34:39.310Z [INFO] TestAgent_ServicesFiltered.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:39.313Z [INFO] TestAgent_ServicesFiltered.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:39.317Z [DEBUG] TestAgent_ServicesFiltered.server: Skipping self join check for node since the cluster is too small: node=Node-f24ccf9f-2f28-d13e-6571-4a85cc466510 writer.go:29: 2021-01-29T19:34:39.320Z [INFO] TestAgent_ServicesFiltered.server: member joined, marking health alive: member=Node-f24ccf9f-2f28-d13e-6571-4a85cc466510 === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.428Z [DEBUG] TestAgent_Services_ExternalConnectProxy: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:39.429Z [DEBUG] TestAgent_Services_ExternalConnectProxy: Node info in sync writer.go:29: 2021-01-29T19:34:39.430Z [DEBUG] TestAgent_Services_ExternalConnectProxy: Node info in sync === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:39.477Z [ERROR] TestAgent_Leave_ACLDeny.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" writer.go:29: 2021-01-29T19:34:39.477Z [ERROR] TestAgent_Leave_ACLDeny.server.autopilot: Error promoting servers: error="error getting server raft protocol versions: No servers found" === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:39.611Z [INFO] TestAgent_Services_ExternalConnectProxy: Requesting shutdown writer.go:29: 2021-01-29T19:34:39.612Z [INFO] TestAgent_Services_ExternalConnectProxy.server: shutting down server writer.go:29: 2021-01-29T19:34:39.613Z [DEBUG] TestAgent_Services_ExternalConnectProxy.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:39.612Z [DEBUG] TestAgent_Services_ExternalConnectProxy: Node info in sync writer.go:29: 2021-01-29T19:34:39.614Z [WARN] TestAgent_Services_ExternalConnectProxy.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:39.614Z [DEBUG] TestAgent_Services_ExternalConnectProxy.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:39.617Z [INFO] TestAgent_Services_ExternalConnectProxy: Synced service: service=db-proxy writer.go:29: 2021-01-29T19:34:39.617Z [WARN] TestAgent_Services_ExternalConnectProxy.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:39.622Z [INFO] TestAgent_Services_ExternalConnectProxy.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:39.623Z [INFO] TestAgent_Services_ExternalConnectProxy: consul server down writer.go:29: 2021-01-29T19:34:39.624Z [INFO] TestAgent_Services_ExternalConnectProxy: shutdown complete writer.go:29: 2021-01-29T19:34:39.624Z [INFO] TestAgent_Services_ExternalConnectProxy: Stopping server: protocol=DNS address=127.0.0.1:29580 network=tcp writer.go:29: 2021-01-29T19:34:39.625Z [INFO] TestAgent_Services_ExternalConnectProxy: Stopping server: protocol=DNS address=127.0.0.1:29580 network=udp writer.go:29: 2021-01-29T19:34:39.628Z [INFO] TestAgent_Services_ExternalConnectProxy: Stopping server: protocol=HTTP address=127.0.0.1:29581 network=tcp writer.go:29: 2021-01-29T19:34:39.630Z [ERROR] TestAgent_Services_ExternalConnectProxy.proxycfg: watch error: id=intentions error="error filling agent cache: No cluster leader" writer.go:29: 2021-01-29T19:34:39.632Z [ERROR] TestAgent_Services_ExternalConnectProxy.proxycfg: watch error: id=intentions error="error filling agent cache: No cluster leader" writer.go:29: 2021-01-29T19:34:39.634Z [ERROR] TestAgent_Services_ExternalConnectProxy.proxycfg: watch error: id=intentions error="error filling agent cache: No cluster leader" writer.go:29: 2021-01-29T19:34:39.644Z [ERROR] TestAgent_Services_ExternalConnectProxy.proxycfg: watch error: id=intentions error="error filling agent cache: No cluster leader" writer.go:29: 2021-01-29T19:34:39.646Z [ERROR] TestAgent_Services_ExternalConnectProxy.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:34:39.672Z [ERROR] TestAgent_Services_ExternalConnectProxy.proxycfg: watch error: id=leaf error="error filling agent cache: internal error: CA provider is nil" writer.go:29: 2021-01-29T19:34:39.673Z [ERROR] TestAgent_Services_ExternalConnectProxy.proxycfg: watch error: id=leaf error="error filling agent cache: No cluster leader" writer.go:29: 2021-01-29T19:34:39.675Z [ERROR] TestAgent_Services_ExternalConnectProxy.proxycfg: watch error: id=leaf error="error filling agent cache: No cluster leader" writer.go:29: 2021-01-29T19:34:39.676Z [ERROR] TestAgent_Services_ExternalConnectProxy.proxycfg: watch error: id=leaf error="error filling agent cache: No cluster leader" === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.707Z [INFO] TestAgent_ServicesFiltered: Requesting shutdown === CONT TestAgent_Services_Sidecar writer.go:29: 2021-01-29T19:34:39.708Z [INFO] TestAgent_Services_Sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:39.710Z [INFO] TestAgent_Services_Sidecar: Endpoints down --- PASS: TestAgent_Services_Sidecar (1.09s) === CONT TestAgent_Services === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.710Z [INFO] TestAgent_ServicesFiltered.server: shutting down server writer.go:29: 2021-01-29T19:34:39.716Z [DEBUG] TestAgent_ServicesFiltered.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:39.717Z [WARN] TestAgent_ServicesFiltered.server.serf.lan: serf: Shutdown without a Leave === CONT TestAgent_Services writer.go:29: 2021-01-29T19:34:39.719Z [WARN] TestAgent_Services: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:39.720Z [DEBUG] TestAgent_Services.tlsutil: Update: version=1 === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.718Z [DEBUG] TestAgent_ServicesFiltered.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:39.722Z [WARN] TestAgent_ServicesFiltered.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_Services writer.go:29: 2021-01-29T19:34:39.721Z [DEBUG] TestAgent_Services.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.728Z [INFO] TestAgent_ServicesFiltered.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:39.728Z [INFO] TestAgent_ServicesFiltered: consul server down writer.go:29: 2021-01-29T19:34:39.732Z [INFO] TestAgent_ServicesFiltered: shutdown complete === CONT TestAgent_Services writer.go:29: 2021-01-29T19:34:39.733Z [INFO] TestAgent_Services.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:04cd9acd-79e3-2856-cb26-1e5768b52df3 Address:127.0.0.1:29615}]" === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.734Z [INFO] TestAgent_ServicesFiltered: Stopping server: protocol=DNS address=127.0.0.1:29598 network=tcp writer.go:29: 2021-01-29T19:34:39.736Z [INFO] TestAgent_ServicesFiltered: Stopping server: protocol=DNS address=127.0.0.1:29598 network=udp === CONT TestAgent_Services writer.go:29: 2021-01-29T19:34:39.735Z [INFO] TestAgent_Services.server.raft: entering follower state: follower="Node at 127.0.0.1:29615 [Follower]" leader= === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:39.737Z [INFO] TestAgent_ServicesFiltered: Stopping server: protocol=HTTP address=127.0.0.1:29599 network=tcp === CONT TestAgent_Services writer.go:29: 2021-01-29T19:34:39.737Z [INFO] TestAgent_Services.server.serf.wan: serf: EventMemberJoin: Node-04cd9acd-79e3-2856-cb26-1e5768b52df3.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:39.756Z [INFO] TestAgent_Services.server.serf.lan: serf: EventMemberJoin: Node-04cd9acd-79e3-2856-cb26-1e5768b52df3 127.0.0.1 writer.go:29: 2021-01-29T19:34:39.761Z [INFO] TestAgent_Services: Started DNS server: address=127.0.0.1:29610 network=udp writer.go:29: 2021-01-29T19:34:39.764Z [INFO] TestAgent_Services: Started DNS server: address=127.0.0.1:29610 network=tcp writer.go:29: 2021-01-29T19:34:39.762Z [INFO] TestAgent_Services.server: Adding LAN server: server="Node-04cd9acd-79e3-2856-cb26-1e5768b52df3 (Addr: tcp/127.0.0.1:29615) (DC: dc1)" writer.go:29: 2021-01-29T19:34:39.762Z [INFO] TestAgent_Services.server: Handled event for server in area: event=member-join server=Node-04cd9acd-79e3-2856-cb26-1e5768b52df3.dc1 area=wan writer.go:29: 2021-01-29T19:34:39.766Z [INFO] TestAgent_Services: Started HTTP server: address=127.0.0.1:29611 network=tcp writer.go:29: 2021-01-29T19:34:39.767Z [INFO] TestAgent_Services: started state syncer writer.go:29: 2021-01-29T19:34:39.795Z [WARN] TestAgent_Services.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:39.797Z [INFO] TestAgent_Services.server.raft: entering candidate state: node="Node at 127.0.0.1:29615 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:39.799Z [DEBUG] TestAgent_Services.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:39.800Z [DEBUG] TestAgent_Services.server.raft: vote granted: from=04cd9acd-79e3-2856-cb26-1e5768b52df3 term=2 tally=1 writer.go:29: 2021-01-29T19:34:39.802Z [INFO] TestAgent_Services.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:39.803Z [INFO] TestAgent_Services.server.raft: entering leader state: leader="Node at 127.0.0.1:29615 [Leader]" writer.go:29: 2021-01-29T19:34:39.805Z [INFO] TestAgent_Services.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:39.806Z [INFO] TestAgent_Services.server: New leader elected: payload=Node-04cd9acd-79e3-2856-cb26-1e5768b52df3 writer.go:29: 2021-01-29T19:34:39.807Z [DEBUG] TestAgent_Services.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29615 writer.go:29: 2021-01-29T19:34:39.811Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:39.824Z [INFO] TestAgent_Services.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:39.825Z [INFO] TestAgent_Services.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:39.827Z [DEBUG] TestAgent_Services.server: Skipping self join check for node since the cluster is too small: node=Node-04cd9acd-79e3-2856-cb26-1e5768b52df3 writer.go:29: 2021-01-29T19:34:39.829Z [INFO] TestAgent_Services.server: member joined, marking health alive: member=Node-04cd9acd-79e3-2856-cb26-1e5768b52df3 writer.go:29: 2021-01-29T19:34:39.840Z [INFO] TestAgent_Services: Requesting shutdown writer.go:29: 2021-01-29T19:34:39.841Z [INFO] TestAgent_Services.server: shutting down server writer.go:29: 2021-01-29T19:34:39.843Z [DEBUG] TestAgent_Services.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:39.845Z [WARN] TestAgent_Services.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:39.845Z [DEBUG] TestAgent_Services.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:39.843Z [ERROR] TestAgent_Services.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:39.848Z [WARN] TestAgent_Services.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:39.862Z [INFO] TestAgent_Services.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:39.862Z [INFO] TestAgent_Services: consul server down writer.go:29: 2021-01-29T19:34:39.863Z [INFO] TestAgent_Services: shutdown complete writer.go:29: 2021-01-29T19:34:39.864Z [INFO] TestAgent_Services: Stopping server: protocol=DNS address=127.0.0.1:29610 network=tcp writer.go:29: 2021-01-29T19:34:39.865Z [INFO] TestAgent_Services: Stopping server: protocol=DNS address=127.0.0.1:29610 network=udp writer.go:29: 2021-01-29T19:34:39.866Z [INFO] TestAgent_Services: Stopping server: protocol=HTTP address=127.0.0.1:29611 network=tcp === CONT TestAgent_Services_ExternalConnectProxy writer.go:29: 2021-01-29T19:34:40.129Z [INFO] TestAgent_Services_ExternalConnectProxy: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:40.131Z [INFO] TestAgent_Services_ExternalConnectProxy: Endpoints down --- PASS: TestAgent_Services_ExternalConnectProxy (1.02s) === CONT TestACL_filterChecks writer.go:29: 2021-01-29T19:34:40.164Z [WARN] TestACL_filterChecks: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.165Z [WARN] TestACL_filterChecks: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:40.167Z [DEBUG] TestACL_filterChecks: dropping check from result due to ACLs: check=my-other writer.go:29: 2021-01-29T19:34:40.168Z [DEBUG] TestACL_filterChecks: dropping check from result due to ACLs: check=my-node writer.go:29: 2021-01-29T19:34:40.171Z [DEBUG] TestACL_filterChecks: dropping check from result due to ACLs: check=my-service writer.go:29: 2021-01-29T19:34:40.172Z [DEBUG] TestACL_filterChecks: dropping check from result due to ACLs: check=my-other --- PASS: TestACL_filterChecks (0.04s) === CONT TestACL_filterServices writer.go:29: 2021-01-29T19:34:40.182Z [WARN] TestACL_filterServices: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.183Z [WARN] TestACL_filterServices: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:40.185Z [DEBUG] TestACL_filterServices: dropping service from result due to ACLs: service=my-other --- PASS: TestACL_filterServices (0.01s) === CONT TestACL_filterMembers writer.go:29: 2021-01-29T19:34:40.195Z [WARN] TestACL_filterMembers: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.195Z [WARN] TestACL_filterMembers: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:40.198Z [DEBUG] TestACL_filterMembers: dropping node from result due to ACLs: node=Nope accessorID=9df2d1a4-2d07-414e-8ead-6053f56ed2eb --- PASS: TestACL_filterMembers (0.01s) === CONT TestACL_vetCheckUpdate writer.go:29: 2021-01-29T19:34:40.220Z [WARN] TestACL_vetCheckUpdate: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.222Z [WARN] TestACL_vetCheckUpdate: bootstrap = true: do not enable unless necessary --- PASS: TestACL_vetCheckUpdate (0.03s) === CONT TestACL_vetCheckRegister === CONT TestAgent_ServicesFiltered writer.go:29: 2021-01-29T19:34:40.253Z [INFO] TestAgent_ServicesFiltered: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:40.254Z [INFO] TestAgent_ServicesFiltered: Endpoints down --- PASS: TestAgent_ServicesFiltered (1.09s) === CONT TestACL_Bootstrap === CONT TestACL_vetCheckRegister writer.go:29: 2021-01-29T19:34:40.257Z [WARN] TestACL_vetCheckRegister: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.258Z [WARN] TestACL_vetCheckRegister: bootstrap = true: do not enable unless necessary === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.264Z [WARN] TestACL_Bootstrap: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.265Z [WARN] TestACL_Bootstrap: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:40.266Z [DEBUG] TestACL_Bootstrap.tlsutil: Update: version=1 --- PASS: TestACL_vetCheckRegister (0.03s) === CONT TestACL_vetServiceRegister === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.268Z [DEBUG] TestACL_Bootstrap.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:40.284Z [INFO] TestACL_Bootstrap.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6606767c-cf41-d223-0e65-97b5c3e00a1c Address:127.0.0.1:29621}]" writer.go:29: 2021-01-29T19:34:40.286Z [INFO] TestACL_Bootstrap.server.raft: entering follower state: follower="Node at 127.0.0.1:29621 [Follower]" leader= writer.go:29: 2021-01-29T19:34:40.286Z [INFO] TestACL_Bootstrap.server.serf.wan: serf: EventMemberJoin: Node-6606767c-cf41-d223-0e65-97b5c3e00a1c.dc1 127.0.0.1 === CONT TestACL_vetServiceRegister writer.go:29: 2021-01-29T19:34:40.289Z [WARN] TestACL_vetServiceRegister: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.290Z [WARN] TestACL_vetServiceRegister: bootstrap = true: do not enable unless necessary === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.291Z [INFO] TestACL_Bootstrap.server.serf.lan: serf: EventMemberJoin: Node-6606767c-cf41-d223-0e65-97b5c3e00a1c 127.0.0.1 writer.go:29: 2021-01-29T19:34:40.295Z [INFO] TestACL_Bootstrap.server: Adding LAN server: server="Node-6606767c-cf41-d223-0e65-97b5c3e00a1c (Addr: tcp/127.0.0.1:29621) (DC: dc1)" --- PASS: TestACL_vetServiceRegister (0.03s) === CONT TestACL_RootAuthorizersDenied === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.296Z [INFO] TestACL_Bootstrap.server: Handled event for server in area: event=member-join server=Node-6606767c-cf41-d223-0e65-97b5c3e00a1c.dc1 area=wan writer.go:29: 2021-01-29T19:34:40.297Z [INFO] TestACL_Bootstrap: Started DNS server: address=127.0.0.1:29616 network=tcp writer.go:29: 2021-01-29T19:34:40.298Z [INFO] TestACL_Bootstrap: Started DNS server: address=127.0.0.1:29616 network=udp writer.go:29: 2021-01-29T19:34:40.300Z [INFO] TestACL_Bootstrap: Started HTTP server: address=127.0.0.1:29617 network=tcp writer.go:29: 2021-01-29T19:34:40.301Z [INFO] TestACL_Bootstrap: started state syncer === CONT TestACL_RootAuthorizersDenied writer.go:29: 2021-01-29T19:34:40.311Z [WARN] TestACL_RootAuthorizersDenied: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.313Z [WARN] TestACL_RootAuthorizersDenied: bootstrap = true: do not enable unless necessary --- PASS: TestACL_RootAuthorizersDenied (0.02s) === CONT TestACL_AgentMasterToken writer.go:29: 2021-01-29T19:34:40.345Z [WARN] TestACL_AgentMasterToken: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.346Z [WARN] TestACL_AgentMasterToken: bootstrap = true: do not enable unless necessary === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.346Z [DEBUG] TestACL_Bootstrap.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= --- PASS: TestACL_AgentMasterToken (0.03s) === CONT TestACL_Version8 === RUN TestACL_Version8/version_8_disabled === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.347Z [WARN] TestACL_Bootstrap.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:40.354Z [INFO] TestACL_Bootstrap.server.raft: entering candidate state: node="Node at 127.0.0.1:29621 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:40.357Z [DEBUG] TestACL_Bootstrap.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:40.358Z [DEBUG] TestACL_Bootstrap.server.raft: vote granted: from=6606767c-cf41-d223-0e65-97b5c3e00a1c term=2 tally=1 writer.go:29: 2021-01-29T19:34:40.360Z [INFO] TestACL_Bootstrap.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:40.362Z [INFO] TestACL_Bootstrap.server.raft: entering leader state: leader="Node at 127.0.0.1:29621 [Leader]" writer.go:29: 2021-01-29T19:34:40.364Z [INFO] TestACL_Bootstrap.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:40.365Z [INFO] TestACL_Bootstrap.server: New leader elected: payload=Node-6606767c-cf41-d223-0e65-97b5c3e00a1c writer.go:29: 2021-01-29T19:34:40.368Z [INFO] TestACL_Bootstrap.server: initializing acls === CONT TestAgent_Services writer.go:29: 2021-01-29T19:34:40.370Z [INFO] TestAgent_Services: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:40.371Z [INFO] TestAgent_Services: Endpoints down --- PASS: TestAgent_Services (0.66s) === CONT TestACL_Authorize === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.374Z [INFO] TestACL_Bootstrap.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:40.392Z [INFO] TestACL_Bootstrap.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:40.393Z [INFO] TestACL_Bootstrap.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:40.394Z [INFO] TestACL_Bootstrap.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:40.395Z [INFO] TestACL_Bootstrap.server.serf.lan: serf: EventMemberUpdate: Node-6606767c-cf41-d223-0e65-97b5c3e00a1c writer.go:29: 2021-01-29T19:34:40.396Z [INFO] TestACL_Bootstrap.server.serf.wan: serf: EventMemberUpdate: Node-6606767c-cf41-d223-0e65-97b5c3e00a1c.dc1 writer.go:29: 2021-01-29T19:34:40.401Z [INFO] TestACL_Bootstrap.server: Updating LAN server: server="Node-6606767c-cf41-d223-0e65-97b5c3e00a1c (Addr: tcp/127.0.0.1:29621) (DC: dc1)" writer.go:29: 2021-01-29T19:34:40.402Z [INFO] TestACL_Bootstrap.server: Handled event for server in area: event=member-update server=Node-6606767c-cf41-d223-0e65-97b5c3e00a1c.dc1 area=wan === CONT TestACL_Version8/version_8_disabled writer.go:29: 2021-01-29T19:34:40.408Z [WARN] TestACL_Version8/version_8_disabled: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.409Z [WARN] TestACL_Version8/version_8_disabled: bootstrap = true: do not enable unless necessary === RUN TestACL_Version8/version_8_enabled === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.432Z [ERROR] TestACL_Bootstrap.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.440Z [WARN] TestACL_Authorize: bootstrap = true: do not enable unless necessary === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.443Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.447Z [DEBUG] TestACL_Authorize.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:40.451Z [DEBUG] TestACL_Authorize.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.460Z [INFO] TestACL_Bootstrap.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:40.461Z [INFO] TestACL_Bootstrap.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:40.462Z [DEBUG] TestACL_Bootstrap.server: Skipping self join check for node since the cluster is too small: node=Node-6606767c-cf41-d223-0e65-97b5c3e00a1c writer.go:29: 2021-01-29T19:34:40.463Z [INFO] TestACL_Bootstrap.server: member joined, marking health alive: member=Node-6606767c-cf41-d223-0e65-97b5c3e00a1c === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.463Z [INFO] TestACL_Authorize.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:26208636-0e11-2314-2ab0-a69686a6a7c0 Address:127.0.0.1:29609}]" === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.465Z [DEBUG] TestACL_Bootstrap.server: Skipping self join check for node since the cluster is too small: node=Node-6606767c-cf41-d223-0e65-97b5c3e00a1c === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.467Z [INFO] TestACL_Authorize.server.raft: entering follower state: follower="Node at 127.0.0.1:29609 [Follower]" leader= writer.go:29: 2021-01-29T19:34:40.470Z [INFO] TestACL_Authorize.server.serf.wan: serf: EventMemberJoin: Node-26208636-0e11-2314-2ab0-a69686a6a7c0.dc1 127.0.0.1 === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.478Z [DEBUG] TestACL_Bootstrap.acl: dropping node from result due to ACLs: node=Node-6606767c-cf41-d223-0e65-97b5c3e00a1c === RUN TestACL_Bootstrap/bootstrap === CONT TestACL_Version8/version_8_enabled writer.go:29: 2021-01-29T19:34:40.488Z [WARN] TestACL_Version8/version_8_enabled: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.489Z [WARN] TestACL_Version8/version_8_enabled: bootstrap = true: do not enable unless necessary === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.487Z [WARN] TestACL_Bootstrap.server.acl: failed to remove bootstrap file: error="remove /tmp/TestACL_Bootstrap-agent007373082/acl-bootstrap-reset: no such file or directory" === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.491Z [INFO] TestACL_Authorize.server.serf.lan: serf: EventMemberJoin: Node-26208636-0e11-2314-2ab0-a69686a6a7c0 127.0.0.1 --- PASS: TestACL_Version8 (0.15s) --- PASS: TestACL_Version8/version_8_disabled (0.06s) --- PASS: TestACL_Version8/version_8_enabled (0.09s) === CONT TestACL_LoginProcedure_HTTP === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.506Z [INFO] TestACL_Bootstrap.server.acl: ACL bootstrap completed === RUN TestACL_Bootstrap/not_again === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.527Z [INFO] TestACL_Bootstrap: Requesting shutdown writer.go:29: 2021-01-29T19:34:40.528Z [INFO] TestACL_Bootstrap.server: shutting down server === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.512Z [INFO] TestACL_Authorize.server: Adding LAN server: server="Node-26208636-0e11-2314-2ab0-a69686a6a7c0 (Addr: tcp/127.0.0.1:29609) (DC: dc1)" === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.530Z [DEBUG] TestACL_Bootstrap.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.515Z [INFO] TestACL_Authorize.server: Handled event for server in area: event=member-join server=Node-26208636-0e11-2314-2ab0-a69686a6a7c0.dc1 area=wan writer.go:29: 2021-01-29T19:34:40.530Z [INFO] TestACL_Authorize: Started DNS server: address=127.0.0.1:29604 network=tcp writer.go:29: 2021-01-29T19:34:40.532Z [INFO] TestACL_Authorize: Started DNS server: address=127.0.0.1:29604 network=udp writer.go:29: 2021-01-29T19:34:40.541Z [WARN] TestACL_Authorize.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.531Z [DEBUG] TestACL_Bootstrap.leader: stopping routine: routine="acl token reaping" === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.544Z [INFO] TestACL_Authorize.server.raft: entering candidate state: node="Node at 127.0.0.1:29609 [Candidate]" term=2 === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.545Z [DEBUG] TestACL_Bootstrap.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:40.547Z [WARN] TestACL_Bootstrap.server.serf.lan: serf: Shutdown without a Leave === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.545Z [INFO] TestACL_Authorize: Started HTTP server: address=127.0.0.1:29605 network=tcp === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.541Z [DEBUG] TestACL_Bootstrap.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.555Z [INFO] TestACL_Authorize: started state syncer === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.556Z [DEBUG] TestACL_Bootstrap.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:40.558Z [DEBUG] TestACL_Bootstrap.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:40.562Z [WARN] TestACL_Bootstrap.server.serf.wan: serf: Shutdown without a Leave === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.597Z [WARN] TestACL_LoginProcedure_HTTP: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:40.599Z [WARN] TestACL_LoginProcedure_HTTP: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:40.602Z [DEBUG] TestACL_LoginProcedure_HTTP.tlsutil: Update: version=1 === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.602Z [INFO] TestACL_Bootstrap.server.router.manager: shutting down === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.606Z [DEBUG] TestACL_Authorize.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:40.621Z [INFO] TestACL_Bootstrap: consul server down writer.go:29: 2021-01-29T19:34:40.622Z [INFO] TestACL_Bootstrap: shutdown complete writer.go:29: 2021-01-29T19:34:40.622Z [INFO] TestACL_Bootstrap: Stopping server: protocol=DNS address=127.0.0.1:29616 network=tcp writer.go:29: 2021-01-29T19:34:40.623Z [INFO] TestACL_Bootstrap: Stopping server: protocol=DNS address=127.0.0.1:29616 network=udp writer.go:29: 2021-01-29T19:34:40.624Z [INFO] TestACL_Bootstrap: Stopping server: protocol=HTTP address=127.0.0.1:29617 network=tcp === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.624Z [DEBUG] TestACL_LoginProcedure_HTTP.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.632Z [DEBUG] TestACL_Authorize.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:40.645Z [DEBUG] TestACL_Authorize.server.raft: vote granted: from=26208636-0e11-2314-2ab0-a69686a6a7c0 term=2 tally=1 writer.go:29: 2021-01-29T19:34:40.651Z [INFO] TestACL_Authorize.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:40.652Z [INFO] TestACL_Authorize.server.raft: entering leader state: leader="Node at 127.0.0.1:29609 [Leader]" writer.go:29: 2021-01-29T19:34:40.654Z [INFO] TestACL_Authorize.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:40.656Z [INFO] TestACL_Authorize.server: New leader elected: payload=Node-26208636-0e11-2314-2ab0-a69686a6a7c0 writer.go:29: 2021-01-29T19:34:40.658Z [INFO] TestACL_Authorize.server: initializing acls === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.657Z [INFO] TestACL_LoginProcedure_HTTP.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:68b2e124-a5c8-b546-efb8-6383b0e17bcc Address:127.0.0.1:29627}]" writer.go:29: 2021-01-29T19:34:40.661Z [INFO] TestACL_LoginProcedure_HTTP.server.raft: entering follower state: follower="Node at 127.0.0.1:29627 [Follower]" leader= writer.go:29: 2021-01-29T19:34:40.665Z [INFO] TestACL_LoginProcedure_HTTP.server.serf.wan: serf: EventMemberJoin: Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc.dc1 127.0.0.1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.673Z [INFO] TestACL_Authorize.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:40.685Z [INFO] TestACL_Authorize.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:40.689Z [INFO] TestACL_Authorize.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:40.690Z [INFO] TestACL_Authorize.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:40.693Z [INFO] TestACL_Authorize.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:40.696Z [INFO] TestACL_Authorize.server.serf.lan: serf: EventMemberUpdate: Node-26208636-0e11-2314-2ab0-a69686a6a7c0 === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.685Z [INFO] TestACL_LoginProcedure_HTTP.server.serf.lan: serf: EventMemberJoin: Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc 127.0.0.1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.703Z [INFO] TestACL_Authorize.server.serf.wan: serf: EventMemberUpdate: Node-26208636-0e11-2314-2ab0-a69686a6a7c0.dc1 writer.go:29: 2021-01-29T19:34:40.712Z [INFO] TestACL_Authorize.server: Handled event for server in area: event=member-update server=Node-26208636-0e11-2314-2ab0-a69686a6a7c0.dc1 area=wan writer.go:29: 2021-01-29T19:34:40.725Z [INFO] TestACL_Authorize.server: Updating LAN server: server="Node-26208636-0e11-2314-2ab0-a69686a6a7c0 (Addr: tcp/127.0.0.1:29609) (DC: dc1)" === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.729Z [INFO] TestACL_LoginProcedure_HTTP: Started DNS server: address=127.0.0.1:29622 network=udp writer.go:29: 2021-01-29T19:34:40.744Z [INFO] TestACL_LoginProcedure_HTTP: Started DNS server: address=127.0.0.1:29622 network=tcp writer.go:29: 2021-01-29T19:34:40.753Z [INFO] TestACL_LoginProcedure_HTTP: Started HTTP server: address=127.0.0.1:29623 network=tcp === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:40.713Z [INFO] TestAgent_Leave_ACLDeny: Requesting shutdown === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.754Z [INFO] TestACL_LoginProcedure_HTTP: started state syncer writer.go:29: 2021-01-29T19:34:40.735Z [WARN] TestACL_LoginProcedure_HTTP.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:40.755Z [INFO] TestAgent_Leave_ACLDeny.server: shutting down server === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.758Z [INFO] TestACL_LoginProcedure_HTTP.server.raft: entering candidate state: node="Node at 127.0.0.1:29627 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:40.740Z [INFO] TestACL_LoginProcedure_HTTP.server: Adding LAN server: server="Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc (Addr: tcp/127.0.0.1:29627) (DC: dc1)" === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:40.759Z [DEBUG] TestAgent_Leave_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.741Z [INFO] TestACL_LoginProcedure_HTTP.server: Handled event for server in area: event=member-join server=Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc.dc1 area=wan === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:40.763Z [DEBUG] TestAgent_Leave_ACLDeny.leader: stopping routine: routine="acl token reaping" === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.765Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:40.766Z [DEBUG] TestAgent_Leave_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:40.777Z [DEBUG] TestAgent_Leave_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:40.766Z [DEBUG] TestAgent_Leave_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.782Z [DEBUG] TestACL_LoginProcedure_HTTP.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:40.767Z [DEBUG] TestAgent_Leave_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:40.781Z [INFO] TestAgent_Leave_ACLDeny: consul server down writer.go:29: 2021-01-29T19:34:40.786Z [INFO] TestAgent_Leave_ACLDeny: shutdown complete === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.787Z [DEBUG] TestACL_LoginProcedure_HTTP.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:40.789Z [DEBUG] TestACL_LoginProcedure_HTTP.server.raft: vote granted: from=68b2e124-a5c8-b546-efb8-6383b0e17bcc term=2 tally=1 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:40.789Z [INFO] TestAgent_Leave_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29430 network=tcp writer.go:29: 2021-01-29T19:34:40.795Z [INFO] TestAgent_Leave_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:29430 network=udp === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.791Z [INFO] TestACL_LoginProcedure_HTTP.server.raft: election won: tally=1 === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:40.798Z [INFO] TestAgent_Leave_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:29431 network=tcp === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.799Z [INFO] TestACL_LoginProcedure_HTTP.server.raft: entering leader state: leader="Node at 127.0.0.1:29627 [Leader]" === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.801Z [INFO] TestACL_Authorize.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:40.805Z [INFO] TestACL_Authorize.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:40.806Z [DEBUG] TestACL_Authorize.server: Skipping self join check for node since the cluster is too small: node=Node-26208636-0e11-2314-2ab0-a69686a6a7c0 writer.go:29: 2021-01-29T19:34:40.806Z [INFO] TestACL_Authorize.server: member joined, marking health alive: member=Node-26208636-0e11-2314-2ab0-a69686a6a7c0 === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.806Z [INFO] TestACL_LoginProcedure_HTTP.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:40.807Z [INFO] TestACL_LoginProcedure_HTTP.server: New leader elected: payload=Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.809Z [DEBUG] TestACL_Authorize.server: Skipping self join check for node since the cluster is too small: node=Node-26208636-0e11-2314-2ab0-a69686a6a7c0 === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.810Z [INFO] TestACL_LoginProcedure_HTTP.server: initializing acls writer.go:29: 2021-01-29T19:34:40.819Z [INFO] TestACL_LoginProcedure_HTTP.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:40.824Z [WARN] TestACL_LoginProcedure_HTTP.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:40.831Z [INFO] TestACL_LoginProcedure_HTTP.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:40.834Z [INFO] TestACL_LoginProcedure_HTTP.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:40.836Z [INFO] TestACL_LoginProcedure_HTTP.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:40.841Z [INFO] TestACL_LoginProcedure_HTTP.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:40.845Z [INFO] TestACL_LoginProcedure_HTTP.server.serf.lan: serf: EventMemberUpdate: Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc writer.go:29: 2021-01-29T19:34:40.850Z [INFO] TestACL_LoginProcedure_HTTP.server: Updating LAN server: server="Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc (Addr: tcp/127.0.0.1:29627) (DC: dc1)" writer.go:29: 2021-01-29T19:34:40.850Z [INFO] TestACL_LoginProcedure_HTTP.server.serf.wan: serf: EventMemberUpdate: Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc.dc1 writer.go:29: 2021-01-29T19:34:40.854Z [INFO] TestACL_LoginProcedure_HTTP.server: Handled event for server in area: event=member-update server=Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc.dc1 area=wan writer.go:29: 2021-01-29T19:34:40.856Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:40.865Z [INFO] TestACL_LoginProcedure_HTTP.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:40.867Z [INFO] TestACL_LoginProcedure_HTTP.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:40.868Z [DEBUG] TestACL_LoginProcedure_HTTP.server: Skipping self join check for node since the cluster is too small: node=Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc writer.go:29: 2021-01-29T19:34:40.870Z [INFO] TestACL_LoginProcedure_HTTP.server: member joined, marking health alive: member=Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc writer.go:29: 2021-01-29T19:34:40.884Z [DEBUG] TestACL_LoginProcedure_HTTP.server: Skipping self join check for node since the cluster is too small: node=Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc writer.go:29: 2021-01-29T19:34:40.895Z [DEBUG] TestACL_LoginProcedure_HTTP.acl: dropping node from result due to ACLs: node=Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc writer.go:29: 2021-01-29T19:34:40.899Z [DEBUG] TestACL_LoginProcedure_HTTP.acl: dropping node from result due to ACLs: node=Node-68b2e124-a5c8-b546-efb8-6383b0e17bcc === RUN TestACL_LoginProcedure_HTTP/AuthMethod === RUN TestACL_LoginProcedure_HTTP/AuthMethod/Create === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:40.910Z [DEBUG] TestACL_LoginProcedure_HTTP: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:40.914Z [INFO] TestACL_LoginProcedure_HTTP: Synced node info writer.go:29: 2021-01-29T19:34:40.915Z [DEBUG] TestACL_LoginProcedure_HTTP: Node info in sync === RUN TestACL_LoginProcedure_HTTP/AuthMethod/Create_other === RUN TestACL_LoginProcedure_HTTP/AuthMethod/Update_Name_URL_Mismatch === RUN TestACL_LoginProcedure_HTTP/AuthMethod/Update === RUN TestACL_LoginProcedure_HTTP/AuthMethod/Invalid_payload === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.939Z [DEBUG] TestACL_Authorize: Skipping remote check since it is managed automatically: check=serfHealth === RUN TestACL_LoginProcedure_HTTP/AuthMethod/List === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.943Z [INFO] TestACL_Authorize: Synced node info writer.go:29: 2021-01-29T19:34:40.945Z [DEBUG] TestACL_Authorize: Node info in sync === RUN TestACL_LoginProcedure_HTTP/AuthMethod/Delete === RUN TestACL_LoginProcedure_HTTP/AuthMethod/Read === RUN TestACL_LoginProcedure_HTTP/BindingRule === RUN TestACL_LoginProcedure_HTTP/BindingRule/Create === RUN TestACL_LoginProcedure_HTTP/BindingRule/Create_other === RUN TestACL_LoginProcedure_HTTP/BindingRule/BindingRule_CRUD_Missing_ID_in_URL === RUN TestACL_LoginProcedure_HTTP/BindingRule/Update === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:40.989Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-26208636-0e11-2314-2ab0-a69686a6a7c0 writer.go:29: 2021-01-29T19:34:41.013Z [WARN] TestACL_Authorize: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:41.017Z [DEBUG] TestACL_Authorize.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:41.021Z [DEBUG] TestACL_Authorize.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:41.029Z [INFO] TestACL_Authorize.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:90798172-bdb7-b836-0d15-84f51727be6a Address:127.0.0.1:29633}]" writer.go:29: 2021-01-29T19:34:41.031Z [INFO] TestACL_Authorize.server.serf.wan: serf: EventMemberJoin: Node-90798172-bdb7-b836-0d15-84f51727be6a.dc2 127.0.0.1 writer.go:29: 2021-01-29T19:34:41.032Z [INFO] TestACL_Authorize.server.raft: entering follower state: follower="Node at 127.0.0.1:29633 [Follower]" leader= writer.go:29: 2021-01-29T19:34:41.037Z [INFO] TestACL_Authorize.server.serf.lan: serf: EventMemberJoin: Node-90798172-bdb7-b836-0d15-84f51727be6a 127.0.0.1 writer.go:29: 2021-01-29T19:34:41.046Z [INFO] TestACL_Authorize.server: Handled event for server in area: event=member-join server=Node-90798172-bdb7-b836-0d15-84f51727be6a.dc2 area=wan === RUN TestACL_LoginProcedure_HTTP/BindingRule/ID_Supplied === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.047Z [INFO] TestACL_Authorize.server: Adding LAN server: server="Node-90798172-bdb7-b836-0d15-84f51727be6a (Addr: tcp/127.0.0.1:29633) (DC: dc2)" writer.go:29: 2021-01-29T19:34:41.048Z [INFO] TestACL_Authorize: Started DNS server: address=127.0.0.1:29628 network=tcp writer.go:29: 2021-01-29T19:34:41.052Z [INFO] TestACL_Authorize: Started DNS server: address=127.0.0.1:29628 network=udp === RUN TestACL_LoginProcedure_HTTP/BindingRule/Invalid_payload === RUN TestACL_LoginProcedure_HTTP/BindingRule/List === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.054Z [INFO] TestACL_Authorize: Started HTTP server: address=127.0.0.1:29629 network=tcp writer.go:29: 2021-01-29T19:34:41.066Z [INFO] TestACL_Authorize: started state syncer writer.go:29: 2021-01-29T19:34:41.083Z [WARN] TestACL_Authorize.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:41.086Z [INFO] TestACL_Authorize.server.raft: entering candidate state: node="Node at 127.0.0.1:29633 [Candidate]" term=2 === RUN TestACL_LoginProcedure_HTTP/BindingRule/Delete === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.091Z [DEBUG] TestACL_Authorize.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:41.093Z [DEBUG] TestACL_Authorize.server.raft: vote granted: from=90798172-bdb7-b836-0d15-84f51727be6a term=2 tally=1 writer.go:29: 2021-01-29T19:34:41.095Z [INFO] TestACL_Authorize.server.raft: election won: tally=1 === RUN TestACL_LoginProcedure_HTTP/BindingRule/Read === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.098Z [INFO] TestACL_Authorize.server.raft: entering leader state: leader="Node at 127.0.0.1:29633 [Leader]" === RUN TestACL_LoginProcedure_HTTP/Login === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.097Z [DEBUG] TestACL_Authorize.server: Cannot upgrade to new ACLs, servers in acl datacenter are not yet upgraded: ACLDatacenter=dc1 mode=1 found=false writer.go:29: 2021-01-29T19:34:41.103Z [INFO] TestACL_Authorize.server: cluster leadership acquired === RUN TestACL_LoginProcedure_HTTP/Login/Create_Token_1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.104Z [INFO] TestACL_Authorize.server: New leader elected: payload=Node-90798172-bdb7-b836-0d15-84f51727be6a writer.go:29: 2021-01-29T19:34:41.105Z [DEBUG] TestACL_Authorize.server: Cannot upgrade to new ACLs, servers in acl datacenter are not yet upgraded: ACLDatacenter=dc1 mode=1 found=false writer.go:29: 2021-01-29T19:34:41.107Z [INFO] TestACL_Authorize.leader: started routine: routine="legacy ACL replication" writer.go:29: 2021-01-29T19:34:41.108Z [INFO] TestACL_Authorize.server: started legacy ACL replication writer.go:29: 2021-01-29T19:34:41.109Z [WARN] TestACL_Authorize.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ACL.List writer.go:29: 2021-01-29T19:34:41.111Z [WARN] TestACL_Authorize.server.replication.acl.legacy: Legacy ACL replication error (will retry if still leader): error="failed to retrieve remote ACLs: No path to datacenter" writer.go:29: 2021-01-29T19:34:41.112Z [WARN] TestACL_Authorize.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ACL.List writer.go:29: 2021-01-29T19:34:41.114Z [WARN] TestACL_Authorize.server.replication.acl.legacy: Legacy ACL replication error (will retry if still leader): error="failed to retrieve remote ACLs: No path to datacenter" writer.go:29: 2021-01-29T19:34:41.112Z [WARN] TestACL_Authorize.server.connect: primary datacenter is configured but unreachable - deferring initialization of the secondary datacenter CA writer.go:29: 2021-01-29T19:34:41.117Z [INFO] TestACL_Authorize.leader: started routine: routine="config entry replication" writer.go:29: 2021-01-29T19:34:41.115Z [WARN] TestACL_Authorize.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ACL.List === CONT TestACL_Bootstrap writer.go:29: 2021-01-29T19:34:41.125Z [INFO] TestACL_Bootstrap: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:41.127Z [INFO] TestACL_Bootstrap: Endpoints down === CONT TestACL_HTTP === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.118Z [WARN] TestACL_Authorize.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ConfigEntry.ListAll writer.go:29: 2021-01-29T19:34:41.118Z [DEBUG] TestACL_Authorize.server.connect: starting Connect CA root replication from primary datacenter: primary=dc1 === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.139Z [WARN] TestACL_HTTP: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.118Z [INFO] TestACL_Authorize.leader: started routine: routine="secondary CA roots watch" === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.140Z [WARN] TestACL_HTTP: bootstrap = true: do not enable unless necessary --- PASS: TestACL_Bootstrap (0.87s) --- PASS: TestACL_Bootstrap/bootstrap (0.04s) --- PASS: TestACL_Bootstrap/not_again (0.00s) === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.141Z [DEBUG] TestACL_HTTP.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:41.143Z [DEBUG] TestACL_HTTP.tlsutil: OutgoingRPCWrapper: version=1 === RUN TestACL_LoginProcedure_HTTP/Login/Create_Token_2 === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.148Z [INFO] TestACL_HTTP.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4e621be6-7be5-57c8-9b9b-f42251ee60d6 Address:127.0.0.1:29651}]" writer.go:29: 2021-01-29T19:34:41.151Z [INFO] TestACL_HTTP.server.serf.wan: serf: EventMemberJoin: Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6.dc1 127.0.0.1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.121Z [WARN] TestACL_Authorize.server.replication.acl.legacy: Legacy ACL replication error (will retry if still leader): error="failed to retrieve remote ACLs: No path to datacenter" === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.154Z [INFO] TestACL_HTTP.server.serf.lan: serf: EventMemberJoin: Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6 127.0.0.1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.153Z [WARN] TestACL_Authorize.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ACL.List writer.go:29: 2021-01-29T19:34:41.155Z [WARN] TestACL_Authorize.server.replication.acl.legacy: Legacy ACL replication error (will retry if still leader): error="failed to retrieve remote ACLs: No path to datacenter" === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.152Z [INFO] TestACL_HTTP.server.raft: entering follower state: follower="Node at 127.0.0.1:29651 [Follower]" leader= writer.go:29: 2021-01-29T19:34:41.157Z [INFO] TestACL_HTTP.server: Handled event for server in area: event=member-join server=Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6.dc1 area=wan === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.157Z [WARN] TestACL_Authorize.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ACL.List === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.158Z [INFO] TestACL_HTTP.server: Adding LAN server: server="Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6 (Addr: tcp/127.0.0.1:29651) (DC: dc1)" writer.go:29: 2021-01-29T19:34:41.159Z [INFO] TestACL_HTTP: Started DNS server: address=127.0.0.1:29646 network=udp === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.159Z [WARN] TestACL_Authorize.server.replication.acl.legacy: Legacy ACL replication error (will retry if still leader): error="failed to retrieve remote ACLs: No path to datacenter" writer.go:29: 2021-01-29T19:34:41.142Z [WARN] TestACL_Authorize.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ConnectCA.Roots writer.go:29: 2021-01-29T19:34:41.166Z [ERROR] TestACL_Authorize.server.connect: CA root replication failed, will retry: routine="secondary CA roots watch" error="Error retrieving the primary datacenter's roots: No path to datacenter" writer.go:29: 2021-01-29T19:34:41.142Z [INFO] TestACL_Authorize.leader: started routine: routine="intention replication" === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.161Z [INFO] TestACL_HTTP: Started DNS server: address=127.0.0.1:29646 network=tcp === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.142Z [DEBUG] TestACL_Authorize.server.connect: starting Connect intention replication from primary datacenter: primary=dc1 === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.178Z [INFO] TestACL_HTTP: Started HTTP server: address=127.0.0.1:29647 network=tcp writer.go:29: 2021-01-29T19:34:41.179Z [INFO] TestACL_HTTP: started state syncer === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.177Z [WARN] TestACL_Authorize.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=Intention.List writer.go:29: 2021-01-29T19:34:41.173Z [INFO] TestACL_Authorize.leader: started routine: routine="secondary cert renew watch" writer.go:29: 2021-01-29T19:34:41.184Z [ERROR] TestACL_Authorize.server.connect: error replicating intentions: routine="intention replication" error="No path to datacenter" writer.go:29: 2021-01-29T19:34:41.188Z [INFO] TestACL_Authorize.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:41.197Z [DEBUG] TestACL_Authorize.server: Cannot upgrade to new ACLs, servers in acl datacenter are not yet upgraded: ACLDatacenter=dc1 mode=1 found=false writer.go:29: 2021-01-29T19:34:41.198Z [DEBUG] TestACL_Authorize.server: Skipping self join check for node since the cluster is too small: node=Node-90798172-bdb7-b836-0d15-84f51727be6a writer.go:29: 2021-01-29T19:34:41.201Z [INFO] TestACL_Authorize.server: member joined, marking health alive: member=Node-90798172-bdb7-b836-0d15-84f51727be6a === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.210Z [DEBUG] TestACL_HTTP.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:41.220Z [WARN] TestACL_HTTP.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:41.221Z [INFO] TestACL_HTTP.server.raft: entering candidate state: node="Node at 127.0.0.1:29651 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:41.224Z [DEBUG] TestACL_HTTP.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:41.225Z [DEBUG] TestACL_HTTP.server.raft: vote granted: from=4e621be6-7be5-57c8-9b9b-f42251ee60d6 term=2 tally=1 writer.go:29: 2021-01-29T19:34:41.225Z [INFO] TestACL_HTTP.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:41.226Z [INFO] TestACL_HTTP.server.raft: entering leader state: leader="Node at 127.0.0.1:29651 [Leader]" writer.go:29: 2021-01-29T19:34:41.227Z [INFO] TestACL_HTTP.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:41.228Z [INFO] TestACL_HTTP.server: New leader elected: payload=Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6 === RUN TestACL_LoginProcedure_HTTP/Login/List_Tokens_by_(incorrect)_Method === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.229Z [INFO] TestACL_HTTP.server: initializing acls === RUN TestACL_LoginProcedure_HTTP/Login/List_Tokens_by_(correct)_Method === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.233Z [INFO] TestACL_HTTP.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:41.234Z [WARN] TestACL_HTTP.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:41.237Z [INFO] TestACL_HTTP.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:41.240Z [INFO] TestACL_HTTP.server: Created ACL anonymous token from configuration === RUN TestACL_LoginProcedure_HTTP/Login/Logout === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.243Z [INFO] TestACL_HTTP.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:41.245Z [INFO] TestACL_HTTP.leader: started routine: routine="acl token reaping" === RUN TestACL_LoginProcedure_HTTP/Login/Token_is_gone_after_Logout === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.247Z [INFO] TestACL_HTTP.server.serf.lan: serf: EventMemberUpdate: Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6 === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:41.251Z [INFO] TestACL_LoginProcedure_HTTP: Requesting shutdown === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.250Z [INFO] TestACL_HTTP.server.serf.wan: serf: EventMemberUpdate: Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6.dc1 === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:41.252Z [INFO] TestACL_LoginProcedure_HTTP.server: shutting down server === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.250Z [INFO] TestACL_HTTP.server: Updating LAN server: server="Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6 (Addr: tcp/127.0.0.1:29651) (DC: dc1)" === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:41.253Z [DEBUG] TestACL_LoginProcedure_HTTP.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:41.255Z [DEBUG] TestACL_LoginProcedure_HTTP.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.253Z [INFO] TestACL_HTTP.server: Handled event for server in area: event=member-update server=Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6.dc1 area=wan === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:41.255Z [DEBUG] TestACL_LoginProcedure_HTTP.leader: stopping routine: routine="acl token reaping" === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.255Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:41.256Z [WARN] TestACL_LoginProcedure_HTTP.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:41.255Z [DEBUG] TestACL_LoginProcedure_HTTP.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:41.255Z [DEBUG] TestACL_LoginProcedure_HTTP.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:41.256Z [DEBUG] TestACL_LoginProcedure_HTTP.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:41.258Z [WARN] TestACL_LoginProcedure_HTTP.server.serf.wan: serf: Shutdown without a Leave === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.262Z [INFO] TestACL_HTTP.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:41.263Z [INFO] TestACL_LoginProcedure_HTTP.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:41.263Z [INFO] TestACL_LoginProcedure_HTTP: consul server down writer.go:29: 2021-01-29T19:34:41.265Z [INFO] TestACL_LoginProcedure_HTTP: shutdown complete === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.265Z [INFO] TestACL_HTTP.leader: started routine: routine="CA root pruning" === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:41.265Z [INFO] TestACL_LoginProcedure_HTTP: Stopping server: protocol=DNS address=127.0.0.1:29622 network=tcp writer.go:29: 2021-01-29T19:34:41.266Z [INFO] TestACL_LoginProcedure_HTTP: Stopping server: protocol=DNS address=127.0.0.1:29622 network=udp === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.267Z [DEBUG] TestACL_HTTP.server: Skipping self join check for node since the cluster is too small: node=Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6 writer.go:29: 2021-01-29T19:34:41.268Z [INFO] TestACL_HTTP.server: member joined, marking health alive: member=Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6 === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:41.268Z [INFO] TestACL_LoginProcedure_HTTP: Stopping server: protocol=HTTP address=127.0.0.1:29623 network=tcp === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.271Z [DEBUG] TestACL_HTTP.server: Skipping self join check for node since the cluster is too small: node=Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6 writer.go:29: 2021-01-29T19:34:41.293Z [DEBUG] TestACL_HTTP.acl: dropping node from result due to ACLs: node=Node-4e621be6-7be5-57c8-9b9b-f42251ee60d6 === RUN TestACL_HTTP/Policy === RUN TestACL_HTTP/Policy/Create === CONT TestAgent_Leave_ACLDeny writer.go:29: 2021-01-29T19:34:41.299Z [INFO] TestAgent_Leave_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:41.301Z [INFO] TestAgent_Leave_ACLDeny: Endpoints down --- PASS: TestAgent_Leave_ACLDeny (11.99s) --- PASS: TestAgent_Leave_ACLDeny/no_token (0.00s) --- PASS: TestAgent_Leave_ACLDeny/read-only_token (0.00s) --- PASS: TestAgent_Leave_ACLDeny/agent_master_token (11.09s) === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.311Z [WARN] TestACL_Legacy_Clone: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:41.311Z [WARN] TestACL_Legacy_Clone: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:41.317Z [DEBUG] TestACL_Legacy_Clone.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:41.321Z [DEBUG] TestACL_Legacy_Clone.tlsutil: OutgoingRPCWrapper: version=1 === RUN TestACL_HTTP/Policy/Minimal === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.329Z [INFO] TestACL_Legacy_Clone.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 Address:127.0.0.1:29645}]" writer.go:29: 2021-01-29T19:34:41.330Z [INFO] TestACL_Legacy_Clone.server.raft: entering follower state: follower="Node at 127.0.0.1:29645 [Follower]" leader= writer.go:29: 2021-01-29T19:34:41.331Z [INFO] TestACL_Legacy_Clone.server.serf.wan: serf: EventMemberJoin: Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:41.336Z [INFO] TestACL_Legacy_Clone.server.serf.lan: serf: EventMemberJoin: Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 127.0.0.1 writer.go:29: 2021-01-29T19:34:41.338Z [INFO] TestACL_Legacy_Clone.server: Adding LAN server: server="Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 (Addr: tcp/127.0.0.1:29645) (DC: dc1)" writer.go:29: 2021-01-29T19:34:41.339Z [INFO] TestACL_Legacy_Clone: Started DNS server: address=127.0.0.1:29640 network=udp writer.go:29: 2021-01-29T19:34:41.339Z [INFO] TestACL_Legacy_Clone.server: Handled event for server in area: event=member-join server=Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6.dc1 area=wan writer.go:29: 2021-01-29T19:34:41.340Z [INFO] TestACL_Legacy_Clone: Started DNS server: address=127.0.0.1:29640 network=tcp writer.go:29: 2021-01-29T19:34:41.342Z [INFO] TestACL_Legacy_Clone: Started HTTP server: address=127.0.0.1:29641 network=tcp writer.go:29: 2021-01-29T19:34:41.343Z [INFO] TestACL_Legacy_Clone: started state syncer === RUN TestACL_HTTP/Policy/Name_Chars === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.352Z [WARN] TestACL_Authorize.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ACL.GetPolicy === RUN TestACL_HTTP/Policy/Update_Name_ID_Mismatch === RUN TestACL_HTTP/Policy/Policy_CRUD_Missing_ID_in_URL === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.356Z [DEBUG] TestACL_Authorize.acl: dropping check from result due to ACLs: check=serfHealth === RUN TestACL_HTTP/Policy/Update === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.360Z [WARN] TestACL_Authorize: Node info update blocked by ACLs: node=90798172-bdb7-b836-0d15-84f51727be6a accessorID= writer.go:29: 2021-01-29T19:34:41.363Z [DEBUG] TestACL_Authorize: Node info in sync === RUN TestACL_HTTP/Policy/ID_Supplied === RUN TestACL_HTTP/Policy/Invalid_payload === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.391Z [DEBUG] TestACL_HTTP: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.390Z [DEBUG] TestACL_Legacy_Clone.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.393Z [INFO] TestACL_HTTP: Synced node info === RUN TestACL_HTTP/Policy/Delete === RUN TestACL_HTTP/Policy/List === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.397Z [DEBUG] TestACL_Authorize.server: Cannot upgrade to new ACLs, servers in acl datacenter are not yet upgraded: ACLDatacenter=dc1 mode=1 found=false === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.405Z [WARN] TestACL_Legacy_Clone.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:41.409Z [INFO] TestACL_Legacy_Clone.server.raft: entering candidate state: node="Node at 127.0.0.1:29645 [Candidate]" term=2 === RUN TestACL_HTTP/Policy/Read === RUN TestACL_HTTP/Policy/Read_Name === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.425Z [DEBUG] TestACL_Legacy_Clone.server.raft: votes: needed=1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.430Z [WARN] TestACL_Authorize.server.rpc: RPC request for DC is currently failing as no path was found: datacenter=dc1 method=ACL.GetPolicy === RUN TestACL_HTTP/Role === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.432Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a writer.go:29: 2021-01-29T19:34:41.433Z [INFO] TestACL_Authorize: (WAN) joining: wan_addresses=[127.0.0.1:29608] === RUN TestACL_HTTP/Role/Create === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.434Z [DEBUG] TestACL_Authorize.server.memberlist.wan: memberlist: Initiating push/pull sync with: 127.0.0.1:29608 === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.429Z [DEBUG] TestACL_Legacy_Clone.server.raft: vote granted: from=48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 term=2 tally=1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.436Z [DEBUG] TestACL_Authorize.server.memberlist.wan: memberlist: Stream connection from=127.0.0.1:59498 writer.go:29: 2021-01-29T19:34:41.438Z [INFO] TestACL_Authorize.server.serf.wan: serf: EventMemberJoin: Node-90798172-bdb7-b836-0d15-84f51727be6a.dc2 127.0.0.1 writer.go:29: 2021-01-29T19:34:41.439Z [INFO] TestACL_Authorize.server: Handled event for server in area: event=member-join server=Node-90798172-bdb7-b836-0d15-84f51727be6a.dc2 area=wan writer.go:29: 2021-01-29T19:34:41.440Z [INFO] TestACL_Authorize.server.serf.wan: serf: EventMemberJoin: Node-26208636-0e11-2314-2ab0-a69686a6a7c0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:41.442Z [INFO] TestACL_Authorize.server: Handled event for server in area: event=member-join server=Node-26208636-0e11-2314-2ab0-a69686a6a7c0.dc1 area=wan === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.437Z [INFO] TestACL_Legacy_Clone.server.raft: election won: tally=1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.445Z [INFO] TestACL_Authorize: (WAN) joined: number_of_nodes=1 === RUN TestACL_HTTP/Role/Name_Chars === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.451Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.443Z [INFO] TestACL_Legacy_Clone.server.raft: entering leader state: leader="Node at 127.0.0.1:29645 [Leader]" writer.go:29: 2021-01-29T19:34:41.456Z [INFO] TestACL_Legacy_Clone.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:41.457Z [INFO] TestACL_Legacy_Clone.server: New leader elected: payload=Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 writer.go:29: 2021-01-29T19:34:41.458Z [INFO] TestACL_Legacy_Clone.server: initializing acls writer.go:29: 2021-01-29T19:34:41.467Z [INFO] TestACL_Legacy_Clone.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:41.470Z [WARN] TestACL_Legacy_Clone.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:41.474Z [INFO] TestACL_Legacy_Clone.server: Bootstrapped ACL master token from configuration === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.475Z [DEBUG] TestACL_Authorize: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:41.476Z [DEBUG] TestACL_Authorize: Node info in sync === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.478Z [INFO] TestACL_Legacy_Clone.server: Created ACL anonymous token from configuration === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.479Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.480Z [INFO] TestACL_Legacy_Clone.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:41.482Z [INFO] TestACL_Legacy_Clone.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:41.484Z [INFO] TestACL_Legacy_Clone.server.serf.lan: serf: EventMemberUpdate: Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 writer.go:29: 2021-01-29T19:34:41.486Z [INFO] TestACL_Legacy_Clone.server.serf.wan: serf: EventMemberUpdate: Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6.dc1 writer.go:29: 2021-01-29T19:34:41.486Z [INFO] TestACL_Legacy_Clone.server: Updating LAN server: server="Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 (Addr: tcp/127.0.0.1:29645) (DC: dc1)" writer.go:29: 2021-01-29T19:34:41.490Z [INFO] TestACL_Legacy_Clone.server: Handled event for server in area: event=member-update server=Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6.dc1 area=wan === RUN TestACL_HTTP/Role/Update_Name_ID_Mismatch === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.492Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === RUN TestACL_HTTP/Role/Role_CRUD_Missing_ID_in_URL === RUN TestACL_HTTP/Role/Update === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.508Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.509Z [INFO] TestACL_Legacy_Clone.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:41.511Z [INFO] TestACL_Legacy_Clone.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:41.512Z [DEBUG] TestACL_Legacy_Clone.server: Skipping self join check for node since the cluster is too small: node=Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 writer.go:29: 2021-01-29T19:34:41.514Z [INFO] TestACL_Legacy_Clone.server: member joined, marking health alive: member=Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 writer.go:29: 2021-01-29T19:34:41.517Z [DEBUG] TestACL_Legacy_Clone.server: Skipping self join check for node since the cluster is too small: node=Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 === RUN TestACL_HTTP/Role/ID_Supplied === RUN TestACL_HTTP/Role/Invalid_payload === RUN TestACL_HTTP/Role/Delete === RUN TestACL_HTTP/Role/List === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.537Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === RUN TestACL_HTTP/Role/Read === RUN TestACL_HTTP/Token === RUN TestACL_HTTP/Token/Create === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.570Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === RUN TestACL_HTTP/Token/Create_Local === RUN TestACL_HTTP/Token/Read === RUN TestACL_HTTP/Token/Self === RUN TestACL_HTTP/Token/Clone === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.603Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === RUN TestACL_HTTP/Token/Update === RUN TestACL_HTTP/Token/CRUD_Missing_Token_Accessor_ID === RUN TestACL_HTTP/Token/Update_Accessor_Mismatch === RUN TestACL_HTTP/Token/Delete === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.634Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === RUN TestACL_HTTP/Token/List === RUN TestACL_HTTP/Token/List_by_Policy === RUN TestACL_HTTP/Token/Create_with_Accessor === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.666Z [DEBUG] TestACL_Legacy_Clone: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:41.669Z [INFO] TestACL_Legacy_Clone: Synced node info writer.go:29: 2021-01-29T19:34:41.671Z [DEBUG] TestACL_Legacy_Clone: Node info in sync === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.673Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === RUN TestACL_HTTP/Token/Create_with_Secret === RUN TestACL_HTTP/Token/Create_with_Accessor_and_Secret === RUN TestACL_HTTP/Token/Create_with_Accessor_Dup === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.703Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === RUN TestACL_HTTP/Token/Create_with_Secret_as_Accessor_Dup === RUN TestACL_HTTP/Token/Create_with_Secret_Dup === RUN TestACL_HTTP/Token/Create_with_Accessor_as_Secret_Dup === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.720Z [DEBUG] TestACL_Authorize.acl: dropping check from result due to ACLs: check=serfHealth writer.go:29: 2021-01-29T19:34:41.723Z [WARN] TestACL_Authorize: Node info update blocked by ACLs: node=90798172-bdb7-b836-0d15-84f51727be6a accessorID= === RUN TestACL_HTTP/Token/Create_with_Reserved_Accessor === RUN TestACL_HTTP/Token/Create_with_Reserved_Secret === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.734Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.734Z [DEBUG] TestACL_Legacy_Clone.acl: dropping node from result due to ACLs: node=Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 writer.go:29: 2021-01-29T19:34:41.739Z [DEBUG] TestACL_Legacy_Clone.acl: dropping node from result due to ACLs: node=Node-48bdad5d-9b1d-9abf-3d8e-e5c0739512b6 === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.757Z [INFO] TestACL_HTTP: Requesting shutdown writer.go:29: 2021-01-29T19:34:41.759Z [INFO] TestACL_HTTP.server: shutting down server writer.go:29: 2021-01-29T19:34:41.761Z [DEBUG] TestACL_HTTP.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:41.763Z [DEBUG] TestACL_HTTP.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.764Z [INFO] TestACL_Legacy_Clone: Requesting shutdown === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.764Z [DEBUG] TestACL_HTTP.leader: stopping routine: routine="acl token reaping" === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.765Z [INFO] TestACL_Legacy_Clone.server: shutting down server === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.764Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.766Z [WARN] TestACL_HTTP.server.serf.lan: serf: Shutdown without a Leave === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.766Z [DEBUG] TestACL_Legacy_Clone.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.767Z [DEBUG] TestACL_HTTP.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.768Z [DEBUG] TestACL_Legacy_Clone.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:41.770Z [DEBUG] TestACL_Legacy_Clone.leader: stopping routine: routine="CA root pruning" === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.767Z [DEBUG] TestACL_HTTP.leader: stopped routine: routine="CA root pruning" === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:41.770Z [INFO] TestACL_LoginProcedure_HTTP: Waiting for endpoints to shut down === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.768Z [DEBUG] TestACL_Legacy_Clone.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.767Z [DEBUG] TestACL_HTTP.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:41.770Z [WARN] TestACL_HTTP.server.serf.wan: serf: Shutdown without a Leave === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.770Z [DEBUG] TestACL_Legacy_Clone.leader: stopped routine: routine="acl token reaping" === CONT TestACL_LoginProcedure_HTTP writer.go:29: 2021-01-29T19:34:41.772Z [INFO] TestACL_LoginProcedure_HTTP: Endpoints down === CONT TestACL_Disabled_Response === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.777Z [INFO] TestACL_HTTP.server.router.manager: shutting down --- PASS: TestACL_LoginProcedure_HTTP (1.27s) --- PASS: TestACL_LoginProcedure_HTTP/AuthMethod (0.06s) --- PASS: TestACL_LoginProcedure_HTTP/AuthMethod/Create (0.02s) --- PASS: TestACL_LoginProcedure_HTTP/AuthMethod/Create_other (0.01s) --- PASS: TestACL_LoginProcedure_HTTP/AuthMethod/Update_Name_URL_Mismatch (0.00s) --- PASS: TestACL_LoginProcedure_HTTP/AuthMethod/Update (0.01s) --- PASS: TestACL_LoginProcedure_HTTP/AuthMethod/Invalid_payload (0.00s) --- PASS: TestACL_LoginProcedure_HTTP/AuthMethod/List (0.01s) --- PASS: TestACL_LoginProcedure_HTTP/AuthMethod/Delete (0.00s) --- PASS: TestACL_LoginProcedure_HTTP/AuthMethod/Read (0.00s) --- PASS: TestACL_LoginProcedure_HTTP/BindingRule (0.14s) --- PASS: TestACL_LoginProcedure_HTTP/BindingRule/Create (0.01s) --- PASS: TestACL_LoginProcedure_HTTP/BindingRule/Create_other (0.01s) --- PASS: TestACL_LoginProcedure_HTTP/BindingRule/BindingRule_CRUD_Missing_ID_in_URL (0.00s) --- PASS: TestACL_LoginProcedure_HTTP/BindingRule/Update (0.06s) --- PASS: TestACL_LoginProcedure_HTTP/BindingRule/ID_Supplied (0.01s) --- PASS: TestACL_LoginProcedure_HTTP/BindingRule/Invalid_payload (0.00s) --- PASS: TestACL_LoginProcedure_HTTP/BindingRule/List (0.03s) --- PASS: TestACL_LoginProcedure_HTTP/BindingRule/Delete (0.01s) --- PASS: TestACL_LoginProcedure_HTTP/BindingRule/Read (0.00s) --- PASS: TestACL_LoginProcedure_HTTP/Login (0.15s) --- PASS: TestACL_LoginProcedure_HTTP/Login/Create_Token_1 (0.04s) --- PASS: TestACL_LoginProcedure_HTTP/Login/Create_Token_2 (0.08s) --- PASS: TestACL_LoginProcedure_HTTP/Login/List_Tokens_by_(incorrect)_Method (0.00s) --- PASS: TestACL_LoginProcedure_HTTP/Login/List_Tokens_by_(correct)_Method (0.01s) --- PASS: TestACL_LoginProcedure_HTTP/Login/Logout (0.00s) --- PASS: TestACL_LoginProcedure_HTTP/Login/Token_is_gone_after_Logout (0.00s) === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.771Z [WARN] TestACL_Legacy_Clone.server.serf.lan: serf: Shutdown without a Leave === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.779Z [INFO] TestACL_HTTP: consul server down === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.772Z [DEBUG] TestACL_Legacy_Clone.leader: stopped routine: routine="CA root pruning" === CONT TestACL_Disabled_Response writer.go:29: 2021-01-29T19:34:41.784Z [WARN] TestACL_Disabled_Response: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:41.785Z [DEBUG] TestACL_Disabled_Response.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:41.787Z [DEBUG] TestACL_Disabled_Response.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.785Z [WARN] TestACL_Legacy_Clone.server.serf.wan: serf: Shutdown without a Leave === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.782Z [INFO] TestACL_HTTP: shutdown complete === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.794Z [DEBUG] TestACL_Authorize.acl: dropping node from result due to ACLs: node=Node-90798172-bdb7-b836-0d15-84f51727be6a === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.790Z [INFO] TestACL_HTTP: Stopping server: protocol=DNS address=127.0.0.1:29646 network=tcp === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.797Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="legacy ACL replication" writer.go:29: 2021-01-29T19:34:41.798Z [INFO] TestACL_Authorize.leader: started routine: routine="ACL policy replication" writer.go:29: 2021-01-29T19:34:41.799Z [INFO] TestACL_Authorize.leader: started routine: routine="ACL role replication" writer.go:29: 2021-01-29T19:34:41.800Z [INFO] TestACL_Authorize.leader: started routine: routine="ACL token replication" writer.go:29: 2021-01-29T19:34:41.801Z [INFO] TestACL_Authorize.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:41.801Z [DEBUG] TestACL_Authorize.server: transitioning out of legacy ACL mode === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.797Z [INFO] TestACL_HTTP: Stopping server: protocol=DNS address=127.0.0.1:29646 network=udp === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.802Z [INFO] TestACL_Authorize.server.serf.lan: serf: EventMemberUpdate: Node-90798172-bdb7-b836-0d15-84f51727be6a writer.go:29: 2021-01-29T19:34:41.801Z [INFO] TestACL_Authorize.server.replication.acl.policy: started ACL Policy replication writer.go:29: 2021-01-29T19:34:41.801Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="legacy ACL replication" === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.804Z [INFO] TestACL_Legacy_Clone.server.router.manager: shutting down === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.801Z [INFO] TestACL_Authorize.server.replication.acl.role: started ACL Role replication === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.807Z [INFO] TestACL_Legacy_Clone: consul server down === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.801Z [INFO] TestACL_Authorize.server.replication.acl.token: started ACL Token replication === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:41.809Z [INFO] TestACL_Legacy_Clone: shutdown complete writer.go:29: 2021-01-29T19:34:41.810Z [INFO] TestACL_Legacy_Clone: Stopping server: protocol=DNS address=127.0.0.1:29640 network=tcp writer.go:29: 2021-01-29T19:34:41.812Z [INFO] TestACL_Legacy_Clone: Stopping server: protocol=DNS address=127.0.0.1:29640 network=udp writer.go:29: 2021-01-29T19:34:41.813Z [INFO] TestACL_Legacy_Clone: Stopping server: protocol=HTTP address=127.0.0.1:29641 network=tcp === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.807Z [DEBUG] TestACL_Authorize.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:41.807Z [INFO] TestACL_Authorize.server.serf.wan: serf: EventMemberUpdate: Node-90798172-bdb7-b836-0d15-84f51727be6a.dc2 === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:41.803Z [INFO] TestACL_HTTP: Stopping server: protocol=HTTP address=127.0.0.1:29647 network=tcp === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.807Z [INFO] TestACL_Authorize.server: Updating LAN server: server="Node-90798172-bdb7-b836-0d15-84f51727be6a (Addr: tcp/127.0.0.1:29633) (DC: dc2)" writer.go:29: 2021-01-29T19:34:41.821Z [DEBUG] TestACL_Authorize.server: Skipping self join check for node since the cluster is too small: node=Node-90798172-bdb7-b836-0d15-84f51727be6a writer.go:29: 2021-01-29T19:34:41.823Z [INFO] TestACL_Authorize.server: Handled event for server in area: event=member-update server=Node-90798172-bdb7-b836-0d15-84f51727be6a.dc2 area=wan === CONT TestACL_Disabled_Response writer.go:29: 2021-01-29T19:34:41.834Z [INFO] TestACL_Disabled_Response.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:063daf4e-db99-bc14-ae90-7d1c8eb1cd70 Address:127.0.0.1:29639}]" writer.go:29: 2021-01-29T19:34:41.837Z [INFO] TestACL_Disabled_Response.server.raft: entering follower state: follower="Node at 127.0.0.1:29639 [Follower]" leader= writer.go:29: 2021-01-29T19:34:41.837Z [INFO] TestACL_Disabled_Response.server.serf.wan: serf: EventMemberJoin: Node-063daf4e-db99-bc14-ae90-7d1c8eb1cd70.dc1 127.0.0.1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.842Z [DEBUG] TestACL_Authorize.server.replication.acl.token: finished fetching acls: amount=3 writer.go:29: 2021-01-29T19:34:41.843Z [DEBUG] TestACL_Authorize.server.replication.acl.token: acl replication: local=0 remote=3 writer.go:29: 2021-01-29T19:34:41.844Z [DEBUG] TestACL_Authorize.server.replication.acl.token: acl replication: deletions=0 updates=3 writer.go:29: 2021-01-29T19:34:41.845Z [DEBUG] TestACL_Authorize.server.replication.acl.token: acl replication - downloaded updates: amount=3 writer.go:29: 2021-01-29T19:34:41.847Z [DEBUG] TestACL_Authorize.server.replication.acl.token: acl replication - performing updates === CONT TestACL_Disabled_Response writer.go:29: 2021-01-29T19:34:41.841Z [INFO] TestACL_Disabled_Response.server.serf.lan: serf: EventMemberJoin: Node-063daf4e-db99-bc14-ae90-7d1c8eb1cd70 127.0.0.1 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.848Z [DEBUG] TestACL_Authorize.server.replication.acl.policy: finished fetching acls: amount=2 writer.go:29: 2021-01-29T19:34:41.850Z [DEBUG] TestACL_Authorize.server.replication.acl.policy: acl replication: local=0 remote=2 === CONT TestACL_Disabled_Response writer.go:29: 2021-01-29T19:34:41.852Z [INFO] TestACL_Disabled_Response: Started DNS server: address=127.0.0.1:29634 network=udp writer.go:29: 2021-01-29T19:34:41.854Z [INFO] TestACL_Disabled_Response: Started DNS server: address=127.0.0.1:29634 network=tcp writer.go:29: 2021-01-29T19:34:41.853Z [INFO] TestACL_Disabled_Response.server: Adding LAN server: server="Node-063daf4e-db99-bc14-ae90-7d1c8eb1cd70 (Addr: tcp/127.0.0.1:29639) (DC: dc1)" writer.go:29: 2021-01-29T19:34:41.853Z [INFO] TestACL_Disabled_Response.server: Handled event for server in area: event=member-join server=Node-063daf4e-db99-bc14-ae90-7d1c8eb1cd70.dc1 area=wan === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.852Z [DEBUG] TestACL_Authorize.server.replication.acl.policy: acl replication: deletions=0 updates=2 === CONT TestACL_Disabled_Response writer.go:29: 2021-01-29T19:34:41.858Z [INFO] TestACL_Disabled_Response: Started HTTP server: address=127.0.0.1:29635 network=tcp writer.go:29: 2021-01-29T19:34:41.861Z [INFO] TestACL_Disabled_Response: started state syncer === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.850Z [DEBUG] TestACL_Authorize.server.replication.acl.role: finished fetching acls: amount=0 writer.go:29: 2021-01-29T19:34:41.865Z [DEBUG] TestACL_Authorize.server.replication.acl.role: acl replication: local=0 remote=0 writer.go:29: 2021-01-29T19:34:41.865Z [DEBUG] TestACL_Authorize.server.replication.acl.policy: acl replication - downloaded updates: amount=2 writer.go:29: 2021-01-29T19:34:41.873Z [DEBUG] TestACL_Authorize.server.replication.acl.policy: acl replication - performing updates writer.go:29: 2021-01-29T19:34:41.867Z [DEBUG] TestACL_Authorize.server.replication.acl.token: acl replication - upserted batch: number_upserted=3 batch_size=442 writer.go:29: 2021-01-29T19:34:41.872Z [DEBUG] TestACL_Authorize.server.replication.acl.role: acl replication: deletions=0 updates=0 writer.go:29: 2021-01-29T19:34:41.877Z [DEBUG] TestACL_Authorize.server.replication.acl.policy: acl replication - upserted batch: number_upserted=2 batch_size=675 writer.go:29: 2021-01-29T19:34:41.880Z [DEBUG] TestACL_Authorize.server.replication.acl.token: acl replication - finished updates writer.go:29: 2021-01-29T19:34:41.884Z [DEBUG] TestACL_Authorize.server.replication.acl.role: ACL replication completed through remote index: index=1 writer.go:29: 2021-01-29T19:34:41.888Z [DEBUG] TestACL_Authorize.server.replication.acl.policy: acl replication - finished updates writer.go:29: 2021-01-29T19:34:41.891Z [DEBUG] TestACL_Authorize.server.replication.acl.token: ACL replication completed through remote index: index=16 writer.go:29: 2021-01-29T19:34:41.898Z [DEBUG] TestACL_Authorize.server.replication.acl.policy: ACL replication completed through remote index: index=15 === CONT TestACL_Disabled_Response writer.go:29: 2021-01-29T19:34:41.904Z [WARN] TestACL_Disabled_Response.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:41.905Z [INFO] TestACL_Disabled_Response.server.raft: entering candidate state: node="Node at 127.0.0.1:29639 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:41.907Z [DEBUG] TestACL_Disabled_Response.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:41.908Z [DEBUG] TestACL_Disabled_Response.server.raft: vote granted: from=063daf4e-db99-bc14-ae90-7d1c8eb1cd70 term=2 tally=1 writer.go:29: 2021-01-29T19:34:41.909Z [INFO] TestACL_Disabled_Response.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:41.910Z [INFO] TestACL_Disabled_Response.server.raft: entering leader state: leader="Node at 127.0.0.1:29639 [Leader]" writer.go:29: 2021-01-29T19:34:41.911Z [INFO] TestACL_Disabled_Response.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:41.913Z [INFO] TestACL_Disabled_Response.server: New leader elected: payload=Node-063daf4e-db99-bc14-ae90-7d1c8eb1cd70 writer.go:29: 2021-01-29T19:34:41.916Z [DEBUG] TestACL_Disabled_Response.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29639 === RUN TestACL_Authorize/master-token === RUN TestACL_Authorize/master-token/dc1 === CONT TestACL_Disabled_Response writer.go:29: 2021-01-29T19:34:41.921Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:41.946Z [INFO] TestACL_Disabled_Response.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:41.947Z [INFO] TestACL_Disabled_Response.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:41.949Z [DEBUG] TestACL_Disabled_Response.server: Skipping self join check for node since the cluster is too small: node=Node-063daf4e-db99-bc14-ae90-7d1c8eb1cd70 writer.go:29: 2021-01-29T19:34:41.950Z [INFO] TestACL_Disabled_Response.server: member joined, marking health alive: member=Node-063daf4e-db99-bc14-ae90-7d1c8eb1cd70 === RUN TestACL_Authorize/master-token/dc2 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:41.969Z [DEBUG] TestACL_Authorize.tlsutil: OutgoingRPCWrapper: version=1 === RUN TestACL_Disabled_Response/ACLBootstrap === RUN TestACL_Disabled_Response/ACLReplicationStatus === RUN TestACL_Disabled_Response/AgentToken === RUN TestACL_Disabled_Response/ACLRulesTranslate === RUN TestACL_Disabled_Response/ACLRulesTranslateLegacyToken === RUN TestACL_Disabled_Response/ACLPolicyList === RUN TestACL_Disabled_Response/ACLPolicyCRUD === RUN TestACL_Disabled_Response/ACLPolicyCreate === RUN TestACL_Disabled_Response/ACLTokenList === RUN TestACL_Disabled_Response/ACLTokenCreate === RUN TestACL_Disabled_Response/ACLTokenSelf === RUN TestACL_Disabled_Response/ACLTokenCRUD === RUN TestACL_Authorize/custom-token === RUN TestACL_Authorize/custom-token/dc1 === RUN TestACL_Disabled_Response/ACLRoleList === RUN TestACL_Disabled_Response/ACLRoleCreate === RUN TestACL_Disabled_Response/ACLRoleCRUD === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:42.044Z [INFO] TestACL_Authorize.server.serf.wan: serf: EventMemberUpdate: Node-90798172-bdb7-b836-0d15-84f51727be6a.dc2 writer.go:29: 2021-01-29T19:34:42.047Z [DEBUG] TestACL_Authorize.server.serf.wan: serf: messageJoinType: Node-90798172-bdb7-b836-0d15-84f51727be6a.dc2 writer.go:29: 2021-01-29T19:34:42.049Z [INFO] TestACL_Authorize.server: Handled event for server in area: event=member-update server=Node-90798172-bdb7-b836-0d15-84f51727be6a.dc2 area=wan === RUN TestACL_Disabled_Response/ACLBindingRuleList === RUN TestACL_Disabled_Response/ACLBindingRuleCreate === RUN TestACL_Disabled_Response/ACLBindingRuleCRUD === RUN TestACL_Disabled_Response/ACLAuthMethodList === RUN TestACL_Disabled_Response/ACLAuthMethodCreate === RUN TestACL_Disabled_Response/ACLAuthMethodCRUD === RUN TestACL_Disabled_Response/ACLLogin === RUN TestACL_Disabled_Response/ACLLogout === RUN TestACL_Authorize/custom-token/dc2 === RUN TestACL_Disabled_Response/ACLAuthorize === CONT TestACL_Disabled_Response writer.go:29: 2021-01-29T19:34:42.117Z [INFO] TestACL_Disabled_Response: Requesting shutdown writer.go:29: 2021-01-29T19:34:42.118Z [INFO] TestACL_Disabled_Response.server: shutting down server writer.go:29: 2021-01-29T19:34:42.120Z [DEBUG] TestACL_Disabled_Response.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:42.120Z [ERROR] TestACL_Disabled_Response.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:42.122Z [WARN] TestACL_Disabled_Response.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:42.122Z [DEBUG] TestACL_Disabled_Response.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:42.125Z [WARN] TestACL_Disabled_Response.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:42.127Z [INFO] TestACL_Disabled_Response.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:42.127Z [INFO] TestACL_Disabled_Response: consul server down writer.go:29: 2021-01-29T19:34:42.129Z [INFO] TestACL_Disabled_Response: shutdown complete writer.go:29: 2021-01-29T19:34:42.131Z [INFO] TestACL_Disabled_Response: Stopping server: protocol=DNS address=127.0.0.1:29634 network=tcp writer.go:29: 2021-01-29T19:34:42.133Z [INFO] TestACL_Disabled_Response: Stopping server: protocol=DNS address=127.0.0.1:29634 network=udp === RUN TestACL_Authorize/too-many-requests === CONT TestACL_Disabled_Response writer.go:29: 2021-01-29T19:34:42.134Z [INFO] TestACL_Disabled_Response: Stopping server: protocol=HTTP address=127.0.0.1:29635 network=tcp === RUN TestACL_Authorize/decode-failure === RUN TestACL_Authorize/acl-not-found === RUN TestACL_Authorize/local-token-in-secondary-dc === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:42.173Z [DEBUG] TestACL_Authorize.server.replication.config_entry: finished fetching config entries: amount=0 writer.go:29: 2021-01-29T19:34:42.175Z [DEBUG] TestACL_Authorize.server.replication.config_entry: Config Entry replication: local=0 remote=0 writer.go:29: 2021-01-29T19:34:42.177Z [DEBUG] TestACL_Authorize.server.replication.config_entry: Config Entry replication: deletions=0 updates=0 writer.go:29: 2021-01-29T19:34:42.179Z [DEBUG] TestACL_Authorize.server.replication.config_entry: replication completed through remote index: index=1 === RUN TestACL_Authorize/local-token-wrong-dc === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:42.212Z [INFO] TestACL_Authorize: Requesting shutdown writer.go:29: 2021-01-29T19:34:42.214Z [INFO] TestACL_Authorize.server: shutting down server writer.go:29: 2021-01-29T19:34:42.216Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="ACL token replication" writer.go:29: 2021-01-29T19:34:42.220Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.224Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="secondary CA roots watch" writer.go:29: 2021-01-29T19:34:42.226Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="intention replication" writer.go:29: 2021-01-29T19:34:42.224Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.228Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="secondary cert renew watch" writer.go:29: 2021-01-29T19:34:42.237Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:42.242Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="ACL role replication" writer.go:29: 2021-01-29T19:34:42.238Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="secondary cert renew watch" writer.go:29: 2021-01-29T19:34:42.242Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:42.244Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="ACL policy replication" writer.go:29: 2021-01-29T19:34:42.250Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="config entry replication" writer.go:29: 2021-01-29T19:34:42.252Z [WARN] TestACL_Authorize.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:42.255Z [WARN] TestACL_Authorize.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:42.258Z [INFO] TestACL_Authorize.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:42.260Z [INFO] TestACL_Authorize.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:42.263Z [INFO] TestACL_Authorize: consul server down writer.go:29: 2021-01-29T19:34:42.266Z [INFO] TestACL_Authorize: shutdown complete writer.go:29: 2021-01-29T19:34:42.264Z [ERROR] TestACL_Authorize.server.rpc: RPC failed to server in DC: server=127.0.0.1:29609 datacenter=dc1 method=ACL.TokenList error="rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:42.264Z [ERROR] TestACL_Authorize.server.rpc: RPC failed to server in DC: server=127.0.0.1:29609 datacenter=dc1 method=ACL.RoleList error="rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:42.264Z [ERROR] TestACL_Authorize.server.rpc: RPC failed to server in DC: server=127.0.0.1:29609 datacenter=dc1 method=ConfigEntry.ListAll error="rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:42.264Z [ERROR] TestACL_Authorize.server.rpc: RPC failed to server in DC: server=127.0.0.1:29609 datacenter=dc1 method=ACL.PolicyList error="rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:42.268Z [INFO] TestACL_Authorize: Stopping server: protocol=DNS address=127.0.0.1:29628 network=tcp writer.go:29: 2021-01-29T19:34:42.271Z [WARN] TestACL_Authorize.server.replication.acl.token: ACL replication error (will retry if still leader): error="failed to retrieve remote ACL tokens: rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:42.274Z [WARN] TestACL_Authorize.server.replication.acl.role: ACL replication error (will retry if still leader): error="failed to retrieve remote ACL roles: rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:42.276Z [INFO] TestACL_Authorize.server.replication.config_entry: stopped replication writer.go:29: 2021-01-29T19:34:42.279Z [WARN] TestACL_Authorize.server.replication.acl.policy: ACL replication error (will retry if still leader): error="failed to retrieve remote ACL policies: rpc error making call: EOF" writer.go:29: 2021-01-29T19:34:42.281Z [INFO] TestACL_Authorize: Stopping server: protocol=DNS address=127.0.0.1:29628 network=udp writer.go:29: 2021-01-29T19:34:42.283Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="ACL token replication" writer.go:29: 2021-01-29T19:34:42.285Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="ACL role replication" writer.go:29: 2021-01-29T19:34:42.286Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="config entry replication" writer.go:29: 2021-01-29T19:34:42.288Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="ACL policy replication" writer.go:29: 2021-01-29T19:34:42.291Z [INFO] TestACL_Authorize: Stopping server: protocol=HTTP address=127.0.0.1:29629 network=tcp === CONT TestACL_Legacy_Clone writer.go:29: 2021-01-29T19:34:42.315Z [INFO] TestACL_Legacy_Clone: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:42.318Z [INFO] TestACL_Legacy_Clone: Endpoints down --- PASS: TestACL_Legacy_Clone (1.02s) === CONT TestACLReplicationStatus === CONT TestACL_HTTP writer.go:29: 2021-01-29T19:34:42.321Z [INFO] TestACL_HTTP: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:42.324Z [INFO] TestACL_HTTP: Endpoints down --- PASS: TestACL_HTTP (1.20s) --- PASS: TestACL_HTTP/Policy (0.14s) --- PASS: TestACL_HTTP/Policy/Create (0.03s) --- PASS: TestACL_HTTP/Policy/Minimal (0.02s) --- PASS: TestACL_HTTP/Policy/Name_Chars (0.01s) --- PASS: TestACL_HTTP/Policy/Update_Name_ID_Mismatch (0.00s) --- PASS: TestACL_HTTP/Policy/Policy_CRUD_Missing_ID_in_URL (0.00s) --- PASS: TestACL_HTTP/Policy/Update (0.03s) --- PASS: TestACL_HTTP/Policy/ID_Supplied (0.00s) --- PASS: TestACL_HTTP/Policy/Invalid_payload (0.00s) --- PASS: TestACL_HTTP/Policy/Delete (0.00s) --- PASS: TestACL_HTTP/Policy/List (0.02s) --- PASS: TestACL_HTTP/Policy/Read (0.01s) --- PASS: TestACL_HTTP/Policy/Read_Name (0.01s) --- PASS: TestACL_HTTP/Role (0.12s) --- PASS: TestACL_HTTP/Role/Create (0.01s) --- PASS: TestACL_HTTP/Role/Name_Chars (0.04s) --- PASS: TestACL_HTTP/Role/Update_Name_ID_Mismatch (0.01s) --- PASS: TestACL_HTTP/Role/Role_CRUD_Missing_ID_in_URL (0.00s) --- PASS: TestACL_HTTP/Role/Update (0.01s) --- PASS: TestACL_HTTP/Role/ID_Supplied (0.01s) --- PASS: TestACL_HTTP/Role/Invalid_payload (0.00s) --- PASS: TestACL_HTTP/Role/Delete (0.00s) --- PASS: TestACL_HTTP/Role/List (0.01s) --- PASS: TestACL_HTTP/Role/Read (0.01s) --- PASS: TestACL_HTTP/Token (0.20s) --- PASS: TestACL_HTTP/Token/Create (0.02s) --- PASS: TestACL_HTTP/Token/Create_Local (0.01s) --- PASS: TestACL_HTTP/Token/Read (0.01s) --- PASS: TestACL_HTTP/Token/Self (0.01s) --- PASS: TestACL_HTTP/Token/Clone (0.01s) --- PASS: TestACL_HTTP/Token/Update (0.01s) --- PASS: TestACL_HTTP/Token/CRUD_Missing_Token_Accessor_ID (0.00s) --- PASS: TestACL_HTTP/Token/Update_Accessor_Mismatch (0.01s) --- PASS: TestACL_HTTP/Token/Delete (0.00s) --- PASS: TestACL_HTTP/Token/List (0.02s) --- PASS: TestACL_HTTP/Token/List_by_Policy (0.01s) --- PASS: TestACL_HTTP/Token/Create_with_Accessor (0.01s) --- PASS: TestACL_HTTP/Token/Create_with_Secret (0.01s) --- PASS: TestACL_HTTP/Token/Create_with_Accessor_and_Secret (0.01s) --- PASS: TestACL_HTTP/Token/Create_with_Accessor_Dup (0.00s) --- PASS: TestACL_HTTP/Token/Create_with_Secret_as_Accessor_Dup (0.00s) --- PASS: TestACL_HTTP/Token/Create_with_Secret_Dup (0.00s) --- PASS: TestACL_HTTP/Token/Create_with_Accessor_as_Secret_Dup (0.00s) --- PASS: TestACL_HTTP/Token/Create_with_Reserved_Accessor (0.00s) --- PASS: TestACL_HTTP/Token/Create_with_Reserved_Secret (0.02s) === CONT TestACL_Legacy_Get === RUN TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.344Z [WARN] TestACL_Legacy_Get/wrong_id: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:42.345Z [WARN] TestACL_Legacy_Get/wrong_id: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:42.348Z [DEBUG] TestACL_Legacy_Get/wrong_id.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:42.351Z [DEBUG] TestACL_Legacy_Get/wrong_id.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:42.360Z [INFO] TestACL_Legacy_Get/wrong_id.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d258f83b-4568-47ce-6f88-f971ef01fbb0 Address:127.0.0.1:29657}]" writer.go:29: 2021-01-29T19:34:42.364Z [INFO] TestACL_Legacy_Get/wrong_id.server.raft: entering follower state: follower="Node at 127.0.0.1:29657 [Follower]" leader= === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.365Z [WARN] TestACLReplicationStatus: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.367Z [INFO] TestACL_Legacy_Get/wrong_id.server.serf.wan: serf: EventMemberJoin: Node-d258f83b-4568-47ce-6f88-f971ef01fbb0.dc1 127.0.0.1 === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.368Z [WARN] TestACLReplicationStatus: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:42.383Z [DEBUG] TestACLReplicationStatus.tlsutil: Update: version=1 === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.384Z [INFO] TestACL_Legacy_Get/wrong_id.server.serf.lan: serf: EventMemberJoin: Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 127.0.0.1 writer.go:29: 2021-01-29T19:34:42.388Z [INFO] TestACL_Legacy_Get/wrong_id.server: Handled event for server in area: event=member-join server=Node-d258f83b-4568-47ce-6f88-f971ef01fbb0.dc1 area=wan writer.go:29: 2021-01-29T19:34:42.390Z [INFO] TestACL_Legacy_Get/wrong_id: Started DNS server: address=127.0.0.1:29652 network=udp writer.go:29: 2021-01-29T19:34:42.393Z [INFO] TestACL_Legacy_Get/wrong_id: Started DNS server: address=127.0.0.1:29652 network=tcp writer.go:29: 2021-01-29T19:34:42.390Z [INFO] TestACL_Legacy_Get/wrong_id.server: Adding LAN server: server="Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 (Addr: tcp/127.0.0.1:29657) (DC: dc1)" writer.go:29: 2021-01-29T19:34:42.397Z [INFO] TestACL_Legacy_Get/wrong_id: Started HTTP server: address=127.0.0.1:29653 network=tcp === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.386Z [DEBUG] TestACLReplicationStatus.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.400Z [INFO] TestACL_Legacy_Get/wrong_id: started state syncer writer.go:29: 2021-01-29T19:34:42.405Z [WARN] TestACL_Legacy_Get/wrong_id.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:42.408Z [INFO] TestACL_Legacy_Get/wrong_id.server.raft: entering candidate state: node="Node at 127.0.0.1:29657 [Candidate]" term=2 === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.412Z [INFO] TestACLReplicationStatus.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9967ccf3-d605-fbdc-a8a3-567391073d3d Address:127.0.0.1:29663}]" writer.go:29: 2021-01-29T19:34:42.416Z [INFO] TestACLReplicationStatus.server.raft: entering follower state: follower="Node at 127.0.0.1:29663 [Follower]" leader= === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.412Z [DEBUG] TestACL_Legacy_Get/wrong_id.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:42.420Z [DEBUG] TestACL_Legacy_Get/wrong_id.server.raft: vote granted: from=d258f83b-4568-47ce-6f88-f971ef01fbb0 term=2 tally=1 === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.417Z [INFO] TestACLReplicationStatus.server.serf.wan: serf: EventMemberJoin: Node-9967ccf3-d605-fbdc-a8a3-567391073d3d.dc1 127.0.0.1 === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.422Z [INFO] TestACL_Legacy_Get/wrong_id.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:42.426Z [INFO] TestACL_Legacy_Get/wrong_id.server.raft: entering leader state: leader="Node at 127.0.0.1:29657 [Leader]" === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.424Z [INFO] TestACLReplicationStatus.server.serf.lan: serf: EventMemberJoin: Node-9967ccf3-d605-fbdc-a8a3-567391073d3d 127.0.0.1 === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.429Z [INFO] TestACL_Legacy_Get/wrong_id.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:42.431Z [INFO] TestACL_Legacy_Get/wrong_id.server: New leader elected: payload=Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.434Z [INFO] TestACLReplicationStatus.server: Adding LAN server: server="Node-9967ccf3-d605-fbdc-a8a3-567391073d3d (Addr: tcp/127.0.0.1:29663) (DC: dc1)" === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.433Z [INFO] TestACL_Legacy_Get/wrong_id.server: initializing acls === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.436Z [INFO] TestACLReplicationStatus: Started DNS server: address=127.0.0.1:29658 network=udp === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.439Z [INFO] TestACL_Legacy_Get/wrong_id.server: initializing acls === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.439Z [INFO] TestACLReplicationStatus: Started DNS server: address=127.0.0.1:29658 network=tcp === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.441Z [WARN] TestACL_Legacy_Get/wrong_id.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:42.440Z [INFO] TestACL_Legacy_Get/wrong_id.server: Created ACL 'global-management' policy === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.436Z [INFO] TestACLReplicationStatus.server: Handled event for server in area: event=member-join server=Node-9967ccf3-d605-fbdc-a8a3-567391073d3d.dc1 area=wan === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.444Z [WARN] TestACL_Legacy_Get/wrong_id.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:42.447Z [INFO] TestACL_Legacy_Get/wrong_id.server: Bootstrapped ACL master token from configuration === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.449Z [INFO] TestACLReplicationStatus: Started HTTP server: address=127.0.0.1:29659 network=tcp === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.447Z [INFO] TestACL_Legacy_Get/wrong_id.server: Bootstrapped ACL master token from configuration === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.452Z [INFO] TestACLReplicationStatus: started state syncer === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.455Z [INFO] TestACL_Legacy_Get/wrong_id.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:42.458Z [INFO] TestACL_Legacy_Get/wrong_id.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:42.461Z [INFO] TestACL_Legacy_Get/wrong_id.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.461Z [INFO] TestACL_Legacy_Get/wrong_id.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:42.465Z [DEBUG] TestACL_Legacy_Get/wrong_id.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:34:42.464Z [INFO] TestACL_Legacy_Get/wrong_id.server.serf.lan: serf: EventMemberUpdate: Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 writer.go:29: 2021-01-29T19:34:42.470Z [INFO] TestACL_Legacy_Get/wrong_id.server: Updating LAN server: server="Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 (Addr: tcp/127.0.0.1:29657) (DC: dc1)" === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.472Z [WARN] TestACLReplicationStatus.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:42.473Z [INFO] TestACLReplicationStatus.server.raft: entering candidate state: node="Node at 127.0.0.1:29663 [Candidate]" term=2 === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.470Z [INFO] TestACL_Legacy_Get/wrong_id.server.serf.wan: serf: EventMemberUpdate: Node-d258f83b-4568-47ce-6f88-f971ef01fbb0.dc1 === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.476Z [DEBUG] TestACLReplicationStatus.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:42.478Z [DEBUG] TestACLReplicationStatus.server.raft: vote granted: from=9967ccf3-d605-fbdc-a8a3-567391073d3d term=2 tally=1 === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.471Z [INFO] TestACL_Legacy_Get/wrong_id.server.serf.lan: serf: EventMemberUpdate: Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.480Z [INFO] TestACLReplicationStatus.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:42.482Z [INFO] TestACLReplicationStatus.server.raft: entering leader state: leader="Node at 127.0.0.1:29663 [Leader]" writer.go:29: 2021-01-29T19:34:42.485Z [DEBUG] TestACLReplicationStatus.server: Cannot upgrade to new ACLs: leaderMode=2 mode=2 found=true leader=127.0.0.1:29663 writer.go:29: 2021-01-29T19:34:42.488Z [INFO] TestACLReplicationStatus.server: cluster leadership acquired === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.481Z [INFO] TestACL_Legacy_Get/wrong_id.server: Updating LAN server: server="Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 (Addr: tcp/127.0.0.1:29657) (DC: dc1)" === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.488Z [INFO] TestACLReplicationStatus.server: New leader elected: payload=Node-9967ccf3-d605-fbdc-a8a3-567391073d3d writer.go:29: 2021-01-29T19:34:42.489Z [INFO] TestACLReplicationStatus.server: initializing acls === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.502Z [INFO] TestACL_Legacy_Get/wrong_id.server: Handled event for server in area: event=member-update server=Node-d258f83b-4568-47ce-6f88-f971ef01fbb0.dc1 area=wan writer.go:29: 2021-01-29T19:34:42.502Z [INFO] TestACL_Legacy_Get/wrong_id.server.serf.wan: serf: EventMemberUpdate: Node-d258f83b-4568-47ce-6f88-f971ef01fbb0.dc1 writer.go:29: 2021-01-29T19:34:42.505Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:42.507Z [INFO] TestACL_Legacy_Get/wrong_id.server: Handled event for server in area: event=member-update server=Node-d258f83b-4568-47ce-6f88-f971ef01fbb0.dc1 area=wan writer.go:29: 2021-01-29T19:34:42.514Z [INFO] TestACL_Legacy_Get/wrong_id.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.509Z [INFO] TestACLReplicationStatus.server: Created ACL 'global-management' policy === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.516Z [INFO] TestACL_Legacy_Get/wrong_id.leader: started routine: routine="CA root pruning" === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.517Z [WARN] TestACLReplicationStatus.server: Configuring a non-UUID master token is deprecated === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.518Z [DEBUG] TestACL_Legacy_Get/wrong_id.server: Skipping self join check for node since the cluster is too small: node=Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 writer.go:29: 2021-01-29T19:34:42.520Z [INFO] TestACL_Legacy_Get/wrong_id.server: member joined, marking health alive: member=Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 writer.go:29: 2021-01-29T19:34:42.523Z [DEBUG] TestACL_Legacy_Get/wrong_id.server: Skipping self join check for node since the cluster is too small: node=Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.522Z [INFO] TestACLReplicationStatus.server: Bootstrapped ACL master token from configuration === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.525Z [DEBUG] TestACL_Legacy_Get/wrong_id.server: Skipping self join check for node since the cluster is too small: node=Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.529Z [INFO] TestACLReplicationStatus.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:42.532Z [INFO] TestACLReplicationStatus.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:42.534Z [INFO] TestACLReplicationStatus.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.538Z [INFO] TestACLReplicationStatus.server.serf.lan: serf: EventMemberUpdate: Node-9967ccf3-d605-fbdc-a8a3-567391073d3d writer.go:29: 2021-01-29T19:34:42.542Z [INFO] TestACLReplicationStatus.server: Updating LAN server: server="Node-9967ccf3-d605-fbdc-a8a3-567391073d3d (Addr: tcp/127.0.0.1:29663) (DC: dc1)" writer.go:29: 2021-01-29T19:34:42.545Z [INFO] TestACLReplicationStatus.server.serf.wan: serf: EventMemberUpdate: Node-9967ccf3-d605-fbdc-a8a3-567391073d3d.dc1 writer.go:29: 2021-01-29T19:34:42.549Z [INFO] TestACLReplicationStatus.server: Handled event for server in area: event=member-update server=Node-9967ccf3-d605-fbdc-a8a3-567391073d3d.dc1 area=wan writer.go:29: 2021-01-29T19:34:42.551Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:42.570Z [INFO] TestACLReplicationStatus.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:42.572Z [INFO] TestACLReplicationStatus.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:42.573Z [DEBUG] TestACLReplicationStatus.server: Skipping self join check for node since the cluster is too small: node=Node-9967ccf3-d605-fbdc-a8a3-567391073d3d writer.go:29: 2021-01-29T19:34:42.575Z [INFO] TestACLReplicationStatus.server: member joined, marking health alive: member=Node-9967ccf3-d605-fbdc-a8a3-567391073d3d writer.go:29: 2021-01-29T19:34:42.578Z [DEBUG] TestACLReplicationStatus.server: Skipping self join check for node since the cluster is too small: node=Node-9967ccf3-d605-fbdc-a8a3-567391073d3d === CONT TestACL_Disabled_Response writer.go:29: 2021-01-29T19:34:42.636Z [INFO] TestACL_Disabled_Response: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:42.638Z [INFO] TestACL_Disabled_Response: Endpoints down --- PASS: TestACL_Disabled_Response (0.86s) --- PASS: TestACL_Disabled_Response/ACLBootstrap (0.00s) --- PASS: TestACL_Disabled_Response/ACLReplicationStatus (0.01s) --- PASS: TestACL_Disabled_Response/AgentToken (0.00s) --- PASS: TestACL_Disabled_Response/ACLRulesTranslate (0.01s) --- PASS: TestACL_Disabled_Response/ACLRulesTranslateLegacyToken (0.00s) --- PASS: TestACL_Disabled_Response/ACLPolicyList (0.00s) --- PASS: TestACL_Disabled_Response/ACLPolicyCRUD (0.00s) --- PASS: TestACL_Disabled_Response/ACLPolicyCreate (0.00s) --- PASS: TestACL_Disabled_Response/ACLTokenList (0.00s) --- PASS: TestACL_Disabled_Response/ACLTokenCreate (0.00s) --- PASS: TestACL_Disabled_Response/ACLTokenSelf (0.00s) --- PASS: TestACL_Disabled_Response/ACLTokenCRUD (0.01s) --- PASS: TestACL_Disabled_Response/ACLRoleList (0.00s) --- PASS: TestACL_Disabled_Response/ACLRoleCreate (0.01s) --- PASS: TestACL_Disabled_Response/ACLRoleCRUD (0.01s) --- PASS: TestACL_Disabled_Response/ACLBindingRuleList (0.00s) --- PASS: TestACL_Disabled_Response/ACLBindingRuleCreate (0.01s) --- PASS: TestACL_Disabled_Response/ACLBindingRuleCRUD (0.00s) --- PASS: TestACL_Disabled_Response/ACLAuthMethodList (0.00s) --- PASS: TestACL_Disabled_Response/ACLAuthMethodCreate (0.00s) --- PASS: TestACL_Disabled_Response/ACLAuthMethodCRUD (0.00s) --- PASS: TestACL_Disabled_Response/ACLLogin (0.00s) --- PASS: TestACL_Disabled_Response/ACLLogout (0.01s) --- PASS: TestACL_Disabled_Response/ACLAuthorize (0.00s) === CONT TestACL_Legacy_Update === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.645Z [DEBUG] TestACL_Legacy_Get/wrong_id: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:42.647Z [INFO] TestACL_Legacy_Get/wrong_id: Synced node info writer.go:29: 2021-01-29T19:34:42.648Z [DEBUG] TestACL_Legacy_Get/wrong_id: Node info in sync === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.662Z [WARN] TestACL_Legacy_Update: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:42.663Z [WARN] TestACL_Legacy_Update: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:42.665Z [DEBUG] TestACL_Legacy_Update.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:42.666Z [DEBUG] TestACL_Legacy_Update.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:42.671Z [INFO] TestACL_Legacy_Update.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4f28edd5-9148-c00f-73a3-5ceedb0b84cb Address:127.0.0.1:29675}]" writer.go:29: 2021-01-29T19:34:42.673Z [INFO] TestACL_Legacy_Update.server.raft: entering follower state: follower="Node at 127.0.0.1:29675 [Follower]" leader= writer.go:29: 2021-01-29T19:34:42.673Z [INFO] TestACL_Legacy_Update.server.serf.wan: serf: EventMemberJoin: Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:42.681Z [INFO] TestACL_Legacy_Update.server.serf.lan: serf: EventMemberJoin: Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb 127.0.0.1 writer.go:29: 2021-01-29T19:34:42.684Z [INFO] TestACL_Legacy_Update.server: Adding LAN server: server="Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb (Addr: tcp/127.0.0.1:29675) (DC: dc1)" writer.go:29: 2021-01-29T19:34:42.685Z [INFO] TestACL_Legacy_Update.server: Handled event for server in area: event=member-join server=Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb.dc1 area=wan writer.go:29: 2021-01-29T19:34:42.687Z [INFO] TestACL_Legacy_Update: Started DNS server: address=127.0.0.1:29670 network=tcp writer.go:29: 2021-01-29T19:34:42.689Z [INFO] TestACL_Legacy_Update: Started DNS server: address=127.0.0.1:29670 network=udp writer.go:29: 2021-01-29T19:34:42.690Z [INFO] TestACL_Legacy_Update: Started HTTP server: address=127.0.0.1:29671 network=tcp writer.go:29: 2021-01-29T19:34:42.691Z [INFO] TestACL_Legacy_Update: started state syncer writer.go:29: 2021-01-29T19:34:42.732Z [WARN] TestACL_Legacy_Update.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:42.733Z [INFO] TestACL_Legacy_Update.server.raft: entering candidate state: node="Node at 127.0.0.1:29675 [Candidate]" term=2 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:42.732Z [DEBUG] TestACL_Authorize.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.735Z [DEBUG] TestACL_Legacy_Update.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:42.735Z [DEBUG] TestACL_Legacy_Update.server.raft: vote granted: from=4f28edd5-9148-c00f-73a3-5ceedb0b84cb term=2 tally=1 writer.go:29: 2021-01-29T19:34:42.736Z [INFO] TestACL_Legacy_Update.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:42.737Z [INFO] TestACL_Legacy_Update.server.raft: entering leader state: leader="Node at 127.0.0.1:29675 [Leader]" writer.go:29: 2021-01-29T19:34:42.736Z [DEBUG] TestACL_Legacy_Update.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:42.741Z [INFO] TestACL_Legacy_Update.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:42.741Z [INFO] TestACL_Legacy_Update.server: New leader elected: payload=Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb writer.go:29: 2021-01-29T19:34:42.743Z [INFO] TestACL_Legacy_Update.server: initializing acls writer.go:29: 2021-01-29T19:34:42.746Z [INFO] TestACL_Legacy_Update.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:42.748Z [WARN] TestACL_Legacy_Update.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:42.755Z [INFO] TestACL_Legacy_Update.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:42.761Z [INFO] TestACL_Legacy_Update.server: Created ACL anonymous token from configuration === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.762Z [DEBUG] TestACLReplicationStatus: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.764Z [INFO] TestACL_Legacy_Update.leader: started routine: routine="legacy ACL token upgrade" === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.767Z [INFO] TestACLReplicationStatus: Synced node info === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.767Z [INFO] TestACL_Legacy_Update.leader: started routine: routine="acl token reaping" === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.769Z [DEBUG] TestACLReplicationStatus: Node info in sync === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.771Z [INFO] TestACL_Legacy_Update.server.serf.lan: serf: EventMemberUpdate: Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb writer.go:29: 2021-01-29T19:34:42.775Z [INFO] TestACL_Legacy_Update.server: Updating LAN server: server="Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb (Addr: tcp/127.0.0.1:29675) (DC: dc1)" writer.go:29: 2021-01-29T19:34:42.776Z [INFO] TestACL_Legacy_Update.server.serf.wan: serf: EventMemberUpdate: Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb.dc1 writer.go:29: 2021-01-29T19:34:42.779Z [INFO] TestACL_Legacy_Update.server: Handled event for server in area: event=member-update server=Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb.dc1 area=wan writer.go:29: 2021-01-29T19:34:42.781Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:42.786Z [INFO] TestACL_Legacy_Update.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:42.788Z [INFO] TestACL_Legacy_Update.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:42.789Z [DEBUG] TestACL_Legacy_Update.server: Skipping self join check for node since the cluster is too small: node=Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.791Z [DEBUG] TestACL_Legacy_Get/wrong_id.acl: dropping node from result due to ACLs: node=Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 writer.go:29: 2021-01-29T19:34:42.799Z [DEBUG] TestACL_Legacy_Get/wrong_id.acl: dropping node from result due to ACLs: node=Node-d258f83b-4568-47ce-6f88-f971ef01fbb0 === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:42.801Z [INFO] TestACL_Authorize: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:42.802Z [INFO] TestACL_Authorize: Endpoints down writer.go:29: 2021-01-29T19:34:42.803Z [INFO] TestACL_Authorize: Requesting shutdown writer.go:29: 2021-01-29T19:34:42.804Z [INFO] TestACL_Authorize.server: shutting down server === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.804Z [INFO] TestACL_Legacy_Get/wrong_id: Requesting shutdown === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:42.805Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:42.805Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.806Z [DEBUG] TestACL_Authorize.leader: stopping routine: routine="CA root pruning" === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.805Z [INFO] TestACL_Legacy_Get/wrong_id.server: shutting down server === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:42.807Z [WARN] TestACL_Authorize.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:42.808Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.807Z [DEBUG] TestACL_Legacy_Get/wrong_id.leader: stopping routine: routine="CA root pruning" === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:42.809Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.809Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="CA root pruning" === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.809Z [DEBUG] TestACL_Legacy_Get/wrong_id.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.791Z [INFO] TestACL_Legacy_Update.server: member joined, marking health alive: member=Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:42.812Z [WARN] TestACL_Authorize.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:42.824Z [INFO] TestACL_Authorize.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:42.826Z [INFO] TestACL_Authorize.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:42.826Z [INFO] TestACL_Authorize: consul server down writer.go:29: 2021-01-29T19:34:42.828Z [INFO] TestACL_Authorize: shutdown complete writer.go:29: 2021-01-29T19:34:42.844Z [INFO] TestACL_Authorize: Stopping server: protocol=DNS address=127.0.0.1:29604 network=tcp writer.go:29: 2021-01-29T19:34:42.845Z [INFO] TestACL_Authorize: Stopping server: protocol=DNS address=127.0.0.1:29604 network=udp writer.go:29: 2021-01-29T19:34:42.846Z [INFO] TestACL_Authorize: Stopping server: protocol=HTTP address=127.0.0.1:29605 network=tcp === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.811Z [DEBUG] TestACL_Legacy_Get/wrong_id.leader: stopping routine: routine="acl token reaping" === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.843Z [INFO] TestACL_Legacy_Update: Synced node info === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.847Z [WARN] TestACL_Legacy_Get/wrong_id.server.serf.lan: serf: Shutdown without a Leave === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.843Z [DEBUG] TestACL_Legacy_Update.server: Skipping self join check for node since the cluster is too small: node=Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.814Z [DEBUG] TestACL_Legacy_Get/wrong_id.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:42.814Z [DEBUG] TestACL_Legacy_Get/wrong_id.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:42.814Z [DEBUG] TestACL_Legacy_Get/wrong_id: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:42.850Z [DEBUG] TestACL_Legacy_Get/wrong_id: Node info in sync === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.853Z [DEBUG] TestACL_Legacy_Update.acl: dropping node from result due to ACLs: node=Node-4f28edd5-9148-c00f-73a3-5ceedb0b84cb === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.848Z [DEBUG] TestACL_Legacy_Get/wrong_id.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.849Z [WARN] TestACL_Legacy_Get/wrong_id.server.serf.wan: serf: Shutdown without a Leave === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.859Z [INFO] TestACL_Legacy_Update: Requesting shutdown === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.860Z [INFO] TestACL_Legacy_Get/wrong_id.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:42.861Z [INFO] TestACL_Legacy_Get/wrong_id: consul server down === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.861Z [INFO] TestACL_Legacy_Update.server: shutting down server === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:42.862Z [INFO] TestACL_Legacy_Get/wrong_id: shutdown complete writer.go:29: 2021-01-29T19:34:42.863Z [INFO] TestACL_Legacy_Get/wrong_id: Stopping server: protocol=DNS address=127.0.0.1:29652 network=tcp writer.go:29: 2021-01-29T19:34:42.864Z [INFO] TestACL_Legacy_Get/wrong_id: Stopping server: protocol=DNS address=127.0.0.1:29652 network=udp writer.go:29: 2021-01-29T19:34:42.866Z [INFO] TestACL_Legacy_Get/wrong_id: Stopping server: protocol=HTTP address=127.0.0.1:29653 network=tcp === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.863Z [DEBUG] TestACL_Legacy_Update.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:42.867Z [DEBUG] TestACL_Legacy_Update.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.871Z [DEBUG] TestACL_Legacy_Update.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:42.867Z [DEBUG] TestACL_Legacy_Update.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:42.871Z [DEBUG] TestACL_Legacy_Update.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.875Z [WARN] TestACL_Legacy_Update.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:42.875Z [DEBUG] TestACL_Legacy_Update.leader: stopped routine: routine="CA root pruning" === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.887Z [DEBUG] TestACLReplicationStatus.acl: dropping node from result due to ACLs: node=Node-9967ccf3-d605-fbdc-a8a3-567391073d3d === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.884Z [WARN] TestACL_Legacy_Update.server.serf.wan: serf: Shutdown without a Leave === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.890Z [DEBUG] TestACLReplicationStatus.acl: dropping node from result due to ACLs: node=Node-9967ccf3-d605-fbdc-a8a3-567391073d3d writer.go:29: 2021-01-29T19:34:42.891Z [INFO] TestACLReplicationStatus: Requesting shutdown === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.891Z [INFO] TestACL_Legacy_Update.server.router.manager: shutting down === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.892Z [INFO] TestACLReplicationStatus.server: shutting down server writer.go:29: 2021-01-29T19:34:42.893Z [DEBUG] TestACLReplicationStatus.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:42.894Z [DEBUG] TestACLReplicationStatus.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.894Z [DEBUG] TestACLReplicationStatus.leader: stopping routine: routine="CA root pruning" === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.894Z [INFO] TestACL_Legacy_Update: consul server down === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.895Z [WARN] TestACLReplicationStatus.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:42.894Z [DEBUG] TestACLReplicationStatus.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.895Z [INFO] TestACL_Legacy_Update: shutdown complete === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.897Z [DEBUG] TestACLReplicationStatus.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:42.898Z [DEBUG] TestACLReplicationStatus.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:42.899Z [WARN] TestACLReplicationStatus.server.serf.wan: serf: Shutdown without a Leave === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.897Z [INFO] TestACL_Legacy_Update: Stopping server: protocol=DNS address=127.0.0.1:29670 network=tcp writer.go:29: 2021-01-29T19:34:42.901Z [INFO] TestACL_Legacy_Update: Stopping server: protocol=DNS address=127.0.0.1:29670 network=udp === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:42.901Z [INFO] TestACLReplicationStatus.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:42.901Z [INFO] TestACLReplicationStatus: consul server down writer.go:29: 2021-01-29T19:34:42.903Z [INFO] TestACLReplicationStatus: shutdown complete writer.go:29: 2021-01-29T19:34:42.904Z [INFO] TestACLReplicationStatus: Stopping server: protocol=DNS address=127.0.0.1:29658 network=tcp writer.go:29: 2021-01-29T19:34:42.905Z [INFO] TestACLReplicationStatus: Stopping server: protocol=DNS address=127.0.0.1:29658 network=udp writer.go:29: 2021-01-29T19:34:42.906Z [INFO] TestACLReplicationStatus: Stopping server: protocol=HTTP address=127.0.0.1:29659 network=tcp === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:42.902Z [INFO] TestACL_Legacy_Update: Stopping server: protocol=HTTP address=127.0.0.1:29671 network=tcp === CONT TestACL_Authorize writer.go:29: 2021-01-29T19:34:43.170Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="secondary CA roots watch" writer.go:29: 2021-01-29T19:34:43.191Z [DEBUG] TestACL_Authorize.leader: stopped routine: routine="intention replication" writer.go:29: 2021-01-29T19:34:43.348Z [INFO] TestACL_Authorize: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:43.351Z [INFO] TestACL_Authorize: Endpoints down --- PASS: TestACL_Authorize (2.99s) --- PASS: TestACL_Authorize/master-token (0.10s) --- PASS: TestACL_Authorize/master-token/dc1 (0.04s) --- PASS: TestACL_Authorize/master-token/dc2 (0.06s) --- PASS: TestACL_Authorize/custom-token (0.11s) --- PASS: TestACL_Authorize/custom-token/dc1 (0.07s) --- PASS: TestACL_Authorize/custom-token/dc2 (0.03s) --- PASS: TestACL_Authorize/too-many-requests (0.00s) --- PASS: TestACL_Authorize/decode-failure (0.00s) --- PASS: TestACL_Authorize/acl-not-found (0.00s) --- PASS: TestACL_Authorize/local-token-in-secondary-dc (0.06s) --- PASS: TestACL_Authorize/local-token-wrong-dc (0.00s) === CONT TestACL_Legacy_UpdateUpsert === CONT TestACL_Legacy_Get/wrong_id writer.go:29: 2021-01-29T19:34:43.367Z [INFO] TestACL_Legacy_Get/wrong_id: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:43.368Z [INFO] TestACL_Legacy_Get/wrong_id: Endpoints down === RUN TestACL_Legacy_Get/right_id === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.377Z [WARN] TestACL_Legacy_UpdateUpsert: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:34:43.379Z [WARN] TestACL_Legacy_UpdateUpsert: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:43.381Z [DEBUG] TestACL_Legacy_UpdateUpsert.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:43.384Z [DEBUG] TestACL_Legacy_UpdateUpsert.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.390Z [WARN] TestACL_Legacy_Get/right_id: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.393Z [INFO] TestACL_Legacy_UpdateUpsert.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:450d4e68-9acd-7366-1684-f495035529ca Address:127.0.0.1:29669}]" === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.392Z [WARN] TestACL_Legacy_Get/right_id: bootstrap = true: do not enable unless necessary === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.395Z [INFO] TestACL_Legacy_UpdateUpsert.server.raft: entering follower state: follower="Node at 127.0.0.1:29669 [Follower]" leader= === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.396Z [DEBUG] TestACL_Legacy_Get/right_id.tlsutil: Update: version=1 === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.398Z [INFO] TestACL_Legacy_UpdateUpsert.server.serf.wan: serf: EventMemberJoin: Node-450d4e68-9acd-7366-1684-f495035529ca.dc1 127.0.0.1 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.400Z [DEBUG] TestACL_Legacy_Get/right_id.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:43.407Z [INFO] TestACLReplicationStatus: Waiting for endpoints to shut down === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.407Z [INFO] TestACL_Legacy_UpdateUpsert.server.serf.lan: serf: EventMemberJoin: Node-450d4e68-9acd-7366-1684-f495035529ca 127.0.0.1 === CONT TestACLReplicationStatus writer.go:29: 2021-01-29T19:34:43.409Z [INFO] TestACLReplicationStatus: Endpoints down --- PASS: TestACLReplicationStatus (1.09s) === CONT TestACL_Legacy_Disabled_Response === CONT TestACL_Legacy_Update writer.go:29: 2021-01-29T19:34:43.412Z [INFO] TestACL_Legacy_Update: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:43.414Z [INFO] TestACL_Legacy_Update: Endpoints down --- PASS: TestACL_Legacy_Update (0.78s) === CONT TestUiServices/No_Filter --- PASS: TestUiServices (0.82s) --- PASS: TestUiServices/Filtered (0.03s) --- PASS: TestUiServices/No_Filter (0.00s) === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.422Z [INFO] TestACL_Legacy_Get/right_id.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:20ec30e1-a61f-1354-3877-64cd0079e2ec Address:127.0.0.1:29693}]" writer.go:29: 2021-01-29T19:34:43.426Z [INFO] TestACL_Legacy_Get/right_id.server.serf.wan: serf: EventMemberJoin: Node-20ec30e1-a61f-1354-3877-64cd0079e2ec.dc1 127.0.0.1 === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.431Z [INFO] TestACL_Legacy_UpdateUpsert: Started DNS server: address=127.0.0.1:29664 network=udp writer.go:29: 2021-01-29T19:34:43.434Z [INFO] TestACL_Legacy_UpdateUpsert.server: Adding LAN server: server="Node-450d4e68-9acd-7366-1684-f495035529ca (Addr: tcp/127.0.0.1:29669) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.435Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: bootstrap = true: do not enable unless necessary === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.430Z [INFO] TestACL_Legacy_Get/right_id.server.serf.lan: serf: EventMemberJoin: Node-20ec30e1-a61f-1354-3877-64cd0079e2ec 127.0.0.1 writer.go:29: 2021-01-29T19:34:43.444Z [INFO] TestACL_Legacy_Get/right_id: Started DNS server: address=127.0.0.1:29688 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.438Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.tlsutil: Update: version=1 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.440Z [INFO] TestACL_Legacy_Get/right_id.server.raft: entering follower state: follower="Node at 127.0.0.1:29693 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.449Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.449Z [INFO] TestACL_Legacy_Get/right_id.server: Adding LAN server: server="Node-20ec30e1-a61f-1354-3877-64cd0079e2ec (Addr: tcp/127.0.0.1:29693) (DC: dc1)" writer.go:29: 2021-01-29T19:34:43.450Z [INFO] TestACL_Legacy_Get/right_id.server: Handled event for server in area: event=member-join server=Node-20ec30e1-a61f-1354-3877-64cd0079e2ec.dc1 area=wan writer.go:29: 2021-01-29T19:34:43.452Z [INFO] TestACL_Legacy_Get/right_id: Started DNS server: address=127.0.0.1:29688 network=tcp === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.457Z [WARN] TestACL_Legacy_Disabled_Response: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:43.460Z [DEBUG] TestACL_Legacy_Disabled_Response.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.463Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5ed52b07-e53b-0293-f81f-113d4d7118f9 Address:127.0.0.1:29681}]" === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.434Z [INFO] TestACL_Legacy_UpdateUpsert.server: Handled event for server in area: event=member-join server=Node-450d4e68-9acd-7366-1684-f495035529ca.dc1 area=wan writer.go:29: 2021-01-29T19:34:43.437Z [INFO] TestACL_Legacy_UpdateUpsert: Started DNS server: address=127.0.0.1:29664 network=tcp writer.go:29: 2021-01-29T19:34:43.487Z [INFO] TestACL_Legacy_UpdateUpsert: Started HTTP server: address=127.0.0.1:29665 network=tcp === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.482Z [DEBUG] TestACL_Legacy_Disabled_Response.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.491Z [INFO] TestACL_Legacy_UpdateUpsert: started state syncer writer.go:29: 2021-01-29T19:34:43.489Z [WARN] TestACL_Legacy_UpdateUpsert.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.475Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: entering follower state: follower="Node at 127.0.0.1:29681 [Follower]" leader= === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.497Z [INFO] TestACL_Legacy_UpdateUpsert.server.raft: entering candidate state: node="Node at 127.0.0.1:29669 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.485Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.495Z [WARN] TestACL_Legacy_Get/right_id.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.489Z [DEBUG] TestACL_Legacy_UpdateUpsert.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.507Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:43.511Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Started DNS server: address=127.0.0.1:29676 network=udp === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.502Z [INFO] TestACL_Legacy_Get/right_id.server.raft: entering candidate state: node="Node at 127.0.0.1:29693 [Candidate]" term=2 === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.511Z [INFO] TestACL_Legacy_Disabled_Response.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8f1fb7e3-80fe-a256-8b3e-6de0d7d95778 Address:127.0.0.1:29699}]" === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.505Z [DEBUG] TestACL_Legacy_Get/right_id.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:43.502Z [INFO] TestACL_Legacy_Get/right_id: Started HTTP server: address=127.0.0.1:29689 network=tcp writer.go:29: 2021-01-29T19:34:43.519Z [INFO] TestACL_Legacy_Get/right_id: started state syncer === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.520Z [INFO] TestACL_Legacy_Disabled_Response.server.raft: entering follower state: follower="Node at 127.0.0.1:29699 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.525Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29681) (DC: dc1)" writer.go:29: 2021-01-29T19:34:43.528Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:43.536Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.532Z [DEBUG] TestACL_Legacy_UpdateUpsert.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.538Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: entering candidate state: node="Node at 127.0.0.1:29681 [Candidate]" term=2 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.545Z [DEBUG] TestACL_Legacy_Get/right_id.server.raft: votes: needed=1 === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.543Z [INFO] TestACL_Legacy_Disabled_Response.server.serf.wan: serf: EventMemberJoin: Node-8f1fb7e3-80fe-a256-8b3e-6de0d7d95778.dc1 127.0.0.1 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.550Z [DEBUG] TestACL_Legacy_Get/right_id.server.raft: vote granted: from=20ec30e1-a61f-1354-3877-64cd0079e2ec term=2 tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.554Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Started DNS server: address=127.0.0.1:29676 network=tcp === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.555Z [INFO] TestACL_Legacy_Get/right_id.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.561Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Started HTTP server: address=127.0.0.1:29677 network=tcp === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.560Z [INFO] TestACL_Legacy_Get/right_id.server.raft: entering leader state: leader="Node at 127.0.0.1:29693 [Leader]" === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.539Z [DEBUG] TestACL_Legacy_UpdateUpsert.server.raft: vote granted: from=450d4e68-9acd-7366-1684-f495035529ca term=2 tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.564Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: votes: needed=1 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.570Z [INFO] TestACL_Legacy_Get/right_id.server: cluster leadership acquired === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.565Z [WARN] TestACL_Legacy_Disabled_Response.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:43.582Z [INFO] TestACL_Legacy_Disabled_Response.server.raft: entering candidate state: node="Node at 127.0.0.1:29699 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.587Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: vote granted: from=5ed52b07-e53b-0293-f81f-113d4d7118f9 term=2 tally=1 === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.566Z [INFO] TestACL_Legacy_Disabled_Response.server.serf.lan: serf: EventMemberJoin: Node-8f1fb7e3-80fe-a256-8b3e-6de0d7d95778 127.0.0.1 === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.588Z [INFO] TestACL_Legacy_UpdateUpsert.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.564Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: started state syncer writer.go:29: 2021-01-29T19:34:43.594Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:43.595Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: entering leader state: leader="Node at 127.0.0.1:29681 [Leader]" writer.go:29: 2021-01-29T19:34:43.596Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: cluster leadership acquired === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.592Z [INFO] TestACL_Legacy_UpdateUpsert.server.raft: entering leader state: leader="Node at 127.0.0.1:29669 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.598Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: New leader elected: payload=test-node === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.591Z [DEBUG] TestACL_Legacy_UpdateUpsert.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:43.605Z [INFO] TestACL_Legacy_UpdateUpsert.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:43.606Z [INFO] TestACL_Legacy_UpdateUpsert.server: New leader elected: payload=Node-450d4e68-9acd-7366-1684-f495035529ca === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.571Z [INFO] TestACL_Legacy_Get/right_id.server: New leader elected: payload=Node-20ec30e1-a61f-1354-3877-64cd0079e2ec === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.608Z [INFO] TestACL_Legacy_Disabled_Response.server: Handled event for server in area: event=member-join server=Node-8f1fb7e3-80fe-a256-8b3e-6de0d7d95778.dc1 area=wan writer.go:29: 2021-01-29T19:34:43.609Z [INFO] TestACL_Legacy_Disabled_Response: Started DNS server: address=127.0.0.1:29694 network=udp === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.587Z [INFO] TestACL_Legacy_Get/right_id.server: initializing acls === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.609Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29681 === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.613Z [INFO] TestACL_Legacy_UpdateUpsert.server: initializing acls === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.621Z [INFO] TestACL_Legacy_Disabled_Response: Started DNS server: address=127.0.0.1:29694 network=tcp === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.630Z [INFO] TestACL_Legacy_UpdateUpsert.server: Created ACL 'global-management' policy === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.608Z [ERROR] TestACL_Legacy_Get/right_id.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.631Z [WARN] TestACL_Legacy_UpdateUpsert.server: Configuring a non-UUID master token is deprecated === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.626Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Synced node info === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.587Z [WARN] TestACL_Legacy_Disabled_Response.server.raft: unable to get address for sever, using fallback address: id=8f1fb7e3-80fe-a256-8b3e-6de0d7d95778 fallback=127.0.0.1:29699 error="Could not find address for server id 8f1fb7e3-80fe-a256-8b3e-6de0d7d95778" === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.614Z [INFO] TestACL_Legacy_Get/right_id.server: initializing acls writer.go:29: 2021-01-29T19:34:43.638Z [WARN] TestACL_Legacy_Get/right_id.server: Configuring a non-UUID master token is deprecated === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.612Z [INFO] TestACL_Legacy_Disabled_Response.server: Adding LAN server: server="Node-8f1fb7e3-80fe-a256-8b3e-6de0d7d95778 (Addr: tcp/127.0.0.1:29699) (DC: dc1)" === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.626Z [INFO] TestACL_Legacy_Get/right_id.server: Created ACL 'global-management' policy === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.642Z [INFO] TestACL_Legacy_UpdateUpsert.server: Bootstrapped ACL master token from configuration === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.643Z [WARN] TestACL_Legacy_Get/right_id.server: Configuring a non-UUID master token is deprecated === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.645Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.643Z [INFO] TestACL_Legacy_Disabled_Response: Started HTTP server: address=127.0.0.1:29695 network=tcp === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.649Z [INFO] TestACL_Legacy_UpdateUpsert.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:43.651Z [INFO] TestACL_Legacy_UpdateUpsert.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:43.653Z [INFO] TestACL_Legacy_UpdateUpsert.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:43.658Z [INFO] TestACL_Legacy_UpdateUpsert.server.serf.lan: serf: EventMemberUpdate: Node-450d4e68-9acd-7366-1684-f495035529ca writer.go:29: 2021-01-29T19:34:43.659Z [INFO] TestACL_Legacy_UpdateUpsert.server.serf.wan: serf: EventMemberUpdate: Node-450d4e68-9acd-7366-1684-f495035529ca.dc1 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.657Z [INFO] TestACL_Legacy_Get/right_id.server: Bootstrapped ACL master token from configuration === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.649Z [INFO] TestACL_Legacy_Disabled_Response: started state syncer === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.662Z [INFO] TestACL_Legacy_UpdateUpsert.server: Updating LAN server: server="Node-450d4e68-9acd-7366-1684-f495035529ca (Addr: tcp/127.0.0.1:29669) (DC: dc1)" writer.go:29: 2021-01-29T19:34:43.664Z [INFO] TestACL_Legacy_UpdateUpsert.server: Handled event for server in area: event=member-update server=Node-450d4e68-9acd-7366-1684-f495035529ca.dc1 area=wan === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.659Z [DEBUG] TestACL_Legacy_Disabled_Response.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:43.668Z [DEBUG] TestACL_Legacy_Disabled_Response.server.raft: vote granted: from=8f1fb7e3-80fe-a256-8b3e-6de0d7d95778 term=2 tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.672Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.676Z [INFO] TestACL_Legacy_Disabled_Response.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.677Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.leader: started routine: routine="CA root pruning" === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.660Z [INFO] TestACL_Legacy_Get/right_id.server: Bootstrapped ACL master token from configuration === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.682Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.685Z [INFO] TestACL_Legacy_Get/right_id.leader: started routine: routine="legacy ACL token upgrade" === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.679Z [INFO] TestACL_Legacy_Disabled_Response.server.raft: entering leader state: leader="Node at 127.0.0.1:29699 [Leader]" writer.go:29: 2021-01-29T19:34:43.688Z [INFO] TestACL_Legacy_Disabled_Response.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:43.686Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: member joined, marking health alive: member=test-node === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.692Z [INFO] TestACL_Legacy_Disabled_Response.server: New leader elected: payload=Node-8f1fb7e3-80fe-a256-8b3e-6de0d7d95778 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.687Z [INFO] TestACL_Legacy_Get/right_id.leader: started routine: routine="acl token reaping" === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.696Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.669Z [INFO] TestACL_Legacy_Get/right_id.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:43.704Z [DEBUG] TestACL_Legacy_Get/right_id.server: transitioning out of legacy ACL mode === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.700Z [DEBUG] TestACL_Legacy_Disabled_Response.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29699 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.708Z [INFO] TestACL_Legacy_Get/right_id.server.serf.lan: serf: EventMemberUpdate: Node-20ec30e1-a61f-1354-3877-64cd0079e2ec writer.go:29: 2021-01-29T19:34:43.713Z [INFO] TestACL_Legacy_Get/right_id.server.serf.wan: serf: EventMemberUpdate: Node-20ec30e1-a61f-1354-3877-64cd0079e2ec.dc1 === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.723Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.724Z [INFO] TestACL_Legacy_Get/right_id.server: Handled event for server in area: event=member-update server=Node-20ec30e1-a61f-1354-3877-64cd0079e2ec.dc1 area=wan writer.go:29: 2021-01-29T19:34:43.721Z [INFO] TestACL_Legacy_Get/right_id.server: Updating LAN server: server="Node-20ec30e1-a61f-1354-3877-64cd0079e2ec (Addr: tcp/127.0.0.1:29693) (DC: dc1)" writer.go:29: 2021-01-29T19:34:43.723Z [INFO] TestACL_Legacy_Get/right_id.server.serf.lan: serf: EventMemberUpdate: Node-20ec30e1-a61f-1354-3877-64cd0079e2ec writer.go:29: 2021-01-29T19:34:43.734Z [INFO] TestACL_Legacy_Get/right_id.server.serf.wan: serf: EventMemberUpdate: Node-20ec30e1-a61f-1354-3877-64cd0079e2ec.dc1 writer.go:29: 2021-01-29T19:34:43.738Z [INFO] TestACL_Legacy_Get/right_id.server: Handled event for server in area: event=member-update server=Node-20ec30e1-a61f-1354-3877-64cd0079e2ec.dc1 area=wan writer.go:29: 2021-01-29T19:34:43.737Z [INFO] TestACL_Legacy_Get/right_id.server: Updating LAN server: server="Node-20ec30e1-a61f-1354-3877-64cd0079e2ec (Addr: tcp/127.0.0.1:29693) (DC: dc1)" === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.739Z [INFO] TestACL_Legacy_UpdateUpsert.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:43.746Z [INFO] TestACL_Legacy_UpdateUpsert.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:43.752Z [DEBUG] TestACL_Legacy_UpdateUpsert.server: Skipping self join check for node since the cluster is too small: node=Node-450d4e68-9acd-7366-1684-f495035529ca writer.go:29: 2021-01-29T19:34:43.754Z [INFO] TestACL_Legacy_UpdateUpsert.server: member joined, marking health alive: member=Node-450d4e68-9acd-7366-1684-f495035529ca writer.go:29: 2021-01-29T19:34:43.759Z [DEBUG] TestACL_Legacy_UpdateUpsert.server: Skipping self join check for node since the cluster is too small: node=Node-450d4e68-9acd-7366-1684-f495035529ca === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.761Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.764Z [INFO] TestACL_Legacy_Disabled_Response.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:43.766Z [INFO] TestACL_Legacy_Disabled_Response.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:43.770Z [DEBUG] TestACL_Legacy_Disabled_Response.server: Skipping self join check for node since the cluster is too small: node=Node-8f1fb7e3-80fe-a256-8b3e-6de0d7d95778 writer.go:29: 2021-01-29T19:34:43.772Z [INFO] TestACL_Legacy_Disabled_Response.server: member joined, marking health alive: member=Node-8f1fb7e3-80fe-a256-8b3e-6de0d7d95778 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.777Z [INFO] TestACL_Legacy_Get/right_id.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:43.778Z [INFO] TestACL_Legacy_Get/right_id.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:43.779Z [DEBUG] TestACL_Legacy_Get/right_id.server: Skipping self join check for node since the cluster is too small: node=Node-20ec30e1-a61f-1354-3877-64cd0079e2ec writer.go:29: 2021-01-29T19:34:43.779Z [INFO] TestACL_Legacy_Get/right_id.server: member joined, marking health alive: member=Node-20ec30e1-a61f-1354-3877-64cd0079e2ec === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.795Z [DEBUG] TestACL_Legacy_UpdateUpsert.acl: dropping node from result due to ACLs: node=Node-450d4e68-9acd-7366-1684-f495035529ca writer.go:29: 2021-01-29T19:34:43.808Z [DEBUG] TestACL_Legacy_UpdateUpsert.acl: dropping node from result due to ACLs: node=Node-450d4e68-9acd-7366-1684-f495035529ca === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.812Z [DEBUG] TestACL_Legacy_Get/right_id.server: Skipping self join check for node since the cluster is too small: node=Node-20ec30e1-a61f-1354-3877-64cd0079e2ec writer.go:29: 2021-01-29T19:34:43.813Z [DEBUG] TestACL_Legacy_Get/right_id.server: Skipping self join check for node since the cluster is too small: node=Node-20ec30e1-a61f-1354-3877-64cd0079e2ec === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.815Z [INFO] TestACL_Legacy_UpdateUpsert: Requesting shutdown writer.go:29: 2021-01-29T19:34:43.816Z [INFO] TestACL_Legacy_UpdateUpsert.server: shutting down server writer.go:29: 2021-01-29T19:34:43.818Z [DEBUG] TestACL_Legacy_UpdateUpsert.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:43.820Z [DEBUG] TestACL_Legacy_UpdateUpsert.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:43.821Z [DEBUG] TestACL_Legacy_UpdateUpsert.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:43.822Z [WARN] TestACL_Legacy_UpdateUpsert.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:43.818Z [ERROR] TestACL_Legacy_UpdateUpsert.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:43.825Z [DEBUG] TestACL_Legacy_UpdateUpsert.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:43.825Z [DEBUG] TestACL_Legacy_UpdateUpsert.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:43.825Z [DEBUG] TestACL_Legacy_UpdateUpsert.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:43.828Z [WARN] TestACL_Legacy_UpdateUpsert.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:43.833Z [INFO] TestACL_Legacy_UpdateUpsert.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:43.839Z [INFO] TestACL_Legacy_UpdateUpsert: consul server down writer.go:29: 2021-01-29T19:34:43.843Z [INFO] TestACL_Legacy_UpdateUpsert: shutdown complete === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.846Z [DEBUG] TestACL_Legacy_Get/right_id.acl: dropping node from result due to ACLs: node=Node-20ec30e1-a61f-1354-3877-64cd0079e2ec === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.847Z [INFO] TestACL_Legacy_UpdateUpsert: Stopping server: protocol=DNS address=127.0.0.1:29664 network=tcp writer.go:29: 2021-01-29T19:34:43.843Z [DEBUG] TestACL_Legacy_UpdateUpsert: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:43.854Z [WARN] TestACL_Legacy_UpdateUpsert: Syncing node info failed.: error="No cluster leader" writer.go:29: 2021-01-29T19:34:43.856Z [ERROR] TestACL_Legacy_UpdateUpsert.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:43.854Z [INFO] TestACL_Legacy_UpdateUpsert: Stopping server: protocol=DNS address=127.0.0.1:29664 network=udp === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.857Z [INFO] TestACL_Legacy_Get/right_id: Requesting shutdown === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:43.860Z [INFO] TestACL_Legacy_UpdateUpsert: Stopping server: protocol=HTTP address=127.0.0.1:29665 network=tcp === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.861Z [INFO] TestACL_Legacy_Get/right_id.server: shutting down server writer.go:29: 2021-01-29T19:34:43.873Z [DEBUG] TestACL_Legacy_Get/right_id.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:43.874Z [DEBUG] TestACL_Legacy_Get/right_id.leader: stopping routine: routine="acl token reaping" === RUN TestACL_Legacy_Disabled_Response/0 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.876Z [DEBUG] TestACL_Legacy_Get/right_id.leader: stopping routine: routine="CA root pruning" === RUN TestACL_Legacy_Disabled_Response/1 === RUN TestACL_Legacy_Disabled_Response/2 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.878Z [WARN] TestACL_Legacy_Get/right_id.server.serf.lan: serf: Shutdown without a Leave === RUN TestACL_Legacy_Disabled_Response/3 === RUN TestACL_Legacy_Disabled_Response/4 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.881Z [DEBUG] TestACL_Legacy_Get/right_id.leader: stopped routine: routine="legacy ACL token upgrade" === RUN TestACL_Legacy_Disabled_Response/5 === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.882Z [DEBUG] TestACL_Legacy_Get/right_id.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:43.883Z [DEBUG] TestACL_Legacy_Get/right_id.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:43.883Z [WARN] TestACL_Legacy_Get/right_id.server.serf.wan: serf: Shutdown without a Leave === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.905Z [INFO] TestACL_Legacy_Disabled_Response: Requesting shutdown === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.910Z [INFO] TestACL_Legacy_Get/right_id.server.router.manager: shutting down === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.910Z [INFO] TestACL_Legacy_Disabled_Response.server: shutting down server === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.913Z [INFO] TestACL_Legacy_Get/right_id: consul server down writer.go:29: 2021-01-29T19:34:43.915Z [INFO] TestACL_Legacy_Get/right_id: shutdown complete === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.913Z [DEBUG] TestACL_Legacy_Disabled_Response.leader: stopping routine: routine="CA root pruning" === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.917Z [INFO] TestACL_Legacy_Get/right_id: Stopping server: protocol=DNS address=127.0.0.1:29688 network=tcp writer.go:29: 2021-01-29T19:34:43.919Z [INFO] TestACL_Legacy_Get/right_id: Stopping server: protocol=DNS address=127.0.0.1:29688 network=udp === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.917Z [WARN] TestACL_Legacy_Disabled_Response.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:43.919Z [ERROR] TestACL_Legacy_Disabled_Response.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:43.920Z [INFO] TestACL_Legacy_Get/right_id: Stopping server: protocol=HTTP address=127.0.0.1:29689 network=tcp === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:43.920Z [DEBUG] TestACL_Legacy_Disabled_Response.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:43.926Z [WARN] TestACL_Legacy_Disabled_Response.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:43.948Z [INFO] TestACL_Legacy_Disabled_Response.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:43.949Z [INFO] TestACL_Legacy_Disabled_Response: consul server down writer.go:29: 2021-01-29T19:34:43.956Z [INFO] TestACL_Legacy_Disabled_Response: shutdown complete writer.go:29: 2021-01-29T19:34:43.959Z [INFO] TestACL_Legacy_Disabled_Response: Stopping server: protocol=DNS address=127.0.0.1:29694 network=tcp writer.go:29: 2021-01-29T19:34:43.962Z [INFO] TestACL_Legacy_Disabled_Response: Stopping server: protocol=DNS address=127.0.0.1:29694 network=udp writer.go:29: 2021-01-29T19:34:43.967Z [INFO] TestACL_Legacy_Disabled_Response: Stopping server: protocol=HTTP address=127.0.0.1:29695 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:44.128Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=951.643µs client=127.0.0.1:56823 client_network=udp writer.go:29: 2021-01-29T19:34:44.130Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=1.003345ms client=127.0.0.1:52202 client_network=udp writer.go:29: 2021-01-29T19:34:44.131Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.dns: request served from client: name=b45fdf77-7ed4-eb6b-aef9-f67dd405ca25.query.consul. type=A class=IN latency=620.501µs client=127.0.0.1:47842 client_network=udp writer.go:29: 2021-01-29T19:34:44.131Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Requesting shutdown writer.go:29: 2021-01-29T19:34:44.134Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: shutting down server writer.go:29: 2021-01-29T19:34:44.135Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:44.136Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:44.136Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:44.138Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:44.142Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:44.143Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: consul server down writer.go:29: 2021-01-29T19:34:44.145Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: shutdown complete writer.go:29: 2021-01-29T19:34:44.147Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Stopping server: protocol=DNS address=127.0.0.1:29676 network=tcp writer.go:29: 2021-01-29T19:34:44.148Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Stopping server: protocol=DNS address=127.0.0.1:29676 network=udp writer.go:29: 2021-01-29T19:34:44.150Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Stopping server: protocol=HTTP address=127.0.0.1:29677 network=tcp === CONT TestACL_Legacy_UpdateUpsert writer.go:29: 2021-01-29T19:34:44.362Z [INFO] TestACL_Legacy_UpdateUpsert: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:44.364Z [INFO] TestACL_Legacy_UpdateUpsert: Endpoints down --- PASS: TestACL_Legacy_UpdateUpsert (1.00s) === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:44.391Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:44.393Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:44.396Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:44.404Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8f0c8048-1e6c-8e63-7a33-b72c6c28f8d3 Address:127.0.0.1:29687}]" writer.go:29: 2021-01-29T19:34:44.408Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29687 [Follower]" leader= writer.go:29: 2021-01-29T19:34:44.409Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:44.413Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:44.417Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:44.418Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29687) (DC: dc1)" writer.go:29: 2021-01-29T19:34:44.418Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Started DNS server: address=127.0.0.1:29682 network=udp writer.go:29: 2021-01-29T19:34:44.421Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Started DNS server: address=127.0.0.1:29682 network=tcp writer.go:29: 2021-01-29T19:34:44.428Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Started HTTP server: address=127.0.0.1:29683 network=tcp writer.go:29: 2021-01-29T19:34:44.429Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: started state syncer === CONT TestACL_Legacy_Get/right_id writer.go:29: 2021-01-29T19:34:44.442Z [INFO] TestACL_Legacy_Get/right_id: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:44.443Z [INFO] TestACL_Legacy_Get/right_id: Endpoints down --- PASS: TestACL_Legacy_Get (2.12s) --- PASS: TestACL_Legacy_Get/wrong_id (1.04s) --- PASS: TestACL_Legacy_Get/right_id (1.07s) === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.455Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:44.455Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.456Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:44.457Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:44.456Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29687 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:44.460Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:44.461Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: vote granted: from=8f0c8048-1e6c-8e63-7a33-b72c6c28f8d3 term=2 tally=1 writer.go:29: 2021-01-29T19:34:44.462Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:44.463Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29687 [Leader]" writer.go:29: 2021-01-29T19:34:44.464Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:44.465Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:44.466Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29687 === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:44.473Z [INFO] TestACL_Legacy_Disabled_Response: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.474Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9f375bea-a468-b4fc-a277-3507c99ab5cb Address:127.0.0.1:29711}]" === CONT TestACL_Legacy_Disabled_Response writer.go:29: 2021-01-29T19:34:44.474Z [INFO] TestACL_Legacy_Disabled_Response: Endpoints down --- PASS: TestACL_Legacy_Disabled_Response (1.06s) --- PASS: TestACL_Legacy_Disabled_Response/0 (0.00s) --- PASS: TestACL_Legacy_Disabled_Response/1 (0.00s) --- PASS: TestACL_Legacy_Disabled_Response/2 (0.00s) --- PASS: TestACL_Legacy_Disabled_Response/3 (0.00s) --- PASS: TestACL_Legacy_Disabled_Response/4 (0.00s) --- PASS: TestACL_Legacy_Disabled_Response/5 (0.00s) === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.476Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29711 [Follower]" leader= writer.go:29: 2021-01-29T19:34:44.477Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:44.487Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:44.489Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.491Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29706 network=udp writer.go:29: 2021-01-29T19:34:44.493Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29711) (DC: dc1)" writer.go:29: 2021-01-29T19:34:44.495Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29706 network=tcp writer.go:29: 2021-01-29T19:34:44.494Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:44.498Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Started HTTP server: address=127.0.0.1:29707 network=tcp writer.go:29: 2021-01-29T19:34:44.500Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:44.502Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:44.504Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:44.506Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:44.508Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.511Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:44.512Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:44.513Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:44.517Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a91fac03-0dbe-73ac-1361-3b908a179a84 Address:127.0.0.1:29705}]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.517Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.519Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29705 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.519Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29711 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.522Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.522Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:44.524Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: vote granted: from=9f375bea-a468-b4fc-a277-3507c99ab5cb term=2 tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.525Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.526Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:44.527Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29711 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.528Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29705) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.528Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.529Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.529Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:44.531Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29711 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.533Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29700 network=tcp writer.go:29: 2021-01-29T19:34:44.535Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29700 network=udp writer.go:29: 2021-01-29T19:34:44.536Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Started HTTP server: address=127.0.0.1:29701 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.537Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.537Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.563Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.579Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.576Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.580Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29705 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:44.581Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.580Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.582Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: vote granted: from=a91fac03-0dbe-73ac-1361-3b908a179a84 term=2 tally=1 writer.go:29: 2021-01-29T19:34:44.583Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:44.583Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29705 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.583Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.584Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:44.585Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:44.585Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29705 writer.go:29: 2021-01-29T19:34:44.593Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:44.599Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:44.601Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:44.602Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:44.604Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: member joined, marking health alive: member=test-node writer.go:29: 2021-01-29T19:34:44.628Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:44.631Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Synced node info writer.go:29: 2021-01-29T19:34:44.633Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:44.646Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:44.649Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:44.652Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:44.654Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.691Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:44.692Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:44.694Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:44.699Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6f13b1d9-5998-12d2-1633-f89df8ddd7fc Address:127.0.0.1:29717}]" writer.go:29: 2021-01-29T19:34:44.702Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29717 [Follower]" leader= writer.go:29: 2021-01-29T19:34:44.703Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:44.709Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:44.715Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29712 network=udp writer.go:29: 2021-01-29T19:34:44.718Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29717) (DC: dc1)" writer.go:29: 2021-01-29T19:34:44.720Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29712 network=tcp writer.go:29: 2021-01-29T19:34:44.720Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:44.732Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Started HTTP server: address=127.0.0.1:29713 network=tcp writer.go:29: 2021-01-29T19:34:44.741Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: started state syncer writer.go:29: 2021-01-29T19:34:44.764Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:44.768Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29717 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:44.772Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:44.773Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: vote granted: from=6f13b1d9-5998-12d2-1633-f89df8ddd7fc term=2 tally=1 writer.go:29: 2021-01-29T19:34:44.774Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:44.775Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29717 [Leader]" writer.go:29: 2021-01-29T19:34:44.776Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:44.777Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:44.778Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.777Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:44.778Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.786Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29717 writer.go:29: 2021-01-29T19:34:44.805Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:44.823Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:44.825Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:44.826Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:44.828Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:44.905Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=699.426µs client=127.0.0.1:36565 client_network=udp writer.go:29: 2021-01-29T19:34:44.907Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=1.456407ms client=127.0.0.1:54008 client_network=udp writer.go:29: 2021-01-29T19:34:44.908Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.dns: request served from client: name=c279c693-ad9a-67b9-ccbb-0b10946e1df4.query.consul. type=A class=IN latency=1.202603ms client=127.0.0.1:34209 client_network=udp writer.go:29: 2021-01-29T19:34:44.908Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:44.914Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:44.916Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.921Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:44.924Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:44.919Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:44.919Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:44.932Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:44.937Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:44.937Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: consul server down writer.go:29: 2021-01-29T19:34:44.939Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:44.940Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Stopping server: protocol=DNS address=127.0.0.1:29682 network=tcp writer.go:29: 2021-01-29T19:34:44.941Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Stopping server: protocol=DNS address=127.0.0.1:29682 network=udp writer.go:29: 2021-01-29T19:34:44.942Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29683 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:44.973Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=1.747343ms client=127.0.0.1:40242 client_network=udp writer.go:29: 2021-01-29T19:34:44.977Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=744.777µs client=127.0.0.1:55534 client_network=udp writer.go:29: 2021-01-29T19:34:44.978Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=b9dd0cce-c957-51e9-22fa-2e8e5a158ed0.query.consul. type=A class=IN latency=651.118µs client=127.0.0.1:56229 client_network=udp writer.go:29: 2021-01-29T19:34:44.978Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:44.993Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:44.994Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:44.994Z [ERROR] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:44.995Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:44.995Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.003Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=5.550179ms client=127.0.0.1:56319 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.008Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:45.010Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:45.014Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: consul server down writer.go:29: 2021-01-29T19:34:45.015Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: shutdown complete === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.015Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=2.264709ms client=127.0.0.1:49459 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.017Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29712 network=tcp writer.go:29: 2021-01-29T19:34:45.018Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29712 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.017Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=711ce5de-2b7f-8fd7-cabf-2c3531ea707c.query.consul. type=ANY class=IN latency=1.853093ms client=127.0.0.1:58433 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.020Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29713 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.017Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:45.022Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:45.023Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:45.028Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:45.028Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:45.039Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:45.054Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:45.054Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: consul server down writer.go:29: 2021-01-29T19:34:45.060Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:45.062Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29706 network=tcp writer.go:29: 2021-01-29T19:34:45.066Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29706 network=udp writer.go:29: 2021-01-29T19:34:45.067Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29707 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.076Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=2.59717ms client=127.0.0.1:39077 client_network=udp writer.go:29: 2021-01-29T19:34:45.077Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=800.773µs client=127.0.0.1:54013 client_network=udp writer.go:29: 2021-01-29T19:34:45.080Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=90aefd06-ae5e-ce0d-cf4c-3dd98dbf5552.query.consul. type=AAAA class=IN latency=2.054372ms client=127.0.0.1:51046 client_network=udp writer.go:29: 2021-01-29T19:34:45.080Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:45.082Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:45.083Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:45.084Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:45.084Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:45.086Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:45.088Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:45.089Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: consul server down writer.go:29: 2021-01-29T19:34:45.090Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:45.091Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29700 network=tcp writer.go:29: 2021-01-29T19:34:45.093Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29700 network=udp writer.go:29: 2021-01-29T19:34:45.094Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29701 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:45.443Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:45.445Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.464Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:45.467Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:45.469Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:45.493Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3b1ae6ee-7442-70c3-3bb9-6db3958bf44a Address:127.0.0.1:29723}]" writer.go:29: 2021-01-29T19:34:45.497Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:45.500Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:45.504Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29723 [Follower]" leader= writer.go:29: 2021-01-29T19:34:45.509Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29718 network=udp writer.go:29: 2021-01-29T19:34:45.515Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29718 network=tcp writer.go:29: 2021-01-29T19:34:45.511Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.526Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:45.532Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.518Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29723) (DC: dc1)" writer.go:29: 2021-01-29T19:34:45.536Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Started HTTP server: address=127.0.0.1:29719 network=tcp writer.go:29: 2021-01-29T19:34:45.540Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: started state syncer writer.go:29: 2021-01-29T19:34:45.561Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:45.565Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29723 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.572Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:45.574Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.577Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:45.578Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:45.580Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.590Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.597Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:45.599Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.594Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: vote granted: from=3b1ae6ee-7442-70c3-3bb9-6db3958bf44a term=2 tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.600Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:45.603Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29723 [Leader]" writer.go:29: 2021-01-29T19:34:45.610Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:45.614Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:45.631Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29723 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.657Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.664Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.667Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1e6f48ac-83ee-453a-b4ca-33d921570ebd Address:127.0.0.1:29735}]" writer.go:29: 2021-01-29T19:34:45.681Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.676Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.695Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29735 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.725Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:45.734Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.735Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.744Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.763Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.769Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2776a64c-7fdd-ba9d-6360-f0a49bdc8b6a Address:127.0.0.1:29729}]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.778Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.780Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29729 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.788Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.786Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.789Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:45.790Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.792Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.803Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:45.810Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29735 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.799Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29724 network=udp writer.go:29: 2021-01-29T19:34:45.810Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29729) (DC: dc1)" writer.go:29: 2021-01-29T19:34:45.811Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.815Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: unable to get address for sever, using fallback address: id=1e6f48ac-83ee-453a-b4ca-33d921570ebd fallback=127.0.0.1:29735 error="Could not find address for server id 1e6f48ac-83ee-453a-b4ca-33d921570ebd" writer.go:29: 2021-01-29T19:34:45.818Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:45.819Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: vote granted: from=1e6f48ac-83ee-453a-b4ca-33d921570ebd term=2 tally=1 writer.go:29: 2021-01-29T19:34:45.823Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.820Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:363310ea-d976-bbf9-1b73-bee74ba66a5c Address:127.0.0.1:29747}]" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.829Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29735 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.834Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29724 network=tcp writer.go:29: 2021-01-29T19:34:45.843Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Started HTTP server: address=127.0.0.1:29725 network=tcp writer.go:29: 2021-01-29T19:34:45.844Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.842Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29747 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.851Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:45.855Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29729 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:45.862Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.861Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.867Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: vote granted: from=2776a64c-7fdd-ba9d-6360-f0a49bdc8b6a term=2 tally=1 writer.go:29: 2021-01-29T19:34:45.881Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:45.882Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29729 [Leader]" writer.go:29: 2021-01-29T19:34:45.883Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.885Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.885Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.888Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29742 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.884Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.891Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29747) (DC: dc1)" writer.go:29: 2021-01-29T19:34:45.892Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.894Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29735) (DC: dc1)" writer.go:29: 2021-01-29T19:34:45.897Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.892Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29729 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.899Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:45.899Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29730 network=tcp writer.go:29: 2021-01-29T19:34:45.903Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Started DNS server: address=127.0.0.1:29730 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.903Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:45.906Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29747 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:45.903Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29742 network=tcp writer.go:29: 2021-01-29T19:34:45.909Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.905Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29735 writer.go:29: 2021-01-29T19:34:45.897Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.912Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: vote granted: from=363310ea-d976-bbf9-1b73-bee74ba66a5c term=2 tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.910Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Started HTTP server: address=127.0.0.1:29731 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.916Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.916Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.919Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.919Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29747 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.917Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.918Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.921Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:45.916Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Started HTTP server: address=127.0.0.1:29743 network=tcp writer.go:29: 2021-01-29T19:34:45.925Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.930Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.926Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29747 writer.go:29: 2021-01-29T19:34:45.925Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.931Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.930Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.933Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.934Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.935Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.936Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.936Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.936Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:45.938Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:45.939Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.942Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:45.943Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.941Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:45.944Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:45.945Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:45.946Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.042Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=1.769943ms client=127.0.0.1:56574 client_network=udp writer.go:29: 2021-01-29T19:34:46.045Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=781.996µs client=127.0.0.1:42938 client_network=udp writer.go:29: 2021-01-29T19:34:46.047Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=41f810fb-0f99-eb1e-dc67-aa9a82eb42ca.query.consul. type=ANY class=IN latency=659.518µs client=127.0.0.1:60917 client_network=udp writer.go:29: 2021-01-29T19:34:46.048Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:46.049Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:46.050Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.051Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:46.052Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.054Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:46.056Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:46.057Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: consul server down writer.go:29: 2021-01-29T19:34:46.060Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:46.062Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29718 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.064Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=764.599µs client=127.0.0.1:55491 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.065Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29718 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.067Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=2.681707ms client=127.0.0.1:46203 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.067Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29719 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.071Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=731.921µs client=127.0.0.1:53102 client_network=udp writer.go:29: 2021-01-29T19:34:46.074Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=2.150005ms client=127.0.0.1:52417 client_network=udp writer.go:29: 2021-01-29T19:34:46.078Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=dcbab4c5-00a6-c7d4-965c-5a28c1b841ea.query.consul. type=ANY class=IN latency=962.568µs client=127.0.0.1:36990 client_network=udp writer.go:29: 2021-01-29T19:34:46.078Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:46.081Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:46.081Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.082Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:46.084Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.086Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:46.090Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:46.091Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.070Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=e1088eee-a2ba-41b3-e9be-35ca3cf9bfc2.query.consul. type=AAAA class=IN latency=766.062µs client=127.0.0.1:60782 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.092Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Node info in sync writer.go:29: 2021-01-29T19:34:46.092Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.router.manager: shutting down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.072Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Requesting shutdown === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.092Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: consul server down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.095Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: shutting down server === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.096Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:46.097Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29742 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.097Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.099Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29742 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.099Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.100Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29743 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.102Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.110Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:46.114Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:46.115Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: consul server down writer.go:29: 2021-01-29T19:34:46.118Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:46.119Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29730 network=tcp writer.go:29: 2021-01-29T19:34:46.121Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29730 network=udp writer.go:29: 2021-01-29T19:34:46.122Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29731 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.206Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=1.056844ms client=127.0.0.1:46758 client_network=udp writer.go:29: 2021-01-29T19:34:46.209Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=2.476562ms client=127.0.0.1:43571 client_network=udp writer.go:29: 2021-01-29T19:34:46.211Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.dns: request served from client: name=92be12cf-10b2-fe68-f506-ac910466234e.query.consul. type=A class=IN latency=877.256µs client=127.0.0.1:57089 client_network=udp writer.go:29: 2021-01-29T19:34:46.211Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:46.215Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:46.216Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.218Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:46.219Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.222Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:46.226Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:46.228Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: consul server down writer.go:29: 2021-01-29T19:34:46.230Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:46.231Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29724 network=tcp writer.go:29: 2021-01-29T19:34:46.233Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=DNS address=127.0.0.1:29724 network=udp writer.go:29: 2021-01-29T19:34:46.234Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29725 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.571Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:46.574Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.592Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:46.593Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:46.596Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.611Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:46.612Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.625Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:46.626Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.634Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1f351f7f-b51f-9b3d-ac47-393a9dcaa5b7 Address:127.0.0.1:29759}]" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.635Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:46.637Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.637Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29759 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.639Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.645Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.646Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:46.649Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:46.652Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.657Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:dc692028-cd99-dfb4-d6d2-6e200ba05cf1 Address:127.0.0.1:29753}]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.657Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3732b10b-b746-dd18-4888-d7e9ebba386c Address:127.0.0.1:29771}]" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.659Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29753 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.660Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29771 [Follower]" leader= writer.go:29: 2021-01-29T19:34:46.662Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.659Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.665Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:46.668Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.668Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.669Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29766 network=udp writer.go:29: 2021-01-29T19:34:46.671Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29766 network=tcp writer.go:29: 2021-01-29T19:34:46.675Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Started HTTP server: address=127.0.0.1:29767 network=tcp writer.go:29: 2021-01-29T19:34:46.677Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.650Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.680Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29771) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.683Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29753) (DC: dc1)" writer.go:29: 2021-01-29T19:34:46.684Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.685Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29759) (DC: dc1)" writer.go:29: 2021-01-29T19:34:46.685Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29754 network=udp writer.go:29: 2021-01-29T19:34:46.687Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:46.693Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.694Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29748 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.694Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29759 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.696Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29748 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.697Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29754 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.700Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Started HTTP server: address=127.0.0.1:29749 network=tcp writer.go:29: 2021-01-29T19:34:46.702Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: started state syncer writer.go:29: 2021-01-29T19:34:46.712Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.701Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Started HTTP server: address=127.0.0.1:29755 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.714Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29753 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.712Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:46.718Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: vote granted: from=1f351f7f-b51f-9b3d-ac47-393a9dcaa5b7 term=2 tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.718Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.720Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:46.721Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29759 [Leader]" writer.go:29: 2021-01-29T19:34:46.716Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.721Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.721Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: vote granted: from=dc692028-cd99-dfb4-d6d2-6e200ba05cf1 term=2 tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.725Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29771 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.725Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.723Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:46.723Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.729Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29753 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.730Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.733Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.733Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: vote granted: from=3732b10b-b746-dd18-4888-d7e9ebba386c term=2 tally=1 writer.go:29: 2021-01-29T19:34:46.735Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.735Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.736Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29771 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.737Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.738Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} writer.go:29: 2021-01-29T19:34:46.739Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.739Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:46.744Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29771 writer.go:29: 2021-01-29T19:34:46.751Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.753Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29753 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.760Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:46.762Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.746Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29759 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.763Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:46.765Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.774Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:46.775Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:46.776Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.802Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.804Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4be1b62a-8603-6109-db7e-8fb0879c23b4 Address:127.0.0.1:29765}]" writer.go:29: 2021-01-29T19:34:46.806Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.806Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.806Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29765 [Follower]" leader= writer.go:29: 2021-01-29T19:34:46.808Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:46.816Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29765) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.817Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.819Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.818Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.820Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.821Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29760 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.822Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.823Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29760 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.824Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.826Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:46.827Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.828Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Started HTTP server: address=127.0.0.1:29761 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.821Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.831Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.855Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:46.858Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Synced node info writer.go:29: 2021-01-29T19:34:46.860Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.875Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:46.878Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29765 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:46.886Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:46.890Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: vote granted: from=4be1b62a-8603-6109-db7e-8fb0879c23b4 term=2 tally=1 writer.go:29: 2021-01-29T19:34:46.892Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:46.894Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29765 [Leader]" writer.go:29: 2021-01-29T19:34:46.896Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:46.897Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:46.901Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29765 writer.go:29: 2021-01-29T19:34:46.920Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:46.935Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:46.938Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.946Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:46.949Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.944Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:46.953Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:46.981Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=1.210795ms client=127.0.0.1:59115 client_network=udp writer.go:29: 2021-01-29T19:34:46.982Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=720.872µs client=127.0.0.1:54524 client_network=udp writer.go:29: 2021-01-29T19:34:46.983Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=c5359258-4775-825b-c1d7-2df60c44b287.query.consul. type=AAAA class=IN latency=652.753µs client=127.0.0.1:60419 client_network=udp writer.go:29: 2021-01-29T19:34:46.983Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:46.987Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:46.987Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.988Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:46.988Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:46.993Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:47.002Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.router.manager: shutting down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.005Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.003Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: consul server down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.007Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.006Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:47.008Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29754 network=tcp writer.go:29: 2021-01-29T19:34:47.010Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29754 network=udp writer.go:29: 2021-01-29T19:34:47.021Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29755 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.035Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:47.040Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Synced node info writer.go:29: 2021-01-29T19:34:47.043Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.472Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:47.474Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.533Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:47.545Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.603Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:47.613Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:47.623Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:47.636Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5372b0ed-fb0f-04d3-0d29-8f91c06e8cca Address:127.0.0.1:29741}]" writer.go:29: 2021-01-29T19:34:47.639Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:47.639Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29741 [Follower]" leader= writer.go:29: 2021-01-29T19:34:47.642Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:47.646Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29736 network=udp writer.go:29: 2021-01-29T19:34:47.651Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:47.653Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Started DNS server: address=127.0.0.1:29736 network=tcp writer.go:29: 2021-01-29T19:34:47.658Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Started HTTP server: address=127.0.0.1:29737 network=tcp writer.go:29: 2021-01-29T19:34:47.661Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: started state syncer writer.go:29: 2021-01-29T19:34:47.661Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29741) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.666Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=759.41µs client=127.0.0.1:59419 client_network=udp writer.go:29: 2021-01-29T19:34:47.667Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=656.222µs client=127.0.0.1:46026 client_network=udp writer.go:29: 2021-01-29T19:34:47.668Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:47.668Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=3a134139-fb0d-af4d-9905-1fb448f3b814.query.consul. type=A class=IN latency=711.493µs client=127.0.0.1:52334 client_network=udp writer.go:29: 2021-01-29T19:34:47.670Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:47.673Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:47.674Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:47.676Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:47.679Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:47.682Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:47.684Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: consul server down writer.go:29: 2021-01-29T19:34:47.686Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:47.688Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29748 network=tcp writer.go:29: 2021-01-29T19:34:47.689Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29748 network=udp writer.go:29: 2021-01-29T19:34:47.691Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29749 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.691Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=1.092413ms client=127.0.0.1:52019 client_network=udp writer.go:29: 2021-01-29T19:34:47.694Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=808.779µs client=127.0.0.1:54449 client_network=udp writer.go:29: 2021-01-29T19:34:47.696Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=a82eb146-74e0-de0a-8a6e-cdb6a298d7bb.query.consul. type=ANY class=IN latency=1.325532ms client=127.0.0.1:45821 client_network=udp writer.go:29: 2021-01-29T19:34:47.696Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:47.700Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: shutting down server === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.701Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.704Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.705Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29741 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:47.710Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.707Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.713Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=798.09µs client=127.0.0.1:34032 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.713Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: vote granted: from=5372b0ed-fb0f-04d3-0d29-8f91c06e8cca term=2 tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.707Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.715Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:47.716Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29741 [Leader]" writer.go:29: 2021-01-29T19:34:47.717Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:47.718Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.722Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=5.545022ms client=127.0.0.1:33064 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.719Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29741 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.726Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.730Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=1c22cac2-aeef-5f14-fb8a-1e3e11b437da.query.consul. type=AAAA class=IN latency=784.337µs client=127.0.0.1:40813 client_network=udp writer.go:29: 2021-01-29T19:34:47.732Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:47.735Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: shutting down server === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.735Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:47.735Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: consul server down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.736Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.736Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.737Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:47.738Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29766 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.738Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.740Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29766 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.738Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.741Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29767 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.742Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:47.743Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:47.744Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.743Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.745Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.748Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:47.748Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: consul server down writer.go:29: 2021-01-29T19:34:47.750Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:47.751Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29760 network=tcp writer.go:29: 2021-01-29T19:34:47.752Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29760 network=udp writer.go:29: 2021-01-29T19:34:47.754Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29761 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:47.856Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:47.859Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Synced node info writer.go:29: 2021-01-29T19:34:47.861Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:48.192Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:48.194Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:48.205Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=842.565µs client=127.0.0.1:41897 client_network=udp writer.go:29: 2021-01-29T19:34:48.213Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=6.899504ms client=127.0.0.1:50354 client_network=udp writer.go:29: 2021-01-29T19:34:48.214Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.dns: request served from client: name=5b312030-df52-dc98-5f5b-1536a1c7b9ad.query.consul. type=A class=IN latency=857.87µs client=127.0.0.1:40113 client_network=udp writer.go:29: 2021-01-29T19:34:48.214Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:48.220Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:48.221Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:48.222Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:48.222Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:48.224Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.212Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:48.227Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:48.228Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:48.230Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: consul server down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.230Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:48.232Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:48.234Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29736 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.236Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bfdd4119-4742-f41e-f288-176acc24ead5 Address:127.0.0.1:29777}]" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:48.238Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=DNS address=127.0.0.1:29736 network=udp writer.go:29: 2021-01-29T19:34:48.239Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29737 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:48.243Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.239Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29777 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:48.245Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.240Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:48.249Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:48.252Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29777) (DC: dc1)" writer.go:29: 2021-01-29T19:34:48.252Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:48.254Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Started DNS server: address=127.0.0.1:29772 network=udp writer.go:29: 2021-01-29T19:34:48.255Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Started DNS server: address=127.0.0.1:29772 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:48.255Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.256Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Started HTTP server: address=127.0.0.1:29773 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:48.257Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.258Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.263Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:48.265Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:48.267Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.269Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:48.271Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:48.273Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.275Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7291f76d-4c37-d263-171d-8d85c2072369 Address:127.0.0.1:29783}]" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.278Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:77cd8f28-7b18-ff67-c773-c7c25cdf64ed Address:127.0.0.1:29789}]" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.278Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29783 [Follower]" leader= writer.go:29: 2021-01-29T19:34:48.279Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.281Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29789 [Follower]" leader= writer.go:29: 2021-01-29T19:34:48.281Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:48.290Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:48.294Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.286Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.294Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29789) (DC: dc1)" writer.go:29: 2021-01-29T19:34:48.296Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Started DNS server: address=127.0.0.1:29784 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.298Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29783) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.298Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Started DNS server: address=127.0.0.1:29784 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.300Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:48.300Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Started DNS server: address=127.0.0.1:29778 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.303Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Started HTTP server: address=127.0.0.1:29785 network=tcp writer.go:29: 2021-01-29T19:34:48.304Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.305Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Started DNS server: address=127.0.0.1:29778 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.307Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.309Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Started HTTP server: address=127.0.0.1:29779 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.309Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29777 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.311Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.313Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:48.314Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: vote granted: from=bfdd4119-4742-f41e-f288-176acc24ead5 term=2 tally=1 writer.go:29: 2021-01-29T19:34:48.316Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:48.318Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29777 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.320Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.321Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:48.321Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.321Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29783 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.322Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29777 writer.go:29: 2021-01-29T19:34:48.327Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.328Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:48.330Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: vote granted: from=7291f76d-4c37-d263-171d-8d85c2072369 term=2 tally=1 writer.go:29: 2021-01-29T19:34:48.331Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.332Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.332Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29783 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.333Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.334Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.334Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.334Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.335Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.338Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29783 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.346Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:48.348Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29789 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:48.351Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.352Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.353Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: vote granted: from=77cd8f28-7b18-ff67-c773-c7c25cdf64ed term=2 tally=1 writer.go:29: 2021-01-29T19:34:48.356Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.359Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:48.361Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.358Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29789 [Leader]" writer.go:29: 2021-01-29T19:34:48.364Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:48.364Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.366Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:48.367Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.365Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29789 writer.go:29: 2021-01-29T19:34:48.371Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:48.387Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:48.389Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:48.391Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:48.393Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: member joined, marking health alive: member=test-node writer.go:29: 2021-01-29T19:34:48.391Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.561Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:48.564Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.621Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=1.119392ms client=127.0.0.1:41259 client_network=udp writer.go:29: 2021-01-29T19:34:48.622Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=783.808µs client=127.0.0.1:48557 client_network=udp writer.go:29: 2021-01-29T19:34:48.625Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.dns: request served from client: name=d0ce186c-b6e5-2865-208d-10bb420c0873.query.consul. type=AAAA class=IN latency=2.443384ms client=127.0.0.1:51222 client_network=udp writer.go:29: 2021-01-29T19:34:48.625Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:48.631Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:48.633Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:48.633Z [ERROR] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:48.635Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:48.635Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:48.641Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:48.644Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:48.645Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: consul server down writer.go:29: 2021-01-29T19:34:48.648Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:48.650Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Stopping server: protocol=DNS address=127.0.0.1:29778 network=tcp writer.go:29: 2021-01-29T19:34:48.652Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Stopping server: protocol=DNS address=127.0.0.1:29778 network=udp writer.go:29: 2021-01-29T19:34:48.674Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29779 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.728Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=1.688004ms client=127.0.0.1:56512 client_network=udp writer.go:29: 2021-01-29T19:34:48.730Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=623.648µs client=127.0.0.1:55539 client_network=udp writer.go:29: 2021-01-29T19:34:48.733Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.dns: request served from client: name=26df83b0-636c-ef34-3b91-324390b8b5e2.query.consul. type=ANY class=IN latency=1.408577ms client=127.0.0.1:49530 client_network=udp writer.go:29: 2021-01-29T19:34:48.734Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:48.734Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:48.735Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:48.736Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:48.738Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} writer.go:29: 2021-01-29T19:34:48.740Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:48.741Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.755Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:48.757Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.754Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.760Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.762Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:48.763Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: consul server down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.766Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9ba0de4e-86ca-d4d5-ce4a-88d0e7bacbb6 Address:127.0.0.1:29801}]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.768Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: shutdown complete === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.770Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29801 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.770Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Stopping server: protocol=DNS address=127.0.0.1:29772 network=tcp writer.go:29: 2021-01-29T19:34:48.780Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Stopping server: protocol=DNS address=127.0.0.1:29772 network=udp writer.go:29: 2021-01-29T19:34:48.782Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29773 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.782Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:48.793Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:48.800Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:48.800Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29801) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.803Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=1.710578ms client=127.0.0.1:51475 client_network=udp writer.go:29: 2021-01-29T19:34:48.807Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=816.308µs client=127.0.0.1:49788 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.805Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Started DNS server: address=127.0.0.1:29796 network=tcp writer.go:29: 2021-01-29T19:34:48.812Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Started DNS server: address=127.0.0.1:29796 network=udp writer.go:29: 2021-01-29T19:34:48.813Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:48.819Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29801 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:48.821Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:48.822Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: vote granted: from=9ba0de4e-86ca-d4d5-ce4a-88d0e7bacbb6 term=2 tally=1 writer.go:29: 2021-01-29T19:34:48.824Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.812Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.dns: request served from client: name=1be83622-03c3-e8b3-cba7-2e557885c652.query.consul. type=A class=IN latency=2.804301ms client=127.0.0.1:59702 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.822Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Started HTTP server: address=127.0.0.1:29797 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.812Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Requesting shutdown === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.825Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29801 [Leader]" writer.go:29: 2021-01-29T19:34:48.827Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.828Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:48.830Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.831Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29801 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.832Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:48.833Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:48.836Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.829Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.840Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}.server.router.manager: shutting down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.829Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.840Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: consul server down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.840Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.843Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: shutdown complete === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.842Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.845Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Stopping server: protocol=DNS address=127.0.0.1:29784 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.844Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:48.847Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Stopping server: protocol=DNS address=127.0.0.1:29784 network=udp writer.go:29: 2021-01-29T19:34:48.849Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29785 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:48.853Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:48.855Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:48.856Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:48.858Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: member joined, marking health alive: member=test-node writer.go:29: 2021-01-29T19:34:49.030Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=2.117954ms client=127.0.0.1:52017 client_network=udp writer.go:29: 2021-01-29T19:34:49.042Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.dns: request served from client: name=68fb4cd6-819a-5066-10be-cc6cd16fea69.query.consul. type=ANY class=IN latency=2.444434ms client=127.0.0.1:54819 client_network=udp writer.go:29: 2021-01-29T19:34:49.039Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=3.802941ms client=127.0.0.1:37624 client_network=udp writer.go:29: 2021-01-29T19:34:49.043Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:49.050Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: shutting down server writer.go:29: 2021-01-29T19:34:49.052Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:49.054Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:49.057Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:49.061Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:49.067Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:49.075Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: consul server down writer.go:29: 2021-01-29T19:34:49.080Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:49.086Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Stopping server: protocol=DNS address=127.0.0.1:29796 network=tcp writer.go:29: 2021-01-29T19:34:49.087Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Stopping server: protocol=DNS address=127.0.0.1:29796 network=udp writer.go:29: 2021-01-29T19:34:49.089Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29797 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:49.182Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:49.185Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:49.288Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:49.291Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.294Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:49.296Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:49.299Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.326Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.326Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:50f3d592-cab3-8473-cf4a-82992712ba4c Address:127.0.0.1:29813}]" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.327Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:49.329Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.330Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:49.335Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:49.341Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Started DNS server: address=127.0.0.1:29808 network=udp writer.go:29: 2021-01-29T19:34:49.337Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29813 [Follower]" leader= writer.go:29: 2021-01-29T19:34:49.343Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29813) (DC: dc1)" writer.go:29: 2021-01-29T19:34:49.344Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:49.361Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.353Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Started DNS server: address=127.0.0.1:29808 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} writer.go:29: 2021-01-29T19:34:49.366Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.367Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a5a11da4-599c-524f-fd99-3ce7e7db1454 Address:127.0.0.1:29807}]" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.370Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Started HTTP server: address=127.0.0.1:29809 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.370Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.372Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.375Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:49.376Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: entering follower state: follower="Node at 127.0.0.1:29807 [Follower]" leader= writer.go:29: 2021-01-29T19:34:49.380Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Started DNS server: address=127.0.0.1:29802 network=udp writer.go:29: 2021-01-29T19:34:49.381Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29807) (DC: dc1)" writer.go:29: 2021-01-29T19:34:49.384Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Started DNS server: address=127.0.0.1:29802 network=tcp writer.go:29: 2021-01-29T19:34:49.381Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:49.388Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Started HTTP server: address=127.0.0.1:29803 network=tcp writer.go:29: 2021-01-29T19:34:49.390Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.403Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:49.405Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:49.407Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.414Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:49.415Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29813 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:49.418Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:49.419Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: vote granted: from=50f3d592-cab3-8473-cf4a-82992712ba4c term=2 tally=1 writer.go:29: 2021-01-29T19:34:49.421Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:49.423Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29813 [Leader]" writer.go:29: 2021-01-29T19:34:49.425Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.424Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c2532840-5b6f-ba0a-a378-e8018abe84ad Address:127.0.0.1:29795}]" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.426Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:49.431Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29813 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.428Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29795 [Follower]" leader= writer.go:29: 2021-01-29T19:34:49.434Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.437Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:49.439Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: entering candidate state: node="Node at 127.0.0.1:29807 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:49.441Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.440Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.443Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: vote granted: from=a5a11da4-599c-524f-fd99-3ce7e7db1454 term=2 tally=1 writer.go:29: 2021-01-29T19:34:49.445Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.445Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29795) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.447Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: entering leader state: leader="Node at 127.0.0.1:29807 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.447Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:49.450Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Started DNS server: address=127.0.0.1:29790 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.449Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:49.452Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.451Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Started DNS server: address=127.0.0.1:29790 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.441Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.454Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Started HTTP server: address=127.0.0.1:29791 network=tcp writer.go:29: 2021-01-29T19:34:49.456Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.454Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29807 writer.go:29: 2021-01-29T19:34:49.461Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:49.473Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.470Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.475Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.476Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.477Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.477Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.480Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.481Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.490Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:49.492Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29795 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:49.496Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:49.498Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: vote granted: from=c2532840-5b6f-ba0a-a378-e8018abe84ad term=2 tally=1 writer.go:29: 2021-01-29T19:34:49.499Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.496Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.501Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29795 [Leader]" writer.go:29: 2021-01-29T19:34:49.502Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:49.505Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.510Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.510Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29795 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.512Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.518Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.523Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:49.527Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.546Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:49.548Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:49.551Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:49.552Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.591Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:49.593Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.615Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:49.617Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:49.619Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:49.627Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:65b18b94-9a82-1bb7-98a6-99edac6438d8 Address:127.0.0.1:29819}]" writer.go:29: 2021-01-29T19:34:49.630Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:49.632Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29819 [Follower]" leader= writer.go:29: 2021-01-29T19:34:49.634Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:49.642Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29819) (DC: dc1)" writer.go:29: 2021-01-29T19:34:49.645Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.648Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:49.651Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Synced node info writer.go:29: 2021-01-29T19:34:49.653Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.653Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Started DNS server: address=127.0.0.1:29814 network=tcp writer.go:29: 2021-01-29T19:34:49.659Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Started DNS server: address=127.0.0.1:29814 network=udp writer.go:29: 2021-01-29T19:34:49.668Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Started HTTP server: address=127.0.0.1:29815 network=tcp writer.go:29: 2021-01-29T19:34:49.673Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: started state syncer writer.go:29: 2021-01-29T19:34:49.706Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:49.710Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29819 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:49.715Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:49.717Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: vote granted: from=65b18b94-9a82-1bb7-98a6-99edac6438d8 term=2 tally=1 writer.go:29: 2021-01-29T19:34:49.719Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:49.721Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29819 [Leader]" writer.go:29: 2021-01-29T19:34:49.724Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:49.724Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:49.728Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29819 writer.go:29: 2021-01-29T19:34:49.741Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.747Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=974.874µs client=127.0.0.1:42761 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.754Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.753Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=1.717486ms client=127.0.0.1:49212 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.756Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.754Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.dns: request served from client: name=700daf2f-18d5-6d1b-056b-3b09ece9e4b0.query.consul. type=A class=IN latency=819.656µs client=127.0.0.1:33954 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.758Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:49.763Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:49.764Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Requesting shutdown writer.go:29: 2021-01-29T19:34:49.770Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: shutting down server writer.go:29: 2021-01-29T19:34:49.777Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:49.784Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:49.784Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:49.791Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:49.796Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:49.800Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: consul server down writer.go:29: 2021-01-29T19:34:49.804Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: shutdown complete writer.go:29: 2021-01-29T19:34:49.807Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Stopping server: protocol=DNS address=127.0.0.1:29802 network=tcp writer.go:29: 2021-01-29T19:34:49.812Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Stopping server: protocol=DNS address=127.0.0.1:29802 network=udp writer.go:29: 2021-01-29T19:34:49.817Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Stopping server: protocol=HTTP address=127.0.0.1:29803 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.834Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:49.841Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Synced node info writer.go:29: 2021-01-29T19:34:49.852Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.874Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=10.630024ms client=127.0.0.1:36943 client_network=udp writer.go:29: 2021-01-29T19:34:49.878Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=3.323571ms client=127.0.0.1:37160 client_network=udp writer.go:29: 2021-01-29T19:34:49.891Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:49.891Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.dns: request served from client: name=bb64ee4b-a48f-d834-309b-9c8ccad7f5ec.query.consul. type=ANY class=IN latency=9.27153ms client=127.0.0.1:59474 client_network=udp writer.go:29: 2021-01-29T19:34:49.902Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: shutting down server writer.go:29: 2021-01-29T19:34:49.906Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:49.909Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.921Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=3.012249ms client=127.0.0.1:48518 client_network=udp writer.go:29: 2021-01-29T19:34:49.922Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=752.028µs client=127.0.0.1:42858 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.910Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.923Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.dns: request served from client: name=5dcdf20e-e8fa-d7c1-6c71-d656472c944b.query.consul. type=AAAA class=IN latency=802.596µs client=127.0.0.1:57833 client_network=udp writer.go:29: 2021-01-29T19:34:49.923Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Requesting shutdown === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.917Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.940Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server: shutting down server === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.942Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.router.manager: shutting down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.942Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.942Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: consul server down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.944Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.945Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: shutdown complete === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.946Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.947Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Stopping server: protocol=DNS address=127.0.0.1:29790 network=tcp writer.go:29: 2021-01-29T19:34:49.948Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Stopping server: protocol=DNS address=127.0.0.1:29790 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.949Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:49.953Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}.server.router.manager: shutting down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:49.950Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Stopping server: protocol=HTTP address=127.0.0.1:29791 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:49.954Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: consul server down writer.go:29: 2021-01-29T19:34:49.965Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: shutdown complete writer.go:29: 2021-01-29T19:34:49.971Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Stopping server: protocol=DNS address=127.0.0.1:29808 network=tcp writer.go:29: 2021-01-29T19:34:49.990Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Stopping server: protocol=DNS address=127.0.0.1:29808 network=udp writer.go:29: 2021-01-29T19:34:49.993Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Stopping server: protocol=HTTP address=127.0.0.1:29809 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:50.107Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=4.508804ms client=127.0.0.1:58138 client_network=udp writer.go:29: 2021-01-29T19:34:50.112Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=3.104198ms client=127.0.0.1:47517 client_network=udp writer.go:29: 2021-01-29T19:34:50.117Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.dns: request served from client: name=b07de585-8d85-c7bb-0edf-a7bb05a4472b.query.consul. type=AAAA class=IN latency=3.402078ms client=127.0.0.1:46495 client_network=udp writer.go:29: 2021-01-29T19:34:50.117Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:50.122Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: shutting down server writer.go:29: 2021-01-29T19:34:50.124Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:50.129Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:50.129Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:50.154Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:50.160Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:50.160Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: consul server down writer.go:29: 2021-01-29T19:34:50.163Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: shutdown complete writer.go:29: 2021-01-29T19:34:50.165Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Stopping server: protocol=DNS address=127.0.0.1:29814 network=tcp writer.go:29: 2021-01-29T19:34:50.167Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Stopping server: protocol=DNS address=127.0.0.1:29814 network=udp writer.go:29: 2021-01-29T19:34:50.168Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Stopping server: protocol=HTTP address=127.0.0.1:29815 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:50.320Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:50.324Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:50.346Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:50.349Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:50.352Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:50.358Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:32a7af2a-0699-1591-b1a9-70adc3026565 Address:127.0.0.1:29837}]" writer.go:29: 2021-01-29T19:34:50.360Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: entering follower state: follower="Node at 127.0.0.1:29837 [Follower]" leader= writer.go:29: 2021-01-29T19:34:50.365Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:50.370Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:50.375Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29837) (DC: dc1)" writer.go:29: 2021-01-29T19:34:50.377Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:50.383Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Started DNS server: address=127.0.0.1:29832 network=tcp writer.go:29: 2021-01-29T19:34:50.386Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Started DNS server: address=127.0.0.1:29832 network=udp writer.go:29: 2021-01-29T19:34:50.388Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Started HTTP server: address=127.0.0.1:29833 network=tcp writer.go:29: 2021-01-29T19:34:50.391Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: started state syncer writer.go:29: 2021-01-29T19:34:50.406Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:50.408Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: entering candidate state: node="Node at 127.0.0.1:29837 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:50.411Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:50.413Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: vote granted: from=32a7af2a-0699-1591-b1a9-70adc3026565 term=2 tally=1 writer.go:29: 2021-01-29T19:34:50.415Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:50.417Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.raft: entering leader state: leader="Node at 127.0.0.1:29837 [Leader]" writer.go:29: 2021-01-29T19:34:50.419Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:50.420Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:50.424Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29837 writer.go:29: 2021-01-29T19:34:50.430Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:50.436Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:50.437Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:50.442Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:50.445Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:50.472Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:50.474Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.501Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} writer.go:29: 2021-01-29T19:34:50.502Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:50.505Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.514Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.519Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.522Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:50.522Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.521Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:50.526Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:50.526Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Synced node info writer.go:29: 2021-01-29T19:34:50.528Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.529Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c889127c-0fff-9d28-9adb-d6264cd33294 Address:127.0.0.1:29825}]" writer.go:29: 2021-01-29T19:34:50.532Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: entering follower state: follower="Node at 127.0.0.1:29825 [Follower]" leader= writer.go:29: 2021-01-29T19:34:50.532Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.537Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ccbbd0f8-439d-19a8-b037-3b1e4f3c5eed Address:127.0.0.1:29843}]" writer.go:29: 2021-01-29T19:34:50.540Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: entering follower state: follower="Node at 127.0.0.1:29843 [Follower]" leader= writer.go:29: 2021-01-29T19:34:50.544Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:50.549Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:50.553Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29843) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.546Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.555Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.556Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29825) (DC: dc1)" writer.go:29: 2021-01-29T19:34:50.558Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:50.562Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Started DNS server: address=127.0.0.1:29820 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.559Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Started DNS server: address=127.0.0.1:29838 network=tcp writer.go:29: 2021-01-29T19:34:50.564Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Started DNS server: address=127.0.0.1:29838 network=udp writer.go:29: 2021-01-29T19:34:50.566Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Started HTTP server: address=127.0.0.1:29839 network=tcp writer.go:29: 2021-01-29T19:34:50.567Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.567Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Started DNS server: address=127.0.0.1:29820 network=udp writer.go:29: 2021-01-29T19:34:50.571Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Started HTTP server: address=127.0.0.1:29821 network=tcp writer.go:29: 2021-01-29T19:34:50.573Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.588Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:50.590Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: entering candidate state: node="Node at 127.0.0.1:29843 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.591Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.592Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.592Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: entering candidate state: node="Node at 127.0.0.1:29825 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.594Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: vote granted: from=ccbbd0f8-439d-19a8-b037-3b1e4f3c5eed term=2 tally=1 writer.go:29: 2021-01-29T19:34:50.595Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:50.597Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: entering leader state: leader="Node at 127.0.0.1:29843 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.599Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:50.601Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: vote granted: from=c889127c-0fff-9d28-9adb-d6264cd33294 term=2 tally=1 writer.go:29: 2021-01-29T19:34:50.602Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:50.604Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: entering leader state: leader="Node at 127.0.0.1:29825 [Leader]" writer.go:29: 2021-01-29T19:34:50.608Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.605Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: cluster leadership acquired === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.608Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.606Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:50.612Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29843 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.613Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29825 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.618Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.626Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.632Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:50.634Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.635Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.636Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:50.637Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.637Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:50.640Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:50.647Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:50.670Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:50.672Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.698Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:50.709Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:50.718Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:50.729Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a93eb208-97e8-b442-8627-36b6a8028895 Address:127.0.0.1:29855}]" writer.go:29: 2021-01-29T19:34:50.733Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: entering follower state: follower="Node at 127.0.0.1:29855 [Follower]" leader= writer.go:29: 2021-01-29T19:34:50.733Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:50.759Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:50.768Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29855) (DC: dc1)" writer.go:29: 2021-01-29T19:34:50.771Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Started DNS server: address=127.0.0.1:29850 network=udp writer.go:29: 2021-01-29T19:34:50.775Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:50.778Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Started DNS server: address=127.0.0.1:29850 network=tcp writer.go:29: 2021-01-29T19:34:50.784Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Started HTTP server: address=127.0.0.1:29851 network=tcp writer.go:29: 2021-01-29T19:34:50.786Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: started state syncer writer.go:29: 2021-01-29T19:34:50.798Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:50.802Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: entering candidate state: node="Node at 127.0.0.1:29855 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:50.818Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:50.821Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: vote granted: from=a93eb208-97e8-b442-8627-36b6a8028895 term=2 tally=1 writer.go:29: 2021-01-29T19:34:50.825Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:50.832Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.raft: entering leader state: leader="Node at 127.0.0.1:29855 [Leader]" writer.go:29: 2021-01-29T19:34:50.837Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:50.838Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:50.840Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29855 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.929Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.929Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.939Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.977Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.980Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:50.985Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:50.996Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:50.999Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.010Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.057Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.090Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=2.532523ms client=127.0.0.1:43325 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.103Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Synced node info writer.go:29: 2021-01-29T19:34:51.104Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.112Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=11.050132ms client=127.0.0.1:43208 client_network=udp writer.go:29: 2021-01-29T19:34:51.123Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.dns: request served from client: name=384bc450-a7f7-71f5-d9e5-0911dbc13ad3.query.consul. type=ANY class=IN latency=9.937792ms client=127.0.0.1:44202 client_network=udp writer.go:29: 2021-01-29T19:34:51.124Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Requesting shutdown writer.go:29: 2021-01-29T19:34:51.150Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: shutting down server writer.go:29: 2021-01-29T19:34:51.152Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:51.153Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:51.159Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:51.179Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:51.189Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:51.190Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: consul server down writer.go:29: 2021-01-29T19:34:51.193Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: shutdown complete writer.go:29: 2021-01-29T19:34:51.195Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Stopping server: protocol=DNS address=127.0.0.1:29820 network=tcp writer.go:29: 2021-01-29T19:34:51.197Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Stopping server: protocol=DNS address=127.0.0.1:29820 network=udp writer.go:29: 2021-01-29T19:34:51.200Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Stopping server: protocol=HTTP address=127.0.0.1:29821 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.240Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=720.437µs client=127.0.0.1:51084 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:51.240Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=1.221695ms client=127.0.0.1:48342 client_network=udp writer.go:29: 2021-01-29T19:34:51.244Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=858.125µs client=127.0.0.1:51748 client_network=udp writer.go:29: 2021-01-29T19:34:51.246Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.dns: request served from client: name=aa69a12c-f83c-8822-368f-c214a09f4e7e.query.consul. type=A class=IN latency=1.843002ms client=127.0.0.1:49106 client_network=udp writer.go:29: 2021-01-29T19:34:51.246Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Requesting shutdown writer.go:29: 2021-01-29T19:34:51.255Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server: shutting down server writer.go:29: 2021-01-29T19:34:51.256Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.252Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=11.504806ms client=127.0.0.1:36918 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:51.257Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:51.257Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.256Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.dns: request served from client: name=030ed252-778d-834e-93a5-6f00d8c6fff6.query.consul. type=AAAA class=IN latency=3.63253ms client=127.0.0.1:33638 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:51.263Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.259Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Requesting shutdown === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:51.265Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:51.265Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: consul server down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.265Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: shutting down server === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:51.267Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: shutdown complete writer.go:29: 2021-01-29T19:34:51.267Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Stopping server: protocol=DNS address=127.0.0.1:29832 network=tcp writer.go:29: 2021-01-29T19:34:51.268Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Stopping server: protocol=DNS address=127.0.0.1:29832 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.267Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:51.270Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Stopping server: protocol=HTTP address=127.0.0.1:29833 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.270Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:51.271Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:51.276Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:51.280Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:51.280Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: consul server down writer.go:29: 2021-01-29T19:34:51.283Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: shutdown complete writer.go:29: 2021-01-29T19:34:51.285Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Stopping server: protocol=DNS address=127.0.0.1:29838 network=tcp writer.go:29: 2021-01-29T19:34:51.287Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Stopping server: protocol=DNS address=127.0.0.1:29838 network=udp writer.go:29: 2021-01-29T19:34:51.288Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Stopping server: protocol=HTTP address=127.0.0.1:29839 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.302Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=2.789353ms client=127.0.0.1:55425 client_network=udp writer.go:29: 2021-01-29T19:34:51.304Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=1.447894ms client=127.0.0.1:43021 client_network=udp writer.go:29: 2021-01-29T19:34:51.305Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.dns: request served from client: name=c08fe037-8fcb-93ff-9c17-6e608fd65de4.query.consul. type=A class=IN latency=923.629µs client=127.0.0.1:37579 client_network=udp writer.go:29: 2021-01-29T19:34:51.305Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Requesting shutdown writer.go:29: 2021-01-29T19:34:51.308Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server: shutting down server writer.go:29: 2021-01-29T19:34:51.309Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:51.310Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:51.310Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:51.312Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:51.320Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:51.320Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: consul server down writer.go:29: 2021-01-29T19:34:51.323Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: shutdown complete writer.go:29: 2021-01-29T19:34:51.325Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Stopping server: protocol=DNS address=127.0.0.1:29850 network=tcp writer.go:29: 2021-01-29T19:34:51.326Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Stopping server: protocol=DNS address=127.0.0.1:29850 network=udp writer.go:29: 2021-01-29T19:34:51.327Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Stopping server: protocol=HTTP address=127.0.0.1:29851 network=tcp writer.go:29: 2021-01-29T19:34:51.333Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:51.335Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.362Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:51.366Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:51.368Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:51.375Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7973d61b-8e9a-df36-f6d0-0a7d878edf9d Address:127.0.0.1:29849}]" writer.go:29: 2021-01-29T19:34:51.378Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: entering follower state: follower="Node at 127.0.0.1:29849 [Follower]" leader= writer.go:29: 2021-01-29T19:34:51.380Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:51.385Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:51.388Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:51.389Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Started DNS server: address=127.0.0.1:29844 network=udp writer.go:29: 2021-01-29T19:34:51.391Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Started DNS server: address=127.0.0.1:29844 network=tcp writer.go:29: 2021-01-29T19:34:51.394Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Started HTTP server: address=127.0.0.1:29845 network=tcp writer.go:29: 2021-01-29T19:34:51.394Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: started state syncer writer.go:29: 2021-01-29T19:34:51.396Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29849) (DC: dc1)" writer.go:29: 2021-01-29T19:34:51.434Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:51.436Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: entering candidate state: node="Node at 127.0.0.1:29849 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:51.439Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:51.441Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: vote granted: from=7973d61b-8e9a-df36-f6d0-0a7d878edf9d term=2 tally=1 writer.go:29: 2021-01-29T19:34:51.443Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:51.445Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: entering leader state: leader="Node at 127.0.0.1:29849 [Leader]" writer.go:29: 2021-01-29T19:34:51.448Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:51.448Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:51.455Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29849 writer.go:29: 2021-01-29T19:34:51.467Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:51.475Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:51.477Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:51.480Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:51.482Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: member joined, marking health alive: member=test-node writer.go:29: 2021-01-29T19:34:51.672Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:51.674Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.702Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:51.703Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.714Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.719Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=2.895579ms client=127.0.0.1:60679 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.715Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:51.721Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.721Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=1.082417ms client=127.0.0.1:57440 client_network=udp writer.go:29: 2021-01-29T19:34:51.723Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.dns: request served from client: name=48f9ca9e-e97a-f8b9-3a86-ec84fbc804ae.query.consul. type=ANY class=IN latency=1.119116ms client=127.0.0.1:35652 client_network=udp writer.go:29: 2021-01-29T19:34:51.723Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Requesting shutdown writer.go:29: 2021-01-29T19:34:51.727Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: shutting down server writer.go:29: 2021-01-29T19:34:51.727Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:51.728Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:51.728Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:51.730Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.wan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.731Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a77cda8c-800b-92dc-aa55-a1c309165a72 Address:127.0.0.1:29831}]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.734Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.router.manager: shutting down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.735Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: entering follower state: follower="Node at 127.0.0.1:29831 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.734Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: consul server down writer.go:29: 2021-01-29T19:34:51.738Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: shutdown complete === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.738Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.739Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Stopping server: protocol=DNS address=127.0.0.1:29844 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.742Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.752Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Stopping server: protocol=DNS address=127.0.0.1:29844 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.753Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Started DNS server: address=127.0.0.1:29826 network=udp writer.go:29: 2021-01-29T19:34:51.754Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29831) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.753Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Stopping server: protocol=HTTP address=127.0.0.1:29845 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.755Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:51.756Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Started DNS server: address=127.0.0.1:29826 network=tcp writer.go:29: 2021-01-29T19:34:51.759Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Started HTTP server: address=127.0.0.1:29827 network=tcp writer.go:29: 2021-01-29T19:34:51.761Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} writer.go:29: 2021-01-29T19:34:51.771Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:51.772Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.787Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:51.788Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} writer.go:29: 2021-01-29T19:34:51.789Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:51.790Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.791Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.806Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.805Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.808Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: entering candidate state: node="Node at 127.0.0.1:29831 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.808Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.810Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.811Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.812Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: vote granted: from=a77cda8c-800b-92dc-aa55-a1c309165a72 term=2 tally=1 writer.go:29: 2021-01-29T19:34:51.814Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:51.816Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.raft: entering leader state: leader="Node at 127.0.0.1:29831 [Leader]" writer.go:29: 2021-01-29T19:34:51.819Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:51.820Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.822Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ed9ad3ad-4030-ba0e-946c-cd958f306c60 Address:127.0.0.1:29861}]" writer.go:29: 2021-01-29T19:34:51.824Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.825Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1f4e024d-e5f8-dcce-3a1f-1a7d7d41f57a Address:127.0.0.1:29879}]" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.824Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: entering follower state: follower="Node at 127.0.0.1:29861 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.828Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: entering follower state: follower="Node at 127.0.0.1:29879 [Follower]" leader= writer.go:29: 2021-01-29T19:34:51.830Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.826Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.832Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29831 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.833Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.834Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:51.834Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29861) (DC: dc1)" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.837Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Started DNS server: address=127.0.0.1:29874 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.839Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.838Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29879) (DC: dc1)" writer.go:29: 2021-01-29T19:34:51.839Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:51.841Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Started DNS server: address=127.0.0.1:29874 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.845Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Started DNS server: address=127.0.0.1:29856 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.850Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Started HTTP server: address=127.0.0.1:29875 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.847Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.853Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.851Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Started DNS server: address=127.0.0.1:29856 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.858Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.860Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Started HTTP server: address=127.0.0.1:29857 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.863Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.865Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:51.869Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:51.871Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.885Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:51.887Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: entering candidate state: node="Node at 127.0.0.1:29861 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:51.891Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:51.893Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: vote granted: from=ed9ad3ad-4030-ba0e-946c-cd958f306c60 term=2 tally=1 writer.go:29: 2021-01-29T19:34:51.895Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: election won: tally=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.898Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:51.900Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: entering candidate state: node="Node at 127.0.0.1:29879 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.897Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: entering leader state: leader="Node at 127.0.0.1:29861 [Leader]" writer.go:29: 2021-01-29T19:34:51.902Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:51.903Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.907Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.912Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29861 writer.go:29: 2021-01-29T19:34:51.918Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.911Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: vote granted: from=1f4e024d-e5f8-dcce-3a1f-1a7d7d41f57a term=2 tally=1 writer.go:29: 2021-01-29T19:34:51.920Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:51.925Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: entering leader state: leader="Node at 127.0.0.1:29879 [Leader]" writer.go:29: 2021-01-29T19:34:51.929Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:51.930Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.927Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.933Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29879 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:51.938Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:51.944Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:51.948Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:51.967Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:51.973Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:51.975Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:51.977Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:51.980Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:52.091Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=2.665961ms client=127.0.0.1:41876 client_network=udp writer.go:29: 2021-01-29T19:34:52.094Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=2.380477ms client=127.0.0.1:36266 client_network=udp writer.go:29: 2021-01-29T19:34:52.097Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.dns: request served from client: name=f9ada9d9-e231-b39c-9933-ca16bb5e7f5c.query.consul. type=AAAA class=IN latency=2.458543ms client=127.0.0.1:49591 client_network=udp writer.go:29: 2021-01-29T19:34:52.097Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Requesting shutdown writer.go:29: 2021-01-29T19:34:52.101Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server: shutting down server writer.go:29: 2021-01-29T19:34:52.102Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:52.104Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:52.104Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:52.124Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:52.127Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:52.127Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: consul server down writer.go:29: 2021-01-29T19:34:52.130Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: shutdown complete writer.go:29: 2021-01-29T19:34:52.132Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Stopping server: protocol=DNS address=127.0.0.1:29826 network=tcp writer.go:29: 2021-01-29T19:34:52.134Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Stopping server: protocol=DNS address=127.0.0.1:29826 network=udp writer.go:29: 2021-01-29T19:34:52.136Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Stopping server: protocol=HTTP address=127.0.0.1:29827 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.167Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=682.552µs client=127.0.0.1:38726 client_network=udp writer.go:29: 2021-01-29T19:34:52.169Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=609.458µs client=127.0.0.1:59862 client_network=udp writer.go:29: 2021-01-29T19:34:52.173Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Requesting shutdown writer.go:29: 2021-01-29T19:34:52.174Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: shutting down server writer.go:29: 2021-01-29T19:34:52.175Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:52.176Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:52.175Z [ERROR] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:52.176Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:52.177Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.dns: request served from client: name=3eb33b18-b4ee-ac3c-ee8d-5866b6b83b6b.query.consul. type=ANY class=IN latency=6.348158ms client=127.0.0.1:54912 client_network=udp writer.go:29: 2021-01-29T19:34:52.177Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:52.184Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:52.184Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: consul server down writer.go:29: 2021-01-29T19:34:52.187Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: shutdown complete writer.go:29: 2021-01-29T19:34:52.187Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Stopping server: protocol=DNS address=127.0.0.1:29874 network=tcp writer.go:29: 2021-01-29T19:34:52.188Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Stopping server: protocol=DNS address=127.0.0.1:29874 network=udp writer.go:29: 2021-01-29T19:34:52.189Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Stopping server: protocol=HTTP address=127.0.0.1:29875 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:52.255Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:52.256Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.268Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:52.270Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:52.273Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.278Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:52.281Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Synced node info writer.go:29: 2021-01-29T19:34:52.283Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.284Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0b5a69b8-71a3-4d5d-d9b4-9f57cd0ff629 Address:127.0.0.1:29885}]" writer.go:29: 2021-01-29T19:34:52.286Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: entering follower state: follower="Node at 127.0.0.1:29885 [Follower]" leader= writer.go:29: 2021-01-29T19:34:52.287Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:52.292Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:52.296Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:52.296Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29885) (DC: dc1)" writer.go:29: 2021-01-29T19:34:52.297Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Started DNS server: address=127.0.0.1:29880 network=tcp writer.go:29: 2021-01-29T19:34:52.298Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Started DNS server: address=127.0.0.1:29880 network=udp writer.go:29: 2021-01-29T19:34:52.299Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Started HTTP server: address=127.0.0.1:29881 network=tcp writer.go:29: 2021-01-29T19:34:52.300Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: started state syncer writer.go:29: 2021-01-29T19:34:52.334Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:52.335Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: entering candidate state: node="Node at 127.0.0.1:29885 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:52.339Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:52.342Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: vote granted: from=0b5a69b8-71a3-4d5d-d9b4-9f57cd0ff629 term=2 tally=1 writer.go:29: 2021-01-29T19:34:52.346Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:52.350Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: entering leader state: leader="Node at 127.0.0.1:29885 [Leader]" writer.go:29: 2021-01-29T19:34:52.354Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:52.355Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:52.357Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29885 writer.go:29: 2021-01-29T19:34:52.396Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:52.494Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:52.499Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:52.495Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Synced node info writer.go:29: 2021-01-29T19:34:52.503Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:52.513Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.617Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=2.179466ms client=127.0.0.1:37760 client_network=udp writer.go:29: 2021-01-29T19:34:52.628Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=9.713345ms client=127.0.0.1:57732 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} writer.go:29: 2021-01-29T19:34:52.644Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:52.645Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.647Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.dns: request served from client: name=315e163c-b302-a0dd-6f1e-6089d4a5f8b6.query.consul. type=ANY class=IN latency=3.320687ms client=127.0.0.1:46612 client_network=udp writer.go:29: 2021-01-29T19:34:52.647Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Requesting shutdown writer.go:29: 2021-01-29T19:34:52.662Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: shutting down server writer.go:29: 2021-01-29T19:34:52.675Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.679Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.679Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:52.679Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.leader: stopped routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.683Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.tlsutil: Update: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.685Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:52.688Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.router.manager: shutting down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.687Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.690Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:52.692Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.688Z [ERROR] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:52.689Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: consul server down writer.go:29: 2021-01-29T19:34:52.697Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: shutdown complete writer.go:29: 2021-01-29T19:34:52.699Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Stopping server: protocol=DNS address=127.0.0.1:29856 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.699Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0b4b6265-e4f3-319e-aac6-23ed3033f882 Address:127.0.0.1:29867}]" writer.go:29: 2021-01-29T19:34:52.702Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: entering follower state: follower="Node at 127.0.0.1:29867 [Follower]" leader= writer.go:29: 2021-01-29T19:34:52.702Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:52.709Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.701Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Stopping server: protocol=DNS address=127.0.0.1:29856 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.723Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.723Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Stopping server: protocol=HTTP address=127.0.0.1:29857 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.725Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Started DNS server: address=127.0.0.1:29862 network=udp writer.go:29: 2021-01-29T19:34:52.725Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29867) (DC: dc1)" writer.go:29: 2021-01-29T19:34:52.727Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Started DNS server: address=127.0.0.1:29862 network=tcp writer.go:29: 2021-01-29T19:34:52.730Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Started HTTP server: address=127.0.0.1:29863 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.734Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: bootstrap = true: do not enable unless necessary === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.751Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.753Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:52.754Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:52.761Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:457fe0f6-4903-ceb0-bb03-8e276ec5c97b Address:127.0.0.1:29873}]" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.762Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:52.764Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: entering candidate state: node="Node at 127.0.0.1:29867 [Candidate]" term=2 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.764Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.767Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: votes: needed=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.767Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.769Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: vote granted: from=0b4b6265-e4f3-319e-aac6-23ed3033f882 term=2 tally=1 writer.go:29: 2021-01-29T19:34:52.771Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:52.773Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.raft: entering leader state: leader="Node at 127.0.0.1:29867 [Leader]" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.774Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=857.452µs client=127.0.0.1:46228 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.775Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:52.777Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: New leader elected: payload=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.778Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=1.941697ms client=127.0.0.1:49357 client_network=udp writer.go:29: 2021-01-29T19:34:52.779Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.dns: request served from client: name=4c96865c-c587-2738-0ce1-848f30dafcb4.query.consul. type=AAAA class=IN latency=836.375µs client=127.0.0.1:35729 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.766Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: entering follower state: follower="Node at 127.0.0.1:29873 [Follower]" leader= writer.go:29: 2021-01-29T19:34:52.770Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:52.770Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29873) (DC: dc1)" writer.go:29: 2021-01-29T19:34:52.780Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Started DNS server: address=127.0.0.1:29868 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.780Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Requesting shutdown writer.go:29: 2021-01-29T19:34:52.786Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: shutting down server === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.785Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Started DNS server: address=127.0.0.1:29868 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.788Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29867 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.790Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.leader: stopping routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.795Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.792Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Started HTTP server: address=127.0.0.1:29869 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.802Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:52.802Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.lan: serf: Shutdown without a Leave === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.807Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.803Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: started state syncer === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.808Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.807Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:52.812Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.router.manager: shutting down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.811Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: Skipping self join check for node since the cluster is too small: node=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.813Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: consul server down === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.814Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:52.815Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: shutdown complete writer.go:29: 2021-01-29T19:34:52.816Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Stopping server: protocol=DNS address=127.0.0.1:29880 network=tcp writer.go:29: 2021-01-29T19:34:52.818Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Stopping server: protocol=DNS address=127.0.0.1:29880 network=udp writer.go:29: 2021-01-29T19:34:52.820Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Stopping server: protocol=HTTP address=127.0.0.1:29881 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:52.828Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:52.831Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: entering candidate state: node="Node at 127.0.0.1:29873 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:52.834Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:52.837Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: vote granted: from=457fe0f6-4903-ceb0-bb03-8e276ec5c97b term=2 tally=1 writer.go:29: 2021-01-29T19:34:52.839Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:52.841Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: entering leader state: leader="Node at 127.0.0.1:29873 [Leader]" writer.go:29: 2021-01-29T19:34:52.844Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:52.846Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:52.852Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29873 writer.go:29: 2021-01-29T19:34:52.859Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:52.868Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:52.870Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:52.872Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:52.874Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: member joined, marking health alive: member=test-node writer.go:29: 2021-01-29T19:34:52.997Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=796.934µs client=127.0.0.1:50520 client_network=udp writer.go:29: 2021-01-29T19:34:53.002Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=1.093887ms client=127.0.0.1:36270 client_network=udp writer.go:29: 2021-01-29T19:34:53.003Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.dns: request served from client: name=e17039f5-7307-29f6-0cec-8d0130ef8ef0.query.consul. type=A class=IN latency=666.879µs client=127.0.0.1:46768 client_network=udp writer.go:29: 2021-01-29T19:34:53.003Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Requesting shutdown writer.go:29: 2021-01-29T19:34:53.008Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: shutting down server writer.go:29: 2021-01-29T19:34:53.010Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.012Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:53.010Z [ERROR] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:53.012Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.016Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:53.024Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:53.026Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: consul server down writer.go:29: 2021-01-29T19:34:53.028Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: shutdown complete writer.go:29: 2021-01-29T19:34:53.029Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Stopping server: protocol=DNS address=127.0.0.1:29868 network=tcp writer.go:29: 2021-01-29T19:34:53.031Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Stopping server: protocol=DNS address=127.0.0.1:29868 network=udp writer.go:29: 2021-01-29T19:34:53.033Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Stopping server: protocol=HTTP address=127.0.0.1:29869 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:53.061Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.dns: request served from client: name=api-tier.service.consul. type=A class=IN latency=780.753µs client=127.0.0.1:57915 client_network=udp writer.go:29: 2021-01-29T19:34:53.063Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.dns: request served from client: name=api-tier.query.consul. type=A class=IN latency=819.409µs client=127.0.0.1:46593 client_network=udp writer.go:29: 2021-01-29T19:34:53.065Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Requesting shutdown writer.go:29: 2021-01-29T19:34:53.065Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.dns: request served from client: name=393b8428-1440-bc8d-96e3-e11d45b44d68.query.consul. type=A class=IN latency=1.046914ms client=127.0.0.1:34190 client_network=udp writer.go:29: 2021-01-29T19:34:53.066Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server: shutting down server writer.go:29: 2021-01-29T19:34:53.070Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.071Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:53.072Z [ERROR] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:53.072Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.076Z [WARN] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:53.080Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:53.081Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: consul server down writer.go:29: 2021-01-29T19:34:53.082Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: shutdown complete writer.go:29: 2021-01-29T19:34:53.083Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Stopping server: protocol=DNS address=127.0.0.1:29862 network=tcp writer.go:29: 2021-01-29T19:34:53.084Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Stopping server: protocol=DNS address=127.0.0.1:29862 network=udp writer.go:29: 2021-01-29T19:34:53.087Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Stopping server: protocol=HTTP address=127.0.0.1:29863 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.226Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:53.227Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.238Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:53.239Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:53.241Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:53.247Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c71d13f6-1901-25c8-e251-f219a893d81e Address:127.0.0.1:29909}]" writer.go:29: 2021-01-29T19:34:53.249Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: entering follower state: follower="Node at 127.0.0.1:29909 [Follower]" leader= writer.go:29: 2021-01-29T19:34:53.250Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:53.253Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:53.255Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29909) (DC: dc1)" writer.go:29: 2021-01-29T19:34:53.255Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Started DNS server: address=127.0.0.1:29904 network=udp writer.go:29: 2021-01-29T19:34:53.256Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:53.257Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Started DNS server: address=127.0.0.1:29904 network=tcp writer.go:29: 2021-01-29T19:34:53.259Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Started HTTP server: address=127.0.0.1:29905 network=tcp writer.go:29: 2021-01-29T19:34:53.260Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: started state syncer writer.go:29: 2021-01-29T19:34:53.301Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:53.303Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: entering candidate state: node="Node at 127.0.0.1:29909 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:53.304Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:53.305Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: vote granted: from=c71d13f6-1901-25c8-e251-f219a893d81e term=2 tally=1 writer.go:29: 2021-01-29T19:34:53.306Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:53.307Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.raft: entering leader state: leader="Node at 127.0.0.1:29909 [Leader]" writer.go:29: 2021-01-29T19:34:53.308Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:53.309Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:53.310Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29909 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:53.321Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:53.322Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.328Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.332Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:53.334Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:53.335Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.336Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.341Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1887784c-fc36-9267-a3ff-f79d1920fc88 Address:127.0.0.1:29903}]" === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.339Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.leader: started routine: routine="CA root pruning" === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.343Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:53.343Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: entering follower state: follower="Node at 127.0.0.1:29903 [Follower]" leader= === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.344Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:53.353Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.345Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:53.366Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29903) (DC: dc1)" writer.go:29: 2021-01-29T19:34:53.366Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:53.370Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Started DNS server: address=127.0.0.1:29898 network=tcp writer.go:29: 2021-01-29T19:34:53.373Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Started DNS server: address=127.0.0.1:29898 network=udp writer.go:29: 2021-01-29T19:34:53.374Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Started HTTP server: address=127.0.0.1:29899 network=tcp writer.go:29: 2021-01-29T19:34:53.376Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: started state syncer writer.go:29: 2021-01-29T19:34:53.389Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:53.392Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: entering candidate state: node="Node at 127.0.0.1:29903 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:53.395Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:53.398Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: vote granted: from=1887784c-fc36-9267-a3ff-f79d1920fc88 term=2 tally=1 writer.go:29: 2021-01-29T19:34:53.401Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:53.403Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: entering leader state: leader="Node at 127.0.0.1:29903 [Leader]" writer.go:29: 2021-01-29T19:34:53.406Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:53.407Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:53.413Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29903 writer.go:29: 2021-01-29T19:34:53.424Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:53.433Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:53.435Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.438Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:53.440Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.536Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:53.537Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.560Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:53.563Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Synced node info writer.go:29: 2021-01-29T19:34:53.564Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.572Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:53.577Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:53.582Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} writer.go:29: 2021-01-29T19:34:53.590Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:53.591Z [INFO] TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.593Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6c108100-fca6-a3c9-1594-60fba64a27fc Address:127.0.0.1:29891}]" writer.go:29: 2021-01-29T19:34:53.595Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: entering follower state: follower="Node at 127.0.0.1:29891 [Follower]" leader= writer.go:29: 2021-01-29T19:34:53.598Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.wan: serf: EventMemberJoin: test-node.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:53.610Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.lan: serf: EventMemberJoin: test-node 127.0.0.1 writer.go:29: 2021-01-29T19:34:53.615Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Adding LAN server: server="test-node (Addr: tcp/127.0.0.1:29891) (DC: dc1)" writer.go:29: 2021-01-29T19:34:53.617Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Handled event for server in area: event=member-join server=test-node.dc1 area=wan writer.go:29: 2021-01-29T19:34:53.619Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Started DNS server: address=127.0.0.1:29886 network=udp writer.go:29: 2021-01-29T19:34:53.620Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Started DNS server: address=127.0.0.1:29886 network=tcp writer.go:29: 2021-01-29T19:34:53.621Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Started HTTP server: address=127.0.0.1:29887 network=tcp writer.go:29: 2021-01-29T19:34:53.623Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: started state syncer writer.go:29: 2021-01-29T19:34:53.642Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:53.643Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: entering candidate state: node="Node at 127.0.0.1:29891 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:53.646Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:53.648Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: vote granted: from=6c108100-fca6-a3c9-1594-60fba64a27fc term=2 tally=1 writer.go:29: 2021-01-29T19:34:53.650Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:53.656Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.raft: entering leader state: leader="Node at 127.0.0.1:29891 [Leader]" writer.go:29: 2021-01-29T19:34:53.659Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:53.660Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: New leader elected: payload=test-node writer.go:29: 2021-01-29T19:34:53.661Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29891 writer.go:29: 2021-01-29T19:34:53.726Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.734Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.735Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.736Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.738Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.741Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: Skipping self join check for node since the cluster is too small: node=test-node writer.go:29: 2021-01-29T19:34:53.743Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: member joined, marking health alive: member=test-node === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.804Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=1.122379ms client=127.0.0.1:35351 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.812Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:53.817Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Synced node info === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.810Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=714.315µs client=127.0.0.1:39806 client_network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.820Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Node info in sync === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:53.815Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.dns: request served from client: name=53fdb4f1-4b0c-d3e4-0180-a7f3582bd0a9.query.consul. type=AAAA class=IN latency=1.562886ms client=127.0.0.1:34216 client_network=udp writer.go:29: 2021-01-29T19:34:53.815Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Requesting shutdown writer.go:29: 2021-01-29T19:34:53.834Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server: shutting down server writer.go:29: 2021-01-29T19:34:53.835Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.837Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:53.837Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.841Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:53.853Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:53.853Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: consul server down writer.go:29: 2021-01-29T19:34:53.856Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: shutdown complete writer.go:29: 2021-01-29T19:34:53.858Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Stopping server: protocol=DNS address=127.0.0.1:29904 network=tcp writer.go:29: 2021-01-29T19:34:53.860Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Stopping server: protocol=DNS address=127.0.0.1:29904 network=udp writer.go:29: 2021-01-29T19:34:53.862Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Stopping server: protocol=HTTP address=127.0.0.1:29905 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:53.962Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.dns: request served from client: name=api-tier.service.consul. type=ANY class=IN latency=1.089628ms client=127.0.0.1:53957 client_network=udp writer.go:29: 2021-01-29T19:34:53.965Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.dns: request served from client: name=api-tier.query.consul. type=ANY class=IN latency=1.981851ms client=127.0.0.1:56327 client_network=udp writer.go:29: 2021-01-29T19:34:53.967Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.dns: request served from client: name=480f53b8-8429-c3ca-86ac-b9ec2e01d82e.query.consul. type=ANY class=IN latency=2.449881ms client=127.0.0.1:54039 client_network=udp writer.go:29: 2021-01-29T19:34:53.967Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Requesting shutdown writer.go:29: 2021-01-29T19:34:53.971Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: shutting down server writer.go:29: 2021-01-29T19:34:53.973Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.975Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.975Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:53.979Z [WARN] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:53.982Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:53.983Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: consul server down writer.go:29: 2021-01-29T19:34:53.985Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: shutdown complete writer.go:29: 2021-01-29T19:34:53.986Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Stopping server: protocol=DNS address=127.0.0.1:29898 network=tcp writer.go:29: 2021-01-29T19:34:53.988Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Stopping server: protocol=DNS address=127.0.0.1:29898 network=udp writer.go:29: 2021-01-29T19:34:53.989Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Stopping server: protocol=HTTP address=127.0.0.1:29899 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:54.300Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.dns: request served from client: name=api-tier.service.consul. type=AAAA class=IN latency=1.111187ms client=127.0.0.1:49187 client_network=udp writer.go:29: 2021-01-29T19:34:54.303Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.dns: request served from client: name=api-tier.query.consul. type=AAAA class=IN latency=1.659609ms client=127.0.0.1:32808 client_network=udp writer.go:29: 2021-01-29T19:34:54.305Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.dns: request served from client: name=0f3ef80c-4ea7-b906-972c-b7fd267bed6a.query.consul. type=AAAA class=IN latency=1.146888ms client=127.0.0.1:32957 client_network=udp writer.go:29: 2021-01-29T19:34:54.305Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Requesting shutdown writer.go:29: 2021-01-29T19:34:54.316Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server: shutting down server writer.go:29: 2021-01-29T19:34:54.318Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:54.320Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:54.322Z [DEBUG] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:54.344Z [WARN] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:54.352Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:54.352Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: consul server down writer.go:29: 2021-01-29T19:34:54.356Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: shutdown complete writer.go:29: 2021-01-29T19:34:54.358Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Stopping server: protocol=DNS address=127.0.0.1:29886 network=tcp writer.go:29: 2021-01-29T19:34:54.359Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Stopping server: protocol=DNS address=127.0.0.1:29886 network=udp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:54.364Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Waiting for endpoints to shut down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:54.365Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Stopping server: protocol=HTTP address=127.0.0.1:29887 network=tcp === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} writer.go:29: 2021-01-29T19:34:54.366Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:54.491Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:54.493Z [INFO] TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Endpoints down === CONT TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} writer.go:29: 2021-01-29T19:34:54.867Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:54.870Z [INFO] TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0}: Endpoints down --- PASS: TestDNS_ServiceLookup_AnswerLimits (0.05s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{0_0_0_0_0_0_0_0_0_0_0} (1.24s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{6_6_6_6_6_6_6_5_6_6_-5} (1.08s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} (0.88s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} (1.13s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{30_30_8_8_6_8_8_5_8_-5_-5} (1.12s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} (1.13s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{9_9_8_8_6_8_8_5_8_8_-5} (1.01s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} (1.10s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{20_20_8_8_6_8_8_5_8_-5_-5} (1.17s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{9_9_8_8_6_8_8_5_8_8_-5} (0.97s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{9_9_8_8_6_8_8_5_8_8_-5} (1.58s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{8_8_8_8_6_8_8_5_8_8_-5} (1.62s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{8_8_8_8_6_8_8_5_8_8_-5} (1.52s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{8_8_8_8_6_8_8_5_8_8_-5} (1.19s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{7_7_7_7_6_7_7_5_7_7_-5} (0.94s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{7_7_7_7_6_7_7_5_7_7_-5} (1.10s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{7_7_7_7_6_7_7_5_7_7_-5} (1.11s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{6_6_6_6_6_6_6_5_6_6_-5} (0.85s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{3_3_3_3_3_3_3_3_3_3_3} (1.03s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{5_5_5_5_5_5_5_5_5_5_5} (1.11s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{6_6_6_6_6_6_6_5_6_6_-5} (1.32s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{5_5_5_5_5_5_5_5_5_5_5} (1.08s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{4_4_4_4_4_4_4_4_4_4_4} (0.66s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{4_4_4_4_4_4_4_4_4_4_4} (1.23s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{5_5_5_5_5_5_5_5_5_5_5} (1.45s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{4_4_4_4_4_4_4_4_4_4_4} (1.29s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{3_3_3_3_3_3_3_3_3_3_3} (0.92s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{3_3_3_3_3_3_3_3_3_3_3} (0.94s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{2_2_2_2_2_2_2_2_2_2_2} (0.90s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{1_1_1_1_1_1_1_1_1_1_1} (1.45s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{2_2_2_2_2_2_2_2_2_2_2} (1.07s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{1_1_1_1_1_1_1_1_1_1_1} (0.85s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/A_lookup_{2_2_2_2_2_2_2_2_2_2_2} (0.95s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{1_1_1_1_1_1_1_1_1_1_1} (1.14s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/ANY_lookup_{0_0_0_0_0_0_0_0_0_0_0} (1.17s) --- PASS: TestDNS_ServiceLookup_AnswerLimits/AAAA_lookup_{0_0_0_0_0_0_0_0_0_0_0} (1.33s) PASS ok github.com/hashicorp/consul/agent 217.149s === RUN TestAE_scaleFactor === PAUSE TestAE_scaleFactor === RUN TestAE_Pause_nestedPauseResume === PAUSE TestAE_Pause_nestedPauseResume === RUN TestAE_Pause_ResumeTriggersSyncChanges --- PASS: TestAE_Pause_ResumeTriggersSyncChanges (0.00s) === RUN TestAE_staggerDependsOnClusterSize --- PASS: TestAE_staggerDependsOnClusterSize (0.00s) === RUN TestAE_Run_SyncFullBeforeChanges --- PASS: TestAE_Run_SyncFullBeforeChanges (0.00s) === RUN TestAE_Run_Quit === RUN TestAE_Run_Quit/Run_panics_without_ClusterSize === RUN TestAE_Run_Quit/runFSM_quits --- PASS: TestAE_Run_Quit (0.00s) --- PASS: TestAE_Run_Quit/Run_panics_without_ClusterSize (0.00s) --- PASS: TestAE_Run_Quit/runFSM_quits (0.00s) === RUN TestAE_FSM === RUN TestAE_FSM/fullSyncState === RUN TestAE_FSM/fullSyncState/Paused_->_retryFullSyncState === RUN TestAE_FSM/fullSyncState/SyncFull()_error_->_retryFullSyncState writer.go:29: 2021-01-29T19:31:01.570Z [ERROR] anti_entropy: failed to sync remote state: error=boom === RUN TestAE_FSM/fullSyncState/SyncFull()_OK_->_partialSyncState === RUN TestAE_FSM/retryFullSyncState === RUN TestAE_FSM/retryFullSyncState/shutdownEvent_->_doneState === RUN TestAE_FSM/retryFullSyncState/syncFullNotifEvent_->_fullSyncState === RUN TestAE_FSM/retryFullSyncState/syncFullTimerEvent_->_fullSyncState === RUN TestAE_FSM/retryFullSyncState/invalid_event_->_panic_ === RUN TestAE_FSM/partialSyncState === RUN TestAE_FSM/partialSyncState/shutdownEvent_->_doneState === RUN TestAE_FSM/partialSyncState/syncFullNotifEvent_->_fullSyncState === RUN TestAE_FSM/partialSyncState/syncFullTimerEvent_->_fullSyncState === RUN TestAE_FSM/partialSyncState/syncChangesEvent+Paused_->_partialSyncState === RUN TestAE_FSM/partialSyncState/syncChangesEvent+SyncChanges()_error_->_partialSyncState writer.go:29: 2021-01-29T19:31:01.573Z [ERROR] anti_entropy: failed to sync changes: error=boom === RUN TestAE_FSM/partialSyncState/syncChangesEvent+SyncChanges()_OK_->_partialSyncState === RUN TestAE_FSM/partialSyncState/invalid_event_->_panic_ === RUN TestAE_FSM/invalid_state_->_panic_ --- PASS: TestAE_FSM (0.00s) --- PASS: TestAE_FSM/fullSyncState (0.00s) --- PASS: TestAE_FSM/fullSyncState/Paused_->_retryFullSyncState (0.00s) --- PASS: TestAE_FSM/fullSyncState/SyncFull()_error_->_retryFullSyncState (0.00s) --- PASS: TestAE_FSM/fullSyncState/SyncFull()_OK_->_partialSyncState (0.00s) --- PASS: TestAE_FSM/retryFullSyncState (0.00s) --- PASS: TestAE_FSM/retryFullSyncState/shutdownEvent_->_doneState (0.00s) --- PASS: TestAE_FSM/retryFullSyncState/syncFullNotifEvent_->_fullSyncState (0.00s) --- PASS: TestAE_FSM/retryFullSyncState/syncFullTimerEvent_->_fullSyncState (0.00s) --- PASS: TestAE_FSM/retryFullSyncState/invalid_event_->_panic_ (0.00s) --- PASS: TestAE_FSM/partialSyncState (0.00s) --- PASS: TestAE_FSM/partialSyncState/shutdownEvent_->_doneState (0.00s) --- PASS: TestAE_FSM/partialSyncState/syncFullNotifEvent_->_fullSyncState (0.00s) --- PASS: TestAE_FSM/partialSyncState/syncFullTimerEvent_->_fullSyncState (0.00s) --- PASS: TestAE_FSM/partialSyncState/syncChangesEvent+Paused_->_partialSyncState (0.00s) --- PASS: TestAE_FSM/partialSyncState/syncChangesEvent+SyncChanges()_error_->_partialSyncState (0.00s) --- PASS: TestAE_FSM/partialSyncState/syncChangesEvent+SyncChanges()_OK_->_partialSyncState (0.00s) --- PASS: TestAE_FSM/partialSyncState/invalid_event_->_panic_ (0.00s) --- PASS: TestAE_FSM/invalid_state_->_panic_ (0.00s) === RUN TestAE_RetrySyncFullEvent === RUN TestAE_RetrySyncFullEvent/trigger_shutdownEvent === RUN TestAE_RetrySyncFullEvent/trigger_shutdownEvent_during_FullNotif === RUN TestAE_RetrySyncFullEvent/trigger_syncFullNotifEvent === RUN TestAE_RetrySyncFullEvent/trigger_syncFullTimerEvent --- PASS: TestAE_RetrySyncFullEvent (0.13s) --- PASS: TestAE_RetrySyncFullEvent/trigger_shutdownEvent (0.00s) --- PASS: TestAE_RetrySyncFullEvent/trigger_shutdownEvent_during_FullNotif (0.10s) --- PASS: TestAE_RetrySyncFullEvent/trigger_syncFullNotifEvent (0.01s) --- PASS: TestAE_RetrySyncFullEvent/trigger_syncFullTimerEvent (0.02s) === RUN TestAE_SyncChangesEvent === RUN TestAE_SyncChangesEvent/trigger_shutdownEvent === RUN TestAE_SyncChangesEvent/trigger_shutdownEvent_during_FullNotif === RUN TestAE_SyncChangesEvent/trigger_syncFullNotifEvent === RUN TestAE_SyncChangesEvent/trigger_syncFullTimerEvent === RUN TestAE_SyncChangesEvent/trigger_syncChangesNotifEvent --- PASS: TestAE_SyncChangesEvent (2.11s) --- PASS: TestAE_SyncChangesEvent/trigger_shutdownEvent (0.00s) --- PASS: TestAE_SyncChangesEvent/trigger_shutdownEvent_during_FullNotif (0.10s) --- PASS: TestAE_SyncChangesEvent/trigger_syncFullNotifEvent (0.01s) --- PASS: TestAE_SyncChangesEvent/trigger_syncFullTimerEvent (2.00s) --- PASS: TestAE_SyncChangesEvent/trigger_syncChangesNotifEvent (0.00s) === CONT TestAE_scaleFactor === CONT TestAE_Pause_nestedPauseResume === RUN TestAE_scaleFactor/100_nodes --- PASS: TestAE_Pause_nestedPauseResume (0.00s) === RUN TestAE_scaleFactor/200_nodes === RUN TestAE_scaleFactor/1000_nodes === RUN TestAE_scaleFactor/10000_nodes --- PASS: TestAE_scaleFactor (0.00s) --- PASS: TestAE_scaleFactor/100_nodes (0.00s) --- PASS: TestAE_scaleFactor/200_nodes (0.00s) --- PASS: TestAE_scaleFactor/1000_nodes (0.00s) --- PASS: TestAE_scaleFactor/10000_nodes (0.00s) PASS ok github.com/hashicorp/consul/agent/ae 2.269s ? github.com/hashicorp/consul/agent/agentpb [no test files] === RUN TestBuilder_validateEnterpriseConfigKeys === RUN TestBuilder_validateEnterpriseConfigKeys/multi === RUN TestBuilder_validateEnterpriseConfigKeys/segment === RUN TestBuilder_validateEnterpriseConfigKeys/segments === RUN TestBuilder_validateEnterpriseConfigKeys/autopilot.redundancy_zone_tag === RUN TestBuilder_validateEnterpriseConfigKeys/autopilot.disable_upgrade_migration === RUN TestBuilder_validateEnterpriseConfigKeys/dns_config.prefer_namespace === RUN TestBuilder_validateEnterpriseConfigKeys/acl.msp_disable_bootstrap === RUN TestBuilder_validateEnterpriseConfigKeys/non_voting_server === RUN TestBuilder_validateEnterpriseConfigKeys/autopilot.upgrade_version_tag === RUN TestBuilder_validateEnterpriseConfigKeys/acl.tokens.managed_service_provider --- PASS: TestBuilder_validateEnterpriseConfigKeys (0.04s) --- PASS: TestBuilder_validateEnterpriseConfigKeys/multi (0.00s) --- PASS: TestBuilder_validateEnterpriseConfigKeys/segment (0.00s) --- PASS: TestBuilder_validateEnterpriseConfigKeys/segments (0.00s) --- PASS: TestBuilder_validateEnterpriseConfigKeys/autopilot.redundancy_zone_tag (0.00s) --- PASS: TestBuilder_validateEnterpriseConfigKeys/autopilot.disable_upgrade_migration (0.00s) --- PASS: TestBuilder_validateEnterpriseConfigKeys/dns_config.prefer_namespace (0.00s) --- PASS: TestBuilder_validateEnterpriseConfigKeys/acl.msp_disable_bootstrap (0.00s) --- PASS: TestBuilder_validateEnterpriseConfigKeys/non_voting_server (0.00s) --- PASS: TestBuilder_validateEnterpriseConfigKeys/autopilot.upgrade_version_tag (0.00s) --- PASS: TestBuilder_validateEnterpriseConfigKeys/acl.tokens.managed_service_provider (0.00s) === RUN TestParseFlags === RUN TestParseFlags/#00 === RUN TestParseFlags/-bind_a === RUN TestParseFlags/-bootstrap === RUN TestParseFlags/-bootstrap=true === RUN TestParseFlags/-bootstrap=false === RUN TestParseFlags/-config-file_a_-config-dir_b_-config-file_c_-config-dir_d === RUN TestParseFlags/-datacenter_a === RUN TestParseFlags/-dns-port_1 === RUN TestParseFlags/-grpc-port_1 === RUN TestParseFlags/-http-port_1 === RUN TestParseFlags/-https-port_1 === RUN TestParseFlags/-serf-lan-port_1 === RUN TestParseFlags/-serf-wan-port_1 === RUN TestParseFlags/-server-port_1 === RUN TestParseFlags/-join_a_-join_b === RUN TestParseFlags/-node-meta_a:b_-node-meta_c:d === RUN TestParseFlags/-bootstrap_true --- PASS: TestParseFlags (0.02s) --- PASS: TestParseFlags/#00 (0.00s) --- PASS: TestParseFlags/-bind_a (0.00s) --- PASS: TestParseFlags/-bootstrap (0.00s) --- PASS: TestParseFlags/-bootstrap=true (0.00s) --- PASS: TestParseFlags/-bootstrap=false (0.00s) --- PASS: TestParseFlags/-config-file_a_-config-dir_b_-config-file_c_-config-dir_d (0.00s) --- PASS: TestParseFlags/-datacenter_a (0.00s) --- PASS: TestParseFlags/-dns-port_1 (0.00s) --- PASS: TestParseFlags/-grpc-port_1 (0.00s) --- PASS: TestParseFlags/-http-port_1 (0.00s) --- PASS: TestParseFlags/-https-port_1 (0.00s) --- PASS: TestParseFlags/-serf-lan-port_1 (0.00s) --- PASS: TestParseFlags/-serf-wan-port_1 (0.00s) --- PASS: TestParseFlags/-server-port_1 (0.00s) --- PASS: TestParseFlags/-join_a_-join_b (0.00s) --- PASS: TestParseFlags/-node-meta_a:b_-node-meta_c:d (0.00s) --- PASS: TestParseFlags/-bootstrap_true (0.00s) === RUN TestMerge === RUN TestMerge/top_level_fields --- PASS: TestMerge (0.00s) --- PASS: TestMerge/top_level_fields (0.00s) === RUN TestConfigFlagsAndEdgecases runtime_test.go:52: DM-skipped --- SKIP: TestConfigFlagsAndEdgecases (0.00s) === RUN TestFullConfig runtime_test.go:3668: DM-skipped --- SKIP: TestFullConfig (0.01s) === RUN TestNonZero === RUN TestNonZero/nil === RUN TestNonZero/zero_bool === RUN TestNonZero/zero_string === RUN TestNonZero/zero_int === RUN TestNonZero/zero_int8 === RUN TestNonZero/zero_int16 === RUN TestNonZero/zero_int32 === RUN TestNonZero/zero_int64 === RUN TestNonZero/zero_uint === RUN TestNonZero/zero_uint8 === RUN TestNonZero/zero_uint16 === RUN TestNonZero/zero_uint32 === RUN TestNonZero/zero_uint64 === RUN TestNonZero/zero_float32 === RUN TestNonZero/zero_float64 === RUN TestNonZero/ptr_to_zero_value === RUN TestNonZero/empty_slice === RUN TestNonZero/slice_with_zero_value === RUN TestNonZero/empty_map === RUN TestNonZero/map_with_zero_value_key === RUN TestNonZero/map_with_zero_value_elem === RUN TestNonZero/struct_with_nil_field === RUN TestNonZero/struct_with_zero_value_field === RUN TestNonZero/struct_with_empty_array --- PASS: TestNonZero (0.01s) --- PASS: TestNonZero/nil (0.00s) --- PASS: TestNonZero/zero_bool (0.00s) --- PASS: TestNonZero/zero_string (0.00s) --- PASS: TestNonZero/zero_int (0.00s) --- PASS: TestNonZero/zero_int8 (0.00s) --- PASS: TestNonZero/zero_int16 (0.00s) --- PASS: TestNonZero/zero_int32 (0.00s) --- PASS: TestNonZero/zero_int64 (0.00s) --- PASS: TestNonZero/zero_uint (0.00s) --- PASS: TestNonZero/zero_uint8 (0.00s) --- PASS: TestNonZero/zero_uint16 (0.00s) --- PASS: TestNonZero/zero_uint32 (0.00s) --- PASS: TestNonZero/zero_uint64 (0.00s) --- PASS: TestNonZero/zero_float32 (0.00s) --- PASS: TestNonZero/zero_float64 (0.00s) --- PASS: TestNonZero/ptr_to_zero_value (0.00s) --- PASS: TestNonZero/empty_slice (0.00s) --- PASS: TestNonZero/slice_with_zero_value (0.00s) --- PASS: TestNonZero/empty_map (0.00s) --- PASS: TestNonZero/map_with_zero_value_key (0.00s) --- PASS: TestNonZero/map_with_zero_value_elem (0.00s) --- PASS: TestNonZero/struct_with_nil_field (0.00s) --- PASS: TestNonZero/struct_with_zero_value_field (0.00s) --- PASS: TestNonZero/struct_with_empty_array (0.00s) === RUN TestConfigDecodeBytes === PAUSE TestConfigDecodeBytes === RUN TestSanitize --- PASS: TestSanitize (0.01s) === RUN TestRuntime_apiAddresses --- PASS: TestRuntime_apiAddresses (0.00s) === RUN TestRuntime_APIConfigHTTPS --- PASS: TestRuntime_APIConfigHTTPS (0.01s) === RUN TestRuntime_APIConfigHTTP --- PASS: TestRuntime_APIConfigHTTP (0.00s) === RUN TestRuntime_APIConfigUNIX --- PASS: TestRuntime_APIConfigUNIX (0.00s) === RUN TestRuntime_APIConfigANYAddrV4 --- PASS: TestRuntime_APIConfigANYAddrV4 (0.00s) === RUN TestRuntime_APIConfigANYAddrV6 --- PASS: TestRuntime_APIConfigANYAddrV6 (0.00s) === RUN TestRuntime_ClientAddress --- PASS: TestRuntime_ClientAddress (0.00s) === RUN TestRuntime_ClientAddressAnyV4 --- PASS: TestRuntime_ClientAddressAnyV4 (0.00s) === RUN TestRuntime_ClientAddressAnyV6 --- PASS: TestRuntime_ClientAddressAnyV6 (0.00s) === RUN TestRuntime_ToTLSUtilConfig --- PASS: TestRuntime_ToTLSUtilConfig (0.00s) === RUN Test_UIPathBuilder --- PASS: Test_UIPathBuilder (0.00s) === RUN TestSegments segment_oss_test.go:13: DM-skipped --- SKIP: TestSegments (0.00s) === CONT TestConfigDecodeBytes --- PASS: TestConfigDecodeBytes (0.00s) PASS ok github.com/hashicorp/consul/agent/config 0.203s === RUN TestCollectHostInfo --- PASS: TestCollectHostInfo (0.00s) PASS ok github.com/hashicorp/consul/agent/debug 0.047s ? github.com/hashicorp/consul/agent/exec [no test files] === RUN TestAgentAntiEntropy_Services === PAUSE TestAgentAntiEntropy_Services === RUN TestAgentAntiEntropy_Services_ConnectProxy === PAUSE TestAgentAntiEntropy_Services_ConnectProxy === RUN TestAgent_ServiceWatchCh === PAUSE TestAgent_ServiceWatchCh === RUN TestAgentAntiEntropy_EnableTagOverride === PAUSE TestAgentAntiEntropy_EnableTagOverride === RUN TestAgentAntiEntropy_Services_WithChecks === PAUSE TestAgentAntiEntropy_Services_WithChecks === RUN TestAgentAntiEntropy_Services_ACLDeny === PAUSE TestAgentAntiEntropy_Services_ACLDeny === RUN TestAgentAntiEntropy_Checks === PAUSE TestAgentAntiEntropy_Checks === RUN TestAgentAntiEntropy_RemovingServiceAndCheck === PAUSE TestAgentAntiEntropy_RemovingServiceAndCheck === RUN TestAgentAntiEntropy_Checks_ACLDeny === PAUSE TestAgentAntiEntropy_Checks_ACLDeny === RUN TestAgent_UpdateCheck_DiscardOutput state_test.go:1380: DM-skipped --- SKIP: TestAgent_UpdateCheck_DiscardOutput (0.00s) === RUN TestAgentAntiEntropy_Check_DeferSync state_test.go:1432: DM-skipped --- SKIP: TestAgentAntiEntropy_Check_DeferSync (0.00s) === RUN TestAgentAntiEntropy_NodeInfo === PAUSE TestAgentAntiEntropy_NodeInfo === RUN TestAgent_ServiceTokens === PAUSE TestAgent_ServiceTokens === RUN TestAgent_CheckTokens === PAUSE TestAgent_CheckTokens === RUN TestAgent_CheckCriticalTime state_test.go:1768: DM-skipped --- SKIP: TestAgent_CheckCriticalTime (0.00s) === RUN TestAgent_AddCheckFailure === PAUSE TestAgent_AddCheckFailure === RUN TestAgent_AliasCheck === PAUSE TestAgent_AliasCheck === RUN TestAgent_sendCoordinate === PAUSE TestAgent_sendCoordinate === RUN TestState_Notify === PAUSE TestState_Notify === RUN TestAliasNotifications_local === PAUSE TestAliasNotifications_local === CONT TestAgentAntiEntropy_Services === CONT TestAgentAntiEntropy_NodeInfo [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestAgentAntiEntropy_Services_ACLDeny === CONT TestAgentAntiEntropy_Checks_ACLDeny [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:15.946Z [WARN] TestAgentAntiEntropy_Services_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. writer.go:29: 2021-01-29T19:31:15.949Z [WARN] TestAgentAntiEntropy_Services_ACLDeny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:15.954Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:15.989Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 2021-01-29T19:31:15.993Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny: The 'acl_datacenter' field is deprecated. Use the 'primary_datacenter' field instead. 2021-01-29T19:31:15.993Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny: bootstrap = true: do not enable unless necessary 2021-01-29T19:31:15.993Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.tlsutil: Update: version=1 === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:15.995Z [WARN] TestAgentAntiEntropy_Services: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:15.996Z [DEBUG] TestAgentAntiEntropy_Services.tlsutil: Update: version=1 2021-01-29T19:31:15.998Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:15.998Z [DEBUG] TestAgentAntiEntropy_Services.tlsutil: OutgoingRPCWrapper: version=1 2021-01-29T19:31:16.001Z [WARN] TestAgentAntiEntropy_NodeInfo: bootstrap = true: do not enable unless necessary 2021-01-29T19:31:16.001Z [DEBUG] TestAgentAntiEntropy_NodeInfo.tlsutil: Update: version=1 2021-01-29T19:31:16.001Z [DEBUG] TestAgentAntiEntropy_NodeInfo.tlsutil: OutgoingRPCWrapper: version=1 2021-01-29T19:31:16.007Z [INFO] TestAgentAntiEntropy_NodeInfo.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:40e4a748-2192-161a-0510-9bf59fe950b5 Address:127.0.0.1:19195}]" 2021-01-29T19:31:16.009Z [INFO] TestAgentAntiEntropy_NodeInfo.server.raft: entering follower state: follower="Node at 127.0.0.1:19195 [Follower]" leader= 2021-01-29T19:31:16.010Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7477723e-990b-2201-46a8-5be1171561af Address:127.0.0.1:19207}]" === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.013Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fc7e5164-fd57-6376-365c-894cedf13b85 Address:127.0.0.1:19201}]" === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.014Z [INFO] TestAgentAntiEntropy_Services.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7f7d30ad-009e-a186-6583-4ec80cf6bd8e Address:127.0.0.1:19213}]" 2021-01-29T19:31:16.017Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-7477723e-990b-2201-46a8-5be1171561af.dc1 127.0.0.1 === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.016Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.serf.wan: serf: EventMemberJoin: Node-fc7e5164-fd57-6376-365c-894cedf13b85.dc1 127.0.0.1 2021-01-29T19:31:16.020Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-7477723e-990b-2201-46a8-5be1171561af 127.0.0.1 writer.go:29: 2021-01-29T19:31:16.019Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.serf.lan: serf: EventMemberJoin: Node-fc7e5164-fd57-6376-365c-894cedf13b85 127.0.0.1 === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.019Z [INFO] TestAgentAntiEntropy_Services.server.serf.wan: serf: EventMemberJoin: Node-7f7d30ad-009e-a186-6583-4ec80cf6bd8e.dc1 127.0.0.1 === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.022Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:19201 [Follower]" leader= 2021-01-29T19:31:16.028Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Started DNS server: address=127.0.0.1:19202 network=udp 2021-01-29T19:31:16.039Z [INFO] TestAgentAntiEntropy_NodeInfo.server.serf.wan: serf: EventMemberJoin: Node-8bbe191d-1144-18b5-5830-f4da62c6478e.dc1 127.0.0.1 2021-01-29T19:31:16.056Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: Adding LAN server: server="Node-7477723e-990b-2201-46a8-5be1171561af (Addr: tcp/127.0.0.1:19207) (DC: dc1)" 2021-01-29T19:31:16.056Z [WARN] TestAgentAntiEntropy_NodeInfo.server.raft: heartbeat timeout reached, starting election: last-leader= 2021-01-29T19:31:16.056Z [INFO] TestAgentAntiEntropy_NodeInfo.server.raft: entering candidate state: node="Node at 127.0.0.1:19195 [Candidate]" term=2 2021-01-29T19:31:16.056Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: Handled event for server in area: event=member-join server=Node-7477723e-990b-2201-46a8-5be1171561af.dc1 area=wan 2021-01-29T19:31:16.056Z [WARN] TestAgentAntiEntropy_NodeInfo.server.raft: unable to get address for sever, using fallback address: id=40e4a748-2192-161a-0510-9bf59fe950b5 fallback=127.0.0.1:19195 error="Could not find address for server id 40e4a748-2192-161a-0510-9bf59fe950b5" 2021-01-29T19:31:16.057Z [DEBUG] TestAgentAntiEntropy_NodeInfo.server.raft: votes: needed=1 2021-01-29T19:31:16.057Z [DEBUG] TestAgentAntiEntropy_NodeInfo.server.raft: vote granted: from=40e4a748-2192-161a-0510-9bf59fe950b5 term=2 tally=1 2021-01-29T19:31:16.057Z [INFO] TestAgentAntiEntropy_NodeInfo.server.raft: election won: tally=1 2021-01-29T19:31:16.057Z [INFO] TestAgentAntiEntropy_NodeInfo.server.raft: entering leader state: leader="Node at 127.0.0.1:19195 [Leader]" writer.go:29: 2021-01-29T19:31:16.056Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: Adding LAN server: server="Node-fc7e5164-fd57-6376-365c-894cedf13b85 (Addr: tcp/127.0.0.1:19201) (DC: dc1)" === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.038Z [INFO] TestAgentAntiEntropy_Services.server.raft: entering follower state: follower="Node at 127.0.0.1:19213 [Follower]" leader= 2021-01-29T19:31:16.059Z [INFO] TestAgentAntiEntropy_NodeInfo.server.serf.lan: serf: EventMemberJoin: Node-8bbe191d-1144-18b5-5830-f4da62c6478e 127.0.0.1 2021-01-29T19:31:16.059Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Started DNS server: address=127.0.0.1:19202 network=tcp 2021-01-29T19:31:16.060Z [INFO] TestAgentAntiEntropy_NodeInfo: Started DNS server: address=127.0.0.1:19190 network=udp 2021-01-29T19:31:16.060Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Started HTTP server: address=127.0.0.1:19203 network=tcp 2021-01-29T19:31:16.060Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: started state syncer === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.058Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: Handled event for server in area: event=member-join server=Node-fc7e5164-fd57-6376-365c-894cedf13b85.dc1 area=wan === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.059Z [INFO] TestAgentAntiEntropy_Services.server.serf.lan: serf: EventMemberJoin: Node-7f7d30ad-009e-a186-6583-4ec80cf6bd8e 127.0.0.1 2021-01-29T19:31:16.062Z [INFO] TestAgentAntiEntropy_NodeInfo.server: Adding LAN server: server="Node-8bbe191d-1144-18b5-5830-f4da62c6478e (Addr: tcp/127.0.0.1:19195) (DC: dc1)" 2021-01-29T19:31:16.062Z [INFO] TestAgentAntiEntropy_NodeInfo.server: cluster leadership acquired 2021-01-29T19:31:16.064Z [INFO] TestAgentAntiEntropy_NodeInfo.server: Handled event for server in area: event=member-join server=Node-8bbe191d-1144-18b5-5830-f4da62c6478e.dc1 area=wan === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.062Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Started DNS server: address=127.0.0.1:19196 network=udp 2021-01-29T19:31:16.066Z [INFO] TestAgentAntiEntropy_NodeInfo: Started DNS server: address=127.0.0.1:19190 network=tcp writer.go:29: 2021-01-29T19:31:16.065Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Started DNS server: address=127.0.0.1:19196 network=tcp 2021-01-29T19:31:16.067Z [INFO] TestAgentAntiEntropy_NodeInfo: Started HTTP server: address=127.0.0.1:19191 network=tcp 2021-01-29T19:31:16.067Z [INFO] TestAgentAntiEntropy_NodeInfo: started state syncer 2021-01-29T19:31:16.071Z [INFO] TestAgentAntiEntropy_NodeInfo.server: New leader elected: payload=Node-8bbe191d-1144-18b5-5830-f4da62c6478e 2021-01-29T19:31:16.068Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.raft: entering follower state: follower="Node at 127.0.0.1:19207 [Follower]" leader= writer.go:29: 2021-01-29T19:31:16.070Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Started HTTP server: address=127.0.0.1:19197 network=tcp writer.go:29: 2021-01-29T19:31:16.072Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: started state syncer === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.063Z [INFO] TestAgentAntiEntropy_Services: Started DNS server: address=127.0.0.1:19208 network=udp 2021-01-29T19:31:16.076Z [DEBUG] TestAgentAntiEntropy_NodeInfo.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19195 writer.go:29: 2021-01-29T19:31:16.074Z [INFO] TestAgentAntiEntropy_Services.server: Adding LAN server: server="Node-7f7d30ad-009e-a186-6583-4ec80cf6bd8e (Addr: tcp/127.0.0.1:19213) (DC: dc1)" writer.go:29: 2021-01-29T19:31:16.074Z [INFO] TestAgentAntiEntropy_Services.server: Handled event for server in area: event=member-join server=Node-7f7d30ad-009e-a186-6583-4ec80cf6bd8e.dc1 area=wan writer.go:29: 2021-01-29T19:31:16.077Z [INFO] TestAgentAntiEntropy_Services: Started DNS server: address=127.0.0.1:19208 network=tcp 2021-01-29T19:31:16.081Z [INFO] TestAgentAntiEntropy_NodeInfo: Synced node info 2021-01-29T19:31:16.081Z [DEBUG] TestAgentAntiEntropy_NodeInfo: Node info in sync writer.go:29: 2021-01-29T19:31:16.082Z [INFO] TestAgentAntiEntropy_Services: Started HTTP server: address=127.0.0.1:19209 network=tcp 2021-01-29T19:31:16.086Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:16.085Z [INFO] TestAgentAntiEntropy_Services: started state syncer === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.093Z [WARN] TestAgentAntiEntropy_Services_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:16.095Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:19201 [Candidate]" term=2 2021-01-29T19:31:16.102Z [INFO] TestAgentAntiEntropy_NodeInfo: Synced node info 2021-01-29T19:31:16.105Z [INFO] TestAgentAntiEntropy_NodeInfo: Synced node info 2021-01-29T19:31:16.107Z [INFO] TestAgentAntiEntropy_NodeInfo: Requesting shutdown 2021-01-29T19:31:16.107Z [INFO] TestAgentAntiEntropy_NodeInfo.server: shutting down server 2021-01-29T19:31:16.107Z [WARN] TestAgentAntiEntropy_NodeInfo.server.serf.lan: serf: Shutdown without a Leave 2021-01-29T19:31:16.109Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= 2021-01-29T19:31:16.109Z [WARN] TestAgentAntiEntropy_NodeInfo.server.serf.wan: serf: Shutdown without a Leave 2021-01-29T19:31:16.110Z [INFO] TestAgentAntiEntropy_NodeInfo.server.router.manager: shutting down 2021-01-29T19:31:16.110Z [INFO] TestAgentAntiEntropy_NodeInfo: consul server down 2021-01-29T19:31:16.110Z [INFO] TestAgentAntiEntropy_NodeInfo: shutdown complete 2021-01-29T19:31:16.110Z [INFO] TestAgentAntiEntropy_NodeInfo: Stopping server: protocol=DNS address=127.0.0.1:19190 network=tcp 2021-01-29T19:31:16.110Z [INFO] TestAgentAntiEntropy_NodeInfo: Stopping server: protocol=DNS address=127.0.0.1:19190 network=udp 2021-01-29T19:31:16.110Z [INFO] TestAgentAntiEntropy_NodeInfo: Stopping server: protocol=HTTP address=127.0.0.1:19191 network=tcp writer.go:29: 2021-01-29T19:31:16.105Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.server.raft: votes: needed=1 2021-01-29T19:31:16.113Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny.server.raft: heartbeat timeout reached, starting election: last-leader= 2021-01-29T19:31:16.113Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.raft: entering candidate state: node="Node at 127.0.0.1:19207 [Candidate]" term=2 2021-01-29T19:31:16.115Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.server.raft: votes: needed=1 2021-01-29T19:31:16.115Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.server.raft: vote granted: from=7477723e-990b-2201-46a8-5be1171561af term=2 tally=1 2021-01-29T19:31:16.115Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.raft: election won: tally=1 2021-01-29T19:31:16.115Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:19207 [Leader]" 2021-01-29T19:31:16.115Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: cluster leadership acquired 2021-01-29T19:31:16.116Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: New leader elected: payload=Node-7477723e-990b-2201-46a8-5be1171561af writer.go:29: 2021-01-29T19:31:16.112Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.server.raft: vote granted: from=fc7e5164-fd57-6376-365c-894cedf13b85 term=2 tally=1 2021-01-29T19:31:16.116Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: initializing acls writer.go:29: 2021-01-29T19:31:16.106Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= 2021-01-29T19:31:16.118Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: Created ACL 'global-management' policy 2021-01-29T19:31:16.118Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny.server: Configuring a non-UUID master token is deprecated 2021-01-29T19:31:16.119Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: Bootstrapped ACL master token from configuration 2021-01-29T19:31:16.120Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: Created ACL anonymous token from configuration 2021-01-29T19:31:16.120Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" 2021-01-29T19:31:16.120Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.leader: started routine: routine="acl token reaping" 2021-01-29T19:31:16.120Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-7477723e-990b-2201-46a8-5be1171561af 2021-01-29T19:31:16.120Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-7477723e-990b-2201-46a8-5be1171561af.dc1 2021-01-29T19:31:16.121Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: Updating LAN server: server="Node-7477723e-990b-2201-46a8-5be1171561af (Addr: tcp/127.0.0.1:19207) (DC: dc1)" 2021-01-29T19:31:16.121Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: Handled event for server in area: event=member-update server=Node-7477723e-990b-2201-46a8-5be1171561af.dc1 area=wan writer.go:29: 2021-01-29T19:31:16.116Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.raft: election won: tally=1 2021-01-29T19:31:16.122Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.121Z [WARN] TestAgentAntiEntropy_Services.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:16.122Z [INFO] TestAgentAntiEntropy_Services.server.raft: entering candidate state: node="Node at 127.0.0.1:19213 [Candidate]" term=2 === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.121Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.raft: entering leader state: leader="Node at 127.0.0.1:19201 [Leader]" === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.124Z [DEBUG] TestAgentAntiEntropy_Services.server.raft: votes: needed=1 === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.124Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: cluster leadership acquired === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.125Z [DEBUG] TestAgentAntiEntropy_Services.server.raft: vote granted: from=7f7d30ad-009e-a186-6583-4ec80cf6bd8e term=2 tally=1 writer.go:29: 2021-01-29T19:31:16.126Z [INFO] TestAgentAntiEntropy_Services.server.raft: election won: tally=1 === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.126Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: New leader elected: payload=Node-fc7e5164-fd57-6376-365c-894cedf13b85 === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.127Z [INFO] TestAgentAntiEntropy_Services.server.raft: entering leader state: leader="Node at 127.0.0.1:19213 [Leader]" writer.go:29: 2021-01-29T19:31:16.128Z [INFO] TestAgentAntiEntropy_Services.server: cluster leadership acquired === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.128Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: initializing acls === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.128Z [INFO] TestAgentAntiEntropy_Services.server: New leader elected: payload=Node-7f7d30ad-009e-a186-6583-4ec80cf6bd8e writer.go:29: 2021-01-29T19:31:16.128Z [DEBUG] TestAgentAntiEntropy_Services.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19213 === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.131Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:16.132Z [WARN] TestAgentAntiEntropy_Services_ACLDeny.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:16.134Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: Bootstrapped ACL master token from configuration === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.136Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.136Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:16.137Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:16.138Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:16.139Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.serf.lan: serf: EventMemberUpdate: Node-fc7e5164-fd57-6376-365c-894cedf13b85 writer.go:29: 2021-01-29T19:31:16.141Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.serf.wan: serf: EventMemberUpdate: Node-fc7e5164-fd57-6376-365c-894cedf13b85.dc1 writer.go:29: 2021-01-29T19:31:16.141Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: Updating LAN server: server="Node-fc7e5164-fd57-6376-365c-894cedf13b85 (Addr: tcp/127.0.0.1:19201) (DC: dc1)" writer.go:29: 2021-01-29T19:31:16.142Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: Handled event for server in area: event=member-update server=Node-fc7e5164-fd57-6376-365c-894cedf13b85.dc1 area=wan 2021-01-29T19:31:16.157Z [ERROR] TestAgentAntiEntropy_NodeInfo.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: raft is already shutdown" writer.go:29: 2021-01-29T19:31:16.161Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true 2021-01-29T19:31:16.175Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul 2021-01-29T19:31:16.175Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.leader: started routine: routine="CA root pruning" 2021-01-29T19:31:16.175Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-7477723e-990b-2201-46a8-5be1171561af 2021-01-29T19:31:16.175Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: member joined, marking health alive: member=Node-7477723e-990b-2201-46a8-5be1171561af 2021-01-29T19:31:16.176Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-7477723e-990b-2201-46a8-5be1171561af writer.go:29: 2021-01-29T19:31:16.176Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:16.177Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.leader: started routine: routine="CA root pruning" === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.175Z [INFO] TestAgentAntiEntropy_Services.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.178Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-fc7e5164-fd57-6376-365c-894cedf13b85 writer.go:29: 2021-01-29T19:31:16.179Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: member joined, marking health alive: member=Node-fc7e5164-fd57-6376-365c-894cedf13b85 === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.179Z [INFO] TestAgentAntiEntropy_Services.leader: started routine: routine="CA root pruning" 2021-01-29T19:31:16.186Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.acl: dropping check from result due to ACLs: check=serfHealth 2021-01-29T19:31:16.186Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny: Node info update blocked by ACLs: node=7477723e-990b-2201-46a8-5be1171561af accessorID= 2021-01-29T19:31:16.186Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny: Node info in sync writer.go:29: 2021-01-29T19:31:16.182Z [DEBUG] TestAgentAntiEntropy_Services.server: Skipping self join check for node since the cluster is too small: node=Node-7f7d30ad-009e-a186-6583-4ec80cf6bd8e === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.186Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.server: Skipping self join check for node since the cluster is too small: node=Node-fc7e5164-fd57-6376-365c-894cedf13b85 === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.186Z [INFO] TestAgentAntiEntropy_Services.server: member joined, marking health alive: member=Node-7f7d30ad-009e-a186-6583-4ec80cf6bd8e === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.192Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.acl: dropping node from result due to ACLs: node=Node-fc7e5164-fd57-6376-365c-894cedf13b85 writer.go:29: 2021-01-29T19:31:16.198Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.acl: dropping check from result due to ACLs: check=serfHealth writer.go:29: 2021-01-29T19:31:16.203Z [WARN] TestAgentAntiEntropy_Services_ACLDeny: Node info update blocked by ACLs: node=fc7e5164-fd57-6376-365c-894cedf13b85 accessorID= writer.go:29: 2021-01-29T19:31:16.205Z [WARN] TestAgentAntiEntropy_Services_ACLDeny: Service registration blocked by ACLs: service=mysql accessorID= writer.go:29: 2021-01-29T19:31:16.210Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Synced service: service=api writer.go:29: 2021-01-29T19:31:16.212Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.acl: dropping check from result due to ACLs: check=serfHealth writer.go:29: 2021-01-29T19:31:16.214Z [WARN] TestAgentAntiEntropy_Services_ACLDeny: Node info update blocked by ACLs: node=fc7e5164-fd57-6376-365c-894cedf13b85 accessorID= writer.go:29: 2021-01-29T19:31:16.215Z [WARN] TestAgentAntiEntropy_Services_ACLDeny: Service registration blocked by ACLs: service=mysql accessorID= writer.go:29: 2021-01-29T19:31:16.217Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Deregistered service: service=api writer.go:29: 2021-01-29T19:31:16.218Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Requesting shutdown writer.go:29: 2021-01-29T19:31:16.220Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server: shutting down server writer.go:29: 2021-01-29T19:31:16.221Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:16.223Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:16.221Z [ERROR] TestAgentAntiEntropy_Services_ACLDeny.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:16.223Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:16.224Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:16.224Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:16.229Z [WARN] TestAgentAntiEntropy_Services_ACLDeny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:16.229Z [DEBUG] TestAgentAntiEntropy_Services_ACLDeny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:16.234Z [WARN] TestAgentAntiEntropy_Services_ACLDeny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:16.244Z [INFO] TestAgentAntiEntropy_Services_ACLDeny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:16.245Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: consul server down writer.go:29: 2021-01-29T19:31:16.246Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: shutdown complete writer.go:29: 2021-01-29T19:31:16.247Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:19196 network=tcp writer.go:29: 2021-01-29T19:31:16.248Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:19196 network=udp writer.go:29: 2021-01-29T19:31:16.249Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:19197 network=tcp === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.371Z [DEBUG] TestAgentAntiEntropy_Services: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:16.373Z [INFO] TestAgentAntiEntropy_Services: Synced node info writer.go:29: 2021-01-29T19:31:16.376Z [INFO] TestAgentAntiEntropy_Services: Synced service: service=cache writer.go:29: 2021-01-29T19:31:16.378Z [INFO] TestAgentAntiEntropy_Services: Deregistered service: service=lb writer.go:29: 2021-01-29T19:31:16.379Z [DEBUG] TestAgentAntiEntropy_Services: Service in sync: service=mysql writer.go:29: 2021-01-29T19:31:16.382Z [INFO] TestAgentAntiEntropy_Services: Synced service: service=redis writer.go:29: 2021-01-29T19:31:16.385Z [INFO] TestAgentAntiEntropy_Services: Synced service: service=web writer.go:29: 2021-01-29T19:31:16.388Z [INFO] TestAgentAntiEntropy_Services: Synced service: service=api writer.go:29: 2021-01-29T19:31:16.393Z [DEBUG] TestAgentAntiEntropy_Services: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:16.396Z [INFO] TestAgentAntiEntropy_Services: Synced node info writer.go:29: 2021-01-29T19:31:16.397Z [DEBUG] TestAgentAntiEntropy_Services: Service in sync: service=redis writer.go:29: 2021-01-29T19:31:16.399Z [DEBUG] TestAgentAntiEntropy_Services: Service in sync: service=web writer.go:29: 2021-01-29T19:31:16.400Z [INFO] TestAgentAntiEntropy_Services: Deregistered service: service=api writer.go:29: 2021-01-29T19:31:16.402Z [DEBUG] TestAgentAntiEntropy_Services: Service in sync: service=cache writer.go:29: 2021-01-29T19:31:16.402Z [DEBUG] TestAgentAntiEntropy_Services: Service in sync: service=mysql writer.go:29: 2021-01-29T19:31:16.404Z [INFO] TestAgentAntiEntropy_Services: Requesting shutdown writer.go:29: 2021-01-29T19:31:16.405Z [INFO] TestAgentAntiEntropy_Services.server: shutting down server writer.go:29: 2021-01-29T19:31:16.406Z [DEBUG] TestAgentAntiEntropy_Services.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:16.406Z [ERROR] TestAgentAntiEntropy_Services.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:16.407Z [WARN] TestAgentAntiEntropy_Services.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:16.407Z [DEBUG] TestAgentAntiEntropy_Services.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:16.411Z [WARN] TestAgentAntiEntropy_Services.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:16.415Z [INFO] TestAgentAntiEntropy_Services.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:16.415Z [INFO] TestAgentAntiEntropy_Services: consul server down writer.go:29: 2021-01-29T19:31:16.418Z [INFO] TestAgentAntiEntropy_Services: shutdown complete writer.go:29: 2021-01-29T19:31:16.420Z [INFO] TestAgentAntiEntropy_Services: Stopping server: protocol=DNS address=127.0.0.1:19208 network=tcp writer.go:29: 2021-01-29T19:31:16.421Z [INFO] TestAgentAntiEntropy_Services: Stopping server: protocol=DNS address=127.0.0.1:19208 network=udp writer.go:29: 2021-01-29T19:31:16.423Z [INFO] TestAgentAntiEntropy_Services: Stopping server: protocol=HTTP address=127.0.0.1:19209 network=tcp 2021-01-29T19:31:16.463Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.acl: dropping node from result due to ACLs: node=Node-7477723e-990b-2201-46a8-5be1171561af 2021-01-29T19:31:16.466Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.acl: dropping node from result due to ACLs: node=Node-7477723e-990b-2201-46a8-5be1171561af 2021-01-29T19:31:16.469Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.acl: dropping check from result due to ACLs: check=serfHealth 2021-01-29T19:31:16.469Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny: Node info update blocked by ACLs: node=7477723e-990b-2201-46a8-5be1171561af accessorID= 2021-01-29T19:31:16.492Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Synced service: service=mysql 2021-01-29T19:31:16.493Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Synced service: service=api 2021-01-29T19:31:16.494Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.acl: dropping check from result due to ACLs: check=serfHealth 2021-01-29T19:31:16.494Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny: Node info update blocked by ACLs: node=7477723e-990b-2201-46a8-5be1171561af accessorID= 2021-01-29T19:31:16.496Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Synced service: service=mysql 2021-01-29T19:31:16.496Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Synced service: service=api 2021-01-29T19:31:16.497Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny: Check registration blocked by ACLs: check=mysql-check accessorID= 2021-01-29T19:31:16.498Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Synced check: check=api-check 2021-01-29T19:31:16.499Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.acl: dropping check from result due to ACLs: check=api-check 2021-01-29T19:31:16.499Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.acl: dropping check from result due to ACLs: check=serfHealth 2021-01-29T19:31:16.499Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny: Node info update blocked by ACLs: node=7477723e-990b-2201-46a8-5be1171561af accessorID= 2021-01-29T19:31:16.500Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Synced service: service=mysql 2021-01-29T19:31:16.501Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Synced service: service=api 2021-01-29T19:31:16.501Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny: Check registration blocked by ACLs: check=mysql-check accessorID= 2021-01-29T19:31:16.505Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Deregistered check: check=api-check 2021-01-29T19:31:16.505Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Requesting shutdown 2021-01-29T19:31:16.505Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server: shutting down server 2021-01-29T19:31:16.505Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.leader: stopping routine: routine="CA root pruning" 2021-01-29T19:31:16.505Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.leader: stopping routine: routine="legacy ACL token upgrade" 2021-01-29T19:31:16.505Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.leader: stopping routine: routine="acl token reaping" 2021-01-29T19:31:16.505Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny.server.serf.lan: serf: Shutdown without a Leave 2021-01-29T19:31:16.505Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.leader: stopped routine: routine="CA root pruning" 2021-01-29T19:31:16.506Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.leader: stopped routine: routine="legacy ACL token upgrade" 2021-01-29T19:31:16.506Z [DEBUG] TestAgentAntiEntropy_Checks_ACLDeny.leader: stopped routine: routine="acl token reaping" 2021-01-29T19:31:16.507Z [WARN] TestAgentAntiEntropy_Checks_ACLDeny.server.serf.wan: serf: Shutdown without a Leave 2021-01-29T19:31:16.507Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny.server.router.manager: shutting down 2021-01-29T19:31:16.508Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: consul server down 2021-01-29T19:31:16.508Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: shutdown complete 2021-01-29T19:31:16.508Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:19202 network=tcp 2021-01-29T19:31:16.508Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Stopping server: protocol=DNS address=127.0.0.1:19202 network=udp 2021-01-29T19:31:16.508Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Stopping server: protocol=HTTP address=127.0.0.1:19203 network=tcp 2021-01-29T19:31:16.611Z [INFO] TestAgentAntiEntropy_NodeInfo: Waiting for endpoints to shut down 2021-01-29T19:31:16.611Z [INFO] TestAgentAntiEntropy_NodeInfo: Endpoints down --- PASS: TestAgentAntiEntropy_NodeInfo (0.74s) === CONT TestAgentAntiEntropy_RemovingServiceAndCheck writer.go:29: 2021-01-29T19:31:16.620Z [WARN] TestAgentAntiEntropy_RemovingServiceAndCheck: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:16.621Z [DEBUG] TestAgentAntiEntropy_RemovingServiceAndCheck.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:16.622Z [DEBUG] TestAgentAntiEntropy_RemovingServiceAndCheck.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:16.628Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:75c0af1f-5e94-b004-a18b-d5624234fe86 Address:127.0.0.1:19219}]" writer.go:29: 2021-01-29T19:31:16.631Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:19219 [Follower]" leader= writer.go:29: 2021-01-29T19:31:16.632Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server.serf.wan: serf: EventMemberJoin: Node-75c0af1f-5e94-b004-a18b-d5624234fe86.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:16.641Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server.serf.lan: serf: EventMemberJoin: Node-75c0af1f-5e94-b004-a18b-d5624234fe86 127.0.0.1 writer.go:29: 2021-01-29T19:31:16.643Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Started DNS server: address=127.0.0.1:19214 network=udp writer.go:29: 2021-01-29T19:31:16.644Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server: Handled event for server in area: event=member-join server=Node-75c0af1f-5e94-b004-a18b-d5624234fe86.dc1 area=wan writer.go:29: 2021-01-29T19:31:16.645Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server: Adding LAN server: server="Node-75c0af1f-5e94-b004-a18b-d5624234fe86 (Addr: tcp/127.0.0.1:19219) (DC: dc1)" writer.go:29: 2021-01-29T19:31:16.646Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Started DNS server: address=127.0.0.1:19214 network=tcp writer.go:29: 2021-01-29T19:31:16.654Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Started HTTP server: address=127.0.0.1:19215 network=tcp writer.go:29: 2021-01-29T19:31:16.659Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: started state syncer writer.go:29: 2021-01-29T19:31:16.695Z [WARN] TestAgentAntiEntropy_RemovingServiceAndCheck.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:16.697Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:19219 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:16.700Z [DEBUG] TestAgentAntiEntropy_RemovingServiceAndCheck.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:16.702Z [DEBUG] TestAgentAntiEntropy_RemovingServiceAndCheck.server.raft: vote granted: from=75c0af1f-5e94-b004-a18b-d5624234fe86 term=2 tally=1 writer.go:29: 2021-01-29T19:31:16.705Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:16.707Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:19219 [Leader]" writer.go:29: 2021-01-29T19:31:16.709Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:16.710Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server: New leader elected: payload=Node-75c0af1f-5e94-b004-a18b-d5624234fe86 writer.go:29: 2021-01-29T19:31:16.711Z [DEBUG] TestAgentAntiEntropy_RemovingServiceAndCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19219 writer.go:29: 2021-01-29T19:31:16.719Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:16.729Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:16.730Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:16.731Z [DEBUG] TestAgentAntiEntropy_RemovingServiceAndCheck.server: Skipping self join check for node since the cluster is too small: node=Node-75c0af1f-5e94-b004-a18b-d5624234fe86 writer.go:29: 2021-01-29T19:31:16.732Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server: member joined, marking health alive: member=Node-75c0af1f-5e94-b004-a18b-d5624234fe86 === CONT TestAgentAntiEntropy_Services_ACLDeny writer.go:29: 2021-01-29T19:31:16.750Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:16.751Z [INFO] TestAgentAntiEntropy_Services_ACLDeny: Endpoints down --- PASS: TestAgentAntiEntropy_Services_ACLDeny (0.88s) === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:16.777Z [WARN] TestAgentAntiEntropy_Checks: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:16.778Z [DEBUG] TestAgentAntiEntropy_Checks.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:16.780Z [DEBUG] TestAgentAntiEntropy_Checks.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:16.835Z [INFO] TestAgentAntiEntropy_Checks.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c9fad03a-3f91-cb5d-b9bb-06db6d57810f Address:127.0.0.1:19225}]" writer.go:29: 2021-01-29T19:31:16.837Z [INFO] TestAgentAntiEntropy_Checks.server.raft: entering follower state: follower="Node at 127.0.0.1:19225 [Follower]" leader= writer.go:29: 2021-01-29T19:31:16.838Z [INFO] TestAgentAntiEntropy_Checks.server.serf.wan: serf: EventMemberJoin: Node-c9fad03a-3f91-cb5d-b9bb-06db6d57810f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:16.842Z [INFO] TestAgentAntiEntropy_Checks.server.serf.lan: serf: EventMemberJoin: Node-c9fad03a-3f91-cb5d-b9bb-06db6d57810f 127.0.0.1 writer.go:29: 2021-01-29T19:31:16.844Z [INFO] TestAgentAntiEntropy_Checks.server: Handled event for server in area: event=member-join server=Node-c9fad03a-3f91-cb5d-b9bb-06db6d57810f.dc1 area=wan writer.go:29: 2021-01-29T19:31:16.845Z [INFO] TestAgentAntiEntropy_Checks.server: Adding LAN server: server="Node-c9fad03a-3f91-cb5d-b9bb-06db6d57810f (Addr: tcp/127.0.0.1:19225) (DC: dc1)" writer.go:29: 2021-01-29T19:31:16.845Z [INFO] TestAgentAntiEntropy_Checks: Started DNS server: address=127.0.0.1:19220 network=udp writer.go:29: 2021-01-29T19:31:16.847Z [INFO] TestAgentAntiEntropy_Checks: Started DNS server: address=127.0.0.1:19220 network=tcp writer.go:29: 2021-01-29T19:31:16.850Z [INFO] TestAgentAntiEntropy_Checks: Started HTTP server: address=127.0.0.1:19221 network=tcp writer.go:29: 2021-01-29T19:31:16.851Z [INFO] TestAgentAntiEntropy_Checks: started state syncer writer.go:29: 2021-01-29T19:31:16.900Z [WARN] TestAgentAntiEntropy_Checks.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:16.902Z [INFO] TestAgentAntiEntropy_Checks.server.raft: entering candidate state: node="Node at 127.0.0.1:19225 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:16.905Z [DEBUG] TestAgentAntiEntropy_Checks.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:16.906Z [DEBUG] TestAgentAntiEntropy_Checks.server.raft: vote granted: from=c9fad03a-3f91-cb5d-b9bb-06db6d57810f term=2 tally=1 writer.go:29: 2021-01-29T19:31:16.908Z [INFO] TestAgentAntiEntropy_Checks.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:16.909Z [INFO] TestAgentAntiEntropy_Checks.server.raft: entering leader state: leader="Node at 127.0.0.1:19225 [Leader]" writer.go:29: 2021-01-29T19:31:16.910Z [INFO] TestAgentAntiEntropy_Checks.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:16.912Z [INFO] TestAgentAntiEntropy_Checks.server: New leader elected: payload=Node-c9fad03a-3f91-cb5d-b9bb-06db6d57810f writer.go:29: 2021-01-29T19:31:16.915Z [DEBUG] TestAgentAntiEntropy_Checks.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19225 === CONT TestAgentAntiEntropy_RemovingServiceAndCheck writer.go:29: 2021-01-29T19:31:16.917Z [DEBUG] TestAgentAntiEntropy_RemovingServiceAndCheck: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.925Z [INFO] TestAgentAntiEntropy_Services: Waiting for endpoints to shut down === CONT TestAgentAntiEntropy_RemovingServiceAndCheck writer.go:29: 2021-01-29T19:31:16.918Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Synced node info === CONT TestAgentAntiEntropy_Services writer.go:29: 2021-01-29T19:31:16.926Z [INFO] TestAgentAntiEntropy_Services: Endpoints down --- PASS: TestAgentAntiEntropy_Services (1.06s) === CONT TestAgent_AliasCheck === CONT TestAgentAntiEntropy_RemovingServiceAndCheck writer.go:29: 2021-01-29T19:31:16.928Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Deregistered service: service=deleted-check-service writer.go:29: 2021-01-29T19:31:16.929Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Deregistered check: check=lb writer.go:29: 2021-01-29T19:31:16.930Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Requesting shutdown writer.go:29: 2021-01-29T19:31:16.932Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server: shutting down server writer.go:29: 2021-01-29T19:31:16.933Z [DEBUG] TestAgentAntiEntropy_RemovingServiceAndCheck.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:16.934Z [WARN] TestAgentAntiEntropy_RemovingServiceAndCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:16.935Z [ERROR] TestAgentAntiEntropy_RemovingServiceAndCheck.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:16.935Z [DEBUG] TestAgentAntiEntropy_RemovingServiceAndCheck.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:16.939Z [WARN] TestAgentAntiEntropy_RemovingServiceAndCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:16.942Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:16.942Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: consul server down writer.go:29: 2021-01-29T19:31:16.944Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: shutdown complete writer.go:29: 2021-01-29T19:31:16.945Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Stopping server: protocol=DNS address=127.0.0.1:19214 network=tcp writer.go:29: 2021-01-29T19:31:16.946Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Stopping server: protocol=DNS address=127.0.0.1:19214 network=udp writer.go:29: 2021-01-29T19:31:16.947Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Stopping server: protocol=HTTP address=127.0.0.1:19215 network=tcp === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:16.946Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:16.952Z [INFO] TestAgentAntiEntropy_Checks.server.connect: initialized primary datacenter CA with provider: provider=consul --- PASS: TestAgent_AliasCheck (0.03s) === CONT TestAliasNotifications_local === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:16.953Z [INFO] TestAgentAntiEntropy_Checks.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:16.955Z [DEBUG] TestAgentAntiEntropy_Checks.server: Skipping self join check for node since the cluster is too small: node=Node-c9fad03a-3f91-cb5d-b9bb-06db6d57810f writer.go:29: 2021-01-29T19:31:16.957Z [INFO] TestAgentAntiEntropy_Checks.server: member joined, marking health alive: member=Node-c9fad03a-3f91-cb5d-b9bb-06db6d57810f === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:16.963Z [WARN] TestAliasNotifications_local: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:16.964Z [DEBUG] TestAliasNotifications_local.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:16.966Z [DEBUG] TestAliasNotifications_local.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:16.977Z [DEBUG] TestAgentAntiEntropy_Checks: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:16.978Z [INFO] TestAgentAntiEntropy_Checks: Synced node info writer.go:29: 2021-01-29T19:31:16.980Z [INFO] TestAgentAntiEntropy_Checks: Synced check: check=cache writer.go:29: 2021-01-29T19:31:16.982Z [INFO] TestAgentAntiEntropy_Checks: Deregistered check: check=lb writer.go:29: 2021-01-29T19:31:16.983Z [DEBUG] TestAgentAntiEntropy_Checks: Check in sync: check=mysql === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:16.983Z [INFO] TestAliasNotifications_local.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:23606949-0a19-d65f-b241-619a10c43f7f Address:127.0.0.1:19231}]" writer.go:29: 2021-01-29T19:31:16.986Z [INFO] TestAliasNotifications_local.server.raft: entering follower state: follower="Node at 127.0.0.1:19231 [Follower]" leader= writer.go:29: 2021-01-29T19:31:16.990Z [INFO] TestAliasNotifications_local.server.serf.wan: serf: EventMemberJoin: Node-23606949-0a19-d65f-b241-619a10c43f7f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:16.994Z [INFO] TestAliasNotifications_local.server.serf.lan: serf: EventMemberJoin: Node-23606949-0a19-d65f-b241-619a10c43f7f 127.0.0.1 writer.go:29: 2021-01-29T19:31:16.997Z [INFO] TestAliasNotifications_local.server: Handled event for server in area: event=member-join server=Node-23606949-0a19-d65f-b241-619a10c43f7f.dc1 area=wan writer.go:29: 2021-01-29T19:31:16.997Z [INFO] TestAliasNotifications_local.server: Adding LAN server: server="Node-23606949-0a19-d65f-b241-619a10c43f7f (Addr: tcp/127.0.0.1:19231) (DC: dc1)" writer.go:29: 2021-01-29T19:31:16.998Z [INFO] TestAliasNotifications_local: Started DNS server: address=127.0.0.1:19226 network=udp writer.go:29: 2021-01-29T19:31:17.000Z [INFO] TestAliasNotifications_local: Started DNS server: address=127.0.0.1:19226 network=tcp === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:16.996Z [INFO] TestAgentAntiEntropy_Checks: Synced check: check=redis 2021-01-29T19:31:17.009Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Waiting for endpoints to shut down 2021-01-29T19:31:17.009Z [INFO] TestAgentAntiEntropy_Checks_ACLDeny: Endpoints down --- PASS: TestAgentAntiEntropy_Checks_ACLDeny (1.14s) === CONT TestState_Notify === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.010Z [INFO] TestAliasNotifications_local: Started HTTP server: address=127.0.0.1:19227 network=tcp writer.go:29: 2021-01-29T19:31:17.012Z [INFO] TestAliasNotifications_local: started state syncer === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.016Z [INFO] TestAgentAntiEntropy_Checks: Synced check: check=web writer.go:29: 2021-01-29T19:31:17.021Z [DEBUG] TestAgentAntiEntropy_Checks: Skipping remote check since it is managed automatically: check=serfHealth --- PASS: TestState_Notify (0.02s) === CONT TestAgent_sendCoordinate === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.030Z [INFO] TestAgentAntiEntropy_Checks: Synced node info === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.035Z [WARN] TestAliasNotifications_local.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.037Z [INFO] TestAgentAntiEntropy_Checks: Deregistered check: check=redis === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.040Z [INFO] TestAliasNotifications_local.server.raft: entering candidate state: node="Node at 127.0.0.1:19231 [Candidate]" term=2 === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.042Z [DEBUG] TestAgentAntiEntropy_Checks: Check in sync: check=web === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.048Z [DEBUG] TestAliasNotifications_local.server.raft: votes: needed=1 === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.047Z [DEBUG] TestAgentAntiEntropy_Checks: Check in sync: check=cache writer.go:29: 2021-01-29T19:31:17.052Z [DEBUG] TestAgentAntiEntropy_Checks: Check in sync: check=mysql writer.go:29: 2021-01-29T19:31:17.057Z [INFO] TestAgentAntiEntropy_Checks: Requesting shutdown writer.go:29: 2021-01-29T19:31:17.059Z [INFO] TestAgentAntiEntropy_Checks.server: shutting down server === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.052Z [DEBUG] TestAliasNotifications_local.server.raft: vote granted: from=23606949-0a19-d65f-b241-619a10c43f7f term=2 tally=1 writer.go:29: 2021-01-29T19:31:17.063Z [INFO] TestAliasNotifications_local.server.raft: election won: tally=1 === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.061Z [DEBUG] TestAgentAntiEntropy_Checks.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.061Z [WARN] TestAgent_sendCoordinate: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:17.065Z [DEBUG] TestAgent_sendCoordinate.tlsutil: Update: version=1 === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.064Z [INFO] TestAliasNotifications_local.server.raft: entering leader state: leader="Node at 127.0.0.1:19231 [Leader]" === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.066Z [DEBUG] TestAgent_sendCoordinate.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.065Z [WARN] TestAgentAntiEntropy_Checks.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:17.063Z [ERROR] TestAgentAntiEntropy_Checks.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.067Z [INFO] TestAliasNotifications_local.server: cluster leadership acquired === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.070Z [DEBUG] TestAgentAntiEntropy_Checks.leader: stopped routine: routine="CA root pruning" === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.070Z [INFO] TestAliasNotifications_local.server: New leader elected: payload=Node-23606949-0a19-d65f-b241-619a10c43f7f === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.074Z [INFO] TestAgent_sendCoordinate.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d580d440-b73f-5ed2-ab28-64177e2c97ae Address:127.0.0.1:19237}]" === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.072Z [WARN] TestAgentAntiEntropy_Checks.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.076Z [INFO] TestAgent_sendCoordinate.server.raft: entering follower state: follower="Node at 127.0.0.1:19237 [Follower]" leader= === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.072Z [DEBUG] TestAliasNotifications_local.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19231 === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.076Z [INFO] TestAgent_sendCoordinate.server.serf.wan: serf: EventMemberJoin: Node-d580d440-b73f-5ed2-ab28-64177e2c97ae.dc1 127.0.0.1 === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.079Z [INFO] TestAgentAntiEntropy_Checks.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:17.079Z [INFO] TestAgentAntiEntropy_Checks: consul server down === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.082Z [INFO] TestAgent_sendCoordinate.server.serf.lan: serf: EventMemberJoin: Node-d580d440-b73f-5ed2-ab28-64177e2c97ae 127.0.0.1 writer.go:29: 2021-01-29T19:31:17.085Z [INFO] TestAgent_sendCoordinate.server: Handled event for server in area: event=member-join server=Node-d580d440-b73f-5ed2-ab28-64177e2c97ae.dc1 area=wan writer.go:29: 2021-01-29T19:31:17.086Z [INFO] TestAgent_sendCoordinate: Started DNS server: address=127.0.0.1:19232 network=udp === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.084Z [INFO] TestAgentAntiEntropy_Checks: shutdown complete writer.go:29: 2021-01-29T19:31:17.088Z [INFO] TestAgentAntiEntropy_Checks: Stopping server: protocol=DNS address=127.0.0.1:19220 network=tcp === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.088Z [INFO] TestAgent_sendCoordinate: Started DNS server: address=127.0.0.1:19232 network=tcp === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.090Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.088Z [INFO] TestAgent_sendCoordinate.server: Adding LAN server: server="Node-d580d440-b73f-5ed2-ab28-64177e2c97ae (Addr: tcp/127.0.0.1:19237) (DC: dc1)" === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.089Z [INFO] TestAgentAntiEntropy_Checks: Stopping server: protocol=DNS address=127.0.0.1:19220 network=udp === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.091Z [INFO] TestAgent_sendCoordinate: Started HTTP server: address=127.0.0.1:19233 network=tcp === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.094Z [INFO] TestAgentAntiEntropy_Checks: Stopping server: protocol=HTTP address=127.0.0.1:19221 network=tcp === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.094Z [INFO] TestAgent_sendCoordinate: started state syncer === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.108Z [INFO] TestAliasNotifications_local.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:17.110Z [INFO] TestAliasNotifications_local.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:17.111Z [DEBUG] TestAliasNotifications_local.server: Skipping self join check for node since the cluster is too small: node=Node-23606949-0a19-d65f-b241-619a10c43f7f writer.go:29: 2021-01-29T19:31:17.112Z [INFO] TestAliasNotifications_local.server: member joined, marking health alive: member=Node-23606949-0a19-d65f-b241-619a10c43f7f === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.126Z [WARN] TestAgent_sendCoordinate.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:17.128Z [INFO] TestAgent_sendCoordinate.server.raft: entering candidate state: node="Node at 127.0.0.1:19237 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:17.130Z [DEBUG] TestAgent_sendCoordinate.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:17.131Z [DEBUG] TestAgent_sendCoordinate.server.raft: vote granted: from=d580d440-b73f-5ed2-ab28-64177e2c97ae term=2 tally=1 writer.go:29: 2021-01-29T19:31:17.132Z [INFO] TestAgent_sendCoordinate.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:17.133Z [INFO] TestAgent_sendCoordinate.server.raft: entering leader state: leader="Node at 127.0.0.1:19237 [Leader]" writer.go:29: 2021-01-29T19:31:17.134Z [INFO] TestAgent_sendCoordinate.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:17.135Z [INFO] TestAgent_sendCoordinate.server: New leader elected: payload=Node-d580d440-b73f-5ed2-ab28-64177e2c97ae writer.go:29: 2021-01-29T19:31:17.136Z [DEBUG] TestAgent_sendCoordinate.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19237 writer.go:29: 2021-01-29T19:31:17.153Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:17.158Z [INFO] TestAgent_sendCoordinate.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:17.159Z [INFO] TestAgent_sendCoordinate.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:17.160Z [DEBUG] TestAgent_sendCoordinate.server: Skipping self join check for node since the cluster is too small: node=Node-d580d440-b73f-5ed2-ab28-64177e2c97ae writer.go:29: 2021-01-29T19:31:17.161Z [INFO] TestAgent_sendCoordinate.server: member joined, marking health alive: member=Node-d580d440-b73f-5ed2-ab28-64177e2c97ae === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.174Z [DEBUG] TestAliasNotifications_local: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:17.177Z [INFO] TestAliasNotifications_local: Synced node info === CONT TestAgent_sendCoordinate state_test.go:1920: 10 1 100ms writer.go:29: 2021-01-29T19:31:17.331Z [INFO] TestAgent_sendCoordinate: Requesting shutdown writer.go:29: 2021-01-29T19:31:17.332Z [INFO] TestAgent_sendCoordinate.server: shutting down server writer.go:29: 2021-01-29T19:31:17.333Z [DEBUG] TestAgent_sendCoordinate.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:17.334Z [WARN] TestAgent_sendCoordinate.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:17.333Z [ERROR] TestAgent_sendCoordinate.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:17.334Z [DEBUG] TestAgent_sendCoordinate.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:17.336Z [WARN] TestAgent_sendCoordinate.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:17.339Z [INFO] TestAgent_sendCoordinate.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:17.339Z [INFO] TestAgent_sendCoordinate: consul server down writer.go:29: 2021-01-29T19:31:17.343Z [INFO] TestAgent_sendCoordinate: shutdown complete writer.go:29: 2021-01-29T19:31:17.344Z [INFO] TestAgent_sendCoordinate: Stopping server: protocol=DNS address=127.0.0.1:19232 network=tcp writer.go:29: 2021-01-29T19:31:17.346Z [INFO] TestAgent_sendCoordinate: Stopping server: protocol=DNS address=127.0.0.1:19232 network=udp writer.go:29: 2021-01-29T19:31:17.347Z [INFO] TestAgent_sendCoordinate: Stopping server: protocol=HTTP address=127.0.0.1:19233 network=tcp === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.436Z [INFO] TestAliasNotifications_local: Requesting shutdown writer.go:29: 2021-01-29T19:31:17.437Z [INFO] TestAliasNotifications_local.server: shutting down server writer.go:29: 2021-01-29T19:31:17.438Z [DEBUG] TestAliasNotifications_local.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:17.439Z [WARN] TestAliasNotifications_local.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:17.439Z [DEBUG] TestAliasNotifications_local.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:17.441Z [WARN] TestAliasNotifications_local.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:17.445Z [INFO] TestAliasNotifications_local.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:17.445Z [INFO] TestAliasNotifications_local: consul server down === CONT TestAgentAntiEntropy_RemovingServiceAndCheck writer.go:29: 2021-01-29T19:31:17.447Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:17.448Z [INFO] TestAgentAntiEntropy_RemovingServiceAndCheck: Endpoints down --- PASS: TestAgentAntiEntropy_RemovingServiceAndCheck (0.84s) === CONT TestAgentAntiEntropy_EnableTagOverride === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.448Z [INFO] TestAliasNotifications_local: shutdown complete writer.go:29: 2021-01-29T19:31:17.449Z [INFO] TestAliasNotifications_local: Stopping server: protocol=DNS address=127.0.0.1:19226 network=tcp writer.go:29: 2021-01-29T19:31:17.450Z [INFO] TestAliasNotifications_local: Stopping server: protocol=DNS address=127.0.0.1:19226 network=udp writer.go:29: 2021-01-29T19:31:17.452Z [INFO] TestAliasNotifications_local: Stopping server: protocol=HTTP address=127.0.0.1:19227 network=tcp === CONT TestAgentAntiEntropy_EnableTagOverride writer.go:29: 2021-01-29T19:31:17.457Z [WARN] TestAgentAntiEntropy_EnableTagOverride: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:17.458Z [DEBUG] TestAgentAntiEntropy_EnableTagOverride.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:17.473Z [DEBUG] TestAgentAntiEntropy_EnableTagOverride.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:17.487Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:50053eb9-d773-fff7-1dbe-ecd77e4b636f Address:127.0.0.1:19243}]" writer.go:29: 2021-01-29T19:31:17.494Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server.raft: entering follower state: follower="Node at 127.0.0.1:19243 [Follower]" leader= writer.go:29: 2021-01-29T19:31:17.497Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server.serf.wan: serf: EventMemberJoin: Node-50053eb9-d773-fff7-1dbe-ecd77e4b636f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:17.508Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server.serf.lan: serf: EventMemberJoin: Node-50053eb9-d773-fff7-1dbe-ecd77e4b636f 127.0.0.1 writer.go:29: 2021-01-29T19:31:17.520Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Started DNS server: address=127.0.0.1:19238 network=udp writer.go:29: 2021-01-29T19:31:17.520Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server: Handled event for server in area: event=member-join server=Node-50053eb9-d773-fff7-1dbe-ecd77e4b636f.dc1 area=wan writer.go:29: 2021-01-29T19:31:17.521Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server: Adding LAN server: server="Node-50053eb9-d773-fff7-1dbe-ecd77e4b636f (Addr: tcp/127.0.0.1:19243) (DC: dc1)" writer.go:29: 2021-01-29T19:31:17.523Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Started DNS server: address=127.0.0.1:19238 network=tcp writer.go:29: 2021-01-29T19:31:17.527Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Started HTTP server: address=127.0.0.1:19239 network=tcp writer.go:29: 2021-01-29T19:31:17.528Z [INFO] TestAgentAntiEntropy_EnableTagOverride: started state syncer writer.go:29: 2021-01-29T19:31:17.560Z [WARN] TestAgentAntiEntropy_EnableTagOverride.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:17.562Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server.raft: entering candidate state: node="Node at 127.0.0.1:19243 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:17.572Z [DEBUG] TestAgentAntiEntropy_EnableTagOverride.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:17.574Z [DEBUG] TestAgentAntiEntropy_EnableTagOverride.server.raft: vote granted: from=50053eb9-d773-fff7-1dbe-ecd77e4b636f term=2 tally=1 writer.go:29: 2021-01-29T19:31:17.575Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:17.576Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server.raft: entering leader state: leader="Node at 127.0.0.1:19243 [Leader]" writer.go:29: 2021-01-29T19:31:17.579Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:17.580Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server: New leader elected: payload=Node-50053eb9-d773-fff7-1dbe-ecd77e4b636f writer.go:29: 2021-01-29T19:31:17.590Z [DEBUG] TestAgentAntiEntropy_EnableTagOverride.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19243 writer.go:29: 2021-01-29T19:31:17.594Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgentAntiEntropy_Checks writer.go:29: 2021-01-29T19:31:17.596Z [INFO] TestAgentAntiEntropy_Checks: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:17.597Z [INFO] TestAgentAntiEntropy_Checks: Endpoints down --- PASS: TestAgentAntiEntropy_Checks (0.85s) === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.608Z [WARN] TestAgentAntiEntropy_Services_WithChecks: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:17.609Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:17.611Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:17.644Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:86da4e68-77a3-7dfd-34ab-4295aa009cf5 Address:127.0.0.1:19249}]" === CONT TestAgentAntiEntropy_EnableTagOverride writer.go:29: 2021-01-29T19:31:17.646Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.647Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server.raft: entering follower state: follower="Node at 127.0.0.1:19249 [Follower]" leader= writer.go:29: 2021-01-29T19:31:17.648Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server.serf.wan: serf: EventMemberJoin: Node-86da4e68-77a3-7dfd-34ab-4295aa009cf5.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:17.651Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server.serf.lan: serf: EventMemberJoin: Node-86da4e68-77a3-7dfd-34ab-4295aa009cf5 127.0.0.1 writer.go:29: 2021-01-29T19:31:17.653Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Started DNS server: address=127.0.0.1:19244 network=udp writer.go:29: 2021-01-29T19:31:17.653Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server: Adding LAN server: server="Node-86da4e68-77a3-7dfd-34ab-4295aa009cf5 (Addr: tcp/127.0.0.1:19249) (DC: dc1)" writer.go:29: 2021-01-29T19:31:17.654Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server: Handled event for server in area: event=member-join server=Node-86da4e68-77a3-7dfd-34ab-4295aa009cf5.dc1 area=wan === CONT TestAgentAntiEntropy_EnableTagOverride writer.go:29: 2021-01-29T19:31:17.648Z [INFO] TestAgentAntiEntropy_EnableTagOverride.leader: started routine: routine="CA root pruning" === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.658Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Started DNS server: address=127.0.0.1:19244 network=tcp === CONT TestAgentAntiEntropy_EnableTagOverride writer.go:29: 2021-01-29T19:31:17.657Z [DEBUG] TestAgentAntiEntropy_EnableTagOverride.server: Skipping self join check for node since the cluster is too small: node=Node-50053eb9-d773-fff7-1dbe-ecd77e4b636f writer.go:29: 2021-01-29T19:31:17.657Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Synced node info writer.go:29: 2021-01-29T19:31:17.661Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server: member joined, marking health alive: member=Node-50053eb9-d773-fff7-1dbe-ecd77e4b636f === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.662Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Started HTTP server: address=127.0.0.1:19245 network=tcp writer.go:29: 2021-01-29T19:31:17.670Z [INFO] TestAgentAntiEntropy_Services_WithChecks: started state syncer writer.go:29: 2021-01-29T19:31:17.691Z [WARN] TestAgentAntiEntropy_Services_WithChecks.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:17.693Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server.raft: entering candidate state: node="Node at 127.0.0.1:19249 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:17.695Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:17.696Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks.server.raft: vote granted: from=86da4e68-77a3-7dfd-34ab-4295aa009cf5 term=2 tally=1 writer.go:29: 2021-01-29T19:31:17.697Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:17.698Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server.raft: entering leader state: leader="Node at 127.0.0.1:19249 [Leader]" writer.go:29: 2021-01-29T19:31:17.699Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:17.699Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server: New leader elected: payload=Node-86da4e68-77a3-7dfd-34ab-4295aa009cf5 writer.go:29: 2021-01-29T19:31:17.715Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19249 writer.go:29: 2021-01-29T19:31:17.723Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:17.729Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:17.730Z [INFO] TestAgentAntiEntropy_Services_WithChecks.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:17.731Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks.server: Skipping self join check for node since the cluster is too small: node=Node-86da4e68-77a3-7dfd-34ab-4295aa009cf5 writer.go:29: 2021-01-29T19:31:17.733Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server: member joined, marking health alive: member=Node-86da4e68-77a3-7dfd-34ab-4295aa009cf5 === CONT TestAgentAntiEntropy_EnableTagOverride writer.go:29: 2021-01-29T19:31:17.784Z [DEBUG] TestAgentAntiEntropy_EnableTagOverride: Node info in sync writer.go:29: 2021-01-29T19:31:17.788Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Synced service: service=svc_id1 writer.go:29: 2021-01-29T19:31:17.791Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Synced service: service=svc_id2 writer.go:29: 2021-01-29T19:31:17.794Z [DEBUG] TestAgentAntiEntropy_EnableTagOverride: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:17.797Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Synced node info writer.go:29: 2021-01-29T19:31:17.800Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Synced service: service=svc_id1 writer.go:29: 2021-01-29T19:31:17.802Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Synced service: service=svc_id2 writer.go:29: 2021-01-29T19:31:17.806Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Requesting shutdown writer.go:29: 2021-01-29T19:31:17.808Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server: shutting down server writer.go:29: 2021-01-29T19:31:17.810Z [DEBUG] TestAgentAntiEntropy_EnableTagOverride.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:17.813Z [WARN] TestAgentAntiEntropy_EnableTagOverride.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:17.813Z [DEBUG] TestAgentAntiEntropy_EnableTagOverride.leader: stopped routine: routine="CA root pruning" === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.815Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgentAntiEntropy_EnableTagOverride writer.go:29: 2021-01-29T19:31:17.815Z [WARN] TestAgentAntiEntropy_EnableTagOverride.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.817Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Synced node info writer.go:29: 2021-01-29T19:31:17.819Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks: Node info in sync === CONT TestAgentAntiEntropy_EnableTagOverride writer.go:29: 2021-01-29T19:31:17.818Z [INFO] TestAgentAntiEntropy_EnableTagOverride.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:17.819Z [INFO] TestAgentAntiEntropy_EnableTagOverride: consul server down writer.go:29: 2021-01-29T19:31:17.821Z [INFO] TestAgentAntiEntropy_EnableTagOverride: shutdown complete writer.go:29: 2021-01-29T19:31:17.823Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Stopping server: protocol=DNS address=127.0.0.1:19238 network=tcp writer.go:29: 2021-01-29T19:31:17.825Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Stopping server: protocol=DNS address=127.0.0.1:19238 network=udp writer.go:29: 2021-01-29T19:31:17.827Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Stopping server: protocol=HTTP address=127.0.0.1:19239 network=tcp === CONT TestAgent_sendCoordinate writer.go:29: 2021-01-29T19:31:17.848Z [INFO] TestAgent_sendCoordinate: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:17.849Z [INFO] TestAgent_sendCoordinate: Endpoints down --- PASS: TestAgent_sendCoordinate (0.82s) === CONT TestAgent_CheckTokens --- PASS: TestAgent_CheckTokens (0.01s) === CONT TestAgent_AddCheckFailure --- PASS: TestAgent_AddCheckFailure (0.01s) === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.905Z [WARN] TestAgent_ServiceWatchCh: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:17.906Z [DEBUG] TestAgent_ServiceWatchCh.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:17.908Z [DEBUG] TestAgent_ServiceWatchCh.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.921Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:17.922Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks: Node info in sync writer.go:29: 2021-01-29T19:31:17.926Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Synced service: service=mysql === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.929Z [INFO] TestAgent_ServiceWatchCh.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:739a914f-84b4-1793-1626-4511d5b75c4a Address:127.0.0.1:19255}]" === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.928Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks: Check in sync: check=mysql writer.go:29: 2021-01-29T19:31:17.931Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.931Z [INFO] TestAgent_ServiceWatchCh.server.raft: entering follower state: follower="Node at 127.0.0.1:19255 [Follower]" leader= === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.933Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks: Node info in sync === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.931Z [INFO] TestAgent_ServiceWatchCh.server.serf.wan: serf: EventMemberJoin: Node-739a914f-84b4-1793-1626-4511d5b75c4a.dc1 127.0.0.1 === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.937Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Synced service: service=mysql === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.939Z [INFO] TestAgent_ServiceWatchCh.server.serf.lan: serf: EventMemberJoin: Node-739a914f-84b4-1793-1626-4511d5b75c4a 127.0.0.1 === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.939Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Synced service: service=redis writer.go:29: 2021-01-29T19:31:17.943Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks: Check in sync: check=mysql writer.go:29: 2021-01-29T19:31:17.944Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks: Check in sync: check=redis:1 writer.go:29: 2021-01-29T19:31:17.946Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks: Check in sync: check=redis:2 writer.go:29: 2021-01-29T19:31:17.947Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Requesting shutdown === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.953Z [INFO] TestAgent_ServiceWatchCh.server: Adding LAN server: server="Node-739a914f-84b4-1793-1626-4511d5b75c4a (Addr: tcp/127.0.0.1:19255) (DC: dc1)" writer.go:29: 2021-01-29T19:31:17.953Z [INFO] TestAgent_ServiceWatchCh.server: Handled event for server in area: event=member-join server=Node-739a914f-84b4-1793-1626-4511d5b75c4a.dc1 area=wan === CONT TestAliasNotifications_local writer.go:29: 2021-01-29T19:31:17.954Z [INFO] TestAliasNotifications_local: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:17.955Z [INFO] TestAliasNotifications_local: Endpoints down --- PASS: TestAliasNotifications_local (1.00s) === CONT TestAgent_ServiceTokens === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.949Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server: shutting down server writer.go:29: 2021-01-29T19:31:17.956Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks.leader: stopping routine: routine="CA root pruning" --- PASS: TestAgent_ServiceTokens (0.00s) === CONT TestAgentAntiEntropy_Services_ConnectProxy === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.957Z [WARN] TestAgentAntiEntropy_Services_WithChecks.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:17.957Z [DEBUG] TestAgentAntiEntropy_Services_WithChecks.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:17.965Z [WARN] TestAgentAntiEntropy_Services_WithChecks.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:17.968Z [WARN] TestAgentAntiEntropy_Services_ConnectProxy: bootstrap = true: do not enable unless necessary === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.963Z [INFO] TestAgent_ServiceWatchCh: Started DNS server: address=127.0.0.1:19250 network=udp === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:17.969Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:17.971Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.969Z [INFO] TestAgent_ServiceWatchCh: Started DNS server: address=127.0.0.1:19250 network=tcp === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.973Z [INFO] TestAgentAntiEntropy_Services_WithChecks.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:17.973Z [INFO] TestAgentAntiEntropy_Services_WithChecks: consul server down writer.go:29: 2021-01-29T19:31:17.976Z [INFO] TestAgentAntiEntropy_Services_WithChecks: shutdown complete writer.go:29: 2021-01-29T19:31:17.977Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Stopping server: protocol=DNS address=127.0.0.1:19244 network=tcp writer.go:29: 2021-01-29T19:31:17.981Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Stopping server: protocol=DNS address=127.0.0.1:19244 network=udp === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:17.981Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7e6a29fd-20d9-d2e9-1527-9a74af20ebb2 Address:127.0.0.1:19261}]" === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:17.982Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Stopping server: protocol=HTTP address=127.0.0.1:19245 network=tcp === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.984Z [INFO] TestAgent_ServiceWatchCh: Started HTTP server: address=127.0.0.1:19251 network=tcp === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:17.983Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server.raft: entering follower state: follower="Node at 127.0.0.1:19261 [Follower]" leader= === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.985Z [INFO] TestAgent_ServiceWatchCh: started state syncer === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:17.983Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server.serf.wan: serf: EventMemberJoin: Node-7e6a29fd-20d9-d2e9-1527-9a74af20ebb2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:17.989Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server.serf.lan: serf: EventMemberJoin: Node-7e6a29fd-20d9-d2e9-1527-9a74af20ebb2 127.0.0.1 === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.991Z [WARN] TestAgent_ServiceWatchCh.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:17.993Z [INFO] TestAgent_ServiceWatchCh.server.raft: entering candidate state: node="Node at 127.0.0.1:19255 [Candidate]" term=2 === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:17.993Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Started DNS server: address=127.0.0.1:19256 network=udp writer.go:29: 2021-01-29T19:31:17.995Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server: Adding LAN server: server="Node-7e6a29fd-20d9-d2e9-1527-9a74af20ebb2 (Addr: tcp/127.0.0.1:19261) (DC: dc1)" writer.go:29: 2021-01-29T19:31:17.997Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server: Handled event for server in area: event=member-join server=Node-7e6a29fd-20d9-d2e9-1527-9a74af20ebb2.dc1 area=wan === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:17.998Z [DEBUG] TestAgent_ServiceWatchCh.server.raft: votes: needed=1 === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:17.998Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Started DNS server: address=127.0.0.1:19256 network=tcp === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.000Z [DEBUG] TestAgent_ServiceWatchCh.server.raft: vote granted: from=739a914f-84b4-1793-1626-4511d5b75c4a term=2 tally=1 writer.go:29: 2021-01-29T19:31:18.003Z [INFO] TestAgent_ServiceWatchCh.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:18.004Z [INFO] TestAgent_ServiceWatchCh.server.raft: entering leader state: leader="Node at 127.0.0.1:19255 [Leader]" === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.007Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Started HTTP server: address=127.0.0.1:19257 network=tcp === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.006Z [INFO] TestAgent_ServiceWatchCh.server: cluster leadership acquired === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.008Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: started state syncer === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.009Z [INFO] TestAgent_ServiceWatchCh.server: New leader elected: payload=Node-739a914f-84b4-1793-1626-4511d5b75c4a === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.026Z [WARN] TestAgentAntiEntropy_Services_ConnectProxy.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:18.027Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server.raft: entering candidate state: node="Node at 127.0.0.1:19261 [Candidate]" term=2 === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.040Z [DEBUG] TestAgent_ServiceWatchCh.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19255 === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.043Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:18.045Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy.server.raft: vote granted: from=7e6a29fd-20d9-d2e9-1527-9a74af20ebb2 term=2 tally=1 writer.go:29: 2021-01-29T19:31:18.050Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server.raft: election won: tally=1 === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.053Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.055Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server.raft: entering leader state: leader="Node at 127.0.0.1:19261 [Leader]" writer.go:29: 2021-01-29T19:31:18.057Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:18.058Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server: New leader elected: payload=Node-7e6a29fd-20d9-d2e9-1527-9a74af20ebb2 === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.068Z [INFO] TestAgent_ServiceWatchCh.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.059Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19261 === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.071Z [INFO] TestAgent_ServiceWatchCh.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:18.083Z [DEBUG] TestAgent_ServiceWatchCh.server: Skipping self join check for node since the cluster is too small: node=Node-739a914f-84b4-1793-1626-4511d5b75c4a === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.091Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.089Z [INFO] TestAgent_ServiceWatchCh.server: member joined, marking health alive: member=Node-739a914f-84b4-1793-1626-4511d5b75c4a === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.103Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:18.104Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:18.107Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy.server: Skipping self join check for node since the cluster is too small: node=Node-7e6a29fd-20d9-d2e9-1527-9a74af20ebb2 writer.go:29: 2021-01-29T19:31:18.108Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server: member joined, marking health alive: member=Node-7e6a29fd-20d9-d2e9-1527-9a74af20ebb2 writer.go:29: 2021-01-29T19:31:18.175Z [ERROR] TestAgentAntiEntropy_Services_ConnectProxy.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:18.180Z [ERROR] TestAgentAntiEntropy_Services_ConnectProxy.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:18.182Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:18.184Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Synced node info writer.go:29: 2021-01-29T19:31:18.184Z [ERROR] TestAgentAntiEntropy_Services_ConnectProxy.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:18.185Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Synced service: service=web-proxy writer.go:29: 2021-01-29T19:31:18.188Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Synced service: service=cache-proxy writer.go:29: 2021-01-29T19:31:18.190Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Deregistered service: service=lb-proxy writer.go:29: 2021-01-29T19:31:18.192Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy: Service in sync: service=mysql-proxy writer.go:29: 2021-01-29T19:31:18.194Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Synced service: service=redis-proxy writer.go:29: 2021-01-29T19:31:18.209Z [ERROR] TestAgentAntiEntropy_Services_ConnectProxy.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:31:18.226Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:18.227Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy: Node info in sync writer.go:29: 2021-01-29T19:31:18.229Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy: Service in sync: service=mysql-proxy writer.go:29: 2021-01-29T19:31:18.230Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy: Service in sync: service=redis-proxy writer.go:29: 2021-01-29T19:31:18.232Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy: Service in sync: service=web-proxy writer.go:29: 2021-01-29T19:31:18.235Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Deregistered service: service=cache-proxy writer.go:29: 2021-01-29T19:31:18.243Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Requesting shutdown writer.go:29: 2021-01-29T19:31:18.252Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server: shutting down server writer.go:29: 2021-01-29T19:31:18.269Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy.leader: stopping routine: routine="CA root pruning" === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.273Z [INFO] TestAgent_ServiceWatchCh: Requesting shutdown writer.go:29: 2021-01-29T19:31:18.274Z [INFO] TestAgent_ServiceWatchCh.server: shutting down server writer.go:29: 2021-01-29T19:31:18.275Z [DEBUG] TestAgent_ServiceWatchCh.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:18.276Z [WARN] TestAgent_ServiceWatchCh.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:18.276Z [ERROR] TestAgent_ServiceWatchCh.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.272Z [WARN] TestAgentAntiEntropy_Services_ConnectProxy.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:18.270Z [ERROR] TestAgentAntiEntropy_Services_ConnectProxy.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.276Z [DEBUG] TestAgent_ServiceWatchCh.leader: stopped routine: routine="CA root pruning" === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.272Z [DEBUG] TestAgentAntiEntropy_Services_ConnectProxy.leader: stopped routine: routine="CA root pruning" === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.278Z [WARN] TestAgent_ServiceWatchCh.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.280Z [WARN] TestAgentAntiEntropy_Services_ConnectProxy.server.serf.wan: serf: Shutdown without a Leave === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.286Z [INFO] TestAgent_ServiceWatchCh.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:18.287Z [INFO] TestAgent_ServiceWatchCh: consul server down === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.287Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy.server.router.manager: shutting down === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.289Z [INFO] TestAgent_ServiceWatchCh: shutdown complete === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.288Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: consul server down === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.289Z [INFO] TestAgent_ServiceWatchCh: Stopping server: protocol=DNS address=127.0.0.1:19250 network=tcp === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.290Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: shutdown complete === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.290Z [INFO] TestAgent_ServiceWatchCh: Stopping server: protocol=DNS address=127.0.0.1:19250 network=udp === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.291Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Stopping server: protocol=DNS address=127.0.0.1:19256 network=tcp === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.291Z [INFO] TestAgent_ServiceWatchCh: Stopping server: protocol=HTTP address=127.0.0.1:19251 network=tcp === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.291Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Stopping server: protocol=DNS address=127.0.0.1:19256 network=udp writer.go:29: 2021-01-29T19:31:18.292Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Stopping server: protocol=HTTP address=127.0.0.1:19257 network=tcp === CONT TestAgentAntiEntropy_EnableTagOverride writer.go:29: 2021-01-29T19:31:18.329Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:18.331Z [INFO] TestAgentAntiEntropy_EnableTagOverride: Endpoints down --- PASS: TestAgentAntiEntropy_EnableTagOverride (0.88s) === CONT TestAgentAntiEntropy_Services_WithChecks writer.go:29: 2021-01-29T19:31:18.483Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:18.485Z [INFO] TestAgentAntiEntropy_Services_WithChecks: Endpoints down --- PASS: TestAgentAntiEntropy_Services_WithChecks (0.89s) === CONT TestAgent_ServiceWatchCh writer.go:29: 2021-01-29T19:31:18.792Z [INFO] TestAgent_ServiceWatchCh: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:18.794Z [INFO] TestAgent_ServiceWatchCh: Endpoints down --- PASS: TestAgent_ServiceWatchCh (0.93s) === CONT TestAgentAntiEntropy_Services_ConnectProxy writer.go:29: 2021-01-29T19:31:18.794Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:18.795Z [INFO] TestAgentAntiEntropy_Services_ConnectProxy: Endpoints down --- PASS: TestAgentAntiEntropy_Services_ConnectProxy (0.84s) PASS ok github.com/hashicorp/consul/agent/local 3.138s === RUN TestBuild === RUN TestBuild/no_version === RUN TestBuild/bad_version === RUN TestBuild/good_version === RUN TestBuild/rc_version === RUN TestBuild/ent_version --- PASS: TestBuild (0.00s) --- PASS: TestBuild/no_version (0.00s) --- PASS: TestBuild/bad_version (0.00s) --- PASS: TestBuild/good_version (0.00s) --- PASS: TestBuild/rc_version (0.00s) --- PASS: TestBuild/ent_version (0.00s) === RUN TestServer_Key_Equal --- PASS: TestServer_Key_Equal (0.00s) === RUN TestServer_Key --- PASS: TestServer_Key (0.00s) === RUN TestServer_Key_params --- PASS: TestServer_Key_params (0.00s) === RUN TestIsConsulServer --- PASS: TestIsConsulServer (0.00s) === RUN TestIsConsulServer_Optional --- PASS: TestIsConsulServer_Optional (0.00s) PASS ok github.com/hashicorp/consul/agent/metadata 0.032s ? github.com/hashicorp/consul/agent/mock [no test files] ? github.com/hashicorp/consul/agent/pool [no test files] === RUN TestManagerInternal_cycleServer --- PASS: TestManagerInternal_cycleServer (0.00s) === RUN TestManagerInternal_getServerList --- PASS: TestManagerInternal_getServerList (0.00s) === RUN TestManagerInternal_New --- PASS: TestManagerInternal_New (0.00s) === RUN TestManagerInternal_reconcileServerList --- PASS: TestManagerInternal_reconcileServerList (0.00s) === RUN TestManagerInternal_refreshServerRebalanceTimer --- PASS: TestManagerInternal_refreshServerRebalanceTimer (0.00s) === RUN TestManagerInternal_saveServerList --- PASS: TestManagerInternal_saveServerList (0.00s) === RUN TestRouter_Shutdown --- PASS: TestRouter_Shutdown (0.00s) === RUN TestRouter_Routing --- PASS: TestRouter_Routing (0.00s) === RUN TestRouter_Routing_Offline writer.go:29: 2021-01-29T19:31:10.770Z [DEBUG] TestRouter_Routing_Offline.router.manager: pinging server failed: server="node4.dc1 (Addr: tcp/127.0.0.5:8300) (DC: dc1)" error= writer.go:29: 2021-01-29T19:31:10.770Z [DEBUG] TestRouter_Routing_Offline.router.manager: pinging server failed: server="node3.dc1 (Addr: tcp/127.0.0.4:8300) (DC: dc1)" error= writer.go:29: 2021-01-29T19:31:10.770Z [DEBUG] TestRouter_Routing_Offline.router.manager: pinging server failed: server="node1.dc1 (Addr: tcp/127.0.0.2:8300) (DC: dc1)" error= writer.go:29: 2021-01-29T19:31:10.770Z [DEBUG] TestRouter_Routing_Offline.router.manager: pinging server failed: server="node2.dc1 (Addr: tcp/127.0.0.3:8300) (DC: dc1)" error= writer.go:29: 2021-01-29T19:31:10.770Z [DEBUG] TestRouter_Routing_Offline.router.manager: No healthy servers during rebalance, aborting --- PASS: TestRouter_Routing_Offline (0.00s) === RUN TestRouter_GetDatacenters --- PASS: TestRouter_GetDatacenters (0.00s) === RUN TestRouter_distanceSorter --- PASS: TestRouter_distanceSorter (0.00s) === RUN TestRouter_GetDatacentersByDistance --- PASS: TestRouter_GetDatacentersByDistance (0.00s) === RUN TestRouter_GetDatacenterMaps --- PASS: TestRouter_GetDatacenterMaps (0.00s) === RUN TestServers_AddServer --- PASS: TestServers_AddServer (0.00s) === RUN TestServers_IsOffline writer.go:29: 2021-01-29T19:31:10.782Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.782Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.782Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.782Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.782Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.783Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.783Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.783Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.783Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.783Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.783Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.783Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.783Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.784Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.784Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.784Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.784Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.784Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.784Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.784Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.784Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.784Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.784Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.785Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.785Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.785Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.785Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.785Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.785Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.785Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.785Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.785Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.786Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.786Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.786Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.786Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.786Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.786Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.786Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.786Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.787Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.787Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.787Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.787Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.787Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.787Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.787Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.787Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.788Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.788Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.788Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.788Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.788Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.788Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.788Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.788Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.788Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.789Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.789Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.789Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.789Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.789Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.789Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.789Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.789Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.789Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.790Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.790Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.790Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.790Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.790Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.790Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.790Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.790Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.790Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.791Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.791Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.791Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.791Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.791Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.791Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.791Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.791Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.791Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.792Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.792Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.792Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.792Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.792Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.792Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.792Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.792Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.792Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.793Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.793Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.793Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.793Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.793Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.793Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.793Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.793Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.794Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.794Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.794Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.794Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.794Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.794Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.794Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.794Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.794Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.795Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.795Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.795Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.795Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.795Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.795Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.795Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.795Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.795Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.796Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.796Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.796Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.796Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.796Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.796Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.796Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.796Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.796Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.797Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.797Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.797Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.797Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.797Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.797Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.797Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.797Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.798Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.798Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.798Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.798Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.798Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.798Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.798Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.798Z [DEBUG] TestServers_IsOffline.manager: Rebalanced servers, new active server: number_of_servers=1 active_server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.798Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.798Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting writer.go:29: 2021-01-29T19:31:10.799Z [DEBUG] TestServers_IsOffline.manager: pinging server failed: server="s1 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.799Z [DEBUG] TestServers_IsOffline.manager: No healthy servers during rebalance, aborting --- PASS: TestServers_IsOffline (0.02s) === RUN TestServers_FindServer writer.go:29: 2021-01-29T19:31:10.799Z [WARN] TestServers_FindServer.manager: No servers available writer.go:29: 2021-01-29T19:31:10.799Z [DEBUG] TestServers_FindServer.manager: cycled away from server: server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.799Z [DEBUG] TestServers_FindServer.manager: cycled away from server: server="s2 (Addr: /) (DC: )" --- PASS: TestServers_FindServer (0.00s) === RUN TestServers_New --- PASS: TestServers_New (0.00s) === RUN TestServers_NotifyFailedServer writer.go:29: 2021-01-29T19:31:10.800Z [DEBUG] TestServers_NotifyFailedServer.manager: cycled away from server: server="s1 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.800Z [DEBUG] TestServers_NotifyFailedServer.manager: cycled away from server: server="s2 (Addr: /) (DC: )" --- PASS: TestServers_NotifyFailedServer (0.00s) === RUN TestServers_NumServers --- PASS: TestServers_NumServers (0.00s) === RUN TestServers_RebalanceServers writer.go:29: 2021-01-29T19:31:10.802Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s66 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.819Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.820Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.821Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.821Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.821Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.821Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.821Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.821Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.821Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.821Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.821Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.821Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.822Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s45 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.844Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s52 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.844Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.860Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.861Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.861Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.861Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s13 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.865Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s23 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s04 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s43 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.875Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.887Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s18 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.887Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s41 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.888Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s50 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.898Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s11 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s92 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s56 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s32 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.906Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s22 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s15 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s78 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s27 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s20 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s38 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s36 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s91 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s97 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s13 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.952Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s47 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s41 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s09 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.953Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.954Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.955Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.956Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.957Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s38 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.969Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.970Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.970Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.970Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.970Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.970Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.970Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.970Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.970Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.971Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.971Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.971Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.971Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.971Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.971Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.971Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.971Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.971Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.972Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.972Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.972Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.972Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.972Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.972Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.972Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.972Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.975Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.975Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.975Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.975Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.975Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.977Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.977Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.977Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.977Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.978Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.978Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.978Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.978Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.978Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s34 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.978Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.978Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.978Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.983Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.983Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.983Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.983Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.983Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.983Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.987Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.987Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.987Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.987Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.987Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.987Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.988Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.988Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.988Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.988Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.989Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.989Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.990Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.990Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.990Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.991Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.991Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.991Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.992Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.992Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.992Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.992Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.993Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.993Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.993Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.993Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.993Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.994Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.994Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.994Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.994Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.995Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.995Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.995Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.995Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.995Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.996Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s35 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:10.996Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.996Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.996Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:10.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.000Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.000Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.000Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.000Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.000Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.001Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.001Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.001Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.001Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.001Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.001Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.003Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.003Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.003Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.003Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.003Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.004Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.004Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.004Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.004Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.004Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.005Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.005Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.005Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.005Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.005Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.007Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.007Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.007Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.007Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.007Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.008Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.008Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.008Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.008Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.008Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.010Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.010Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.010Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.010Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.010Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.012Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.012Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.012Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.012Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.012Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.012Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.013Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.013Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.013Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.013Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.013Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.013Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.014Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.014Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.014Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.014Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.019Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.019Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.019Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.019Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.019Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.019Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.019Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.019Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.027Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.027Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.027Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.027Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.027Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.027Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s01 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.028Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.028Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.028Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.028Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.028Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.028Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.029Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.029Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.029Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.029Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.029Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.029Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.041Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.051Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.051Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.051Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.051Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.051Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.052Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.052Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.052Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.052Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.053Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.053Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.053Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.053Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.053Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.054Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.054Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.054Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.054Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s65 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.055Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s61 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.055Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s92 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.055Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s31 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.055Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.056Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.056Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.056Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.056Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.057Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.057Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.057Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.057Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.058Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.058Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.058Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.058Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.059Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.059Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.059Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.059Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.059Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.060Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.060Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.060Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.060Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.060Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.061Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.061Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.061Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.061Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.062Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.062Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.062Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.062Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.062Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.063Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.063Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.063Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.063Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.064Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.064Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.064Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.064Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.064Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.065Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.065Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.065Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.065Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.066Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.066Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.066Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.066Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.067Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.067Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.067Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.067Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.067Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.068Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.068Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.068Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.068Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.069Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.069Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.069Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.069Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.069Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.070Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.070Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.070Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.070Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.070Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.071Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.071Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.071Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.071Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.071Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.072Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.072Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.072Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.072Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.073Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.073Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.073Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.073Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.073Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.074Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.074Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.074Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.074Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.074Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.075Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.075Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.075Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.075Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.075Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.076Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.076Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.076Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.076Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.076Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.077Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.077Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.077Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.077Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.078Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s07 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.078Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s64 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.078Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s90 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.078Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.079Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.079Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.079Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.079Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.080Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.080Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.080Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.080Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.081Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.081Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.081Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.081Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.082Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.082Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.082Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.082Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.083Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.083Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.083Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.083Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.083Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.085Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.085Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.085Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.085Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.086Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.086Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.086Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.086Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.101Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s39 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.102Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.112Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.112Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.112Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.112Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.117Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.117Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.117Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.117Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.118Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.118Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.118Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.118Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.125Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.125Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.125Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.125Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.125Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s90 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.126Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.126Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.126Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.134Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.134Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.134Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.136Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.136Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.136Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.136Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.148Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.148Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.148Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.148Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.148Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.149Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.149Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.149Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.149Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.150Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.150Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.150Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.150Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.151Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.151Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.151Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.152Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.152Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.152Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.153Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.153Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.153Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.153Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.154Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.154Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.154Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.154Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.154Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.155Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.155Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.155Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.156Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.156Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.156Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.156Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.156Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.157Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.164Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.164Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.165Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.166Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.166Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.166Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.167Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.167Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.168Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.168Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.168Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.169Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.169Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.169Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.169Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.170Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.170Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.170Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.171Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.172Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.172Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.173Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.173Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.173Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.174Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.174Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.174Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.175Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.175Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.175Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.176Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.176Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.176Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.177Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.177Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.178Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.178Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.179Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.179Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.179Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.180Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.180Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.181Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.181Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.181Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.182Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.182Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.183Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.183Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.183Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.184Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.184Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.184Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.184Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.185Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.185Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.185Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.185Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.185Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.186Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.186Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.186Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.186Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.187Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.187Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.187Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.187Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.187Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.188Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.188Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.188Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.188Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.189Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.189Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s66 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.189Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.189Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.190Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.190Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.190Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.191Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.191Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.191Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.191Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.192Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.192Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.192Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.192Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.193Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.193Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.193Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.194Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.194Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.195Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.195Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.196Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.196Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.197Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.197Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.198Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.198Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.198Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.198Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.211Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.212Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.213Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.214Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.215Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.216Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.217Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.218Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.219Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.220Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.221Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.231Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.231Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.232Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.233Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.234Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s70 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.235Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.236Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.237Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.238Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.245Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.246Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.247Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s70 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.248Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.249Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.250Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.251Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.256Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.256Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.256Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.256Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.257Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.258Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.259Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.260Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.261Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.262Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.263Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s47 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.264Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.265Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.266Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.267Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.268Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.269Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.270Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.271Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.272Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.273Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.274Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.275Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.276Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.277Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.277Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.277Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.277Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.277Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.278Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.278Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.278Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.278Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.278Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.279Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.279Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.279Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.279Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.279Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.279Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.280Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.280Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.280Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.280Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.281Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.281Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.281Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.281Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.281Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.282Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.282Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.282Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.282Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.282Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.283Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.283Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.283Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.283Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.283Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.284Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.284Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.284Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.284Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.284Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.285Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.285Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.285Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.285Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.285Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.286Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.286Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.286Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.286Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.287Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.287Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.287Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.287Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.288Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.288Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.288Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.288Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.289Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.289Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.289Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.289Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.289Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.289Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.290Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.290Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.290Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.291Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.291Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.291Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.291Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.291Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.291Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.292Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.292Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.292Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.292Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.292Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.293Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.293Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.293Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.293Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.294Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.294Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.294Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.294Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.294Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.295Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.295Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.295Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.295Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.296Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.296Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.296Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.296Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.296Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.297Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.297Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.297Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.297Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.297Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.298Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.298Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.298Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.298Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.299Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.299Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.299Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.299Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.299Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.299Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.300Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.300Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.300Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.300Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.300Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.301Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.301Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.301Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.301Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.301Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.302Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.302Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.302Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.302Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.302Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.303Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.303Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.303Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.303Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.303Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.303Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.304Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.304Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.304Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.304Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.305Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.305Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.305Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.306Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.306Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.306Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.306Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.306Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.307Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.307Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.307Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.307Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.307Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.308Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.308Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.308Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.308Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.308Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.309Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.309Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.309Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.309Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.309Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.310Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.310Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.310Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.310Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.310Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.310Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.311Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.311Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.311Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.311Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.311Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.312Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.312Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.312Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.312Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.312Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.313Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.313Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.313Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.313Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.313Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.314Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.314Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.314Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.314Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.314Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.315Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.315Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.315Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.315Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.316Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.316Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.316Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.316Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.316Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.317Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.317Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.317Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.317Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.317Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.318Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.318Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.318Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.318Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s92 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.319Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.319Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.319Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.319Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.319Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.320Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.320Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.320Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.320Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.320Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.321Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.321Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.321Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.321Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.321Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.322Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.322Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.322Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.322Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.323Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.323Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.323Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.323Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.323Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.324Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.324Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.324Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.324Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.324Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.324Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.325Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.325Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.325Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.325Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.325Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.325Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.326Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.326Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.326Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.326Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.326Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.327Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.327Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.327Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.327Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.327Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.327Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.328Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.328Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.328Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.328Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.328Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.328Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.329Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.329Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.329Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.329Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.329Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.330Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.330Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.330Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.331Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.331Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.331Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.331Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.332Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.332Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.332Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.332Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.333Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.333Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.333Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.334Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.334Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.334Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.335Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.335Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.335Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.336Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.336Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.337Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.337Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.337Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.338Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.338Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.338Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.338Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.339Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.339Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.339Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.339Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.339Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.340Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.340Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.340Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.340Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.340Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.340Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.341Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.341Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.341Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.341Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.341Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.341Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.341Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.342Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.342Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.342Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.342Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.342Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.342Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.343Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.343Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.343Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.343Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.343Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.343Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.343Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.343Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.343Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.344Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.344Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.344Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.344Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.344Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.344Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.344Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.345Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.345Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.345Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.345Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.345Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.345Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.345Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.346Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.346Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.346Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.346Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.346Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.346Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.346Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.347Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.347Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.347Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.347Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.347Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.347Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.348Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.348Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.348Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.348Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.348Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.348Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.348Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.348Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.349Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.349Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.349Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.349Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.349Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.349Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.349Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.349Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.350Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.350Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.350Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.350Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.350Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.350Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.350Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.350Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.351Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.351Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.351Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.351Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.351Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.351Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.351Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.351Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.351Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.357Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.357Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.357Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.358Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.358Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.358Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.358Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.358Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.358Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.359Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.359Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.359Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.359Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.359Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.360Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.360Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.360Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.360Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.361Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.361Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.361Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.361Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.361Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.362Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.362Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.362Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.362Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.362Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.362Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.362Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.362Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.363Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.363Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.363Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.363Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.363Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.363Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.363Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.363Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.363Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.364Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.364Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.364Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.364Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.364Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.364Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.364Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.364Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.365Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.365Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.365Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.365Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.365Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.365Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.365Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.365Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.366Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.366Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.366Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.366Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.366Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.366Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.366Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.367Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.367Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.367Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.367Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.367Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.367Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.367Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.368Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.368Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.368Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.368Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.368Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.368Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.368Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.368Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.368Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.369Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.369Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.369Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.369Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.369Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.369Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.369Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.369Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.370Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.370Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.370Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.370Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.370Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.370Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.370Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.370Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.371Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.371Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.371Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.371Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.371Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.371Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.371Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.371Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.372Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.372Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.372Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.372Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.372Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.372Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.372Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.372Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.372Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.373Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.373Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.373Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.373Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.373Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.373Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.373Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.373Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.373Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.374Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.374Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.374Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s22 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.374Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.374Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.375Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.375Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.375Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.375Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.375Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.376Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.376Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.376Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.376Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.376Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.376Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.376Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.376Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.377Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.377Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.377Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.377Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.377Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.377Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.377Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.377Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.378Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.378Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.378Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.378Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.378Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.378Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.378Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.378Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.379Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.379Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.379Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.379Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.379Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.379Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.379Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.380Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.380Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.380Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.380Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.380Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.380Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.380Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.381Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.381Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.381Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.381Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.381Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.381Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.381Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.382Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.382Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.382Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.382Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.382Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.382Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.382Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.382Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.383Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.383Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.383Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.383Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.383Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.383Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.383Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.383Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.383Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.384Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.384Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.384Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.384Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.384Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.384Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.384Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.384Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.385Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.385Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.385Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.385Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.385Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.385Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.385Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.386Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.386Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.386Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.386Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.386Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.386Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.386Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.386Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.387Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.387Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.387Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.387Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.387Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.387Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.387Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.387Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.388Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.388Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.388Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.388Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.388Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.388Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.389Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.389Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.389Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.389Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.389Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.389Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.389Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.389Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.390Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.390Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.390Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.390Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.390Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.390Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.390Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.390Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.391Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.391Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.391Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.391Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.391Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.391Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.391Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.391Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.391Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.392Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.392Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.392Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.392Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.392Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.392Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.392Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.393Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.393Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.393Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.393Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.393Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.393Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.393Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.393Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.393Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.394Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.394Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.394Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.394Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.394Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.394Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.394Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.394Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.394Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.395Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.395Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.395Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.395Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.395Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.395Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.395Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.395Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.396Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.396Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.396Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.396Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.396Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.396Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.396Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.396Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.397Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.397Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.397Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.397Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.397Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.397Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.397Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.397Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.397Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.398Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.398Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.398Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.398Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.398Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.398Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.398Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.398Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.399Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.399Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.399Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.399Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.399Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.399Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.399Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.399Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.399Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.400Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.400Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.400Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.400Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.400Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s96 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.400Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.400Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.401Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.401Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.401Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.401Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.401Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.401Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.401Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.401Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.402Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.402Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.402Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.402Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.402Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.402Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.402Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.402Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.402Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.403Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.403Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.403Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.403Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.403Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.403Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.403Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.403Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.404Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.404Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.404Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.404Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.404Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.404Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.404Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.404Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.404Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.405Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.405Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.405Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.405Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.405Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.405Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.405Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.405Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.406Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.406Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.406Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.406Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.406Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.406Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.406Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.406Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.407Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.407Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.407Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.407Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.407Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.407Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.407Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.407Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.407Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.408Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.408Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.408Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.408Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.408Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.408Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.408Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.408Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.409Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.409Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.409Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.409Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.409Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.409Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.409Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.409Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.409Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.410Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.410Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.410Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.410Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.410Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.410Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.411Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.411Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.411Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.411Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.411Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.411Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.411Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.411Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.412Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.412Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.412Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.412Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.412Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.412Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.412Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.412Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.413Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.413Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.413Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.413Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.413Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.413Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.413Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.413Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.414Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.414Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.414Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.414Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.414Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.414Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.414Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.414Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.415Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.415Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.415Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.415Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.415Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.415Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.415Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.416Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.416Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.416Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.416Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.416Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.416Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.416Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.417Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.417Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.417Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.417Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.417Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.417Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.418Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.418Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.418Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.418Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.418Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.418Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.418Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.418Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.419Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.419Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.419Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.419Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.419Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.419Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.419Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.420Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.420Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.420Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.420Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.420Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.420Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.420Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.421Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.421Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.421Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.421Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.421Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.421Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.421Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.421Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.422Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.422Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.422Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.422Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.422Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.422Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.423Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.423Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.423Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.423Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.423Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.423Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.423Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.423Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.424Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.424Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.424Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.424Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.424Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.424Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.424Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.425Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.425Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.425Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.425Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.427Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.427Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.427Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.427Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.427Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.427Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.427Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.427Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.427Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.427Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.428Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.428Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s61 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.428Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.428Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.428Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.428Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.429Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.429Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.429Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.429Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.429Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.429Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.429Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.430Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.430Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.430Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.430Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.430Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.430Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.430Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.431Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.431Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.431Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.431Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.431Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.431Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.431Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.431Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.432Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.432Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.432Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.432Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.432Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.432Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.432Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.433Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.433Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.433Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.433Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.433Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.433Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.433Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.434Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.434Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.434Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.434Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.434Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.434Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.434Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.434Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.435Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.435Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.435Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.435Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.435Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.435Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.435Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.436Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.436Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.436Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.436Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.436Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.436Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.436Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.437Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.437Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.437Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.437Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.437Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.437Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.437Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.438Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.438Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.438Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.438Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.438Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.438Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.438Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.438Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.439Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.439Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.439Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.439Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.439Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.439Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.439Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.439Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.440Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.440Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.440Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.440Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.440Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.440Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.440Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.441Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.441Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.441Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.441Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.441Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.441Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.441Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.441Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.441Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.442Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s05 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.442Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.442Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.442Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.442Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.442Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.442Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.443Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.443Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.443Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.443Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.443Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.443Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.443Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.444Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.444Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.444Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.445Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.445Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.445Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.445Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.445Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.445Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.446Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.446Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.446Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.447Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.447Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.447Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.447Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.447Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.447Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.447Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.447Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.448Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.448Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.448Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.448Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.448Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.448Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.448Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.448Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.449Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.449Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.449Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.449Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.449Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.449Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.450Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.450Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.450Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.450Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.450Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.450Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.450Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.450Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.451Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.451Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.451Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.451Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.451Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.451Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.451Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.451Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.452Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.452Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.452Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.452Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.452Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.452Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.452Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.453Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.453Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.453Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.453Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.453Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.453Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.453Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.453Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.453Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.454Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.454Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.454Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.454Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.454Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.454Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.454Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.454Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.454Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.454Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.455Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.455Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.455Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.455Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.455Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.455Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.455Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.455Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.456Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.456Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.456Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.456Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.456Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.456Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.456Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.457Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.457Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.457Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.457Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.457Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.457Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.457Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.457Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.458Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.458Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.458Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.458Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.458Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.458Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.458Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.458Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.458Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.459Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.459Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.459Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.459Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.459Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.459Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.459Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.459Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.460Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.460Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.460Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.460Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.460Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.460Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.460Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.460Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.461Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.461Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.461Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.461Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.461Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.461Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.461Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.461Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.461Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.461Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.462Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.462Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.462Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.462Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.462Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.462Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.462Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.462Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.462Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.462Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.463Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.464Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.464Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.464Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.464Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.464Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.464Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.464Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.464Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.464Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.464Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.465Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.465Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.465Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.465Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.465Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.465Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.465Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.465Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.465Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.466Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.466Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.466Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.466Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.466Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.466Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.466Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.466Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.467Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.467Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.467Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.467Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.467Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.467Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.467Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.468Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.468Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.468Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.468Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.468Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.468Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.468Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.468Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.468Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.468Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.469Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.469Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.469Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.469Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.469Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.469Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.469Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.469Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.469Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.469Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.470Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.482Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.482Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.482Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.482Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.482Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.482Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.482Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.482Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.482Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.483Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.483Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.483Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.483Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.483Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.483Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.483Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.483Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.483Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.483Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.484Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.484Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.484Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.484Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.484Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.484Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.484Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.485Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.485Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.485Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.485Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.485Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.485Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.485Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.486Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.486Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.486Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.486Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.486Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.487Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.487Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.487Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.487Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.487Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.488Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.488Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.488Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.489Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.489Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.489Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.489Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.490Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.490Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.490Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.491Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.491Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.491Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.491Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.492Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.492Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.493Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.493Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.493Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.493Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.494Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.494Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.494Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.494Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.494Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.494Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.495Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.495Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.495Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.495Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.495Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.496Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.496Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.496Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.496Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.496Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.497Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.497Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s55 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.497Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.498Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.498Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.498Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.498Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.498Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.498Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.499Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.499Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.499Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.499Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.500Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.500Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.500Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.500Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.501Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.501Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.501Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.501Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.502Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.502Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.502Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.502Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.502Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.503Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.503Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.504Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.504Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.504Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.504Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.504Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.505Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.505Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.505Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.505Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.505Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.506Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.506Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.506Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.506Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.506Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.507Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.507Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.507Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.507Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.508Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.508Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.508Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.508Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.509Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.509Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.509Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.509Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.510Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.510Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.510Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.510Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.510Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.510Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.511Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.511Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.511Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.511Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.512Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.512Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.512Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.512Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.512Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.512Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.513Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.513Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.513Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.513Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.514Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.514Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.514Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.514Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.514Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.515Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.515Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.515Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.515Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.516Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.516Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.516Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.516Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.516Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.517Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.517Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.517Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.517Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.518Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.518Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.518Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.518Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.518Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.519Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.519Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.519Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.519Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.519Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.520Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.520Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.520Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.520Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.521Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.521Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.521Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.521Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.521Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.522Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.522Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.522Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.522Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.523Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.523Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.523Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.523Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.523Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.523Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.524Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.524Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.524Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.524Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.525Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.525Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.525Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.525Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.526Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.526Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.526Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.526Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.526Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.527Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.527Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.527Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.527Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.528Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.528Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.528Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.528Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.528Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.529Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.529Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.529Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.529Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.529Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.530Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.530Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.530Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.530Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.530Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.531Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.531Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.531Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.531Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.531Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.532Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.532Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.532Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.532Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.532Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.533Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.533Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.533Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.533Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.533Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.534Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.534Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.534Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.534Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.534Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.535Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.535Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.535Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.535Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.536Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.536Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.536Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.536Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.536Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.537Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.537Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.537Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.537Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.538Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.538Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.538Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.538Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.539Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.539Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.539Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.539Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.539Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.540Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.540Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.540Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.540Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.541Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.541Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.541Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.541Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.542Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s34 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.542Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.542Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.542Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.543Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.543Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.543Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.543Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.544Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.544Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.544Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.544Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.544Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.545Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.545Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.545Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.545Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.545Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.546Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.546Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.546Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.546Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.546Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.547Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.547Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.547Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.548Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.548Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.548Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.548Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.548Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.549Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.549Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.549Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.549Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.550Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.550Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.550Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.550Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.550Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.551Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.551Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.551Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.551Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.552Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.552Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.552Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.552Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.552Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.553Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.553Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.553Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.553Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.553Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.554Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.554Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.554Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.554Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.554Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.555Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.555Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.555Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.555Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.555Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.556Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.556Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.556Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.556Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.557Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.557Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.557Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.557Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.557Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.558Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.558Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.558Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.558Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.558Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.559Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.559Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.559Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.559Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.559Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.560Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.560Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.560Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.560Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.560Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.561Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.561Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.561Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.561Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.562Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.562Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.562Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.562Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.563Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.563Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.563Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.563Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.564Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.564Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.564Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s04 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.564Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s05 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.565Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s28 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.565Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s86 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.565Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.566Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.566Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.566Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.566Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.567Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.567Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.567Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.567Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.567Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.568Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.568Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.568Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.568Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.569Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.569Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.569Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.569Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.570Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.570Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.570Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.570Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.571Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.571Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.571Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.571Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.571Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.572Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.572Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.572Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.572Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.573Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.573Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.573Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.573Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.573Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.574Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.574Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.574Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.574Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.574Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.574Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.574Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.574Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.574Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.575Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.575Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.575Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.575Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.575Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.575Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.575Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.575Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.575Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.576Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.576Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.576Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.576Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.576Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.576Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.576Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.576Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.577Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.577Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.577Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.577Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.577Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.577Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.577Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.577Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.577Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.578Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.578Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.582Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.582Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.582Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.582Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.583Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.583Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.583Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.583Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.583Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.584Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.584Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.584Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.584Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.585Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.585Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.585Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.585Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.586Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.586Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.586Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.586Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.586Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.587Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.587Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.587Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.587Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.587Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.588Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.588Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.588Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.588Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.588Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.589Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.589Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.589Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.589Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.589Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.590Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.590Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.590Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.590Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.590Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.591Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.591Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.591Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.591Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.591Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.592Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.592Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.592Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.592Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.593Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.593Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.593Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.593Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.593Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.594Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.594Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.594Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.594Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.594Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.595Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.595Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.595Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.595Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.595Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.596Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.596Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.596Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.596Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.596Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.597Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.597Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.597Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.597Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.597Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.598Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.598Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.598Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.598Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.598Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.599Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.599Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.599Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.599Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.599Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.599Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.599Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.600Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.600Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.600Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.600Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.600Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.600Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.600Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.600Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.601Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.601Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.601Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.601Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.601Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.601Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.601Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.601Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.601Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.602Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.602Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.602Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.602Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.602Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.602Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.602Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.603Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.603Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.603Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.603Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.603Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.603Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.603Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.603Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.604Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.604Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.604Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.604Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.604Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.604Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.604Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.604Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.604Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.605Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.605Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.605Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.605Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.605Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.605Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.605Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.605Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.606Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.606Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.606Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.606Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.606Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.606Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.606Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.606Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.607Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.607Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.607Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.607Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.607Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.607Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.607Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.607Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.608Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.608Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.608Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.608Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.608Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.608Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.608Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.608Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.609Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.609Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.609Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.609Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.609Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.609Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.609Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.609Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.610Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.610Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.610Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.610Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.610Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.610Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.610Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.610Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.610Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.611Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.611Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.611Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.611Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.611Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.611Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.611Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.611Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.611Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.612Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.612Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.612Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.612Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.612Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.612Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.612Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.612Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.613Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.613Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.613Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.613Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.613Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.613Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.613Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.613Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.613Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.614Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.614Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.614Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.614Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.614Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.614Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.614Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.614Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.614Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.615Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.615Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.615Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.615Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.615Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.615Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.615Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.615Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.616Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.616Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.616Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.616Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.616Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.616Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.616Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.616Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.617Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.617Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.617Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.617Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.617Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.617Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.617Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.617Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.618Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.618Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.618Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.618Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.618Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.618Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.618Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.618Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.618Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.619Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.619Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.619Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.619Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.619Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.619Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.619Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.619Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.619Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.620Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.620Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.620Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.620Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.620Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.620Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.620Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.620Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.620Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.621Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.621Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.621Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.621Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.621Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.621Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.621Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.621Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.621Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.622Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.623Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.624Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.625Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.626Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.626Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.626Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.626Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.626Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.626Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s33 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.626Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.626Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.626Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.627Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.628Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.629Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.630Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.631Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.632Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.633Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s10 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.634Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.635Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.636Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.637Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.638Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.639Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.640Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.641Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.642Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.643Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.644Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.645Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.645Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.645Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.645Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.645Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.645Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.645Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.645Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.645Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.646Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.647Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.648Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.649Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.650Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.651Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.652Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.653Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.654Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.655Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s13 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s31 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s69 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s41 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.656Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.657Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.658Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.659Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.660Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.661Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.662Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.663Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s25 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s30 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.664Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.665Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.666Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.667Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.668Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.669Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.670Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s87 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.671Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.672Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.673Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.674Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.675Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.676Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.677Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.678Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.679Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.680Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.681Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.682Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.682Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.682Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.682Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.682Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.682Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.682Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.682Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.682Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.683Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.683Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.683Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.683Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.683Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.683Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.683Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.683Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.683Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.684Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.684Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.684Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.684Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.684Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.684Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.684Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.684Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.684Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.684Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.685Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.685Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.685Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.685Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.685Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.685Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.685Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.685Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.685Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.686Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.686Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.686Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.686Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.686Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.686Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.686Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.686Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.686Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.686Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.687Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.687Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.687Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.687Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.687Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.687Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.687Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.687Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.688Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.688Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.688Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.688Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.688Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.688Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.688Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.688Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.688Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.689Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.689Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.689Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.689Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.689Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.689Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.689Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.689Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.689Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.689Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.690Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.690Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.690Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.690Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.690Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.690Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.690Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.690Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.690Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.691Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.691Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.691Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.691Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.691Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.691Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.691Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.691Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.691Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.692Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.692Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.692Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.692Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.692Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.692Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.692Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.692Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.692Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.692Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.693Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.693Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.693Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.693Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.693Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.693Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.693Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.693Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.693Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.694Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.694Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.694Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.694Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.694Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.694Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.694Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.694Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.694Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.694Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.695Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.695Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.695Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.695Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.695Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.695Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.695Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.695Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.696Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.696Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.696Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.696Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.696Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.696Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.696Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.696Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.696Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.696Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.697Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.697Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.697Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.697Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.697Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.711Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.712Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.713Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.714Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.715Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.716Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.717Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.718Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.718Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.718Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.718Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.718Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.718Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.718Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.718Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.718Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.718Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.719Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.719Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.719Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.719Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.719Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.719Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.719Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.719Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.720Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.720Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.720Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.720Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.720Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.720Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.720Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.720Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.720Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.721Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.721Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.721Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.721Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.721Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.721Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.721Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.722Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.722Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.722Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.722Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.722Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.722Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.722Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.722Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.722Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.723Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.723Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.723Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.723Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.723Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.723Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.723Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.723Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.723Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.724Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.724Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.724Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.724Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.724Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.724Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.724Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.724Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.725Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.725Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.725Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.725Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.725Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.725Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.725Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.725Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.725Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.726Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.726Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.726Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.726Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.726Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.726Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.726Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.726Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.726Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.727Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.727Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.727Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.727Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.727Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.727Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.727Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.727Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.727Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.728Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.728Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.728Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.728Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.728Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.728Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.728Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.728Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.729Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.729Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.729Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.729Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.729Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.729Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.729Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.729Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.729Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.730Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.730Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.730Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.730Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.730Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.730Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.730Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.730Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.730Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.730Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.731Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.731Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.731Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.731Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.731Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.731Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.731Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.741Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.742Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.742Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.742Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.742Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.743Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.743Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.743Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.743Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.743Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.743Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.744Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.744Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.744Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.744Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.745Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.745Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.745Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.746Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.746Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.746Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.746Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.747Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.747Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.747Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.748Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.748Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.748Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.748Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.749Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.749Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.749Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.750Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.750Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.750Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.750Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.751Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.751Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.751Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.751Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.751Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.751Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.752Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.752Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.752Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.752Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.752Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.753Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.753Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.753Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.753Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.753Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.753Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.754Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.754Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.754Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.754Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.754Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.755Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.755Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.755Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.755Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.755Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.755Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.756Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.756Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.756Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.756Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.756Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.757Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.757Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.757Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.757Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.759Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.759Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.759Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.759Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.760Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.760Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.760Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.760Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.760Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.760Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.761Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.761Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.761Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.761Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.761Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.762Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.762Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.762Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.763Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.763Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.763Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.763Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.764Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.764Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.764Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.764Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.764Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.765Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.765Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.765Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.765Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.765Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.766Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.766Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.766Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.766Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.766Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.767Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.767Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.767Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.767Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.768Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.768Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.768Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.768Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.768Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.769Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.769Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.769Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.769Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.769Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.770Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.770Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.770Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.770Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.770Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.771Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.771Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.771Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.771Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.771Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.772Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.772Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.772Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.772Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.772Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.772Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.772Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.772Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.773Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.773Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.773Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.773Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.773Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.773Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.773Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.773Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.773Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.774Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.774Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.774Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.774Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.774Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.774Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.774Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.774Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.775Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.775Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.775Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.775Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.775Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.775Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.775Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.776Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.776Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.776Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.776Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.776Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.776Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.776Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.776Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.777Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.777Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.777Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.777Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.777Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.777Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.777Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.777Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.778Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.778Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.778Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.778Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.778Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.778Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.778Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.778Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.779Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.779Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.779Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.779Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.779Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.779Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s15 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.779Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.779Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.780Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.780Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.780Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.780Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.780Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.780Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.780Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.780Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.781Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.781Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.781Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.781Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.781Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.781Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.781Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.781Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.781Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.782Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.782Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.782Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.782Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.782Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.782Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.782Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.782Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.783Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.783Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.783Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.783Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.783Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.783Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.783Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.783Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.783Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.784Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.784Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.784Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.784Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.784Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.784Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.784Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.784Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.785Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.785Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.785Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.785Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.785Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.785Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.785Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.786Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.786Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.786Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.786Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.786Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.786Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.786Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.786Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.786Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.787Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.787Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.787Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.787Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.787Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.787Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.787Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.787Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.788Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.788Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.788Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.788Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.788Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.788Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.788Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.788Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.788Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.789Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.789Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.789Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.789Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.789Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.789Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.789Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.789Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.789Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.790Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.790Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.790Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.790Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.790Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.790Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.790Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.790Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.791Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.791Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.791Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.791Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.791Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.791Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.791Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.791Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s33 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.792Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s90 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.792Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.792Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.792Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.792Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.792Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.792Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.792Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.793Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.793Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.793Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.793Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.793Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.793Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.793Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.793Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.794Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.794Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.794Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.794Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.794Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.794Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.794Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.794Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.794Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.795Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.795Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.795Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.795Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.795Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.795Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.795Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.795Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.796Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.796Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.796Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.796Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.796Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.796Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.796Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.796Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.796Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.797Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.797Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.797Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.797Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.797Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.797Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.797Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.797Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.798Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.798Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.798Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.798Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.798Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.798Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.798Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.798Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.798Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.799Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.799Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.799Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.799Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.799Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.799Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.800Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.800Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.800Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.800Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.800Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.800Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.800Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.801Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.801Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.801Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.801Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.801Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.801Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.801Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.801Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.801Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.802Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.803Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.804Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.804Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.805Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.806Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.807Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.808Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.809Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.810Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.811Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.812Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.813Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.814Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.815Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.816Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.816Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.817Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.818Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.823Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.824Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.825Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.826Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.827Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s80 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s51 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.828Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.829Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.830Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.831Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.832Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.833Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.834Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.835Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.836Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s59 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.836Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s88 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.836Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.836Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.837Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.838Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.839Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.840Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.841Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.842Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.843Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.844Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.845Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.846Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.847Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.848Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s55 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.848Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.848Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.849Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.850Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.851Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.852Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.853Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.854Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.855Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.856Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.857Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.858Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.859Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.859Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.859Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.859Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.859Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.859Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.859Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.859Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.860Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.860Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.860Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.860Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.860Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.860Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.860Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.860Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.861Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s92 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.861Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.861Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.861Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.861Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.861Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.861Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.862Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.863Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.864Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.865Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.866Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.867Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.868Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.869Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.870Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.871Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.872Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.873Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.874Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.874Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.875Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.876Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.877Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.878Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.879Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.880Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.881Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.882Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.883Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.884Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.885Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.886Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.887Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.888Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.889Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s10 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.890Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.891Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.892Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.893Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.894Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.895Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.896Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s17 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s14 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.897Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.898Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.899Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.900Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.901Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.902Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.903Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s08 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.904Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.905Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.906Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.907Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.908Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.909Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.910Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.911Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.912Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s58 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s29 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s77 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s86 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s67 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.913Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.914Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.915Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.916Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.917Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.918Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.919Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s39 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s89 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.920Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.921Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.922Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.923Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.924Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.925Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.926Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s45 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s54 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s52 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.927Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.928Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.929Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.930Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.931Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.932Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.933Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.934Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.935Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.936Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.937Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.938Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.939Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.940Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.941Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.942Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.943Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.944Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.945Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s05 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.945Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s88 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.945Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.945Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.946Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.947Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.948Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.949Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.950Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.951Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.958Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.959Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.960Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.961Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.962Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.963Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s08 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.963Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.963Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.964Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.964Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.973Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.974Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.975Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.975Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.975Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.975Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.976Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.977Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.977Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.977Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.977Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.977Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.978Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.978Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.978Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.978Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.979Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.980Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.981Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.982Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.983Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.983Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.983Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.983Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.984Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.985Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.986Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.987Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.987Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.987Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.988Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.988Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.988Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.988Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.988Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.989Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.989Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.989Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.989Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.990Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.990Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.990Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.990Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.990Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.991Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.991Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.991Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.991Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.991Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.992Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.992Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.992Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.992Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.992Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.993Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.993Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.993Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.993Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.993Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.994Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.994Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.994Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.994Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.994Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.995Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.995Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s47 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.995Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s08 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.996Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s89 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.996Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s79 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:11.996Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.997Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.998Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:11.999Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.000Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.000Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.000Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.000Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.000Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.001Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.001Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.001Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.001Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.002Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.003Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.003Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.003Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.003Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.003Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.004Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.004Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.004Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.004Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.005Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.005Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.005Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.005Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.005Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.006Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.007Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.007Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.007Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.007Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.008Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.008Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.008Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.008Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.009Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.010Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.010Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.010Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.010Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.010Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.011Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.012Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.012Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.012Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.012Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.013Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.013Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.013Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.013Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.014Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.014Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.014Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.015Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.016Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.017Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.018Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.019Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s83 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.019Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.019Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.019Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.020Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.021Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.022Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.023Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.024Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.025Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.026Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.027Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.027Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.027Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.027Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.028Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.028Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.028Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.028Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.028Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.029Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.029Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.029Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.029Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.030Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.031Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.032Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.033Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.034Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.035Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.036Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.037Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.038Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.039Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.040Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.041Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.041Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.042Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.043Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.044Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.045Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.046Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.047Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.048Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.049Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.050Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.051Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.051Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.051Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.051Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.052Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.052Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.052Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.052Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.053Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.053Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.053Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.053Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.053Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.054Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.054Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.054Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.054Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.055Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.055Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.055Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.055Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.056Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.056Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.056Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.056Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.057Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.057Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.057Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.057Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.057Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.058Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.058Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.058Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.059Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.059Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.059Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.059Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.060Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.060Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.060Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.060Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.060Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.061Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.061Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.061Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.061Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.062Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.062Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.062Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.062Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.062Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.063Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.063Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.063Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.063Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.064Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.064Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.064Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.064Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s29 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.064Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s80 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.065Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s65 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.065Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.065Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.065Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.066Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.066Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.066Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.066Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.067Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.067Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.067Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.067Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.068Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.068Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.068Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.068Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.069Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.069Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.069Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.069Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.069Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.070Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.070Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.070Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.071Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.071Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.071Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.071Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.071Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.072Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.072Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.072Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.072Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.072Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.073Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.073Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.073Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.073Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.073Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.074Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.074Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.074Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.074Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.074Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.075Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.075Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.076Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.076Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.076Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.076Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.076Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.077Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.077Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.077Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.077Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.077Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.078Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.078Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.078Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.078Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.078Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.079Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.079Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.079Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.079Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.079Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.080Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.080Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.080Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.080Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.080Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.081Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.081Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.081Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.081Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.081Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.082Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.082Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.082Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.082Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.082Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.083Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.083Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.083Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.083Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.083Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.084Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.085Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.085Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.085Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.085Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.085Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.085Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.086Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.086Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.086Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s54 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.086Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.086Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.086Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.086Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.086Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.087Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.088Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.089Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.090Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.091Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.092Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.093Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.094Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.095Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.096Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.097Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.097Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.098Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.099Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.100Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.101Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.102Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.103Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.104Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.105Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.106Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.107Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.108Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.109Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s47 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.109Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.109Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.110Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.111Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.112Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.112Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.112Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.112Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.112Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.112Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.113Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.114Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.115Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.116Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.117Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.117Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.117Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.117Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.117Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.117Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.119Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.120Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.121Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.122Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.123Z [DEBUG] TestServers_RebalanceServers.manager: pinging server failed: server="s76 (Addr: /) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.123Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.123Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.124Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.125Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.125Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.125Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.125Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.125Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.125Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.126Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.126Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.126Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.126Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.126Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.126Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.126Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.126Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.127Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.128Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.129Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.130Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.131Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.132Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.133Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.134Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.134Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.134Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.134Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.134Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.134Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.134Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.135Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.136Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.136Z [DEBUG] TestServers_RebalanceServers.manager: Rebalanced servers, new active server: number_of_servers=100 active_server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.136Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.136Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.136Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s87 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s32 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s82 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s83 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s60 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s46 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s55 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s67 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.137Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s27 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s88 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s84 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s56 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s89 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s37 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s81 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.138Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s86 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s85 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s20 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s53 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s00 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s45 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s95 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.139Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s49 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s63 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s30 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s61 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s94 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s25 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s59 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s97 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.140Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s76 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s99 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s35 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s24 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s74 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s48 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.141Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s64 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s43 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s65 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s62 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s72 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s77 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s80 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s54 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.142Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s96 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s68 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s21 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s75 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s28 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s31 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.143Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s23 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s70 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s39 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s79 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s90 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s26 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s92 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.144Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s12 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s40 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s52 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s44 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s36 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s34 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s42 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s73 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s58 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.145Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s51 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s69 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s38 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s78 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s41 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s29 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s66 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s50 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.146Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s93 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s98 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s10 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s47 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s33 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s57 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s91 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s22 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.147Z [DEBUG] TestServers_RebalanceServers.manager: cycled away from server: server="s71 (Addr: /) (DC: )" --- PASS: TestServers_RebalanceServers (1.35s) === RUN TestServers_RebalanceServers_AvoidFailed writer.go:29: 2021-01-29T19:31:12.148Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.148Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.148Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.148Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.149Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.149Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.149Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.149Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.149Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.149Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.149Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.150Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.150Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.150Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.150Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.150Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.150Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.150Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.151Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.151Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.151Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.151Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.151Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.151Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.151Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.151Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.152Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.152Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.152Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.152Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.152Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.152Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.152Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.152Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.153Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.153Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.153Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.153Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.153Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.153Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.153Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.154Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.154Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.154Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.154Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.154Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.154Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.154Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.155Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.155Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.155Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.155Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.155Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.155Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.155Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.156Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.156Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.156Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.156Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.156Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.156Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.156Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.156Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.157Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.157Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.157Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.157Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.157Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.157Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.157Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.158Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.158Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.158Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.158Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.158Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.158Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.159Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.159Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.159Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.159Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.159Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.159Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.160Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.160Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.160Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.160Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.160Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.160Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.160Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.160Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.161Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.161Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.161Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.161Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.161Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.161Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.161Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.161Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.162Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.162Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.162Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.162Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.162Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.162Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.162Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.162Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.163Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.163Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.163Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.163Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.163Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.163Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.163Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.163Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.163Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.164Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.164Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.164Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.164Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.164Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.164Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.164Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.165Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.165Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.165Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.165Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.165Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.165Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.165Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s3 (Addr: faux/s3) (DC: )" writer.go:29: 2021-01-29T19:31:12.165Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.166Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.166Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" writer.go:29: 2021-01-29T19:31:12.166Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: pinging server failed: server="s2 (Addr: faux/s2) (DC: )" error= writer.go:29: 2021-01-29T19:31:12.166Z [DEBUG] TestServers_RebalanceServers_AvoidFailed.manager: Rebalanced servers, new active server: number_of_servers=3 active_server="s1 (Addr: faux/s1) (DC: )" --- PASS: TestServers_RebalanceServers_AvoidFailed (0.02s) === RUN TestManager_RemoveServer writer.go:29: 2021-01-29T19:31:12.166Z [DEBUG] TestManager_RemoveServer.manager: Rebalanced servers, new active server: number_of_servers=19 active_server="s15 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.167Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s11 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.167Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s16 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.167Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s08 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.167Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s03 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.167Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.167Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s19 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.167Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.167Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s13 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.168Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.169Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s18 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.169Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.169Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s09 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.169Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.169Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s07 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.169Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.169Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s01 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.169Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.170Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s04 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.170Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.170Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s02 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.170Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.170Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s06 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.170Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.170Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s05 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.170Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s14 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.170Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s17 (Addr: /) (DC: )" writer.go:29: 2021-01-29T19:31:12.170Z [DEBUG] TestManager_RemoveServer.manager: cycled away from server: server="s14 (Addr: /) (DC: )" --- PASS: TestManager_RemoveServer (0.00s) PASS ok github.com/hashicorp/consul/agent/router 1.430s === RUN TestStructs_ACLCaches === PAUSE TestStructs_ACLCaches === RUN TestStructs_ACL_IsSame --- PASS: TestStructs_ACL_IsSame (0.00s) === RUN TestStructs_ACL_Convert === PAUSE TestStructs_ACL_Convert === RUN TestStructs_ACLToken_Convert === PAUSE TestStructs_ACLToken_Convert === RUN TestStructs_ACLToken_PolicyIDs === PAUSE TestStructs_ACLToken_PolicyIDs === RUN TestStructs_ACLToken_EmbeddedPolicy === PAUSE TestStructs_ACLToken_EmbeddedPolicy === RUN TestStructs_ACLServiceIdentity_SyntheticPolicy === PAUSE TestStructs_ACLServiceIdentity_SyntheticPolicy === RUN TestStructs_ACLToken_SetHash === PAUSE TestStructs_ACLToken_SetHash === RUN TestStructs_ACLToken_EstimateSize === PAUSE TestStructs_ACLToken_EstimateSize === RUN TestStructs_ACLToken_Stub === PAUSE TestStructs_ACLToken_Stub === RUN TestStructs_ACLTokens_Sort === PAUSE TestStructs_ACLTokens_Sort === RUN TestStructs_ACLTokenListStubs_Sort === PAUSE TestStructs_ACLTokenListStubs_Sort === RUN TestStructs_ACLPolicy_Stub === PAUSE TestStructs_ACLPolicy_Stub === RUN TestStructs_ACLPolicy_SetHash === PAUSE TestStructs_ACLPolicy_SetHash === RUN TestStructs_ACLPolicy_EstimateSize === PAUSE TestStructs_ACLPolicy_EstimateSize === RUN TestStructs_ACLPolicies_Sort === PAUSE TestStructs_ACLPolicies_Sort === RUN TestStructs_ACLPolicyListStubs_Sort === PAUSE TestStructs_ACLPolicyListStubs_Sort === RUN TestStructs_ACLPolicies_resolveWithCache === PAUSE TestStructs_ACLPolicies_resolveWithCache === RUN TestStructs_ACLPolicies_Compile === PAUSE TestStructs_ACLPolicies_Compile === RUN TestCheckDefinition_Defaults === PAUSE TestCheckDefinition_Defaults === RUN TestCheckDefinition_CheckType === PAUSE TestCheckDefinition_CheckType === RUN TestCheckDefinitionToCheckType === PAUSE TestCheckDefinitionToCheckType === RUN TestConfigEntries_ListRelatedServices_AndACLs === PAUSE TestConfigEntries_ListRelatedServices_AndACLs === RUN TestServiceResolverConfigEntry === PAUSE TestServiceResolverConfigEntry === RUN TestServiceSplitterConfigEntry === PAUSE TestServiceSplitterConfigEntry === RUN TestServiceRouterConfigEntry === PAUSE TestServiceRouterConfigEntry === RUN TestValidateServiceSubset === RUN TestValidateServiceSubset/a === RUN TestValidateServiceSubset/aa === RUN TestValidateServiceSubset/2a === RUN TestValidateServiceSubset/a2 === RUN TestValidateServiceSubset/a2a === RUN TestValidateServiceSubset/a22a === RUN TestValidateServiceSubset/1 === RUN TestValidateServiceSubset/11 === RUN TestValidateServiceSubset/10 === RUN TestValidateServiceSubset/01 === RUN TestValidateServiceSubset/a-a === RUN TestValidateServiceSubset/a--a === RUN TestValidateServiceSubset/a--a--a === RUN TestValidateServiceSubset/0-0 === RUN TestValidateServiceSubset/0--0 === RUN TestValidateServiceSubset/0--0--0 === RUN TestValidateServiceSubset/aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa === RUN TestValidateServiceSubset/A === RUN TestValidateServiceSubset/AA === RUN TestValidateServiceSubset/2A === RUN TestValidateServiceSubset/A2 === RUN TestValidateServiceSubset/A2A === RUN TestValidateServiceSubset/A22A === RUN TestValidateServiceSubset/A-A === RUN TestValidateServiceSubset/A--A === RUN TestValidateServiceSubset/A--A--A === RUN TestValidateServiceSubset/_ === RUN TestValidateServiceSubset/_a === RUN TestValidateServiceSubset/a_ === RUN TestValidateServiceSubset/a_a === RUN TestValidateServiceSubset/_#01 === RUN TestValidateServiceSubset/_a#01 === RUN TestValidateServiceSubset/a_#01 === RUN TestValidateServiceSubset/a_a#01 === RUN TestValidateServiceSubset/. === RUN TestValidateServiceSubset/.a === RUN TestValidateServiceSubset/a. === RUN TestValidateServiceSubset/a.a === RUN TestValidateServiceSubset/- === RUN TestValidateServiceSubset/-a === RUN TestValidateServiceSubset/a- === RUN TestValidateServiceSubset/aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa --- PASS: TestValidateServiceSubset (0.02s) --- PASS: TestValidateServiceSubset/a (0.01s) --- PASS: TestValidateServiceSubset/aa (0.00s) --- PASS: TestValidateServiceSubset/2a (0.00s) --- PASS: TestValidateServiceSubset/a2 (0.00s) --- PASS: TestValidateServiceSubset/a2a (0.00s) --- PASS: TestValidateServiceSubset/a22a (0.00s) --- PASS: TestValidateServiceSubset/1 (0.00s) --- PASS: TestValidateServiceSubset/11 (0.00s) --- PASS: TestValidateServiceSubset/10 (0.00s) --- PASS: TestValidateServiceSubset/01 (0.00s) --- PASS: TestValidateServiceSubset/a-a (0.00s) --- PASS: TestValidateServiceSubset/a--a (0.00s) --- PASS: TestValidateServiceSubset/a--a--a (0.00s) --- PASS: TestValidateServiceSubset/0-0 (0.00s) --- PASS: TestValidateServiceSubset/0--0 (0.00s) --- PASS: TestValidateServiceSubset/0--0--0 (0.00s) --- PASS: TestValidateServiceSubset/aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa (0.00s) --- PASS: TestValidateServiceSubset/A (0.00s) --- PASS: TestValidateServiceSubset/AA (0.00s) --- PASS: TestValidateServiceSubset/2A (0.00s) --- PASS: TestValidateServiceSubset/A2 (0.00s) --- PASS: TestValidateServiceSubset/A2A (0.00s) --- PASS: TestValidateServiceSubset/A22A (0.00s) --- PASS: TestValidateServiceSubset/A-A (0.00s) --- PASS: TestValidateServiceSubset/A--A (0.00s) --- PASS: TestValidateServiceSubset/A--A--A (0.00s) --- PASS: TestValidateServiceSubset/_ (0.00s) --- PASS: TestValidateServiceSubset/_a (0.00s) --- PASS: TestValidateServiceSubset/a_ (0.00s) --- PASS: TestValidateServiceSubset/a_a (0.00s) --- PASS: TestValidateServiceSubset/_#01 (0.00s) --- PASS: TestValidateServiceSubset/_a#01 (0.00s) --- PASS: TestValidateServiceSubset/a_#01 (0.00s) --- PASS: TestValidateServiceSubset/a_a#01 (0.00s) --- PASS: TestValidateServiceSubset/. (0.00s) --- PASS: TestValidateServiceSubset/.a (0.00s) --- PASS: TestValidateServiceSubset/a. (0.00s) --- PASS: TestValidateServiceSubset/a.a (0.00s) --- PASS: TestValidateServiceSubset/- (0.00s) --- PASS: TestValidateServiceSubset/-a (0.00s) --- PASS: TestValidateServiceSubset/a- (0.00s) --- PASS: TestValidateServiceSubset/aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa (0.00s) === RUN TestDecodeConfigEntry === PAUSE TestDecodeConfigEntry === RUN TestServiceConfigResponse_MsgPack --- PASS: TestServiceConfigResponse_MsgPack (0.00s) === RUN TestConfigEntryResponseMarshalling === PAUSE TestConfigEntryResponseMarshalling === RUN TestCAConfiguration_GetCommonConfig === RUN TestCAConfiguration_GetCommonConfig/basic_defaults === RUN TestCAConfiguration_GetCommonConfig/basic_defaults_after_encoding_fun --- PASS: TestCAConfiguration_GetCommonConfig (0.00s) --- PASS: TestCAConfiguration_GetCommonConfig/basic_defaults (0.00s) --- PASS: TestCAConfiguration_GetCommonConfig/basic_defaults_after_encoding_fun (0.00s) === RUN TestCAProviderConfig_Validate === RUN TestCAProviderConfig_Validate/defaults === RUN TestCAProviderConfig_Validate/intermediate_cert_ttl_too_short === RUN TestCAProviderConfig_Validate/intermediate_cert_ttl_too_short#01 === RUN TestCAProviderConfig_Validate/good_intermediate_and_leaf_cert_TTL --- PASS: TestCAProviderConfig_Validate (0.00s) --- PASS: TestCAProviderConfig_Validate/defaults (0.00s) --- PASS: TestCAProviderConfig_Validate/intermediate_cert_ttl_too_short (0.00s) --- PASS: TestCAProviderConfig_Validate/intermediate_cert_ttl_too_short#01 (0.00s) --- PASS: TestCAProviderConfig_Validate/good_intermediate_and_leaf_cert_TTL (0.00s) === RUN TestConnectProxyConfig_ToAPI === RUN TestConnectProxyConfig_ToAPI/service --- PASS: TestConnectProxyConfig_ToAPI (0.00s) --- PASS: TestConnectProxyConfig_ToAPI/service (0.00s) === RUN TestUpstream_MarshalJSON === RUN TestUpstream_MarshalJSON/service === RUN TestUpstream_MarshalJSON/pq --- PASS: TestUpstream_MarshalJSON (0.00s) --- PASS: TestUpstream_MarshalJSON/service (0.00s) --- PASS: TestUpstream_MarshalJSON/pq (0.00s) === RUN TestUpstream_UnmarshalJSON === RUN TestUpstream_UnmarshalJSON/service === RUN TestUpstream_UnmarshalJSON/pq --- PASS: TestUpstream_UnmarshalJSON (0.00s) --- PASS: TestUpstream_UnmarshalJSON/service (0.00s) --- PASS: TestUpstream_UnmarshalJSON/pq (0.00s) === RUN TestMeshGatewayConfig_OverlayWith === RUN TestMeshGatewayConfig_OverlayWith/_overlaid_with_ === RUN TestMeshGatewayConfig_OverlayWith/_overlaid_with_none === RUN TestMeshGatewayConfig_OverlayWith/_overlaid_with_remote === RUN TestMeshGatewayConfig_OverlayWith/_overlaid_with_local === RUN TestMeshGatewayConfig_OverlayWith/none_overlaid_with_ === RUN TestMeshGatewayConfig_OverlayWith/none_overlaid_with_none === RUN TestMeshGatewayConfig_OverlayWith/none_overlaid_with_remote === RUN TestMeshGatewayConfig_OverlayWith/none_overlaid_with_local === RUN TestMeshGatewayConfig_OverlayWith/remote_overlaid_with_ === RUN TestMeshGatewayConfig_OverlayWith/remote_overlaid_with_none === RUN TestMeshGatewayConfig_OverlayWith/remote_overlaid_with_remote === RUN TestMeshGatewayConfig_OverlayWith/remote_overlaid_with_local === RUN TestMeshGatewayConfig_OverlayWith/local_overlaid_with_ === RUN TestMeshGatewayConfig_OverlayWith/local_overlaid_with_none === RUN TestMeshGatewayConfig_OverlayWith/local_overlaid_with_remote === RUN TestMeshGatewayConfig_OverlayWith/local_overlaid_with_local --- PASS: TestMeshGatewayConfig_OverlayWith (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/_overlaid_with_ (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/_overlaid_with_none (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/_overlaid_with_remote (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/_overlaid_with_local (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/none_overlaid_with_ (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/none_overlaid_with_none (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/none_overlaid_with_remote (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/none_overlaid_with_local (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/remote_overlaid_with_ (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/remote_overlaid_with_none (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/remote_overlaid_with_remote (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/remote_overlaid_with_local (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/local_overlaid_with_ (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/local_overlaid_with_none (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/local_overlaid_with_remote (0.00s) --- PASS: TestMeshGatewayConfig_OverlayWith/local_overlaid_with_local (0.00s) === RUN TestValidateMeshGatewayMode === RUN TestValidateMeshGatewayMode/none_(constant) === RUN TestValidateMeshGatewayMode/none_(explicit) === RUN TestValidateMeshGatewayMode/_(constant) === RUN TestValidateMeshGatewayMode/_(explicit) === RUN TestValidateMeshGatewayMode/local_(constant) === RUN TestValidateMeshGatewayMode/local_(explicit) === RUN TestValidateMeshGatewayMode/remote_(constant) === RUN TestValidateMeshGatewayMode/remote_(explicit) --- PASS: TestValidateMeshGatewayMode (0.00s) --- PASS: TestValidateMeshGatewayMode/none_(constant) (0.00s) --- PASS: TestValidateMeshGatewayMode/none_(explicit) (0.00s) --- PASS: TestValidateMeshGatewayMode/_(constant) (0.00s) --- PASS: TestValidateMeshGatewayMode/_(explicit) (0.00s) --- PASS: TestValidateMeshGatewayMode/local_(constant) (0.00s) --- PASS: TestValidateMeshGatewayMode/local_(explicit) (0.00s) --- PASS: TestValidateMeshGatewayMode/remote_(constant) (0.00s) --- PASS: TestValidateMeshGatewayMode/remote_(explicit) (0.00s) === RUN TestIntention_ACLs === PAUSE TestIntention_ACLs === RUN TestIntentionValidate === RUN TestIntentionValidate/long_description === RUN TestIntentionValidate/no_action_set === RUN TestIntentionValidate/no_SourceNS === RUN TestIntentionValidate/no_SourceName === RUN TestIntentionValidate/no_DestinationNS === RUN TestIntentionValidate/no_DestinationName === RUN TestIntentionValidate/SourceNS_partial_wildcard === RUN TestIntentionValidate/SourceName_partial_wildcard === RUN TestIntentionValidate/SourceName_exact_following_wildcard === RUN TestIntentionValidate/DestinationNS_partial_wildcard === RUN TestIntentionValidate/DestinationName_partial_wildcard === RUN TestIntentionValidate/DestinationName_exact_following_wildcard === RUN TestIntentionValidate/SourceType_is_not_set === RUN TestIntentionValidate/SourceType_is_other --- PASS: TestIntentionValidate (0.00s) --- PASS: TestIntentionValidate/long_description (0.00s) --- PASS: TestIntentionValidate/no_action_set (0.00s) --- PASS: TestIntentionValidate/no_SourceNS (0.00s) --- PASS: TestIntentionValidate/no_SourceName (0.00s) --- PASS: TestIntentionValidate/no_DestinationNS (0.00s) --- PASS: TestIntentionValidate/no_DestinationName (0.00s) --- PASS: TestIntentionValidate/SourceNS_partial_wildcard (0.00s) --- PASS: TestIntentionValidate/SourceName_partial_wildcard (0.00s) --- PASS: TestIntentionValidate/SourceName_exact_following_wildcard (0.00s) --- PASS: TestIntentionValidate/DestinationNS_partial_wildcard (0.00s) --- PASS: TestIntentionValidate/DestinationName_partial_wildcard (0.00s) --- PASS: TestIntentionValidate/DestinationName_exact_following_wildcard (0.00s) --- PASS: TestIntentionValidate/SourceType_is_not_set (0.00s) --- PASS: TestIntentionValidate/SourceType_is_other (0.00s) === RUN TestIntentionPrecedenceSorter === RUN TestIntentionPrecedenceSorter/exhaustive_list === RUN TestIntentionPrecedenceSorter/tiebreak_deterministically --- PASS: TestIntentionPrecedenceSorter (0.00s) --- PASS: TestIntentionPrecedenceSorter/exhaustive_list (0.00s) --- PASS: TestIntentionPrecedenceSorter/tiebreak_deterministically (0.00s) === RUN TestStructs_PreparedQuery_GetACLPrefix --- PASS: TestStructs_PreparedQuery_GetACLPrefix (0.00s) === RUN TestAgentStructs_CheckTypes === PAUSE TestAgentStructs_CheckTypes === RUN TestServiceDefinitionValidate === RUN TestServiceDefinitionValidate/valid --- PASS: TestServiceDefinitionValidate (0.00s) --- PASS: TestServiceDefinitionValidate/valid (0.00s) === RUN TestStructs_FilterFieldConfigurations === PAUSE TestStructs_FilterFieldConfigurations === RUN TestEncodeDecode --- PASS: TestEncodeDecode (0.00s) === RUN TestStructs_Implements --- PASS: TestStructs_Implements (0.00s) === RUN TestStructs_RegisterRequest_ChangesNode --- PASS: TestStructs_RegisterRequest_ChangesNode (0.00s) === RUN TestNode_IsSame --- PASS: TestNode_IsSame (0.00s) === RUN TestStructs_ServiceNode_IsSameService --- PASS: TestStructs_ServiceNode_IsSameService (0.00s) === RUN TestStructs_ServiceNode_PartialClone --- PASS: TestStructs_ServiceNode_PartialClone (0.00s) === RUN TestStructs_ServiceNode_Conversions --- PASS: TestStructs_ServiceNode_Conversions (0.00s) === RUN TestStructs_NodeService_ValidateMeshGateway === RUN TestStructs_NodeService_ValidateMeshGateway/proxy-local-address === RUN TestStructs_NodeService_ValidateMeshGateway/proxy-local-port === RUN TestStructs_NodeService_ValidateMeshGateway/proxy-upstreams === RUN TestStructs_NodeService_ValidateMeshGateway/valid === RUN TestStructs_NodeService_ValidateMeshGateway/zero-port === RUN TestStructs_NodeService_ValidateMeshGateway/sidecar-service === RUN TestStructs_NodeService_ValidateMeshGateway/proxy-destination-name === RUN TestStructs_NodeService_ValidateMeshGateway/proxy-destination-id --- PASS: TestStructs_NodeService_ValidateMeshGateway (0.00s) --- PASS: TestStructs_NodeService_ValidateMeshGateway/proxy-local-address (0.00s) --- PASS: TestStructs_NodeService_ValidateMeshGateway/proxy-local-port (0.00s) --- PASS: TestStructs_NodeService_ValidateMeshGateway/proxy-upstreams (0.00s) --- PASS: TestStructs_NodeService_ValidateMeshGateway/valid (0.00s) --- PASS: TestStructs_NodeService_ValidateMeshGateway/zero-port (0.00s) --- PASS: TestStructs_NodeService_ValidateMeshGateway/sidecar-service (0.00s) --- PASS: TestStructs_NodeService_ValidateMeshGateway/proxy-destination-name (0.00s) --- PASS: TestStructs_NodeService_ValidateMeshGateway/proxy-destination-id (0.00s) === RUN TestStructs_NodeService_ValidateExposeConfig === RUN TestStructs_NodeService_ValidateExposeConfig/protocol_not_supported === RUN TestStructs_NodeService_ValidateExposeConfig/valid === RUN TestStructs_NodeService_ValidateExposeConfig/empty_path === RUN TestStructs_NodeService_ValidateExposeConfig/invalid_port_negative === RUN TestStructs_NodeService_ValidateExposeConfig/invalid_port_too_large === RUN TestStructs_NodeService_ValidateExposeConfig/duplicate_paths === RUN TestStructs_NodeService_ValidateExposeConfig/duplicate_ports --- PASS: TestStructs_NodeService_ValidateExposeConfig (0.00s) --- PASS: TestStructs_NodeService_ValidateExposeConfig/protocol_not_supported (0.00s) --- PASS: TestStructs_NodeService_ValidateExposeConfig/valid (0.00s) --- PASS: TestStructs_NodeService_ValidateExposeConfig/empty_path (0.00s) --- PASS: TestStructs_NodeService_ValidateExposeConfig/invalid_port_negative (0.00s) --- PASS: TestStructs_NodeService_ValidateExposeConfig/invalid_port_too_large (0.00s) --- PASS: TestStructs_NodeService_ValidateExposeConfig/duplicate_paths (0.00s) --- PASS: TestStructs_NodeService_ValidateExposeConfig/duplicate_ports (0.00s) === RUN TestStructs_NodeService_ValidateConnectProxy === RUN TestStructs_NodeService_ValidateConnectProxy/valid === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_no_Proxy.DestinationServiceName === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_whitespace_Proxy.DestinationServiceName === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_valid_Proxy.DestinationServiceName === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_no_port_set === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_ConnectNative_set === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_upstream_missing_type_(defaulted) === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_upstream_invalid_type === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_upstream_empty_name === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_upstream_empty_bind_port === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_almost-but-not-quite-duplicated_in_various_ways === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_port === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_ip_and_port === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_ip_and_port_with_ip_defaulted_in_one === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_name === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_name_and_datacenter === RUN TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_name_and_namespace --- PASS: TestStructs_NodeService_ValidateConnectProxy (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/valid (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_no_Proxy.DestinationServiceName (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_whitespace_Proxy.DestinationServiceName (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_valid_Proxy.DestinationServiceName (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_no_port_set (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_ConnectNative_set (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_upstream_missing_type_(defaulted) (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_upstream_invalid_type (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_upstream_empty_name (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_upstream_empty_bind_port (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_almost-but-not-quite-duplicated_in_various_ways (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_port (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_ip_and_port (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_ip_and_port_with_ip_defaulted_in_one (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_name (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_name_and_datacenter (0.00s) --- PASS: TestStructs_NodeService_ValidateConnectProxy/connect-proxy:_Upstreams_duplicated_by_name_and_namespace (0.00s) === RUN TestStructs_NodeService_ValidateSidecarService === RUN TestStructs_NodeService_ValidateSidecarService/valid === RUN TestStructs_NodeService_ValidateSidecarService/ID_can't_be_set === RUN TestStructs_NodeService_ValidateSidecarService/Nested_sidecar_can't_be_set --- PASS: TestStructs_NodeService_ValidateSidecarService (0.00s) --- PASS: TestStructs_NodeService_ValidateSidecarService/valid (0.00s) --- PASS: TestStructs_NodeService_ValidateSidecarService/ID_can't_be_set (0.00s) --- PASS: TestStructs_NodeService_ValidateSidecarService/Nested_sidecar_can't_be_set (0.00s) === RUN TestStructs_NodeService_IsSame --- PASS: TestStructs_NodeService_IsSame (0.00s) === RUN TestStructs_HealthCheck_IsSame --- PASS: TestStructs_HealthCheck_IsSame (0.00s) === RUN TestStructs_HealthCheck_Marshalling --- PASS: TestStructs_HealthCheck_Marshalling (0.00s) === RUN TestStructs_HealthCheck_Clone --- PASS: TestStructs_HealthCheck_Clone (0.00s) === RUN TestStructs_CheckServiceNodes_Shuffle --- PASS: TestStructs_CheckServiceNodes_Shuffle (0.00s) === RUN TestStructs_CheckServiceNodes_Filter --- PASS: TestStructs_CheckServiceNodes_Filter (0.00s) === RUN TestStructs_DirEntry_Clone --- PASS: TestStructs_DirEntry_Clone (0.00s) === RUN TestStructs_ValidateMetadata --- PASS: TestStructs_ValidateMetadata (0.00s) === RUN TestStructs_validateMetaPair --- PASS: TestStructs_validateMetaPair (0.00s) === RUN TestSpecificServiceRequest_CacheInfo === RUN TestSpecificServiceRequest_CacheInfo/basic_params === RUN TestSpecificServiceRequest_CacheInfo/name_should_be_considered === RUN TestSpecificServiceRequest_CacheInfo/node_meta_should_be_considered === RUN TestSpecificServiceRequest_CacheInfo/address_should_be_considered === RUN TestSpecificServiceRequest_CacheInfo/tag_filter_should_be_considered === RUN TestSpecificServiceRequest_CacheInfo/connect_should_be_considered === RUN TestSpecificServiceRequest_CacheInfo/tags_should_be_different === RUN TestSpecificServiceRequest_CacheInfo/tags_should_not_depend_on_order === RUN TestSpecificServiceRequest_CacheInfo/legacy_requests_with_singular_tag_should_be_different --- PASS: TestSpecificServiceRequest_CacheInfo (0.00s) --- PASS: TestSpecificServiceRequest_CacheInfo/basic_params (0.00s) --- PASS: TestSpecificServiceRequest_CacheInfo/name_should_be_considered (0.00s) --- PASS: TestSpecificServiceRequest_CacheInfo/node_meta_should_be_considered (0.00s) --- PASS: TestSpecificServiceRequest_CacheInfo/address_should_be_considered (0.00s) --- PASS: TestSpecificServiceRequest_CacheInfo/tag_filter_should_be_considered (0.00s) --- PASS: TestSpecificServiceRequest_CacheInfo/connect_should_be_considered (0.00s) --- PASS: TestSpecificServiceRequest_CacheInfo/tags_should_be_different (0.00s) --- PASS: TestSpecificServiceRequest_CacheInfo/tags_should_not_depend_on_order (0.00s) --- PASS: TestSpecificServiceRequest_CacheInfo/legacy_requests_with_singular_tag_should_be_different (0.00s) === RUN TestNodeService_JSON_OmitTaggedAdddresses === PAUSE TestNodeService_JSON_OmitTaggedAdddresses === RUN TestServiceNode_JSON_OmitServiceTaggedAdddresses === PAUSE TestServiceNode_JSON_OmitServiceTaggedAdddresses === RUN TestNode_BestAddress === PAUSE TestNode_BestAddress === RUN TestNodeService_BestAddress === PAUSE TestNodeService_BestAddress === RUN TestCheckServiceNode_BestAddress === PAUSE TestCheckServiceNode_BestAddress === RUN TestNodeService_JSON_Marshal --- PASS: TestNodeService_JSON_Marshal (0.00s) === RUN TestServiceNode_JSON_Marshal --- PASS: TestServiceNode_JSON_Marshal (0.00s) === RUN TestStructs_MsgpackEncodeDecode_Monolith === RUN TestStructs_MsgpackEncodeDecode_Monolith/monster === RUN TestStructs_MsgpackEncodeDecode_Monolith/monster/old_encoder_and_old_decoder === RUN TestStructs_MsgpackEncodeDecode_Monolith/monster/old_encoder_and_new_decoder === RUN TestStructs_MsgpackEncodeDecode_Monolith/monster/new_encoder_and_old_decoder === RUN TestStructs_MsgpackEncodeDecode_Monolith/monster/new_encoder_and_new_decoder === RUN TestStructs_MsgpackEncodeDecode_Monolith/frankenstein === RUN TestStructs_MsgpackEncodeDecode_Monolith/frankenstein/old_encoder_and_old_decoder === RUN TestStructs_MsgpackEncodeDecode_Monolith/frankenstein/old_encoder_and_new_decoder === RUN TestStructs_MsgpackEncodeDecode_Monolith/frankenstein/new_encoder_and_old_decoder === RUN TestStructs_MsgpackEncodeDecode_Monolith/frankenstein/new_encoder_and_new_decoder --- PASS: TestStructs_MsgpackEncodeDecode_Monolith (0.00s) --- PASS: TestStructs_MsgpackEncodeDecode_Monolith/monster (0.00s) --- PASS: TestStructs_MsgpackEncodeDecode_Monolith/monster/old_encoder_and_old_decoder (0.00s) --- PASS: TestStructs_MsgpackEncodeDecode_Monolith/monster/old_encoder_and_new_decoder (0.00s) --- PASS: TestStructs_MsgpackEncodeDecode_Monolith/monster/new_encoder_and_old_decoder (0.00s) --- PASS: TestStructs_MsgpackEncodeDecode_Monolith/monster/new_encoder_and_new_decoder (0.00s) --- PASS: TestStructs_MsgpackEncodeDecode_Monolith/frankenstein (0.00s) --- PASS: TestStructs_MsgpackEncodeDecode_Monolith/frankenstein/old_encoder_and_old_decoder (0.00s) --- PASS: TestStructs_MsgpackEncodeDecode_Monolith/frankenstein/old_encoder_and_new_decoder (0.00s) --- PASS: TestStructs_MsgpackEncodeDecode_Monolith/frankenstein/new_encoder_and_old_decoder (0.00s) --- PASS: TestStructs_MsgpackEncodeDecode_Monolith/frankenstein/new_encoder_and_new_decoder (0.00s) === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode/request === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode/request/old_encoder_and_old_decoder === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode/request/old_encoder_and_new_decoder === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode/request/new_encoder_and_old_decoder === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode/request/new_encoder_and_new_decoder === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode/response === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode/response/old_encoder_and_old_decoder === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode/response/old_encoder_and_new_decoder === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode/response/new_encoder_and_old_decoder === RUN TestSnapshotRequestResponse_MsgpackEncodeDecode/response/new_encoder_and_new_decoder --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode (0.00s) --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode/request (0.00s) --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode/request/old_encoder_and_old_decoder (0.00s) --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode/request/old_encoder_and_new_decoder (0.00s) --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode/request/new_encoder_and_old_decoder (0.00s) --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode/request/new_encoder_and_new_decoder (0.00s) --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode/response (0.00s) --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode/response/old_encoder_and_old_decoder (0.00s) --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode/response/old_encoder_and_new_decoder (0.00s) --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode/response/new_encoder_and_old_decoder (0.00s) --- PASS: TestSnapshotRequestResponse_MsgpackEncodeDecode/response/new_encoder_and_new_decoder (0.00s) === CONT TestStructs_ACLCaches === CONT TestCheckDefinition_Defaults --- PASS: TestCheckDefinition_Defaults (0.00s) === CONT TestCheckServiceNode_BestAddress === RUN TestStructs_ACLCaches/New === PAUSE TestStructs_ACLCaches/New === RUN TestStructs_ACLCaches/Identities === RUN TestCheckServiceNode_BestAddress/service-wan-address-node-lan === PAUSE TestStructs_ACLCaches/Identities === CONT TestStructs_ACLTokens_Sort === PAUSE TestCheckServiceNode_BestAddress/service-wan-address-node-lan === RUN TestCheckServiceNode_BestAddress/node-address === PAUSE TestCheckServiceNode_BestAddress/node-address === RUN TestCheckServiceNode_BestAddress/node-wan-address === PAUSE TestCheckServiceNode_BestAddress/node-wan-address === RUN TestCheckServiceNode_BestAddress/service-address --- PASS: TestStructs_ACLTokens_Sort (0.00s) === CONT TestStructs_ACLToken_EstimateSize === PAUSE TestCheckServiceNode_BestAddress/service-address === RUN TestCheckServiceNode_BestAddress/service-wan-address --- PASS: TestStructs_ACLToken_EstimateSize (0.00s) === CONT TestStructs_ACLToken_SetHash === PAUSE TestCheckServiceNode_BestAddress/service-wan-address === RUN TestStructs_ACLToken_SetHash/Nil_Hash_-_Generate === RUN TestCheckServiceNode_BestAddress/service-wan-address-default-port === PAUSE TestCheckServiceNode_BestAddress/service-wan-address-default-port === CONT TestStructs_ACLToken_Stub === RUN TestStructs_ACLToken_Stub/Basic === PAUSE TestStructs_ACLToken_Stub/Basic === RUN TestStructs_ACLToken_Stub/Legacy === PAUSE TestStructs_ACLToken_Stub/Legacy === CONT TestStructs_ACLServiceIdentity_SyntheticPolicy === RUN TestStructs_ACLToken_SetHash/Hash_Set_-_Dont_Generate === RUN TestStructs_ACLServiceIdentity_SyntheticPolicy/web === RUN TestStructs_ACLToken_SetHash/Hash_Set_-_Generate === RUN TestStructs_ACLServiceIdentity_SyntheticPolicy/companion-cube-99_[dc1,_dc2] --- PASS: TestStructs_ACLToken_SetHash (0.00s) --- PASS: TestStructs_ACLToken_SetHash/Nil_Hash_-_Generate (0.00s) --- PASS: TestStructs_ACLToken_SetHash/Hash_Set_-_Dont_Generate (0.00s) --- PASS: TestStructs_ACLToken_SetHash/Hash_Set_-_Generate (0.00s) === CONT TestStructs_ACLToken_PolicyIDs === RUN TestStructs_ACLToken_PolicyIDs/Basic === PAUSE TestStructs_ACLToken_PolicyIDs/Basic === CONT TestStructs_ACLToken_EmbeddedPolicy === RUN TestStructs_ACLToken_PolicyIDs/Legacy_Management === PAUSE TestStructs_ACLToken_PolicyIDs/Legacy_Management === RUN TestStructs_ACLToken_EmbeddedPolicy/No_Rules === RUN TestStructs_ACLToken_PolicyIDs/Legacy_Management_With_Rules === PAUSE TestStructs_ACLToken_EmbeddedPolicy/No_Rules === RUN TestStructs_ACLToken_EmbeddedPolicy/Legacy_Client === PAUSE TestStructs_ACLToken_PolicyIDs/Legacy_Management_With_Rules === PAUSE TestStructs_ACLToken_EmbeddedPolicy/Legacy_Client === RUN TestStructs_ACLToken_EmbeddedPolicy/Same_Policy_for_Tokens_with_same_Rules === PAUSE TestStructs_ACLToken_EmbeddedPolicy/Same_Policy_for_Tokens_with_same_Rules --- PASS: TestStructs_ACLServiceIdentity_SyntheticPolicy (0.00s) --- PASS: TestStructs_ACLServiceIdentity_SyntheticPolicy/web (0.00s) --- PASS: TestStructs_ACLServiceIdentity_SyntheticPolicy/companion-cube-99_[dc1,_dc2] (0.00s) === RUN TestStructs_ACLToken_PolicyIDs/No_Policies === PAUSE TestStructs_ACLToken_PolicyIDs/No_Policies === CONT TestStructs_ACL_Convert --- PASS: TestStructs_ACL_Convert (0.00s) === CONT TestNodeService_BestAddress === RUN TestNodeService_BestAddress/service-address === PAUSE TestNodeService_BestAddress/service-address === RUN TestNodeService_BestAddress/service-wan-address === PAUSE TestNodeService_BestAddress/service-wan-address === RUN TestNodeService_BestAddress/service-wan-address-default-port === PAUSE TestNodeService_BestAddress/service-wan-address-default-port === RUN TestNodeService_BestAddress/service-wan-address-node-lan === PAUSE TestNodeService_BestAddress/service-wan-address-node-lan === RUN TestNodeService_BestAddress/no-address === PAUSE TestNodeService_BestAddress/no-address === RUN TestStructs_ACLCaches/Policies === PAUSE TestStructs_ACLCaches/Policies === RUN TestStructs_ACLCaches/ParsedPolicies === PAUSE TestStructs_ACLCaches/ParsedPolicies === RUN TestStructs_ACLCaches/Authorizers === PAUSE TestStructs_ACLCaches/Authorizers === RUN TestStructs_ACLCaches/Roles === PAUSE TestStructs_ACLCaches/Roles === CONT TestConfigEntryResponseMarshalling === RUN TestConfigEntryResponseMarshalling/service-default_entry === PAUSE TestConfigEntryResponseMarshalling/service-default_entry === RUN TestConfigEntryResponseMarshalling/nil_entry === PAUSE TestConfigEntryResponseMarshalling/nil_entry === RUN TestConfigEntryResponseMarshalling/proxy-default_entry === PAUSE TestConfigEntryResponseMarshalling/proxy-default_entry === CONT TestNode_BestAddress === RUN TestNode_BestAddress/address === PAUSE TestNode_BestAddress/address === RUN TestNode_BestAddress/wan-address === PAUSE TestNode_BestAddress/wan-address === CONT TestStructs_ACLToken_Convert === RUN TestStructs_ACLToken_Convert/Management === PAUSE TestStructs_ACLToken_Convert/Management === RUN TestStructs_ACLToken_Convert/Client === PAUSE TestStructs_ACLToken_Convert/Client === RUN TestStructs_ACLToken_Convert/Unconvertible === PAUSE TestStructs_ACLToken_Convert/Unconvertible === CONT TestServiceNode_JSON_OmitServiceTaggedAdddresses === RUN TestServiceNode_JSON_OmitServiceTaggedAdddresses/nil === PAUSE TestServiceNode_JSON_OmitServiceTaggedAdddresses/nil === RUN TestServiceNode_JSON_OmitServiceTaggedAdddresses/empty === PAUSE TestServiceNode_JSON_OmitServiceTaggedAdddresses/empty === CONT TestNodeService_JSON_OmitTaggedAdddresses === RUN TestNodeService_JSON_OmitTaggedAdddresses/nil === PAUSE TestNodeService_JSON_OmitTaggedAdddresses/nil === RUN TestNodeService_JSON_OmitTaggedAdddresses/empty === PAUSE TestNodeService_JSON_OmitTaggedAdddresses/empty === CONT TestStructs_FilterFieldConfigurations === RUN TestStructs_FilterFieldConfigurations/Node === RUN TestStructs_FilterFieldConfigurations/NodeService === CONT TestIntention_ACLs === RUN TestIntention_ACLs/deny-write-with-source-write === CONT TestAgentStructs_CheckTypes --- PASS: TestAgentStructs_CheckTypes (0.00s) === CONT TestStructs_ACLPolicies_Compile === CONT TestStructs_ACLPolicies_Sort === RUN TestStructs_ACLPolicies_Compile/Cache_Miss --- PASS: TestStructs_ACLPolicies_Sort (0.00s) === CONT TestStructs_ACLPolicies_resolveWithCache === RUN TestStructs_ACLPolicies_resolveWithCache/Cache_Misses === RUN TestStructs_ACLPolicies_Compile/Check_Cache === RUN TestStructs_ACLPolicies_Compile/Cache_Hit === RUN TestIntention_ACLs/deny-wildcard-write-allow-read --- PASS: TestStructs_ACLPolicies_Compile (0.00s) --- PASS: TestStructs_ACLPolicies_Compile/Cache_Miss (0.00s) --- PASS: TestStructs_ACLPolicies_Compile/Check_Cache (0.00s) --- PASS: TestStructs_ACLPolicies_Compile/Cache_Hit (0.00s) === CONT TestStructs_ACLPolicyListStubs_Sort --- PASS: TestStructs_ACLPolicyListStubs_Sort (0.00s) === CONT TestStructs_ACLPolicy_SetHash === RUN TestStructs_ACLPolicy_SetHash/Nil_Hash_-_Generate === RUN TestStructs_ACLPolicies_resolveWithCache/Check_Cache === RUN TestStructs_ACLPolicy_SetHash/Hash_Set_-_Dont_Generate === RUN TestIntention_ACLs/allow-wildcard-write === RUN TestStructs_ACLPolicy_SetHash/Hash_Set_-_Generate --- PASS: TestStructs_ACLPolicy_SetHash (0.00s) --- PASS: TestStructs_ACLPolicy_SetHash/Nil_Hash_-_Generate (0.00s) --- PASS: TestStructs_ACLPolicy_SetHash/Hash_Set_-_Dont_Generate (0.00s) --- PASS: TestStructs_ACLPolicy_SetHash/Hash_Set_-_Generate (0.00s) === CONT TestStructs_ACLPolicy_EstimateSize --- PASS: TestStructs_ACLPolicy_EstimateSize (0.00s) === CONT TestServiceResolverConfigEntry === RUN TestServiceResolverConfigEntry/nil === RUN TestServiceResolverConfigEntry/no_name === RUN TestIntention_ACLs/allow-wildcard-read === RUN TestServiceResolverConfigEntry/empty === RUN TestStructs_ACLPolicies_resolveWithCache/Cache_Hits === RUN TestServiceResolverConfigEntry/empty_subset_name === RUN TestServiceResolverConfigEntry/default_subset_does_not_exist === RUN TestIntention_ACLs/all-denied --- PASS: TestStructs_ACLPolicies_resolveWithCache (0.00s) --- PASS: TestStructs_ACLPolicies_resolveWithCache/Cache_Misses (0.00s) --- PASS: TestStructs_ACLPolicies_resolveWithCache/Check_Cache (0.00s) --- PASS: TestStructs_ACLPolicies_resolveWithCache/Cache_Hits (0.00s) === RUN TestServiceResolverConfigEntry/default_subset_does_exist === CONT TestDecodeConfigEntry === RUN TestIntention_ACLs/deny-write-read-dest === RUN TestDecodeConfigEntry/proxy-defaults:_extra_fields_or_typo_(snake_case) === RUN TestServiceResolverConfigEntry/empty_redirect === RUN TestDecodeConfigEntry/proxy-defaults:_extra_fields_or_typo_(camel_case) === RUN TestIntention_ACLs/deny-write-read-source === RUN TestServiceResolverConfigEntry/redirect_subset_with_no_service === RUN TestDecodeConfigEntry/proxy-defaults_(snake_case) === RUN TestServiceResolverConfigEntry/redirect_namespace_with_no_service === RUN TestIntention_ACLs/allow-write-with-dest-write === RUN TestDecodeConfigEntry/proxy-defaults_(camel_case) === RUN TestServiceResolverConfigEntry/self_redirect_with_invalid_subset --- PASS: TestIntention_ACLs (0.01s) --- PASS: TestIntention_ACLs/deny-write-with-source-write (0.00s) --- PASS: TestIntention_ACLs/deny-wildcard-write-allow-read (0.00s) --- PASS: TestIntention_ACLs/allow-wildcard-write (0.00s) --- PASS: TestIntention_ACLs/allow-wildcard-read (0.00s) --- PASS: TestIntention_ACLs/all-denied (0.00s) --- PASS: TestIntention_ACLs/deny-write-read-dest (0.00s) --- PASS: TestIntention_ACLs/deny-write-read-source (0.00s) --- PASS: TestIntention_ACLs/allow-write-with-dest-write (0.00s) === RUN TestServiceResolverConfigEntry/self_redirect_with_valid_subset === RUN TestServiceResolverConfigEntry/simple_wildcard_failover === RUN TestServiceResolverConfigEntry/failover_for_missing_subset === RUN TestServiceResolverConfigEntry/failover_for_present_subset === RUN TestServiceResolverConfigEntry/failover_empty === RUN TestServiceResolverConfigEntry/failover_to_self_using_invalid_subset === RUN TestServiceResolverConfigEntry/failover_to_self_using_valid_subset === RUN TestServiceResolverConfigEntry/failover_with_empty_datacenters_in_list === RUN TestServiceResolverConfigEntry/bad_connect_timeout === RUN TestServiceResolverConfigEntry/invalid_subset_name:_A === RUN TestServiceResolverConfigEntry/invalid_subset_name:_AA === RUN TestServiceResolverConfigEntry/invalid_subset_name:_2A === RUN TestServiceResolverConfigEntry/invalid_subset_name:_A2 === RUN TestServiceResolverConfigEntry/invalid_subset_name:_A2A === RUN TestServiceResolverConfigEntry/invalid_subset_name:_A22A === RUN TestServiceResolverConfigEntry/invalid_subset_name:_A-A === RUN TestServiceResolverConfigEntry/invalid_subset_name:_A--A === RUN TestServiceResolverConfigEntry/invalid_subset_name:_A--A--A === RUN TestServiceResolverConfigEntry/invalid_subset_name:__ === RUN TestServiceResolverConfigEntry/invalid_subset_name:__a === RUN TestServiceResolverConfigEntry/invalid_subset_name:_a_ === RUN TestServiceResolverConfigEntry/invalid_subset_name:_a_a === RUN TestServiceResolverConfigEntry/invalid_subset_name:__#01 === RUN TestServiceResolverConfigEntry/invalid_subset_name:__a#01 === RUN TestServiceResolverConfigEntry/invalid_subset_name:_a_#01 === RUN TestServiceResolverConfigEntry/invalid_subset_name:_a_a#01 === RUN TestServiceResolverConfigEntry/invalid_subset_name:_. === CONT TestServiceRouterConfigEntry === RUN TestServiceRouterConfigEntry/nil === RUN TestDecodeConfigEntry/service-defaults_(snake_case) === RUN TestServiceRouterConfigEntry/no_name === RUN TestDecodeConfigEntry/service-defaults_(camel_case) === RUN TestServiceRouterConfigEntry/empty === RUN TestDecodeConfigEntry/service-router:_kitchen_sink_(snake_case) === RUN TestServiceRouterConfigEntry/1_empty_route === RUN TestServiceResolverConfigEntry/invalid_subset_name:_.a === RUN TestServiceRouterConfigEntry/route_with_path_exact === RUN TestServiceRouterConfigEntry/route_with_bad_path_exact === RUN TestServiceRouterConfigEntry/route_with_path_prefix === RUN TestServiceRouterConfigEntry/route_with_bad_path_prefix === RUN TestServiceRouterConfigEntry/route_with_path_regex === RUN TestServiceRouterConfigEntry/route_with_path_exact_and_prefix === RUN TestServiceRouterConfigEntry/route_with_path_exact_and_regex === RUN TestServiceRouterConfigEntry/route_with_path_prefix_and_regex === RUN TestServiceRouterConfigEntry/route_with_path_exact,_prefix,_and_regex === RUN TestServiceRouterConfigEntry/route_with_no_name_header === RUN TestServiceRouterConfigEntry/route_with_header_present === RUN TestServiceRouterConfigEntry/route_with_header_not_present === RUN TestServiceRouterConfigEntry/route_with_header_exact === RUN TestServiceRouterConfigEntry/route_with_header_regex === RUN TestServiceRouterConfigEntry/route_with_header_prefix === RUN TestServiceRouterConfigEntry/route_with_header_suffix === RUN TestServiceRouterConfigEntry/route_with_header_present_and_exact === RUN TestServiceRouterConfigEntry/route_with_header_present_and_regex === RUN TestServiceRouterConfigEntry/route_with_header_present_and_prefix === RUN TestServiceRouterConfigEntry/route_with_header_present_and_suffix === RUN TestServiceRouterConfigEntry/route_with_no_name_query_param === RUN TestServiceRouterConfigEntry/route_with_query_param_exact_match === RUN TestServiceRouterConfigEntry/route_with_query_param_regex_match === RUN TestServiceRouterConfigEntry/route_with_query_param_present_match === RUN TestServiceRouterConfigEntry/route_with_query_param_exact_and_regex_match === RUN TestServiceRouterConfigEntry/route_with_query_param_exact_and_present_match === RUN TestServiceRouterConfigEntry/route_with_query_param_regex_and_present_match === RUN TestServiceRouterConfigEntry/route_with_query_param_exact,_regex,_and_present_match === RUN TestServiceRouterConfigEntry/route_with_no_match_and_prefix_rewrite === RUN TestDecodeConfigEntry/service-router:_kitchen_sink_(camel_case) === RUN TestServiceRouterConfigEntry/route_with_path_prefix_match_and_prefix_rewrite === RUN TestServiceRouterConfigEntry/route_with_path_exact_match_and_prefix_rewrite === RUN TestServiceRouterConfigEntry/route_with_path_regex_match_and_prefix_rewrite === RUN TestServiceResolverConfigEntry/invalid_subset_name:_a. === RUN TestServiceRouterConfigEntry/route_with_header_match_and_prefix_rewrite === RUN TestServiceResolverConfigEntry/invalid_subset_name:_a.a === RUN TestServiceRouterConfigEntry/route_with_header_match_and_prefix_rewrite#01 === RUN TestServiceRouterConfigEntry/route_with_method_matches === RUN TestServiceResolverConfigEntry/invalid_subset_name:_- === RUN TestServiceRouterConfigEntry/route_with_method_matches_repeated === RUN TestServiceResolverConfigEntry/invalid_subset_name:_-a --- PASS: TestServiceRouterConfigEntry (0.03s) --- PASS: TestServiceRouterConfigEntry/nil (0.00s) --- PASS: TestServiceRouterConfigEntry/no_name (0.00s) --- PASS: TestServiceRouterConfigEntry/empty (0.00s) --- PASS: TestServiceRouterConfigEntry/1_empty_route (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_path_exact (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_bad_path_exact (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_path_prefix (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_bad_path_prefix (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_path_regex (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_path_exact_and_prefix (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_path_exact_and_regex (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_path_prefix_and_regex (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_path_exact,_prefix,_and_regex (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_no_name_header (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_present (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_not_present (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_exact (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_regex (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_prefix (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_suffix (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_present_and_exact (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_present_and_regex (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_present_and_prefix (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_present_and_suffix (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_no_name_query_param (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_query_param_exact_match (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_query_param_regex_match (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_query_param_present_match (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_query_param_exact_and_regex_match (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_query_param_exact_and_present_match (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_query_param_regex_and_present_match (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_query_param_exact,_regex,_and_present_match (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_no_match_and_prefix_rewrite (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_path_prefix_match_and_prefix_rewrite (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_path_exact_match_and_prefix_rewrite (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_path_regex_match_and_prefix_rewrite (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_match_and_prefix_rewrite (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_header_match_and_prefix_rewrite#01 (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_method_matches (0.00s) --- PASS: TestServiceRouterConfigEntry/route_with_method_matches_repeated (0.00s) === CONT TestServiceSplitterConfigEntry === RUN TestServiceSplitterConfigEntry/nil === RUN TestServiceResolverConfigEntry/invalid_subset_name:_a- === RUN TestServiceSplitterConfigEntry/no_name === RUN TestServiceResolverConfigEntry/invalid_subset_name:_aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa === RUN TestServiceResolverConfigEntry/valid_subset_name:_a === RUN TestServiceResolverConfigEntry/valid_subset_name:_aa === RUN TestServiceResolverConfigEntry/valid_subset_name:_2a === RUN TestServiceResolverConfigEntry/valid_subset_name:_a2 === RUN TestServiceResolverConfigEntry/valid_subset_name:_a2a === RUN TestServiceResolverConfigEntry/valid_subset_name:_a22a === RUN TestServiceResolverConfigEntry/valid_subset_name:_1 === RUN TestServiceResolverConfigEntry/valid_subset_name:_11 === RUN TestServiceResolverConfigEntry/valid_subset_name:_10 === RUN TestServiceResolverConfigEntry/valid_subset_name:_01 === RUN TestServiceResolverConfigEntry/valid_subset_name:_a-a === RUN TestServiceResolverConfigEntry/valid_subset_name:_a--a === RUN TestServiceResolverConfigEntry/valid_subset_name:_a--a--a === RUN TestServiceResolverConfigEntry/valid_subset_name:_0-0 === RUN TestServiceResolverConfigEntry/valid_subset_name:_0--0 === RUN TestStructs_FilterFieldConfigurations/ServiceNode === RUN TestServiceResolverConfigEntry/valid_subset_name:_0--0--0 === RUN TestServiceResolverConfigEntry/valid_subset_name:_aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa --- PASS: TestServiceResolverConfigEntry (0.05s) --- PASS: TestServiceResolverConfigEntry/nil (0.00s) --- PASS: TestServiceResolverConfigEntry/no_name (0.00s) --- PASS: TestServiceResolverConfigEntry/empty (0.00s) --- PASS: TestServiceResolverConfigEntry/empty_subset_name (0.00s) --- PASS: TestServiceResolverConfigEntry/default_subset_does_not_exist (0.00s) --- PASS: TestServiceResolverConfigEntry/default_subset_does_exist (0.00s) --- PASS: TestServiceResolverConfigEntry/empty_redirect (0.00s) --- PASS: TestServiceResolverConfigEntry/redirect_subset_with_no_service (0.00s) --- PASS: TestServiceResolverConfigEntry/redirect_namespace_with_no_service (0.00s) --- PASS: TestServiceResolverConfigEntry/self_redirect_with_invalid_subset (0.00s) --- PASS: TestServiceResolverConfigEntry/self_redirect_with_valid_subset (0.00s) --- PASS: TestServiceResolverConfigEntry/simple_wildcard_failover (0.00s) --- PASS: TestServiceResolverConfigEntry/failover_for_missing_subset (0.00s) --- PASS: TestServiceResolverConfigEntry/failover_for_present_subset (0.00s) --- PASS: TestServiceResolverConfigEntry/failover_empty (0.00s) --- PASS: TestServiceResolverConfigEntry/failover_to_self_using_invalid_subset (0.00s) --- PASS: TestServiceResolverConfigEntry/failover_to_self_using_valid_subset (0.00s) --- PASS: TestServiceResolverConfigEntry/failover_with_empty_datacenters_in_list (0.00s) --- PASS: TestServiceResolverConfigEntry/bad_connect_timeout (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_A (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_AA (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_2A (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_A2 (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_A2A (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_A22A (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_A-A (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_A--A (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_A--A--A (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:__ (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:__a (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_a_ (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_a_a (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:__#01 (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:__a#01 (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_a_#01 (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_a_a#01 (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_. (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_.a (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_a. (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_a.a (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_- (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_-a (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_a- (0.00s) --- PASS: TestServiceResolverConfigEntry/invalid_subset_name:_aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_a (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_aa (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_2a (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_a2 (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_a2a (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_a22a (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_1 (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_11 (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_10 (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_01 (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_a-a (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_a--a (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_a--a--a (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_0-0 (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_0--0 (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_0--0--0 (0.00s) --- PASS: TestServiceResolverConfigEntry/valid_subset_name:_aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa (0.00s) === CONT TestStructs_ACLPolicy_Stub --- PASS: TestStructs_ACLPolicy_Stub (0.00s) === CONT TestStructs_ACLTokenListStubs_Sort === RUN TestDecodeConfigEntry/service-splitter:_kitchen_sink_(snake_case) === RUN TestServiceSplitterConfigEntry/empty === RUN TestDecodeConfigEntry/service-splitter:_kitchen_sink_(camel_case) === RUN TestServiceSplitterConfigEntry/1_split === RUN TestServiceSplitterConfigEntry/1_split_not_enough_weight === RUN TestDecodeConfigEntry/service-resolver:_subsets_with_failover_(snake_case) === RUN TestServiceSplitterConfigEntry/1_split_too_much_weight === RUN TestServiceSplitterConfigEntry/2_splits === RUN TestServiceSplitterConfigEntry/2_splits_-_rounded_up_to_smallest_units === RUN TestDecodeConfigEntry/service-resolver:_subsets_with_failover_(camel_case) === RUN TestServiceSplitterConfigEntry/2_splits_not_enough_weight === RUN TestServiceSplitterConfigEntry/2_splits_too_much_weight === RUN TestDecodeConfigEntry/service-resolver:_redirect_(snake_case) === RUN TestServiceSplitterConfigEntry/3_splits === CONT TestCheckDefinitionToCheckType --- PASS: TestStructs_ACLTokenListStubs_Sort (0.00s) --- PASS: TestCheckDefinitionToCheckType (0.00s) === CONT TestConfigEntries_ListRelatedServices_AndACLs === RUN TestServiceSplitterConfigEntry/3_splits_one_duplicated_same_weights === RUN TestDecodeConfigEntry/service-resolver:_redirect_(camel_case) === RUN TestServiceSplitterConfigEntry/3_splits_one_duplicated_diff_weights --- PASS: TestServiceSplitterConfigEntry (0.02s) --- PASS: TestServiceSplitterConfigEntry/nil (0.00s) --- PASS: TestServiceSplitterConfigEntry/no_name (0.01s) --- PASS: TestServiceSplitterConfigEntry/empty (0.00s) --- PASS: TestServiceSplitterConfigEntry/1_split (0.00s) --- PASS: TestServiceSplitterConfigEntry/1_split_not_enough_weight (0.00s) --- PASS: TestServiceSplitterConfigEntry/1_split_too_much_weight (0.00s) --- PASS: TestServiceSplitterConfigEntry/2_splits (0.00s) --- PASS: TestServiceSplitterConfigEntry/2_splits_-_rounded_up_to_smallest_units (0.00s) --- PASS: TestServiceSplitterConfigEntry/2_splits_not_enough_weight (0.00s) --- PASS: TestServiceSplitterConfigEntry/2_splits_too_much_weight (0.00s) --- PASS: TestServiceSplitterConfigEntry/3_splits (0.00s) --- PASS: TestServiceSplitterConfigEntry/3_splits_one_duplicated_same_weights (0.00s) --- PASS: TestServiceSplitterConfigEntry/3_splits_one_duplicated_diff_weights (0.00s) === CONT TestCheckDefinition_CheckType === RUN TestDecodeConfigEntry/service-resolver:_default_(snake_case) --- PASS: TestCheckDefinition_CheckType (0.00s) === CONT TestCheckServiceNode_BestAddress/service-wan-address-node-lan === CONT TestStructs_ACLToken_Stub/Basic === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_self === RUN TestDecodeConfigEntry/service-resolver:_default_(camel_case) === CONT TestCheckServiceNode_BestAddress/service-wan-address-default-port === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_self/deny === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_self/can_read_test === CONT TestCheckServiceNode_BestAddress/service-wan-address --- PASS: TestDecodeConfigEntry (0.06s) --- PASS: TestDecodeConfigEntry/proxy-defaults:_extra_fields_or_typo_(snake_case) (0.00s) --- PASS: TestDecodeConfigEntry/proxy-defaults:_extra_fields_or_typo_(camel_case) (0.00s) --- PASS: TestDecodeConfigEntry/proxy-defaults_(snake_case) (0.00s) --- PASS: TestDecodeConfigEntry/proxy-defaults_(camel_case) (0.00s) --- PASS: TestDecodeConfigEntry/service-defaults_(snake_case) (0.00s) --- PASS: TestDecodeConfigEntry/service-defaults_(camel_case) (0.00s) --- PASS: TestDecodeConfigEntry/service-router:_kitchen_sink_(snake_case) (0.02s) --- PASS: TestDecodeConfigEntry/service-router:_kitchen_sink_(camel_case) (0.02s) --- PASS: TestDecodeConfigEntry/service-splitter:_kitchen_sink_(snake_case) (0.00s) --- PASS: TestDecodeConfigEntry/service-splitter:_kitchen_sink_(camel_case) (0.00s) --- PASS: TestDecodeConfigEntry/service-resolver:_subsets_with_failover_(snake_case) (0.00s) --- PASS: TestDecodeConfigEntry/service-resolver:_subsets_with_failover_(camel_case) (0.00s) --- PASS: TestDecodeConfigEntry/service-resolver:_redirect_(snake_case) (0.00s) --- PASS: TestDecodeConfigEntry/service-resolver:_redirect_(camel_case) (0.00s) --- PASS: TestDecodeConfigEntry/service-resolver:_default_(snake_case) (0.00s) --- PASS: TestDecodeConfigEntry/service-resolver:_default_(camel_case) (0.00s) === CONT TestCheckServiceNode_BestAddress/service-address === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_self/can_write_test === CONT TestCheckServiceNode_BestAddress/node-wan-address === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_redirect === CONT TestCheckServiceNode_BestAddress/node-address === CONT TestStructs_ACLToken_Stub/Legacy === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_redirect/deny --- PASS: TestCheckServiceNode_BestAddress (0.00s) --- PASS: TestCheckServiceNode_BestAddress/service-wan-address-node-lan (0.00s) --- PASS: TestCheckServiceNode_BestAddress/service-wan-address-default-port (0.00s) --- PASS: TestCheckServiceNode_BestAddress/service-wan-address (0.00s) --- PASS: TestCheckServiceNode_BestAddress/service-address (0.00s) --- PASS: TestCheckServiceNode_BestAddress/node-wan-address (0.00s) --- PASS: TestCheckServiceNode_BestAddress/node-address (0.00s) === CONT TestStructs_ACLToken_EmbeddedPolicy/No_Rules === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_redirect/can_read_test === CONT TestStructs_ACLToken_PolicyIDs/Basic --- PASS: TestStructs_ACLToken_Stub (0.00s) --- PASS: TestStructs_ACLToken_Stub/Basic (0.00s) --- PASS: TestStructs_ACLToken_Stub/Legacy (0.00s) === CONT TestNodeService_BestAddress/service-address === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_redirect/cannot_write_test === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_redirect/can_write_test_(with_other:read) === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_failover === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_failover/deny === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_failover/can_read_test === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_failover/cannot_write_test === RUN TestConfigEntries_ListRelatedServices_AndACLs/resolver:_failover/can_write_test_(with_other1:read_and_other2:read) === RUN TestConfigEntries_ListRelatedServices_AndACLs/splitter:_self === RUN TestConfigEntries_ListRelatedServices_AndACLs/splitter:_self/deny === RUN TestConfigEntries_ListRelatedServices_AndACLs/splitter:_self/can_read_test === RUN TestConfigEntries_ListRelatedServices_AndACLs/splitter:_self/can_write_test === RUN TestConfigEntries_ListRelatedServices_AndACLs/splitter:_some === RUN TestConfigEntries_ListRelatedServices_AndACLs/splitter:_some/deny === RUN TestConfigEntries_ListRelatedServices_AndACLs/splitter:_some/can_read_test === RUN TestConfigEntries_ListRelatedServices_AndACLs/splitter:_some/cannot_write_test === RUN TestConfigEntries_ListRelatedServices_AndACLs/splitter:_some/can_write_test_(with_a:read,_b:read,_and_c:read) === RUN TestConfigEntries_ListRelatedServices_AndACLs/router:_self === RUN TestConfigEntries_ListRelatedServices_AndACLs/router:_self/deny === RUN TestConfigEntries_ListRelatedServices_AndACLs/router:_self/can_read_test === RUN TestConfigEntries_ListRelatedServices_AndACLs/router:_self/can_write_test === RUN TestConfigEntries_ListRelatedServices_AndACLs/router:_some === CONT TestStructs_ACLToken_EmbeddedPolicy/Legacy_Client === CONT TestStructs_ACLToken_EmbeddedPolicy/Same_Policy_for_Tokens_with_same_Rules === RUN TestConfigEntries_ListRelatedServices_AndACLs/router:_some/deny === CONT TestStructs_ACLCaches/New === RUN TestStructs_ACLCaches/New/Valid_Sizes === RUN TestConfigEntries_ListRelatedServices_AndACLs/router:_some/can_read_test === PAUSE TestStructs_ACLCaches/New/Valid_Sizes === RUN TestStructs_ACLCaches/New/Zero_Sizes === PAUSE TestStructs_ACLCaches/New/Zero_Sizes === CONT TestConfigEntryResponseMarshalling/service-default_entry === RUN TestConfigEntries_ListRelatedServices_AndACLs/router:_some/cannot_write_test === CONT TestStructs_ACLToken_PolicyIDs/No_Policies === CONT TestStructs_ACLToken_PolicyIDs/Legacy_Management_With_Rules --- PASS: TestStructs_ACLToken_EmbeddedPolicy (0.00s) --- PASS: TestStructs_ACLToken_EmbeddedPolicy/No_Rules (0.00s) --- PASS: TestStructs_ACLToken_EmbeddedPolicy/Same_Policy_for_Tokens_with_same_Rules (0.00s) --- PASS: TestStructs_ACLToken_EmbeddedPolicy/Legacy_Client (0.00s) === RUN TestConfigEntries_ListRelatedServices_AndACLs/router:_some/can_write_test_(with_foo:read_and_bar:read) === CONT TestStructs_ACLToken_PolicyIDs/Legacy_Management --- PASS: TestConfigEntries_ListRelatedServices_AndACLs (0.01s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_self (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_self/deny (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_self/can_read_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_self/can_write_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_redirect (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_redirect/deny (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_redirect/can_read_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_redirect/cannot_write_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_redirect/can_write_test_(with_other:read) (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_failover (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_failover/deny (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_failover/can_read_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_failover/cannot_write_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/resolver:_failover/can_write_test_(with_other1:read_and_other2:read) (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/splitter:_self (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/splitter:_self/deny (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/splitter:_self/can_read_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/splitter:_self/can_write_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/splitter:_some (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/splitter:_some/deny (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/splitter:_some/can_read_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/splitter:_some/cannot_write_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/splitter:_some/can_write_test_(with_a:read,_b:read,_and_c:read) (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/router:_self (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/router:_self/deny (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/router:_self/can_read_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/router:_self/can_write_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/router:_some (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/router:_some/deny (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/router:_some/can_read_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/router:_some/cannot_write_test (0.00s) --- PASS: TestConfigEntries_ListRelatedServices_AndACLs/router:_some/can_write_test_(with_foo:read_and_bar:read) (0.00s) === CONT TestNode_BestAddress/address --- PASS: TestStructs_ACLToken_PolicyIDs (0.00s) --- PASS: TestStructs_ACLToken_PolicyIDs/Basic (0.01s) --- PASS: TestStructs_ACLToken_PolicyIDs/No_Policies (0.00s) --- PASS: TestStructs_ACLToken_PolicyIDs/Legacy_Management_With_Rules (0.00s) --- PASS: TestStructs_ACLToken_PolicyIDs/Legacy_Management (0.00s) === CONT TestNodeService_BestAddress/no-address === CONT TestNodeService_BestAddress/service-wan-address-node-lan === CONT TestNodeService_BestAddress/service-wan-address-default-port === CONT TestNodeService_BestAddress/service-wan-address === CONT TestStructs_ACLToken_Convert/Management --- PASS: TestNodeService_BestAddress (0.00s) --- PASS: TestNodeService_BestAddress/service-address (0.01s) --- PASS: TestNodeService_BestAddress/service-wan-address-node-lan (0.00s) --- PASS: TestNodeService_BestAddress/no-address (0.00s) --- PASS: TestNodeService_BestAddress/service-wan-address-default-port (0.00s) --- PASS: TestNodeService_BestAddress/service-wan-address (0.00s) === CONT TestServiceNode_JSON_OmitServiceTaggedAdddresses/nil === CONT TestStructs_ACLCaches/Roles === CONT TestStructs_ACLCaches/Authorizers === CONT TestStructs_ACLCaches/ParsedPolicies === CONT TestStructs_ACLCaches/Policies === CONT TestStructs_ACLCaches/Identities === CONT TestNodeService_JSON_OmitTaggedAdddresses/nil === CONT TestConfigEntryResponseMarshalling/proxy-default_entry === CONT TestConfigEntryResponseMarshalling/nil_entry === CONT TestStructs_ACLToken_Convert/Client === CONT TestNode_BestAddress/wan-address --- PASS: TestConfigEntryResponseMarshalling (0.00s) --- PASS: TestConfigEntryResponseMarshalling/service-default_entry (0.00s) --- PASS: TestConfigEntryResponseMarshalling/proxy-default_entry (0.00s) --- PASS: TestConfigEntryResponseMarshalling/nil_entry (0.00s) === CONT TestStructs_ACLToken_Convert/Unconvertible --- PASS: TestNode_BestAddress (0.00s) --- PASS: TestNode_BestAddress/address (0.00s) --- PASS: TestNode_BestAddress/wan-address (0.00s) === CONT TestServiceNode_JSON_OmitServiceTaggedAdddresses/empty --- PASS: TestStructs_ACLToken_Convert (0.00s) --- PASS: TestStructs_ACLToken_Convert/Management (0.00s) --- PASS: TestStructs_ACLToken_Convert/Client (0.00s) --- PASS: TestStructs_ACLToken_Convert/Unconvertible (0.00s) === CONT TestNodeService_JSON_OmitTaggedAdddresses/empty --- PASS: TestServiceNode_JSON_OmitServiceTaggedAdddresses (0.00s) --- PASS: TestServiceNode_JSON_OmitServiceTaggedAdddresses/nil (0.00s) --- PASS: TestServiceNode_JSON_OmitServiceTaggedAdddresses/empty (0.00s) === CONT TestStructs_ACLCaches/New/Valid_Sizes --- PASS: TestNodeService_JSON_OmitTaggedAdddresses (0.00s) --- PASS: TestNodeService_JSON_OmitTaggedAdddresses/nil (0.00s) --- PASS: TestNodeService_JSON_OmitTaggedAdddresses/empty (0.00s) === CONT TestStructs_ACLCaches/New/Zero_Sizes --- PASS: TestStructs_ACLCaches (0.01s) --- PASS: TestStructs_ACLCaches/Roles (0.00s) --- PASS: TestStructs_ACLCaches/Policies (0.00s) --- PASS: TestStructs_ACLCaches/ParsedPolicies (0.00s) --- PASS: TestStructs_ACLCaches/Identities (0.00s) --- PASS: TestStructs_ACLCaches/New (0.00s) --- PASS: TestStructs_ACLCaches/New/Valid_Sizes (0.00s) --- PASS: TestStructs_ACLCaches/New/Zero_Sizes (0.00s) --- PASS: TestStructs_ACLCaches/Authorizers (0.00s) === RUN TestStructs_FilterFieldConfigurations/HealthCheck === RUN TestStructs_FilterFieldConfigurations/CheckServiceNode === RUN TestStructs_FilterFieldConfigurations/NodeInfo === RUN TestStructs_FilterFieldConfigurations/api.AgentService --- PASS: TestStructs_FilterFieldConfigurations (0.16s) --- PASS: TestStructs_FilterFieldConfigurations/Node (0.00s) --- PASS: TestStructs_FilterFieldConfigurations/NodeService (0.06s) --- PASS: TestStructs_FilterFieldConfigurations/ServiceNode (0.03s) --- PASS: TestStructs_FilterFieldConfigurations/HealthCheck (0.00s) --- PASS: TestStructs_FilterFieldConfigurations/CheckServiceNode (0.02s) --- PASS: TestStructs_FilterFieldConfigurations/NodeInfo (0.01s) --- PASS: TestStructs_FilterFieldConfigurations/api.AgentService (0.03s) PASS ok github.com/hashicorp/consul/agent/structs 0.245s ? github.com/hashicorp/consul/agent/systemd [no test files] === RUN TestStore_RegularTokens === PAUSE TestStore_RegularTokens === RUN TestStore_AgentMasterToken === PAUSE TestStore_AgentMasterToken === CONT TestStore_RegularTokens === RUN TestStore_RegularTokens/set_user_-_config === PAUSE TestStore_RegularTokens/set_user_-_config === RUN TestStore_RegularTokens/set_user_-_api === PAUSE TestStore_RegularTokens/set_user_-_api === RUN TestStore_RegularTokens/set_agent_-_config === PAUSE TestStore_RegularTokens/set_agent_-_config === RUN TestStore_RegularTokens/set_agent_-_api === PAUSE TestStore_RegularTokens/set_agent_-_api === RUN TestStore_RegularTokens/set_user_and_agent === PAUSE TestStore_RegularTokens/set_user_and_agent === RUN TestStore_RegularTokens/set_repl_-_config === PAUSE TestStore_RegularTokens/set_repl_-_config === RUN TestStore_RegularTokens/set_repl_-_api === PAUSE TestStore_RegularTokens/set_repl_-_api === RUN TestStore_RegularTokens/set_master_-_config === PAUSE TestStore_RegularTokens/set_master_-_config === RUN TestStore_RegularTokens/set_master_-_api === PAUSE TestStore_RegularTokens/set_master_-_api === RUN TestStore_RegularTokens/set_all === PAUSE TestStore_RegularTokens/set_all === CONT TestStore_RegularTokens/set_user_-_config === CONT TestStore_AgentMasterToken --- PASS: TestStore_AgentMasterToken (0.00s) === CONT TestStore_RegularTokens/set_all === CONT TestStore_RegularTokens/set_master_-_api === CONT TestStore_RegularTokens/set_master_-_config === CONT TestStore_RegularTokens/set_repl_-_api === CONT TestStore_RegularTokens/set_repl_-_config === CONT TestStore_RegularTokens/set_user_and_agent === CONT TestStore_RegularTokens/set_agent_-_api === CONT TestStore_RegularTokens/set_agent_-_config === CONT TestStore_RegularTokens/set_user_-_api --- PASS: TestStore_RegularTokens (0.00s) --- PASS: TestStore_RegularTokens/set_user_-_config (0.00s) --- PASS: TestStore_RegularTokens/set_all (0.00s) --- PASS: TestStore_RegularTokens/set_master_-_api (0.00s) --- PASS: TestStore_RegularTokens/set_master_-_config (0.00s) --- PASS: TestStore_RegularTokens/set_repl_-_api (0.00s) --- PASS: TestStore_RegularTokens/set_repl_-_config (0.00s) --- PASS: TestStore_RegularTokens/set_user_and_agent (0.00s) --- PASS: TestStore_RegularTokens/set_agent_-_api (0.00s) --- PASS: TestStore_RegularTokens/set_agent_-_config (0.00s) --- PASS: TestStore_RegularTokens/set_user_-_api (0.00s) PASS ok github.com/hashicorp/consul/agent/token 0.004s === RUN TestClustersFromSnapshot === RUN TestClustersFromSnapshot/defaults writer.go:29: 2021-01-29T19:31:22.234Z [DEBUG] TestClustersFromSnapshot/defaults: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/custom-local-app writer.go:29: 2021-01-29T19:31:22.250Z [DEBUG] TestClustersFromSnapshot/custom-local-app: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/custom-local-app-typed writer.go:29: 2021-01-29T19:31:22.264Z [DEBUG] TestClustersFromSnapshot/custom-local-app-typed: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/custom-upstream writer.go:29: 2021-01-29T19:31:22.273Z [DEBUG] TestClustersFromSnapshot/custom-upstream: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/custom-upstream-default-chain writer.go:29: 2021-01-29T19:31:22.280Z [DEBUG] TestClustersFromSnapshot/custom-upstream-default-chain: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/custom-upstream-typed writer.go:29: 2021-01-29T19:31:22.287Z [DEBUG] TestClustersFromSnapshot/custom-upstream-typed: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/custom-upstream-ignores-tls writer.go:29: 2021-01-29T19:31:22.294Z [DEBUG] TestClustersFromSnapshot/custom-upstream-ignores-tls: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/custom-timeouts writer.go:29: 2021-01-29T19:31:22.300Z [DEBUG] TestClustersFromSnapshot/custom-timeouts: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/custom-limits-max-connections-only writer.go:29: 2021-01-29T19:31:22.308Z [DEBUG] TestClustersFromSnapshot/custom-limits-max-connections-only: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/custom-limits-set-to-zero writer.go:29: 2021-01-29T19:31:22.319Z [DEBUG] TestClustersFromSnapshot/custom-limits-set-to-zero: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/custom-limits writer.go:29: 2021-01-29T19:31:22.326Z [DEBUG] TestClustersFromSnapshot/custom-limits: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-chain writer.go:29: 2021-01-29T19:31:22.338Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-chain: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-chain-external-sni writer.go:29: 2021-01-29T19:31:22.349Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-chain-external-sni: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-chain-and-overrides writer.go:29: 2021-01-29T19:31:22.359Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-chain-and-overrides: generating cluster for: cluster=a236e964~db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-chain-and-failover writer.go:29: 2021-01-29T19:31:22.371Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-chain-and-failover: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway writer.go:29: 2021-01-29T19:31:22.397Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway-triggered writer.go:29: 2021-01-29T19:31:22.414Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway-triggered: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-remote-gateway writer.go:29: 2021-01-29T19:31:22.421Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-remote-gateway: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-remote-gateway-triggered writer.go:29: 2021-01-29T19:31:22.437Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-remote-gateway-triggered: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway writer.go:29: 2021-01-29T19:31:22.449Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway-triggered writer.go:29: 2021-01-29T19:31:22.478Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway-triggered: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-local-gateway writer.go:29: 2021-01-29T19:31:22.492Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-local-gateway: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-local-gateway-triggered writer.go:29: 2021-01-29T19:31:22.513Z [DEBUG] TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-local-gateway-triggered: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/splitter-with-resolver-redirect writer.go:29: 2021-01-29T19:31:22.529Z [DEBUG] TestClustersFromSnapshot/splitter-with-resolver-redirect: generating cluster for: cluster=v1.db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul writer.go:29: 2021-01-29T19:31:22.530Z [DEBUG] TestClustersFromSnapshot/splitter-with-resolver-redirect: generating cluster for: cluster=v2.db.default.dc2.internal.11111111-2222-3333-4444-555555555555.consul === RUN TestClustersFromSnapshot/expose-paths-local-app-paths === RUN TestClustersFromSnapshot/expose-paths-new-cluster-http2 === RUN TestClustersFromSnapshot/mesh-gateway === RUN TestClustersFromSnapshot/mesh-gateway-no-services === RUN TestClustersFromSnapshot/mesh-gateway-service-subsets === RUN TestClustersFromSnapshot/mesh-gateway-ignore-extra-resolvers === RUN TestClustersFromSnapshot/mesh-gateway-service-timeouts --- PASS: TestClustersFromSnapshot (0.41s) --- PASS: TestClustersFromSnapshot/defaults (0.06s) --- PASS: TestClustersFromSnapshot/custom-local-app (0.01s) --- PASS: TestClustersFromSnapshot/custom-local-app-typed (0.01s) --- PASS: TestClustersFromSnapshot/custom-upstream (0.01s) --- PASS: TestClustersFromSnapshot/custom-upstream-default-chain (0.01s) --- PASS: TestClustersFromSnapshot/custom-upstream-typed (0.01s) --- PASS: TestClustersFromSnapshot/custom-upstream-ignores-tls (0.01s) --- PASS: TestClustersFromSnapshot/custom-timeouts (0.01s) --- PASS: TestClustersFromSnapshot/custom-limits-max-connections-only (0.01s) --- PASS: TestClustersFromSnapshot/custom-limits-set-to-zero (0.01s) --- PASS: TestClustersFromSnapshot/custom-limits (0.01s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-chain (0.01s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-chain-external-sni (0.01s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-chain-and-overrides (0.01s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-chain-and-failover (0.01s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway (0.02s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway-triggered (0.01s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-remote-gateway (0.01s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-remote-gateway-triggered (0.01s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway (0.03s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway-triggered (0.01s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-local-gateway (0.01s) --- PASS: TestClustersFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-local-gateway-triggered (0.02s) --- PASS: TestClustersFromSnapshot/splitter-with-resolver-redirect (0.03s) --- PASS: TestClustersFromSnapshot/expose-paths-local-app-paths (0.01s) --- PASS: TestClustersFromSnapshot/expose-paths-new-cluster-http2 (0.01s) --- PASS: TestClustersFromSnapshot/mesh-gateway (0.00s) --- PASS: TestClustersFromSnapshot/mesh-gateway-no-services (0.00s) --- PASS: TestClustersFromSnapshot/mesh-gateway-service-subsets (0.00s) --- PASS: TestClustersFromSnapshot/mesh-gateway-ignore-extra-resolvers (0.00s) --- PASS: TestClustersFromSnapshot/mesh-gateway-service-timeouts (0.00s) === RUN TestParseProxyConfig === RUN TestParseProxyConfig/defaults_-_nil === RUN TestParseProxyConfig/defaults_-_empty === RUN TestParseProxyConfig/defaults_-_other_stuff === RUN TestParseProxyConfig/protocol_override === RUN TestParseProxyConfig/protocol_uppercase_override === RUN TestParseProxyConfig/bind_address_override,_string === RUN TestParseProxyConfig/bind_port_override,_string === RUN TestParseProxyConfig/bind_port_override,_int === RUN TestParseProxyConfig/local_connect_timeout_override,_string === RUN TestParseProxyConfig/local_connect_timeout_override,_float_ === RUN TestParseProxyConfig/local_connect_timeout_override,_int_ --- PASS: TestParseProxyConfig (0.01s) --- PASS: TestParseProxyConfig/defaults_-_nil (0.00s) --- PASS: TestParseProxyConfig/defaults_-_empty (0.00s) --- PASS: TestParseProxyConfig/defaults_-_other_stuff (0.00s) --- PASS: TestParseProxyConfig/protocol_override (0.00s) --- PASS: TestParseProxyConfig/protocol_uppercase_override (0.00s) --- PASS: TestParseProxyConfig/bind_address_override,_string (0.00s) --- PASS: TestParseProxyConfig/bind_port_override,_string (0.00s) --- PASS: TestParseProxyConfig/bind_port_override,_int (0.00s) --- PASS: TestParseProxyConfig/local_connect_timeout_override,_string (0.00s) --- PASS: TestParseProxyConfig/local_connect_timeout_override,_float_ (0.00s) --- PASS: TestParseProxyConfig/local_connect_timeout_override,_int_ (0.00s) === RUN TestParseUpstreamConfig === RUN TestParseUpstreamConfig/defaults_-_nil === RUN TestParseUpstreamConfig/defaults_-_empty === RUN TestParseUpstreamConfig/defaults_-_other_stuff === RUN TestParseUpstreamConfig/protocol_override === RUN TestParseUpstreamConfig/connect_timeout_override,_string === RUN TestParseUpstreamConfig/connect_timeout_override,_float_ === RUN TestParseUpstreamConfig/connect_timeout_override,_int_ === RUN TestParseUpstreamConfig/connect_limits_map === RUN TestParseUpstreamConfig/connect_limits_map_zero --- PASS: TestParseUpstreamConfig (0.01s) --- PASS: TestParseUpstreamConfig/defaults_-_nil (0.00s) --- PASS: TestParseUpstreamConfig/defaults_-_empty (0.00s) --- PASS: TestParseUpstreamConfig/defaults_-_other_stuff (0.00s) --- PASS: TestParseUpstreamConfig/protocol_override (0.00s) --- PASS: TestParseUpstreamConfig/connect_timeout_override,_string (0.00s) --- PASS: TestParseUpstreamConfig/connect_timeout_override,_float_ (0.00s) --- PASS: TestParseUpstreamConfig/connect_timeout_override,_int_ (0.00s) --- PASS: TestParseUpstreamConfig/connect_limits_map (0.00s) --- PASS: TestParseUpstreamConfig/connect_limits_map_zero (0.00s) === RUN Test_makeLoadAssignment === RUN Test_makeLoadAssignment/no_instances === RUN Test_makeLoadAssignment/instances,_no_weights === RUN Test_makeLoadAssignment/instances,_healthy_weights === RUN Test_makeLoadAssignment/instances,_warning_weights --- PASS: Test_makeLoadAssignment (0.01s) --- PASS: Test_makeLoadAssignment/no_instances (0.00s) --- PASS: Test_makeLoadAssignment/instances,_no_weights (0.00s) --- PASS: Test_makeLoadAssignment/instances,_healthy_weights (0.00s) --- PASS: Test_makeLoadAssignment/instances,_warning_weights (0.00s) === RUN Test_endpointsFromSnapshot === RUN Test_endpointsFromSnapshot/defaults === RUN Test_endpointsFromSnapshot/mesh-gateway === RUN Test_endpointsFromSnapshot/mesh-gateway-no-services === RUN Test_endpointsFromSnapshot/connect-proxy-with-chain === RUN Test_endpointsFromSnapshot/connect-proxy-with-chain-external-sni === RUN Test_endpointsFromSnapshot/connect-proxy-with-chain-and-overrides === RUN Test_endpointsFromSnapshot/connect-proxy-with-chain-and-failover === RUN Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway === RUN Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway-triggered === RUN Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-remote-gateway === RUN Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-remote-gateway-triggered === RUN Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway === RUN Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway-triggered === RUN Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-local-gateway === RUN Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-local-gateway-triggered === RUN Test_endpointsFromSnapshot/splitter-with-resolver-redirect === RUN Test_endpointsFromSnapshot/mesh-gateway-service-subsets === RUN Test_endpointsFromSnapshot/mesh-gateway-default-service-subset --- PASS: Test_endpointsFromSnapshot (0.16s) --- PASS: Test_endpointsFromSnapshot/defaults (0.01s) --- PASS: Test_endpointsFromSnapshot/mesh-gateway (0.00s) --- PASS: Test_endpointsFromSnapshot/mesh-gateway-no-services (0.00s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-chain (0.00s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-chain-external-sni (0.01s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-chain-and-overrides (0.01s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-chain-and-failover (0.01s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway (0.01s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway-triggered (0.01s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-remote-gateway (0.01s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-remote-gateway-triggered (0.01s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway (0.01s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway-triggered (0.01s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-local-gateway (0.01s) --- PASS: Test_endpointsFromSnapshot/connect-proxy-with-tcp-chain-double-failover-through-local-gateway-triggered (0.01s) --- PASS: Test_endpointsFromSnapshot/splitter-with-resolver-redirect (0.01s) --- PASS: Test_endpointsFromSnapshot/mesh-gateway-service-subsets (0.01s) --- PASS: Test_endpointsFromSnapshot/mesh-gateway-default-service-subset (0.01s) === RUN TestFirstHealthyTarget === RUN TestFirstHealthyTarget/all-ok.default.dc1_and_[] === RUN TestFirstHealthyTarget/all-warn.default.dc1_and_[] === RUN TestFirstHealthyTarget/all-fail.default.dc1_and_[] === RUN TestFirstHealthyTarget/all-warn-onlypassing.default.dc1_and_[] === RUN TestFirstHealthyTarget/all-fail-onlypassing.default.dc1_and_[] === RUN TestFirstHealthyTarget/all-ok.default.dc1_and_[all-warn.default.dc1] === RUN TestFirstHealthyTarget/all-warn.default.dc1_and_[all-ok.default.dc1] === RUN TestFirstHealthyTarget/all-warn-onlypassing.default.dc1_and_[all-ok.default.dc1] === RUN TestFirstHealthyTarget/all-fail.default.dc1_and_[all-ok.default.dc1] === RUN TestFirstHealthyTarget/all-fail-onlypassing.default.dc1_and_[all-ok.default.dc1] === RUN TestFirstHealthyTarget/all-fail.default.dc1_and_[all-warn-onlypassing.default.dc1_all-warn.default.dc1_all-ok.default.dc1] --- PASS: TestFirstHealthyTarget (0.02s) --- PASS: TestFirstHealthyTarget/all-ok.default.dc1_and_[] (0.00s) --- PASS: TestFirstHealthyTarget/all-warn.default.dc1_and_[] (0.00s) --- PASS: TestFirstHealthyTarget/all-fail.default.dc1_and_[] (0.00s) --- PASS: TestFirstHealthyTarget/all-warn-onlypassing.default.dc1_and_[] (0.00s) --- PASS: TestFirstHealthyTarget/all-fail-onlypassing.default.dc1_and_[] (0.00s) --- PASS: TestFirstHealthyTarget/all-ok.default.dc1_and_[all-warn.default.dc1] (0.00s) --- PASS: TestFirstHealthyTarget/all-warn.default.dc1_and_[all-ok.default.dc1] (0.00s) --- PASS: TestFirstHealthyTarget/all-warn-onlypassing.default.dc1_and_[all-ok.default.dc1] (0.00s) --- PASS: TestFirstHealthyTarget/all-fail.default.dc1_and_[all-ok.default.dc1] (0.00s) --- PASS: TestFirstHealthyTarget/all-fail-onlypassing.default.dc1_and_[all-ok.default.dc1] (0.00s) --- PASS: TestFirstHealthyTarget/all-fail.default.dc1_and_[all-warn-onlypassing.default.dc1_all-warn.default.dc1_all-ok.default.dc1] (0.00s) === RUN TestListenersFromSnapshot === RUN TestListenersFromSnapshot/defaults === RUN TestListenersFromSnapshot/listener-bind-address === RUN TestListenersFromSnapshot/listener-bind-port === RUN TestListenersFromSnapshot/listener-bind-address-port === RUN TestListenersFromSnapshot/http-public-listener === RUN TestListenersFromSnapshot/http-upstream === RUN TestListenersFromSnapshot/custom-public-listener === RUN TestListenersFromSnapshot/custom-public-listener-typed === RUN TestListenersFromSnapshot/custom-public-listener-ignores-tls === RUN TestListenersFromSnapshot/custom-upstream === RUN TestListenersFromSnapshot/custom-upstream-typed === RUN TestListenersFromSnapshot/splitter-with-resolver-redirect === RUN TestListenersFromSnapshot/connect-proxy-with-tcp-chain === RUN TestListenersFromSnapshot/connect-proxy-with-http-chain === RUN TestListenersFromSnapshot/connect-proxy-with-http2-chain === RUN TestListenersFromSnapshot/connect-proxy-with-grpc-chain === RUN TestListenersFromSnapshot/connect-proxy-with-chain-external-sni === RUN TestListenersFromSnapshot/connect-proxy-with-chain-and-overrides === RUN TestListenersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway === RUN TestListenersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway === RUN TestListenersFromSnapshot/expose-paths-local-app-paths === RUN TestListenersFromSnapshot/expose-paths-new-cluster-http2 === RUN TestListenersFromSnapshot/mesh-gateway === RUN TestListenersFromSnapshot/mesh-gateway-no-services === RUN TestListenersFromSnapshot/mesh-gateway-tagged-addresses === RUN TestListenersFromSnapshot/mesh-gateway-custom-addresses --- PASS: TestListenersFromSnapshot (0.27s) --- PASS: TestListenersFromSnapshot/defaults (0.03s) --- PASS: TestListenersFromSnapshot/listener-bind-address (0.02s) --- PASS: TestListenersFromSnapshot/listener-bind-port (0.01s) --- PASS: TestListenersFromSnapshot/listener-bind-address-port (0.01s) --- PASS: TestListenersFromSnapshot/http-public-listener (0.01s) --- PASS: TestListenersFromSnapshot/http-upstream (0.01s) --- PASS: TestListenersFromSnapshot/custom-public-listener (0.01s) --- PASS: TestListenersFromSnapshot/custom-public-listener-typed (0.01s) --- PASS: TestListenersFromSnapshot/custom-public-listener-ignores-tls (0.01s) --- PASS: TestListenersFromSnapshot/custom-upstream (0.01s) --- PASS: TestListenersFromSnapshot/custom-upstream-typed (0.01s) --- PASS: TestListenersFromSnapshot/splitter-with-resolver-redirect (0.01s) --- PASS: TestListenersFromSnapshot/connect-proxy-with-tcp-chain (0.01s) --- PASS: TestListenersFromSnapshot/connect-proxy-with-http-chain (0.01s) --- PASS: TestListenersFromSnapshot/connect-proxy-with-http2-chain (0.01s) --- PASS: TestListenersFromSnapshot/connect-proxy-with-grpc-chain (0.01s) --- PASS: TestListenersFromSnapshot/connect-proxy-with-chain-external-sni (0.01s) --- PASS: TestListenersFromSnapshot/connect-proxy-with-chain-and-overrides (0.01s) --- PASS: TestListenersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-remote-gateway (0.01s) --- PASS: TestListenersFromSnapshot/connect-proxy-with-tcp-chain-failover-through-local-gateway (0.01s) --- PASS: TestListenersFromSnapshot/expose-paths-local-app-paths (0.01s) --- PASS: TestListenersFromSnapshot/expose-paths-new-cluster-http2 (0.01s) --- PASS: TestListenersFromSnapshot/mesh-gateway (0.00s) --- PASS: TestListenersFromSnapshot/mesh-gateway-no-services (0.00s) --- PASS: TestListenersFromSnapshot/mesh-gateway-tagged-addresses (0.01s) --- PASS: TestListenersFromSnapshot/mesh-gateway-custom-addresses (0.01s) === RUN TestRoutesFromSnapshot === RUN TestRoutesFromSnapshot/defaults-no-chain === RUN TestRoutesFromSnapshot/connect-proxy-with-chain === RUN TestRoutesFromSnapshot/connect-proxy-with-chain-external-sni === RUN TestRoutesFromSnapshot/connect-proxy-with-chain-and-overrides === RUN TestRoutesFromSnapshot/splitter-with-resolver-redirect === RUN TestRoutesFromSnapshot/connect-proxy-with-chain-and-splitter === RUN TestRoutesFromSnapshot/connect-proxy-with-grpc-router === RUN TestRoutesFromSnapshot/connect-proxy-with-chain-and-router --- PASS: TestRoutesFromSnapshot (0.09s) --- PASS: TestRoutesFromSnapshot/defaults-no-chain (0.01s) --- PASS: TestRoutesFromSnapshot/connect-proxy-with-chain (0.01s) --- PASS: TestRoutesFromSnapshot/connect-proxy-with-chain-external-sni (0.01s) --- PASS: TestRoutesFromSnapshot/connect-proxy-with-chain-and-overrides (0.01s) --- PASS: TestRoutesFromSnapshot/splitter-with-resolver-redirect (0.01s) --- PASS: TestRoutesFromSnapshot/connect-proxy-with-chain-and-splitter (0.01s) --- PASS: TestRoutesFromSnapshot/connect-proxy-with-grpc-router (0.01s) --- PASS: TestRoutesFromSnapshot/connect-proxy-with-chain-and-router (0.02s) === RUN TestServer_StreamAggregatedResources_BasicProtocol writer.go:29: 2021-01-29T19:31:23.186Z [DEBUG] TestServer_StreamAggregatedResources_BasicProtocol.envoy: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul writer.go:29: 2021-01-29T19:31:23.218Z [DEBUG] TestServer_StreamAggregatedResources_BasicProtocol.envoy: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul writer.go:29: 2021-01-29T19:31:23.251Z [DEBUG] TestServer_StreamAggregatedResources_BasicProtocol.envoy: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul --- PASS: TestServer_StreamAggregatedResources_BasicProtocol (0.09s) === RUN TestServer_StreamAggregatedResources_ACLEnforcement server_test.go:396: DM-skipped --- SKIP: TestServer_StreamAggregatedResources_ACLEnforcement (0.00s) === RUN TestServer_StreamAggregatedResources_ACLTokenDeleted_StreamTerminatedDuringDiscoveryRequest writer.go:29: 2021-01-29T19:31:23.283Z [DEBUG] TestServer_StreamAggregatedResources_ACLTokenDeleted_StreamTerminatedDuringDiscoveryRequest.envoy: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul writer.go:29: 2021-01-29T19:31:23.296Z [DEBUG] TestServer_StreamAggregatedResources_ACLTokenDeleted_StreamTerminatedDuringDiscoveryRequest.envoy: Error handling ADS stream: error="rpc error: code = Unauthenticated desc = unauthenticated: ACL not found" --- PASS: TestServer_StreamAggregatedResources_ACLTokenDeleted_StreamTerminatedDuringDiscoveryRequest (0.04s) === RUN TestServer_StreamAggregatedResources_ACLTokenDeleted_StreamTerminatedInBackground writer.go:29: 2021-01-29T19:31:23.328Z [DEBUG] TestServer_StreamAggregatedResources_ACLTokenDeleted_StreamTerminatedInBackground.envoy: generating cluster for: cluster=db.default.dc1.internal.11111111-2222-3333-4444-555555555555.consul writer.go:29: 2021-01-29T19:31:23.432Z [DEBUG] TestServer_StreamAggregatedResources_ACLTokenDeleted_StreamTerminatedInBackground.envoy: Error handling ADS stream: error="rpc error: code = Unauthenticated desc = unauthenticated: ACL not found" --- PASS: TestServer_StreamAggregatedResources_ACLTokenDeleted_StreamTerminatedInBackground (0.14s) === RUN TestServer_Check === RUN TestServer_Check/auth_allowed writer.go:29: 2021-01-29T19:31:23.445Z [DEBUG] TestServer_Check/auth_allowed.envoy: Connect AuthZ ALLOWED: source=spiffe://11111111-2222-3333-4444-555555555555.consul/ns/default/dc/dc1/svc/web destination=spiffe://11111111-2222-3333-4444-555555555555.consul/ns/default/dc/dc1/svc/db reason="default allow" === RUN TestServer_Check/auth_denied writer.go:29: 2021-01-29T19:31:23.451Z [DEBUG] TestServer_Check/auth_denied.envoy: Connect AuthZ DENIED: source=spiffe://11111111-2222-3333-4444-555555555555.consul/ns/default/dc/dc1/svc/web destination=spiffe://11111111-2222-3333-4444-555555555555.consul/ns/default/dc/dc1/svc/db reason="default deny" === RUN TestServer_Check/no_source === RUN TestServer_Check/no_dest === RUN TestServer_Check/dest_invalid_format writer.go:29: 2021-01-29T19:31:23.473Z [DEBUG] TestServer_Check/dest_invalid_format.envoy: Connect AuthZ DENIED: bad destination URI: source=spiffe://11111111-2222-3333-4444-555555555555.consul/ns/default/dc/dc1/svc/web destination=not-a-spiffe-id === RUN TestServer_Check/dest_not_a_service_URI writer.go:29: 2021-01-29T19:31:23.486Z [DEBUG] TestServer_Check/dest_not_a_service_URI.envoy: Connect AuthZ DENIED: bad destination service ID: source=spiffe://11111111-2222-3333-4444-555555555555.consul/ns/default/dc/dc1/svc/web destination=spiffe://trust-domain.consul === RUN TestServer_Check/ACL_not_got_permission_for_authz_call writer.go:29: 2021-01-29T19:31:23.496Z [DEBUG] TestServer_Check/ACL_not_got_permission_for_authz_call.envoy: Connect AuthZ failed ACL check: error="Permission denied" source=spiffe://11111111-2222-3333-4444-555555555555.consul/ns/default/dc/dc1/svc/web dest=spiffe://11111111-2222-3333-4444-555555555555.consul/ns/default/dc/dc1/svc/db === RUN TestServer_Check/Random_error_running_authz writer.go:29: 2021-01-29T19:31:23.504Z [DEBUG] TestServer_Check/Random_error_running_authz.envoy: Connect AuthZ failed: error="gremlin attack" source=spiffe://11111111-2222-3333-4444-555555555555.consul/ns/default/dc/dc1/svc/web destination=spiffe://11111111-2222-3333-4444-555555555555.consul/ns/default/dc/dc1/svc/db --- PASS: TestServer_Check (0.07s) --- PASS: TestServer_Check/auth_allowed (0.01s) --- PASS: TestServer_Check/auth_denied (0.01s) --- PASS: TestServer_Check/no_source (0.01s) --- PASS: TestServer_Check/no_dest (0.00s) --- PASS: TestServer_Check/dest_invalid_format (0.01s) --- PASS: TestServer_Check/dest_not_a_service_URI (0.01s) --- PASS: TestServer_Check/ACL_not_got_permission_for_authz_call (0.00s) --- PASS: TestServer_Check/Random_error_running_authz (0.01s) PASS ok github.com/hashicorp/consul/agent/xds 1.401s ? github.com/hashicorp/consul/command [no test files] ? github.com/hashicorp/consul/command/acl [no test files] === RUN TestAgentTokensCommand_noTabs === PAUSE TestAgentTokensCommand_noTabs === RUN TestAgentTokensCommand === PAUSE TestAgentTokensCommand === CONT TestAgentTokensCommand_noTabs === CONT TestAgentTokensCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestAgentTokensCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestAgentTokensCommand writer.go:29: 2021-01-29T19:31:24.701Z [WARN] TestAgentTokensCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:24.704Z [DEBUG] TestAgentTokensCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:24.708Z [DEBUG] TestAgentTokensCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:24.715Z [INFO] TestAgentTokensCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5f8b86a6-d596-04d0-2d0c-ebc584b957db Address:127.0.0.1:15111}]" writer.go:29: 2021-01-29T19:31:24.720Z [INFO] TestAgentTokensCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:15111 [Follower]" leader= writer.go:29: 2021-01-29T19:31:24.723Z [INFO] TestAgentTokensCommand.server.serf.wan: serf: EventMemberJoin: Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:24.729Z [INFO] TestAgentTokensCommand.server.serf.lan: serf: EventMemberJoin: Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db 127.0.0.1 writer.go:29: 2021-01-29T19:31:24.734Z [INFO] TestAgentTokensCommand.server: Adding LAN server: server="Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db (Addr: tcp/127.0.0.1:15111) (DC: dc1)" writer.go:29: 2021-01-29T19:31:24.734Z [INFO] TestAgentTokensCommand: Started DNS server: address=127.0.0.1:15106 network=udp writer.go:29: 2021-01-29T19:31:24.737Z [INFO] TestAgentTokensCommand: Started DNS server: address=127.0.0.1:15106 network=tcp writer.go:29: 2021-01-29T19:31:24.741Z [INFO] TestAgentTokensCommand: Started HTTP server: address=127.0.0.1:15107 network=tcp writer.go:29: 2021-01-29T19:31:24.734Z [INFO] TestAgentTokensCommand.server: Handled event for server in area: event=member-join server=Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db.dc1 area=wan writer.go:29: 2021-01-29T19:31:24.749Z [INFO] TestAgentTokensCommand: started state syncer writer.go:29: 2021-01-29T19:31:24.783Z [WARN] TestAgentTokensCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:24.785Z [INFO] TestAgentTokensCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:15111 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:24.786Z [DEBUG] TestAgentTokensCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:24.787Z [DEBUG] TestAgentTokensCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:24.790Z [DEBUG] TestAgentTokensCommand.server.raft: vote granted: from=5f8b86a6-d596-04d0-2d0c-ebc584b957db term=2 tally=1 writer.go:29: 2021-01-29T19:31:24.792Z [INFO] TestAgentTokensCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:24.793Z [INFO] TestAgentTokensCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:15111 [Leader]" writer.go:29: 2021-01-29T19:31:24.795Z [INFO] TestAgentTokensCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:24.796Z [INFO] TestAgentTokensCommand.server: New leader elected: payload=Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db writer.go:29: 2021-01-29T19:31:24.799Z [INFO] TestAgentTokensCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:24.803Z [INFO] TestAgentTokensCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:24.804Z [WARN] TestAgentTokensCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:24.807Z [INFO] TestAgentTokensCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:24.809Z [INFO] TestAgentTokensCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:24.811Z [INFO] TestAgentTokensCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:24.819Z [INFO] TestAgentTokensCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:24.820Z [INFO] TestAgentTokensCommand.server.serf.lan: serf: EventMemberUpdate: Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db writer.go:29: 2021-01-29T19:31:24.823Z [INFO] TestAgentTokensCommand.server.serf.wan: serf: EventMemberUpdate: Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db.dc1 writer.go:29: 2021-01-29T19:31:24.823Z [INFO] TestAgentTokensCommand.server: Updating LAN server: server="Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db (Addr: tcp/127.0.0.1:15111) (DC: dc1)" writer.go:29: 2021-01-29T19:31:24.825Z [INFO] TestAgentTokensCommand.server: Handled event for server in area: event=member-update server=Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db.dc1 area=wan writer.go:29: 2021-01-29T19:31:24.826Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:24.876Z [INFO] TestAgentTokensCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:24.878Z [INFO] TestAgentTokensCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:24.879Z [DEBUG] TestAgentTokensCommand.server: Skipping self join check for node since the cluster is too small: node=Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db writer.go:29: 2021-01-29T19:31:24.880Z [INFO] TestAgentTokensCommand.server: member joined, marking health alive: member=Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db writer.go:29: 2021-01-29T19:31:24.882Z [DEBUG] TestAgentTokensCommand.server: Skipping self join check for node since the cluster is too small: node=Node-5f8b86a6-d596-04d0-2d0c-ebc584b957db writer.go:29: 2021-01-29T19:31:24.910Z [DEBUG] TestAgentTokensCommand.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:35962 latency=3.884441ms writer.go:29: 2021-01-29T19:31:24.960Z [INFO] TestAgentTokensCommand: Updated agent's ACL token: token=default writer.go:29: 2021-01-29T19:31:24.964Z [DEBUG] TestAgentTokensCommand.http: Request finished: method=PUT url=/v1/agent/token/default from=127.0.0.1:35964 latency=5.635647ms writer.go:29: 2021-01-29T19:31:24.972Z [INFO] TestAgentTokensCommand: Updated agent's ACL token: token=agent writer.go:29: 2021-01-29T19:31:24.973Z [DEBUG] TestAgentTokensCommand.http: Request finished: method=PUT url=/v1/agent/token/agent from=127.0.0.1:35966 latency=1.585347ms writer.go:29: 2021-01-29T19:31:24.978Z [INFO] TestAgentTokensCommand: Updated agent's ACL token: token=agent_master writer.go:29: 2021-01-29T19:31:24.980Z [DEBUG] TestAgentTokensCommand.http: Request finished: method=PUT url=/v1/agent/token/agent_master from=127.0.0.1:35968 latency=1.367057ms writer.go:29: 2021-01-29T19:31:24.984Z [INFO] TestAgentTokensCommand: Updated agent's ACL token: token=replication writer.go:29: 2021-01-29T19:31:24.988Z [DEBUG] TestAgentTokensCommand.http: Request finished: method=PUT url=/v1/agent/token/replication from=127.0.0.1:35970 latency=3.897561ms writer.go:29: 2021-01-29T19:31:24.996Z [INFO] TestAgentTokensCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:24.997Z [INFO] TestAgentTokensCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:24.999Z [DEBUG] TestAgentTokensCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:25.001Z [DEBUG] TestAgentTokensCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:25.003Z [DEBUG] TestAgentTokensCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:25.001Z [ERROR] TestAgentTokensCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:25.001Z [DEBUG] TestAgentTokensCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:25.003Z [DEBUG] TestAgentTokensCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:25.005Z [WARN] TestAgentTokensCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:25.005Z [DEBUG] TestAgentTokensCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:25.011Z [WARN] TestAgentTokensCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:25.016Z [INFO] TestAgentTokensCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:25.016Z [INFO] TestAgentTokensCommand: consul server down writer.go:29: 2021-01-29T19:31:25.020Z [INFO] TestAgentTokensCommand: shutdown complete writer.go:29: 2021-01-29T19:31:25.021Z [INFO] TestAgentTokensCommand: Stopping server: protocol=DNS address=127.0.0.1:15106 network=tcp writer.go:29: 2021-01-29T19:31:25.023Z [INFO] TestAgentTokensCommand: Stopping server: protocol=DNS address=127.0.0.1:15106 network=udp writer.go:29: 2021-01-29T19:31:25.025Z [INFO] TestAgentTokensCommand: Stopping server: protocol=HTTP address=127.0.0.1:15107 network=tcp writer.go:29: 2021-01-29T19:31:25.527Z [INFO] TestAgentTokensCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:25.528Z [INFO] TestAgentTokensCommand: Endpoints down --- PASS: TestAgentTokensCommand (0.87s) PASS ok github.com/hashicorp/consul/command/acl/agenttokens 1.078s ? github.com/hashicorp/consul/command/acl/authmethod [no test files] === RUN TestAuthMethodCreateCommand_noTabs === PAUSE TestAuthMethodCreateCommand_noTabs === RUN TestAuthMethodCreateCommand === PAUSE TestAuthMethodCreateCommand === RUN TestAuthMethodCreateCommand_JSON === PAUSE TestAuthMethodCreateCommand_JSON === RUN TestAuthMethodCreateCommand_k8s === PAUSE TestAuthMethodCreateCommand_k8s === CONT TestAuthMethodCreateCommand_noTabs --- PASS: TestAuthMethodCreateCommand_noTabs (0.00s) === CONT TestAuthMethodCreateCommand_k8s === CONT TestAuthMethodCreateCommand_JSON === CONT TestAuthMethodCreateCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.116Z [WARN] TestAuthMethodCreateCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:26.118Z [DEBUG] TestAuthMethodCreateCommand_JSON.tlsutil: Update: version=1 === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.123Z [WARN] TestAuthMethodCreateCommand: bootstrap = true: do not enable unless necessary === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.121Z [DEBUG] TestAuthMethodCreateCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.128Z [DEBUG] TestAuthMethodCreateCommand.tlsutil: Update: version=1 === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.131Z [INFO] TestAuthMethodCreateCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b648673e-3912-36cf-454f-6bbd8f779f36 Address:127.0.0.1:31459}]" === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.131Z [DEBUG] TestAuthMethodCreateCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.133Z [WARN] TestAuthMethodCreateCommand_k8s: bootstrap = true: do not enable unless necessary === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.133Z [INFO] TestAuthMethodCreateCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-b648673e-3912-36cf-454f-6bbd8f779f36.dc1 127.0.0.1 === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.138Z [INFO] TestAuthMethodCreateCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f78cf2df-715b-d46a-9035-a76977fda77d Address:127.0.0.1:31453}]" === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.135Z [INFO] TestAuthMethodCreateCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:31459 [Follower]" leader= === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.143Z [INFO] TestAuthMethodCreateCommand.server.serf.wan: serf: EventMemberJoin: Node-f78cf2df-715b-d46a-9035-a76977fda77d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:26.146Z [INFO] TestAuthMethodCreateCommand.server.serf.lan: serf: EventMemberJoin: Node-f78cf2df-715b-d46a-9035-a76977fda77d 127.0.0.1 writer.go:29: 2021-01-29T19:31:26.151Z [INFO] TestAuthMethodCreateCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:31453 [Follower]" leader= === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.152Z [DEBUG] TestAuthMethodCreateCommand_k8s.tlsutil: Update: version=1 === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.151Z [INFO] TestAuthMethodCreateCommand: Started DNS server: address=127.0.0.1:31448 network=udp === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.154Z [DEBUG] TestAuthMethodCreateCommand_k8s.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.143Z [INFO] TestAuthMethodCreateCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-b648673e-3912-36cf-454f-6bbd8f779f36 127.0.0.1 writer.go:29: 2021-01-29T19:31:26.160Z [INFO] TestAuthMethodCreateCommand_JSON: Started DNS server: address=127.0.0.1:31454 network=udp writer.go:29: 2021-01-29T19:31:26.161Z [INFO] TestAuthMethodCreateCommand_JSON.server: Handled event for server in area: event=member-join server=Node-b648673e-3912-36cf-454f-6bbd8f779f36.dc1 area=wan writer.go:29: 2021-01-29T19:31:26.163Z [INFO] TestAuthMethodCreateCommand_JSON.server: Adding LAN server: server="Node-b648673e-3912-36cf-454f-6bbd8f779f36 (Addr: tcp/127.0.0.1:31459) (DC: dc1)" === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.166Z [INFO] TestAuthMethodCreateCommand.server: Adding LAN server: server="Node-f78cf2df-715b-d46a-9035-a76977fda77d (Addr: tcp/127.0.0.1:31453) (DC: dc1)" writer.go:29: 2021-01-29T19:31:26.171Z [INFO] TestAuthMethodCreateCommand.server: Handled event for server in area: event=member-join server=Node-f78cf2df-715b-d46a-9035-a76977fda77d.dc1 area=wan === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.174Z [INFO] TestAuthMethodCreateCommand_JSON: Started DNS server: address=127.0.0.1:31454 network=tcp writer.go:29: 2021-01-29T19:31:26.180Z [INFO] TestAuthMethodCreateCommand_JSON: Started HTTP server: address=127.0.0.1:31455 network=tcp === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.181Z [INFO] TestAuthMethodCreateCommand: Started DNS server: address=127.0.0.1:31448 network=tcp === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.182Z [INFO] TestAuthMethodCreateCommand_JSON: started state syncer === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.187Z [INFO] TestAuthMethodCreateCommand: Started HTTP server: address=127.0.0.1:31449 network=tcp writer.go:29: 2021-01-29T19:31:26.196Z [INFO] TestAuthMethodCreateCommand: started state syncer writer.go:29: 2021-01-29T19:31:26.194Z [WARN] TestAuthMethodCreateCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:26.199Z [INFO] TestAuthMethodCreateCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:31453 [Candidate]" term=2 === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.202Z [WARN] TestAuthMethodCreateCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:26.203Z [INFO] TestAuthMethodCreateCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:31459 [Candidate]" term=2 === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.210Z [DEBUG] TestAuthMethodCreateCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.210Z [INFO] TestAuthMethodCreateCommand_k8s.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ef969aab-1304-dad5-93de-f96294b3232d Address:127.0.0.1:31447}]" === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.211Z [DEBUG] TestAuthMethodCreateCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:26.212Z [DEBUG] TestAuthMethodCreateCommand_JSON.server.raft: vote granted: from=b648673e-3912-36cf-454f-6bbd8f779f36 term=2 tally=1 writer.go:29: 2021-01-29T19:31:26.213Z [INFO] TestAuthMethodCreateCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:26.214Z [INFO] TestAuthMethodCreateCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:31459 [Leader]" === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.214Z [INFO] TestAuthMethodCreateCommand_k8s.server.serf.wan: serf: EventMemberJoin: Node-ef969aab-1304-dad5-93de-f96294b3232d.dc1 127.0.0.1 === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.216Z [INFO] TestAuthMethodCreateCommand_JSON.server: cluster leadership acquired === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.217Z [INFO] TestAuthMethodCreateCommand_k8s.server.raft: entering follower state: follower="Node at 127.0.0.1:31447 [Follower]" leader= === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.218Z [DEBUG] TestAuthMethodCreateCommand.server.raft: votes: needed=1 === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.219Z [INFO] TestAuthMethodCreateCommand_JSON.server: New leader elected: payload=Node-b648673e-3912-36cf-454f-6bbd8f779f36 === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.220Z [DEBUG] TestAuthMethodCreateCommand.server.raft: vote granted: from=f78cf2df-715b-d46a-9035-a76977fda77d term=2 tally=1 === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.221Z [INFO] TestAuthMethodCreateCommand_k8s.server.serf.lan: serf: EventMemberJoin: Node-ef969aab-1304-dad5-93de-f96294b3232d 127.0.0.1 === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.221Z [INFO] TestAuthMethodCreateCommand_JSON.server: initializing acls === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.223Z [INFO] TestAuthMethodCreateCommand.server.raft: election won: tally=1 === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.224Z [INFO] TestAuthMethodCreateCommand_k8s: Started DNS server: address=127.0.0.1:31442 network=udp === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.223Z [INFO] TestAuthMethodCreateCommand_JSON.server: initializing acls === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.224Z [INFO] TestAuthMethodCreateCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:31453 [Leader]" writer.go:29: 2021-01-29T19:31:26.226Z [INFO] TestAuthMethodCreateCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:26.235Z [INFO] TestAuthMethodCreateCommand.server: New leader elected: payload=Node-f78cf2df-715b-d46a-9035-a76977fda77d writer.go:29: 2021-01-29T19:31:26.236Z [INFO] TestAuthMethodCreateCommand.server: initializing acls === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.236Z [INFO] TestAuthMethodCreateCommand_JSON.server: Created ACL 'global-management' policy === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.235Z [INFO] TestAuthMethodCreateCommand_k8s.server: Handled event for server in area: event=member-join server=Node-ef969aab-1304-dad5-93de-f96294b3232d.dc1 area=wan === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.237Z [WARN] TestAuthMethodCreateCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.239Z [INFO] TestAuthMethodCreateCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:26.240Z [WARN] TestAuthMethodCreateCommand.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.235Z [INFO] TestAuthMethodCreateCommand_k8s: Started DNS server: address=127.0.0.1:31442 network=tcp === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.236Z [INFO] TestAuthMethodCreateCommand_JSON.server: Created ACL 'global-management' policy === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.243Z [INFO] TestAuthMethodCreateCommand.server: Bootstrapped ACL master token from configuration === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.235Z [INFO] TestAuthMethodCreateCommand_k8s.server: Adding LAN server: server="Node-ef969aab-1304-dad5-93de-f96294b3232d (Addr: tcp/127.0.0.1:31447) (DC: dc1)" writer.go:29: 2021-01-29T19:31:26.245Z [INFO] TestAuthMethodCreateCommand_k8s: Started HTTP server: address=127.0.0.1:31443 network=tcp === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.245Z [INFO] TestAuthMethodCreateCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:26.246Z [INFO] TestAuthMethodCreateCommand.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.246Z [INFO] TestAuthMethodCreateCommand_k8s: started state syncer === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.247Z [INFO] TestAuthMethodCreateCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:26.248Z [INFO] TestAuthMethodCreateCommand.server.serf.lan: serf: EventMemberUpdate: Node-f78cf2df-715b-d46a-9035-a76977fda77d writer.go:29: 2021-01-29T19:31:26.249Z [INFO] TestAuthMethodCreateCommand.server.serf.wan: serf: EventMemberUpdate: Node-f78cf2df-715b-d46a-9035-a76977fda77d.dc1 writer.go:29: 2021-01-29T19:31:26.251Z [INFO] TestAuthMethodCreateCommand.server: Updating LAN server: server="Node-f78cf2df-715b-d46a-9035-a76977fda77d (Addr: tcp/127.0.0.1:31453) (DC: dc1)" writer.go:29: 2021-01-29T19:31:26.251Z [INFO] TestAuthMethodCreateCommand.server: Handled event for server in area: event=member-update server=Node-f78cf2df-715b-d46a-9035-a76977fda77d.dc1 area=wan writer.go:29: 2021-01-29T19:31:26.258Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.243Z [WARN] TestAuthMethodCreateCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:26.241Z [INFO] TestAuthMethodCreateCommand_JSON.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:26.255Z [ERROR] TestAuthMethodCreateCommand_JSON.anti_entropy: failed to sync remote state: error="ACL not found" writer.go:29: 2021-01-29T19:31:26.267Z [INFO] TestAuthMethodCreateCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:26.269Z [INFO] TestAuthMethodCreateCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:26.268Z [INFO] TestAuthMethodCreateCommand_JSON.server: Created ACL anonymous token from configuration === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.272Z [INFO] TestAuthMethodCreateCommand: Synced node info writer.go:29: 2021-01-29T19:31:26.273Z [DEBUG] TestAuthMethodCreateCommand: Node info in sync === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.271Z [INFO] TestAuthMethodCreateCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:26.276Z [DEBUG] TestAuthMethodCreateCommand_JSON.server: transitioning out of legacy ACL mode === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.280Z [WARN] TestAuthMethodCreateCommand_k8s.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.276Z [INFO] TestAuthMethodCreateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-b648673e-3912-36cf-454f-6bbd8f779f36 === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.280Z [INFO] TestAuthMethodCreateCommand_k8s.server.raft: entering candidate state: node="Node at 127.0.0.1:31447 [Candidate]" term=2 === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.281Z [INFO] TestAuthMethodCreateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-b648673e-3912-36cf-454f-6bbd8f779f36.dc1 writer.go:29: 2021-01-29T19:31:26.281Z [INFO] TestAuthMethodCreateCommand_JSON.server: Updating LAN server: server="Node-b648673e-3912-36cf-454f-6bbd8f779f36 (Addr: tcp/127.0.0.1:31459) (DC: dc1)" === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.283Z [DEBUG] TestAuthMethodCreateCommand_k8s.server.raft: votes: needed=1 === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.282Z [INFO] TestAuthMethodCreateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-b648673e-3912-36cf-454f-6bbd8f779f36 writer.go:29: 2021-01-29T19:31:26.285Z [INFO] TestAuthMethodCreateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-b648673e-3912-36cf-454f-6bbd8f779f36.dc1 area=wan === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.286Z [DEBUG] TestAuthMethodCreateCommand_k8s.server.raft: vote granted: from=ef969aab-1304-dad5-93de-f96294b3232d term=2 tally=1 === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.288Z [INFO] TestAuthMethodCreateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-b648673e-3912-36cf-454f-6bbd8f779f36.dc1 writer.go:29: 2021-01-29T19:31:26.288Z [INFO] TestAuthMethodCreateCommand_JSON.server: Updating LAN server: server="Node-b648673e-3912-36cf-454f-6bbd8f779f36 (Addr: tcp/127.0.0.1:31459) (DC: dc1)" writer.go:29: 2021-01-29T19:31:26.289Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === RUN TestAuthMethodCreateCommand/type_required === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.290Z [INFO] TestAuthMethodCreateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-b648673e-3912-36cf-454f-6bbd8f779f36.dc1 area=wan === RUN TestAuthMethodCreateCommand/name_required === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.289Z [INFO] TestAuthMethodCreateCommand_k8s.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:26.285Z [DEBUG] TestAuthMethodCreateCommand_k8s.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === RUN TestAuthMethodCreateCommand/invalid_type === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.299Z [INFO] TestAuthMethodCreateCommand_k8s.server.raft: entering leader state: leader="Node at 127.0.0.1:31447 [Leader]" writer.go:29: 2021-01-29T19:31:26.301Z [INFO] TestAuthMethodCreateCommand_k8s.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:26.303Z [INFO] TestAuthMethodCreateCommand_k8s.server: New leader elected: payload=Node-ef969aab-1304-dad5-93de-f96294b3232d writer.go:29: 2021-01-29T19:31:26.303Z [INFO] TestAuthMethodCreateCommand_k8s.server: initializing acls === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.304Z [ERROR] TestAuthMethodCreateCommand.http: Request error: method=PUT url=/v1/acl/auth-method from=127.0.0.1:39212 error="Invalid Auth Method: Type should be one of: [kubernetes testing]" writer.go:29: 2021-01-29T19:31:26.307Z [DEBUG] TestAuthMethodCreateCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:39212 latency=4.237906ms === RUN TestAuthMethodCreateCommand/create_testing === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.311Z [INFO] TestAuthMethodCreateCommand_k8s.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:26.312Z [WARN] TestAuthMethodCreateCommand_k8s.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:26.313Z [INFO] TestAuthMethodCreateCommand_k8s.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:26.316Z [INFO] TestAuthMethodCreateCommand_k8s.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:26.318Z [INFO] TestAuthMethodCreateCommand_k8s.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:26.320Z [INFO] TestAuthMethodCreateCommand_k8s.leader: started routine: routine="acl token reaping" === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.327Z [DEBUG] TestAuthMethodCreateCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:39214 latency=9.618186ms === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.322Z [INFO] TestAuthMethodCreateCommand_k8s.server.serf.lan: serf: EventMemberUpdate: Node-ef969aab-1304-dad5-93de-f96294b3232d writer.go:29: 2021-01-29T19:31:26.329Z [INFO] TestAuthMethodCreateCommand_k8s.server.serf.wan: serf: EventMemberUpdate: Node-ef969aab-1304-dad5-93de-f96294b3232d.dc1 === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.329Z [INFO] TestAuthMethodCreateCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:26.331Z [INFO] TestAuthMethodCreateCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.332Z [INFO] TestAuthMethodCreateCommand_k8s.server: Updating LAN server: server="Node-ef969aab-1304-dad5-93de-f96294b3232d (Addr: tcp/127.0.0.1:31447) (DC: dc1)" === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.332Z [DEBUG] TestAuthMethodCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-b648673e-3912-36cf-454f-6bbd8f779f36 === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.333Z [INFO] TestAuthMethodCreateCommand_k8s.server: Handled event for server in area: event=member-update server=Node-ef969aab-1304-dad5-93de-f96294b3232d.dc1 area=wan === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.337Z [INFO] TestAuthMethodCreateCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:26.339Z [INFO] TestAuthMethodCreateCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:26.341Z [DEBUG] TestAuthMethodCreateCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:26.343Z [DEBUG] TestAuthMethodCreateCommand.leader: stopping routine: routine="acl token reaping" === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.334Z [INFO] TestAuthMethodCreateCommand_JSON.server: member joined, marking health alive: member=Node-b648673e-3912-36cf-454f-6bbd8f779f36 writer.go:29: 2021-01-29T19:31:26.354Z [DEBUG] TestAuthMethodCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-b648673e-3912-36cf-454f-6bbd8f779f36 === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.348Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.355Z [DEBUG] TestAuthMethodCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-b648673e-3912-36cf-454f-6bbd8f779f36 === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.347Z [WARN] TestAuthMethodCreateCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:26.347Z [DEBUG] TestAuthMethodCreateCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:26.353Z [DEBUG] TestAuthMethodCreateCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:26.356Z [INFO] TestAuthMethodCreateCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:26.364Z [INFO] TestAuthMethodCreateCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.365Z [DEBUG] TestAuthMethodCreateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-f78cf2df-715b-d46a-9035-a76977fda77d writer.go:29: 2021-01-29T19:31:26.367Z [INFO] TestAuthMethodCreateCommand.server: member joined, marking health alive: member=Node-f78cf2df-715b-d46a-9035-a76977fda77d writer.go:29: 2021-01-29T19:31:26.371Z [WARN] TestAuthMethodCreateCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:26.374Z [INFO] TestAuthMethodCreateCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:26.375Z [DEBUG] TestAuthMethodCreateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-f78cf2df-715b-d46a-9035-a76977fda77d writer.go:29: 2021-01-29T19:31:26.377Z [DEBUG] TestAuthMethodCreateCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.375Z [INFO] TestAuthMethodCreateCommand: consul server down writer.go:29: 2021-01-29T19:31:26.378Z [INFO] TestAuthMethodCreateCommand: shutdown complete === RUN TestAuthMethodCreateCommand_JSON/type_required === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.387Z [INFO] TestAuthMethodCreateCommand_k8s.server.connect: initialized primary datacenter CA with provider: provider=consul === RUN TestAuthMethodCreateCommand_JSON/create_testing === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.388Z [INFO] TestAuthMethodCreateCommand_k8s.leader: started routine: routine="CA root pruning" === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.382Z [DEBUG] TestAuthMethodCreateCommand.leader: stopped routine: routine="CA root pruning" === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.389Z [DEBUG] TestAuthMethodCreateCommand_k8s.server: Skipping self join check for node since the cluster is too small: node=Node-ef969aab-1304-dad5-93de-f96294b3232d === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.387Z [INFO] TestAuthMethodCreateCommand: Stopping server: protocol=DNS address=127.0.0.1:31448 network=tcp === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.392Z [INFO] TestAuthMethodCreateCommand_k8s.server: member joined, marking health alive: member=Node-ef969aab-1304-dad5-93de-f96294b3232d writer.go:29: 2021-01-29T19:31:26.402Z [DEBUG] TestAuthMethodCreateCommand_k8s.server: Skipping self join check for node since the cluster is too small: node=Node-ef969aab-1304-dad5-93de-f96294b3232d === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.401Z [INFO] TestAuthMethodCreateCommand: Stopping server: protocol=DNS address=127.0.0.1:31448 network=udp writer.go:29: 2021-01-29T19:31:26.406Z [INFO] TestAuthMethodCreateCommand: Stopping server: protocol=HTTP address=127.0.0.1:31449 network=tcp === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.423Z [DEBUG] TestAuthMethodCreateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:44768 latency=11.669683ms writer.go:29: 2021-01-29T19:31:26.429Z [INFO] TestAuthMethodCreateCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:31:26.437Z [INFO] TestAuthMethodCreateCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:31:26.438Z [DEBUG] TestAuthMethodCreateCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:26.439Z [DEBUG] TestAuthMethodCreateCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:26.441Z [DEBUG] TestAuthMethodCreateCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.440Z [DEBUG] TestAuthMethodCreateCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:26.441Z [DEBUG] TestAuthMethodCreateCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:26.442Z [DEBUG] TestAuthMethodCreateCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.442Z [WARN] TestAuthMethodCreateCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:26.453Z [WARN] TestAuthMethodCreateCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:26.461Z [INFO] TestAuthMethodCreateCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:26.461Z [INFO] TestAuthMethodCreateCommand_JSON: consul server down writer.go:29: 2021-01-29T19:31:26.464Z [INFO] TestAuthMethodCreateCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:31:26.465Z [INFO] TestAuthMethodCreateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:31454 network=tcp writer.go:29: 2021-01-29T19:31:26.466Z [INFO] TestAuthMethodCreateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:31454 network=udp writer.go:29: 2021-01-29T19:31:26.468Z [INFO] TestAuthMethodCreateCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:31455 network=tcp === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.496Z [DEBUG] TestAuthMethodCreateCommand_k8s: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:26.501Z [INFO] TestAuthMethodCreateCommand_k8s: Synced node info === RUN TestAuthMethodCreateCommand_k8s/k8s_host_required === RUN TestAuthMethodCreateCommand_k8s/k8s_ca_cert_required === RUN TestAuthMethodCreateCommand_k8s/k8s_jwt_required === RUN TestAuthMethodCreateCommand_k8s/create_k8s === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.584Z [DEBUG] TestAuthMethodCreateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:56402 latency=5.658147ms === RUN TestAuthMethodCreateCommand_k8s/create_k8s_with_cert_file === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:26.598Z [DEBUG] TestAuthMethodCreateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:56404 latency=3.278382ms writer.go:29: 2021-01-29T19:31:26.602Z [INFO] TestAuthMethodCreateCommand_k8s: Requesting shutdown writer.go:29: 2021-01-29T19:31:26.603Z [INFO] TestAuthMethodCreateCommand_k8s.server: shutting down server writer.go:29: 2021-01-29T19:31:26.604Z [DEBUG] TestAuthMethodCreateCommand_k8s.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:26.606Z [DEBUG] TestAuthMethodCreateCommand_k8s.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:26.606Z [DEBUG] TestAuthMethodCreateCommand_k8s.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.606Z [DEBUG] TestAuthMethodCreateCommand_k8s.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:26.607Z [DEBUG] TestAuthMethodCreateCommand_k8s.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:26.607Z [WARN] TestAuthMethodCreateCommand_k8s.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:26.607Z [DEBUG] TestAuthMethodCreateCommand_k8s.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:26.611Z [WARN] TestAuthMethodCreateCommand_k8s.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:26.613Z [INFO] TestAuthMethodCreateCommand_k8s.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:26.613Z [INFO] TestAuthMethodCreateCommand_k8s: consul server down writer.go:29: 2021-01-29T19:31:26.615Z [INFO] TestAuthMethodCreateCommand_k8s: shutdown complete writer.go:29: 2021-01-29T19:31:26.616Z [INFO] TestAuthMethodCreateCommand_k8s: Stopping server: protocol=DNS address=127.0.0.1:31442 network=tcp writer.go:29: 2021-01-29T19:31:26.617Z [INFO] TestAuthMethodCreateCommand_k8s: Stopping server: protocol=DNS address=127.0.0.1:31442 network=udp writer.go:29: 2021-01-29T19:31:26.618Z [INFO] TestAuthMethodCreateCommand_k8s: Stopping server: protocol=HTTP address=127.0.0.1:31443 network=tcp === CONT TestAuthMethodCreateCommand writer.go:29: 2021-01-29T19:31:26.912Z [INFO] TestAuthMethodCreateCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:26.914Z [INFO] TestAuthMethodCreateCommand: Endpoints down --- PASS: TestAuthMethodCreateCommand (0.90s) --- PASS: TestAuthMethodCreateCommand/type_required (0.00s) --- PASS: TestAuthMethodCreateCommand/name_required (0.00s) --- PASS: TestAuthMethodCreateCommand/invalid_type (0.01s) --- PASS: TestAuthMethodCreateCommand/create_testing (0.03s) === CONT TestAuthMethodCreateCommand_JSON writer.go:29: 2021-01-29T19:31:26.969Z [INFO] TestAuthMethodCreateCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:26.981Z [INFO] TestAuthMethodCreateCommand_JSON: Endpoints down --- PASS: TestAuthMethodCreateCommand_JSON (0.97s) --- PASS: TestAuthMethodCreateCommand_JSON/type_required (0.00s) --- PASS: TestAuthMethodCreateCommand_JSON/create_testing (0.04s) === CONT TestAuthMethodCreateCommand_k8s writer.go:29: 2021-01-29T19:31:27.120Z [INFO] TestAuthMethodCreateCommand_k8s: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:27.122Z [INFO] TestAuthMethodCreateCommand_k8s: Endpoints down --- PASS: TestAuthMethodCreateCommand_k8s (1.11s) --- PASS: TestAuthMethodCreateCommand_k8s/k8s_host_required (0.00s) --- PASS: TestAuthMethodCreateCommand_k8s/k8s_ca_cert_required (0.00s) --- PASS: TestAuthMethodCreateCommand_k8s/k8s_jwt_required (0.00s) --- PASS: TestAuthMethodCreateCommand_k8s/create_k8s (0.02s) --- PASS: TestAuthMethodCreateCommand_k8s/create_k8s_with_cert_file (0.01s) PASS ok github.com/hashicorp/consul/command/acl/authmethod/create 1.331s === RUN TestAuthMethodDeleteCommand_noTabs === PAUSE TestAuthMethodDeleteCommand_noTabs === RUN TestAuthMethodDeleteCommand === PAUSE TestAuthMethodDeleteCommand === CONT TestAuthMethodDeleteCommand_noTabs === CONT TestAuthMethodDeleteCommand --- PASS: TestAuthMethodDeleteCommand_noTabs (0.00s) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestAuthMethodDeleteCommand writer.go:29: 2021-01-29T19:31:30.339Z [WARN] TestAuthMethodDeleteCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:30.341Z [DEBUG] TestAuthMethodDeleteCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:30.343Z [DEBUG] TestAuthMethodDeleteCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:30.348Z [INFO] TestAuthMethodDeleteCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bf79ec87-1aa6-aca3-1be2-b55c6d6510ae Address:127.0.0.1:28384}]" writer.go:29: 2021-01-29T19:31:30.350Z [INFO] TestAuthMethodDeleteCommand.server.serf.wan: serf: EventMemberJoin: Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:30.351Z [INFO] TestAuthMethodDeleteCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:28384 [Follower]" leader= writer.go:29: 2021-01-29T19:31:30.355Z [INFO] TestAuthMethodDeleteCommand.server.serf.lan: serf: EventMemberJoin: Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae 127.0.0.1 writer.go:29: 2021-01-29T19:31:30.364Z [INFO] TestAuthMethodDeleteCommand.server: Adding LAN server: server="Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae (Addr: tcp/127.0.0.1:28384) (DC: dc1)" writer.go:29: 2021-01-29T19:31:30.365Z [INFO] TestAuthMethodDeleteCommand.server: Handled event for server in area: event=member-join server=Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae.dc1 area=wan writer.go:29: 2021-01-29T19:31:30.368Z [INFO] TestAuthMethodDeleteCommand: Started DNS server: address=127.0.0.1:28379 network=udp writer.go:29: 2021-01-29T19:31:30.369Z [INFO] TestAuthMethodDeleteCommand: Started DNS server: address=127.0.0.1:28379 network=tcp writer.go:29: 2021-01-29T19:31:30.371Z [INFO] TestAuthMethodDeleteCommand: Started HTTP server: address=127.0.0.1:28380 network=tcp writer.go:29: 2021-01-29T19:31:30.373Z [INFO] TestAuthMethodDeleteCommand: started state syncer writer.go:29: 2021-01-29T19:31:30.394Z [WARN] TestAuthMethodDeleteCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:30.396Z [INFO] TestAuthMethodDeleteCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:28384 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:30.412Z [DEBUG] TestAuthMethodDeleteCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:30.414Z [DEBUG] TestAuthMethodDeleteCommand.server.raft: vote granted: from=bf79ec87-1aa6-aca3-1be2-b55c6d6510ae term=2 tally=1 writer.go:29: 2021-01-29T19:31:30.419Z [INFO] TestAuthMethodDeleteCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:30.417Z [DEBUG] TestAuthMethodDeleteCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:30.423Z [INFO] TestAuthMethodDeleteCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:28384 [Leader]" writer.go:29: 2021-01-29T19:31:30.432Z [INFO] TestAuthMethodDeleteCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:30.433Z [INFO] TestAuthMethodDeleteCommand.server: New leader elected: payload=Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae writer.go:29: 2021-01-29T19:31:30.434Z [INFO] TestAuthMethodDeleteCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:30.447Z [INFO] TestAuthMethodDeleteCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:30.449Z [WARN] TestAuthMethodDeleteCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:30.472Z [INFO] TestAuthMethodDeleteCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:30.475Z [INFO] TestAuthMethodDeleteCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:30.477Z [INFO] TestAuthMethodDeleteCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:30.480Z [INFO] TestAuthMethodDeleteCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:30.483Z [INFO] TestAuthMethodDeleteCommand.server.serf.lan: serf: EventMemberUpdate: Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae writer.go:29: 2021-01-29T19:31:30.486Z [INFO] TestAuthMethodDeleteCommand.server: Updating LAN server: server="Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae (Addr: tcp/127.0.0.1:28384) (DC: dc1)" writer.go:29: 2021-01-29T19:31:30.486Z [INFO] TestAuthMethodDeleteCommand.server.serf.wan: serf: EventMemberUpdate: Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae.dc1 writer.go:29: 2021-01-29T19:31:30.489Z [INFO] TestAuthMethodDeleteCommand.server: Handled event for server in area: event=member-update server=Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae.dc1 area=wan writer.go:29: 2021-01-29T19:31:30.490Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:30.568Z [INFO] TestAuthMethodDeleteCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:30.570Z [INFO] TestAuthMethodDeleteCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:30.572Z [DEBUG] TestAuthMethodDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae writer.go:29: 2021-01-29T19:31:30.576Z [INFO] TestAuthMethodDeleteCommand.server: member joined, marking health alive: member=Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae writer.go:29: 2021-01-29T19:31:30.589Z [DEBUG] TestAuthMethodDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-bf79ec87-1aa6-aca3-1be2-b55c6d6510ae === RUN TestAuthMethodDeleteCommand/name_required === RUN TestAuthMethodDeleteCommand/delete_notfound === CONT TestAuthMethodDeleteCommand writer.go:29: 2021-01-29T19:31:30.630Z [DEBUG] TestAuthMethodDeleteCommand.http: Request finished: method=DELETE url=/v1/acl/auth-method/notfound from=127.0.0.1:58378 latency=6.802112ms === RUN TestAuthMethodDeleteCommand/delete_works === CONT TestAuthMethodDeleteCommand writer.go:29: 2021-01-29T19:31:30.661Z [DEBUG] TestAuthMethodDeleteCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:58380 latency=5.393824ms writer.go:29: 2021-01-29T19:31:30.668Z [DEBUG] TestAuthMethodDeleteCommand.http: Request finished: method=DELETE url=/v1/acl/auth-method/test-e7fb2160-2190-f861-4465-4d47ff5649e0 from=127.0.0.1:58382 latency=2.370194ms writer.go:29: 2021-01-29T19:31:30.672Z [DEBUG] TestAuthMethodDeleteCommand.http: Request finished: method=GET url=/v1/acl/auth-method/test-e7fb2160-2190-f861-4465-4d47ff5649e0 from=127.0.0.1:58380 latency=49.871µs writer.go:29: 2021-01-29T19:31:30.675Z [INFO] TestAuthMethodDeleteCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:30.685Z [INFO] TestAuthMethodDeleteCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:30.687Z [DEBUG] TestAuthMethodDeleteCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:30.689Z [DEBUG] TestAuthMethodDeleteCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:30.687Z [ERROR] TestAuthMethodDeleteCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:30.689Z [DEBUG] TestAuthMethodDeleteCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:30.690Z [DEBUG] TestAuthMethodDeleteCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:30.690Z [DEBUG] TestAuthMethodDeleteCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:30.693Z [DEBUG] TestAuthMethodDeleteCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:30.693Z [WARN] TestAuthMethodDeleteCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:30.701Z [WARN] TestAuthMethodDeleteCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:30.704Z [INFO] TestAuthMethodDeleteCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:30.704Z [INFO] TestAuthMethodDeleteCommand: consul server down writer.go:29: 2021-01-29T19:31:30.706Z [INFO] TestAuthMethodDeleteCommand: shutdown complete writer.go:29: 2021-01-29T19:31:30.707Z [INFO] TestAuthMethodDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:28379 network=tcp writer.go:29: 2021-01-29T19:31:30.708Z [INFO] TestAuthMethodDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:28379 network=udp writer.go:29: 2021-01-29T19:31:30.709Z [INFO] TestAuthMethodDeleteCommand: Stopping server: protocol=HTTP address=127.0.0.1:28380 network=tcp writer.go:29: 2021-01-29T19:31:31.210Z [INFO] TestAuthMethodDeleteCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:31.212Z [INFO] TestAuthMethodDeleteCommand: Endpoints down --- PASS: TestAuthMethodDeleteCommand (0.92s) --- PASS: TestAuthMethodDeleteCommand/name_required (0.00s) --- PASS: TestAuthMethodDeleteCommand/delete_notfound (0.02s) --- PASS: TestAuthMethodDeleteCommand/delete_works (0.03s) PASS ok github.com/hashicorp/consul/command/acl/authmethod/delete 1.045s === RUN TestAuthMethodListCommand_noTabs === PAUSE TestAuthMethodListCommand_noTabs === RUN TestAuthMethodListCommand === PAUSE TestAuthMethodListCommand === RUN TestAuthMethodListCommand_JSON === PAUSE TestAuthMethodListCommand_JSON === CONT TestAuthMethodListCommand_noTabs === CONT TestAuthMethodListCommand_JSON === CONT TestAuthMethodListCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestAuthMethodListCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.436Z [WARN] TestAuthMethodListCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:32.438Z [DEBUG] TestAuthMethodListCommand.tlsutil: Update: version=1 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.446Z [WARN] TestAuthMethodListCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:32.448Z [DEBUG] TestAuthMethodListCommand_JSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:32.450Z [DEBUG] TestAuthMethodListCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.453Z [DEBUG] TestAuthMethodListCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.460Z [INFO] TestAuthMethodListCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a26e1eb7-b6db-ffec-334a-28bb6b101ad4 Address:127.0.0.1:14096}]" writer.go:29: 2021-01-29T19:31:32.463Z [INFO] TestAuthMethodListCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4.dc1 127.0.0.1 === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.468Z [INFO] TestAuthMethodListCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:473a16f2-0125-c173-ad89-1b77b50959f7 Address:127.0.0.1:14090}]" writer.go:29: 2021-01-29T19:31:32.472Z [INFO] TestAuthMethodListCommand.server.serf.wan: serf: EventMemberJoin: Node-473a16f2-0125-c173-ad89-1b77b50959f7.dc1 127.0.0.1 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.467Z [INFO] TestAuthMethodListCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:14096 [Follower]" leader= === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.472Z [INFO] TestAuthMethodListCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:14090 [Follower]" leader= writer.go:29: 2021-01-29T19:31:32.482Z [INFO] TestAuthMethodListCommand.server.serf.lan: serf: EventMemberJoin: Node-473a16f2-0125-c173-ad89-1b77b50959f7 127.0.0.1 writer.go:29: 2021-01-29T19:31:32.500Z [INFO] TestAuthMethodListCommand.server: Adding LAN server: server="Node-473a16f2-0125-c173-ad89-1b77b50959f7 (Addr: tcp/127.0.0.1:14090) (DC: dc1)" === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.475Z [INFO] TestAuthMethodListCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 127.0.0.1 writer.go:29: 2021-01-29T19:31:32.522Z [INFO] TestAuthMethodListCommand_JSON: Started DNS server: address=127.0.0.1:14091 network=udp === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.520Z [INFO] TestAuthMethodListCommand: Started DNS server: address=127.0.0.1:14085 network=udp writer.go:29: 2021-01-29T19:31:32.525Z [INFO] TestAuthMethodListCommand: Started DNS server: address=127.0.0.1:14085 network=tcp writer.go:29: 2021-01-29T19:31:32.524Z [INFO] TestAuthMethodListCommand.server: Handled event for server in area: event=member-join server=Node-473a16f2-0125-c173-ad89-1b77b50959f7.dc1 area=wan === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.529Z [WARN] TestAuthMethodListCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:32.533Z [INFO] TestAuthMethodListCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:14096 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:32.531Z [INFO] TestAuthMethodListCommand_JSON.server: Adding LAN server: server="Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 (Addr: tcp/127.0.0.1:14096) (DC: dc1)" writer.go:29: 2021-01-29T19:31:32.531Z [INFO] TestAuthMethodListCommand_JSON.server: Handled event for server in area: event=member-join server=Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4.dc1 area=wan writer.go:29: 2021-01-29T19:31:32.531Z [INFO] TestAuthMethodListCommand_JSON: Started DNS server: address=127.0.0.1:14091 network=tcp writer.go:29: 2021-01-29T19:31:32.534Z [WARN] TestAuthMethodListCommand_JSON.server.raft: unable to get address for sever, using fallback address: id=a26e1eb7-b6db-ffec-334a-28bb6b101ad4 fallback=127.0.0.1:14096 error="Could not find address for server id a26e1eb7-b6db-ffec-334a-28bb6b101ad4" === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.525Z [WARN] TestAuthMethodListCommand.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.542Z [INFO] TestAuthMethodListCommand_JSON: Started HTTP server: address=127.0.0.1:14092 network=tcp writer.go:29: 2021-01-29T19:31:32.544Z [DEBUG] TestAuthMethodListCommand_JSON.server.raft: votes: needed=1 === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.542Z [INFO] TestAuthMethodListCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:14090 [Candidate]" term=2 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.546Z [DEBUG] TestAuthMethodListCommand_JSON.server.raft: vote granted: from=a26e1eb7-b6db-ffec-334a-28bb6b101ad4 term=2 tally=1 === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.533Z [INFO] TestAuthMethodListCommand: Started HTTP server: address=127.0.0.1:14086 network=tcp === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.549Z [INFO] TestAuthMethodListCommand_JSON.server.raft: election won: tally=1 === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.550Z [INFO] TestAuthMethodListCommand: started state syncer === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.547Z [INFO] TestAuthMethodListCommand_JSON: started state syncer === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.551Z [DEBUG] TestAuthMethodListCommand.server.raft: votes: needed=1 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.551Z [INFO] TestAuthMethodListCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:14096 [Leader]" === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.554Z [DEBUG] TestAuthMethodListCommand.server.raft: vote granted: from=473a16f2-0125-c173-ad89-1b77b50959f7 term=2 tally=1 writer.go:29: 2021-01-29T19:31:32.555Z [INFO] TestAuthMethodListCommand.server.raft: election won: tally=1 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.555Z [INFO] TestAuthMethodListCommand_JSON.server: cluster leadership acquired === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.557Z [INFO] TestAuthMethodListCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:14090 [Leader]" === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.556Z [INFO] TestAuthMethodListCommand_JSON.server: New leader elected: payload=Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.559Z [INFO] TestAuthMethodListCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:32.559Z [INFO] TestAuthMethodListCommand.server: New leader elected: payload=Node-473a16f2-0125-c173-ad89-1b77b50959f7 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.567Z [ERROR] TestAuthMethodListCommand_JSON.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.573Z [INFO] TestAuthMethodListCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:32.579Z [INFO] TestAuthMethodListCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:32.581Z [WARN] TestAuthMethodListCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:32.580Z [INFO] TestAuthMethodListCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:32.585Z [WARN] TestAuthMethodListCommand.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.577Z [INFO] TestAuthMethodListCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:31:32.578Z [INFO] TestAuthMethodListCommand_JSON.server: initializing acls === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.590Z [INFO] TestAuthMethodListCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:32.592Z [INFO] TestAuthMethodListCommand.server: Bootstrapped ACL master token from configuration === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.592Z [INFO] TestAuthMethodListCommand_JSON.server: Created ACL 'global-management' policy === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.593Z [INFO] TestAuthMethodListCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:32.595Z [INFO] TestAuthMethodListCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:32.596Z [DEBUG] TestAuthMethodListCommand.server: transitioning out of legacy ACL mode === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.593Z [WARN] TestAuthMethodListCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:32.592Z [INFO] TestAuthMethodListCommand_JSON.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:32.602Z [WARN] TestAuthMethodListCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.592Z [INFO] TestAuthMethodListCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:32.598Z [INFO] TestAuthMethodListCommand.server.serf.lan: serf: EventMemberUpdate: Node-473a16f2-0125-c173-ad89-1b77b50959f7 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.601Z [INFO] TestAuthMethodListCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.611Z [INFO] TestAuthMethodListCommand.server: Updating LAN server: server="Node-473a16f2-0125-c173-ad89-1b77b50959f7 (Addr: tcp/127.0.0.1:14090) (DC: dc1)" writer.go:29: 2021-01-29T19:31:32.611Z [INFO] TestAuthMethodListCommand.server.serf.wan: serf: EventMemberUpdate: Node-473a16f2-0125-c173-ad89-1b77b50959f7.dc1 writer.go:29: 2021-01-29T19:31:32.614Z [INFO] TestAuthMethodListCommand.server: Handled event for server in area: event=member-update server=Node-473a16f2-0125-c173-ad89-1b77b50959f7.dc1 area=wan === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.612Z [INFO] TestAuthMethodListCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.612Z [INFO] TestAuthMethodListCommand.server.serf.lan: serf: EventMemberUpdate: Node-473a16f2-0125-c173-ad89-1b77b50959f7 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.604Z [INFO] TestAuthMethodListCommand_JSON.server: Created ACL anonymous token from configuration === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.617Z [INFO] TestAuthMethodListCommand.server: Updating LAN server: server="Node-473a16f2-0125-c173-ad89-1b77b50959f7 (Addr: tcp/127.0.0.1:14090) (DC: dc1)" === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.616Z [INFO] TestAuthMethodListCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:32.619Z [DEBUG] TestAuthMethodListCommand_JSON.server: transitioning out of legacy ACL mode === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.617Z [INFO] TestAuthMethodListCommand.server.serf.wan: serf: EventMemberUpdate: Node-473a16f2-0125-c173-ad89-1b77b50959f7.dc1 writer.go:29: 2021-01-29T19:31:32.621Z [INFO] TestAuthMethodListCommand.server: Handled event for server in area: event=member-update server=Node-473a16f2-0125-c173-ad89-1b77b50959f7.dc1 area=wan === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.619Z [INFO] TestAuthMethodListCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 writer.go:29: 2021-01-29T19:31:32.625Z [INFO] TestAuthMethodListCommand_JSON.server: Updating LAN server: server="Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 (Addr: tcp/127.0.0.1:14096) (DC: dc1)" writer.go:29: 2021-01-29T19:31:32.625Z [INFO] TestAuthMethodListCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 writer.go:29: 2021-01-29T19:31:32.625Z [INFO] TestAuthMethodListCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4.dc1 === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.630Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.627Z [INFO] TestAuthMethodListCommand_JSON.server: Updating LAN server: server="Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 (Addr: tcp/127.0.0.1:14096) (DC: dc1)" writer.go:29: 2021-01-29T19:31:32.631Z [INFO] TestAuthMethodListCommand_JSON.server: Handled event for server in area: event=member-update server=Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4.dc1 area=wan writer.go:29: 2021-01-29T19:31:32.631Z [INFO] TestAuthMethodListCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4.dc1 writer.go:29: 2021-01-29T19:31:32.635Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:32.638Z [INFO] TestAuthMethodListCommand_JSON.server: Handled event for server in area: event=member-update server=Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4.dc1 area=wan === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.669Z [INFO] TestAuthMethodListCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.669Z [INFO] TestAuthMethodListCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.670Z [INFO] TestAuthMethodListCommand.leader: started routine: routine="CA root pruning" === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.671Z [INFO] TestAuthMethodListCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.672Z [DEBUG] TestAuthMethodListCommand.server: Skipping self join check for node since the cluster is too small: node=Node-473a16f2-0125-c173-ad89-1b77b50959f7 writer.go:29: 2021-01-29T19:31:32.673Z [INFO] TestAuthMethodListCommand.server: member joined, marking health alive: member=Node-473a16f2-0125-c173-ad89-1b77b50959f7 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.672Z [DEBUG] TestAuthMethodListCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 writer.go:29: 2021-01-29T19:31:32.673Z [INFO] TestAuthMethodListCommand_JSON.server: member joined, marking health alive: member=Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.676Z [DEBUG] TestAuthMethodListCommand.server: Skipping self join check for node since the cluster is too small: node=Node-473a16f2-0125-c173-ad89-1b77b50959f7 writer.go:29: 2021-01-29T19:31:32.677Z [DEBUG] TestAuthMethodListCommand.server: Skipping self join check for node since the cluster is too small: node=Node-473a16f2-0125-c173-ad89-1b77b50959f7 === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.677Z [DEBUG] TestAuthMethodListCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 writer.go:29: 2021-01-29T19:31:32.680Z [DEBUG] TestAuthMethodListCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-a26e1eb7-b6db-ffec-334a-28bb6b101ad4 writer.go:29: 2021-01-29T19:31:32.713Z [DEBUG] TestAuthMethodListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:46286 latency=2.930585ms === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.747Z [DEBUG] TestAuthMethodListCommand: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.730Z [DEBUG] TestAuthMethodListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:46286 latency=4.562056ms writer.go:29: 2021-01-29T19:31:32.759Z [DEBUG] TestAuthMethodListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:46286 latency=2.842468ms === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.758Z [INFO] TestAuthMethodListCommand: Synced node info writer.go:29: 2021-01-29T19:31:32.762Z [DEBUG] TestAuthMethodListCommand: Node info in sync === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.767Z [DEBUG] TestAuthMethodListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:46286 latency=3.906066ms writer.go:29: 2021-01-29T19:31:32.775Z [DEBUG] TestAuthMethodListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:46286 latency=3.037318ms === RUN TestAuthMethodListCommand_JSON/found_some === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.795Z [DEBUG] TestAuthMethodListCommand_JSON.http: Request finished: method=GET url=/v1/acl/auth-methods from=127.0.0.1:46288 latency=7.818155ms writer.go:29: 2021-01-29T19:31:32.807Z [INFO] TestAuthMethodListCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:31:32.809Z [INFO] TestAuthMethodListCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:31:32.810Z [DEBUG] TestAuthMethodListCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:32.812Z [DEBUG] TestAuthMethodListCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:32.814Z [DEBUG] TestAuthMethodListCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.818Z [WARN] TestAuthMethodListCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:32.814Z [DEBUG] TestAuthMethodListCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:32.814Z [DEBUG] TestAuthMethodListCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:32.818Z [DEBUG] TestAuthMethodListCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.821Z [WARN] TestAuthMethodListCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:32.830Z [INFO] TestAuthMethodListCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:32.830Z [INFO] TestAuthMethodListCommand_JSON: consul server down === RUN TestAuthMethodListCommand/found_none === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:32.834Z [INFO] TestAuthMethodListCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:31:32.836Z [INFO] TestAuthMethodListCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:14091 network=tcp writer.go:29: 2021-01-29T19:31:32.838Z [INFO] TestAuthMethodListCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:14091 network=udp writer.go:29: 2021-01-29T19:31:32.841Z [INFO] TestAuthMethodListCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:14092 network=tcp === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.844Z [DEBUG] TestAuthMethodListCommand.http: Request finished: method=GET url=/v1/acl/auth-methods from=127.0.0.1:50918 latency=2.929011ms writer.go:29: 2021-01-29T19:31:32.879Z [DEBUG] TestAuthMethodListCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:50920 latency=25.438448ms writer.go:29: 2021-01-29T19:31:32.885Z [DEBUG] TestAuthMethodListCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:50920 latency=1.26324ms writer.go:29: 2021-01-29T19:31:32.904Z [DEBUG] TestAuthMethodListCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:50920 latency=13.049185ms writer.go:29: 2021-01-29T19:31:32.909Z [DEBUG] TestAuthMethodListCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:50920 latency=1.027157ms writer.go:29: 2021-01-29T19:31:32.914Z [DEBUG] TestAuthMethodListCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:50920 latency=778.854µs === RUN TestAuthMethodListCommand/found_some === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:32.922Z [DEBUG] TestAuthMethodListCommand.http: Request finished: method=GET url=/v1/acl/auth-methods from=127.0.0.1:50922 latency=131.881µs writer.go:29: 2021-01-29T19:31:32.928Z [INFO] TestAuthMethodListCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:32.930Z [INFO] TestAuthMethodListCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:32.932Z [DEBUG] TestAuthMethodListCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:32.933Z [DEBUG] TestAuthMethodListCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:32.935Z [DEBUG] TestAuthMethodListCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.933Z [DEBUG] TestAuthMethodListCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:32.935Z [DEBUG] TestAuthMethodListCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:32.937Z [WARN] TestAuthMethodListCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:32.937Z [DEBUG] TestAuthMethodListCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:32.943Z [WARN] TestAuthMethodListCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:32.948Z [INFO] TestAuthMethodListCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:32.948Z [INFO] TestAuthMethodListCommand: consul server down writer.go:29: 2021-01-29T19:31:32.952Z [INFO] TestAuthMethodListCommand: shutdown complete writer.go:29: 2021-01-29T19:31:32.954Z [INFO] TestAuthMethodListCommand: Stopping server: protocol=DNS address=127.0.0.1:14085 network=tcp writer.go:29: 2021-01-29T19:31:32.956Z [INFO] TestAuthMethodListCommand: Stopping server: protocol=DNS address=127.0.0.1:14085 network=udp writer.go:29: 2021-01-29T19:31:32.958Z [INFO] TestAuthMethodListCommand: Stopping server: protocol=HTTP address=127.0.0.1:14086 network=tcp === CONT TestAuthMethodListCommand_JSON writer.go:29: 2021-01-29T19:31:33.343Z [INFO] TestAuthMethodListCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:33.344Z [INFO] TestAuthMethodListCommand_JSON: Endpoints down --- PASS: TestAuthMethodListCommand_JSON (0.96s) --- PASS: TestAuthMethodListCommand_JSON/found_some (0.02s) === CONT TestAuthMethodListCommand writer.go:29: 2021-01-29T19:31:33.461Z [INFO] TestAuthMethodListCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:33.462Z [INFO] TestAuthMethodListCommand: Endpoints down --- PASS: TestAuthMethodListCommand (1.08s) --- PASS: TestAuthMethodListCommand/found_none (0.01s) --- PASS: TestAuthMethodListCommand/found_some (0.01s) PASS ok github.com/hashicorp/consul/command/acl/authmethod/list 1.288s === RUN TestAuthMethodReadCommand_noTabs === PAUSE TestAuthMethodReadCommand_noTabs === RUN TestAuthMethodReadCommand === PAUSE TestAuthMethodReadCommand === RUN TestAuthMethodReadCommand_JSON === PAUSE TestAuthMethodReadCommand_JSON === CONT TestAuthMethodReadCommand_noTabs === CONT TestAuthMethodReadCommand_JSON === CONT TestAuthMethodReadCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestAuthMethodReadCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.610Z [WARN] TestAuthMethodReadCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:34.621Z [DEBUG] TestAuthMethodReadCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:34.623Z [DEBUG] TestAuthMethodReadCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.625Z [WARN] TestAuthMethodReadCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:34.627Z [DEBUG] TestAuthMethodReadCommand_JSON.tlsutil: Update: version=1 === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.629Z [INFO] TestAuthMethodReadCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd Address:127.0.0.1:18180}]" === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.629Z [DEBUG] TestAuthMethodReadCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.630Z [INFO] TestAuthMethodReadCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:18180 [Follower]" leader= writer.go:29: 2021-01-29T19:31:34.633Z [INFO] TestAuthMethodReadCommand.server.serf.wan: serf: EventMemberJoin: Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd.dc1 127.0.0.1 === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.639Z [INFO] TestAuthMethodReadCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1a3f1f28-31b3-811f-a698-82a6702b112d Address:127.0.0.1:18174}]" === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.638Z [INFO] TestAuthMethodReadCommand.server.serf.lan: serf: EventMemberJoin: Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd 127.0.0.1 === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.641Z [INFO] TestAuthMethodReadCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:18174 [Follower]" leader= === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.644Z [INFO] TestAuthMethodReadCommand.server: Adding LAN server: server="Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd (Addr: tcp/127.0.0.1:18180) (DC: dc1)" writer.go:29: 2021-01-29T19:31:34.644Z [INFO] TestAuthMethodReadCommand.server: Handled event for server in area: event=member-join server=Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd.dc1 area=wan writer.go:29: 2021-01-29T19:31:34.646Z [INFO] TestAuthMethodReadCommand: Started DNS server: address=127.0.0.1:18175 network=udp writer.go:29: 2021-01-29T19:31:34.648Z [INFO] TestAuthMethodReadCommand: Started DNS server: address=127.0.0.1:18175 network=tcp writer.go:29: 2021-01-29T19:31:34.649Z [INFO] TestAuthMethodReadCommand: Started HTTP server: address=127.0.0.1:18176 network=tcp === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.642Z [INFO] TestAuthMethodReadCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-1a3f1f28-31b3-811f-a698-82a6702b112d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:34.650Z [INFO] TestAuthMethodReadCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-1a3f1f28-31b3-811f-a698-82a6702b112d 127.0.0.1 writer.go:29: 2021-01-29T19:31:34.653Z [INFO] TestAuthMethodReadCommand_JSON.server: Adding LAN server: server="Node-1a3f1f28-31b3-811f-a698-82a6702b112d (Addr: tcp/127.0.0.1:18174) (DC: dc1)" writer.go:29: 2021-01-29T19:31:34.653Z [INFO] TestAuthMethodReadCommand_JSON: Started DNS server: address=127.0.0.1:18169 network=udp === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.650Z [INFO] TestAuthMethodReadCommand: started state syncer === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.653Z [INFO] TestAuthMethodReadCommand_JSON.server: Handled event for server in area: event=member-join server=Node-1a3f1f28-31b3-811f-a698-82a6702b112d.dc1 area=wan writer.go:29: 2021-01-29T19:31:34.654Z [INFO] TestAuthMethodReadCommand_JSON: Started DNS server: address=127.0.0.1:18169 network=tcp writer.go:29: 2021-01-29T19:31:34.657Z [INFO] TestAuthMethodReadCommand_JSON: Started HTTP server: address=127.0.0.1:18170 network=tcp writer.go:29: 2021-01-29T19:31:34.658Z [INFO] TestAuthMethodReadCommand_JSON: started state syncer === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.687Z [WARN] TestAuthMethodReadCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:34.689Z [INFO] TestAuthMethodReadCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:18180 [Candidate]" term=2 === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.691Z [WARN] TestAuthMethodReadCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:34.693Z [INFO] TestAuthMethodReadCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:18174 [Candidate]" term=2 === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.691Z [DEBUG] TestAuthMethodReadCommand.server.raft: votes: needed=1 === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.696Z [DEBUG] TestAuthMethodReadCommand_JSON.server.raft: votes: needed=1 === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.699Z [DEBUG] TestAuthMethodReadCommand.server.raft: vote granted: from=3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd term=2 tally=1 writer.go:29: 2021-01-29T19:31:34.695Z [DEBUG] TestAuthMethodReadCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.700Z [DEBUG] TestAuthMethodReadCommand_JSON.server.raft: vote granted: from=1a3f1f28-31b3-811f-a698-82a6702b112d term=2 tally=1 === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.704Z [INFO] TestAuthMethodReadCommand.server.raft: election won: tally=1 === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.705Z [INFO] TestAuthMethodReadCommand_JSON.server.raft: election won: tally=1 === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.708Z [INFO] TestAuthMethodReadCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:18180 [Leader]" writer.go:29: 2021-01-29T19:31:34.713Z [INFO] TestAuthMethodReadCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:34.713Z [INFO] TestAuthMethodReadCommand.server: New leader elected: payload=Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.704Z [DEBUG] TestAuthMethodReadCommand_JSON.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:34.708Z [INFO] TestAuthMethodReadCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:18174 [Leader]" writer.go:29: 2021-01-29T19:31:34.719Z [INFO] TestAuthMethodReadCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:34.719Z [INFO] TestAuthMethodReadCommand_JSON.server: New leader elected: payload=Node-1a3f1f28-31b3-811f-a698-82a6702b112d writer.go:29: 2021-01-29T19:31:34.720Z [INFO] TestAuthMethodReadCommand_JSON.server: initializing acls === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.720Z [INFO] TestAuthMethodReadCommand.server: initializing acls === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.735Z [INFO] TestAuthMethodReadCommand_JSON.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:34.739Z [WARN] TestAuthMethodReadCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:34.743Z [INFO] TestAuthMethodReadCommand_JSON.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:34.746Z [INFO] TestAuthMethodReadCommand_JSON.server: Created ACL anonymous token from configuration === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.748Z [INFO] TestAuthMethodReadCommand.server: Created ACL 'global-management' policy === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.749Z [INFO] TestAuthMethodReadCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.752Z [WARN] TestAuthMethodReadCommand.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.754Z [INFO] TestAuthMethodReadCommand_JSON.leader: started routine: routine="acl token reaping" === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.758Z [INFO] TestAuthMethodReadCommand.server: Bootstrapped ACL master token from configuration === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.757Z [INFO] TestAuthMethodReadCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-1a3f1f28-31b3-811f-a698-82a6702b112d === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.762Z [INFO] TestAuthMethodReadCommand.server: Created ACL anonymous token from configuration === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.763Z [INFO] TestAuthMethodReadCommand_JSON.server: Updating LAN server: server="Node-1a3f1f28-31b3-811f-a698-82a6702b112d (Addr: tcp/127.0.0.1:18174) (DC: dc1)" writer.go:29: 2021-01-29T19:31:34.764Z [INFO] TestAuthMethodReadCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-1a3f1f28-31b3-811f-a698-82a6702b112d.dc1 === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.765Z [INFO] TestAuthMethodReadCommand.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.768Z [INFO] TestAuthMethodReadCommand_JSON.server: Handled event for server in area: event=member-update server=Node-1a3f1f28-31b3-811f-a698-82a6702b112d.dc1 area=wan === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.769Z [INFO] TestAuthMethodReadCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:34.772Z [INFO] TestAuthMethodReadCommand.server.serf.lan: serf: EventMemberUpdate: Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd writer.go:29: 2021-01-29T19:31:34.773Z [INFO] TestAuthMethodReadCommand.server.serf.wan: serf: EventMemberUpdate: Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd.dc1 writer.go:29: 2021-01-29T19:31:34.781Z [INFO] TestAuthMethodReadCommand.server: Handled event for server in area: event=member-update server=Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd.dc1 area=wan === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.782Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.784Z [INFO] TestAuthMethodReadCommand.server: Updating LAN server: server="Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd (Addr: tcp/127.0.0.1:18180) (DC: dc1)" writer.go:29: 2021-01-29T19:31:34.786Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:34.804Z [INFO] TestAuthMethodReadCommand: Synced node info === RUN TestAuthMethodReadCommand/name_required === RUN TestAuthMethodReadCommand/not_found === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.841Z [DEBUG] TestAuthMethodReadCommand.http: Request finished: method=GET url=/v1/acl/auth-method/notfound from=127.0.0.1:36870 latency=3.054992ms === RUN TestAuthMethodReadCommand/read_by_name === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.859Z [INFO] TestAuthMethodReadCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.859Z [DEBUG] TestAuthMethodReadCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:36872 latency=6.149837ms === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.860Z [INFO] TestAuthMethodReadCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.861Z [INFO] TestAuthMethodReadCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:34.864Z [INFO] TestAuthMethodReadCommand.leader: started routine: routine="CA root pruning" === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.863Z [DEBUG] TestAuthMethodReadCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-1a3f1f28-31b3-811f-a698-82a6702b112d === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.866Z [DEBUG] TestAuthMethodReadCommand.server: Skipping self join check for node since the cluster is too small: node=Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd writer.go:29: 2021-01-29T19:31:34.868Z [INFO] TestAuthMethodReadCommand.server: member joined, marking health alive: member=Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.867Z [INFO] TestAuthMethodReadCommand_JSON.server: member joined, marking health alive: member=Node-1a3f1f28-31b3-811f-a698-82a6702b112d writer.go:29: 2021-01-29T19:31:34.873Z [DEBUG] TestAuthMethodReadCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-1a3f1f28-31b3-811f-a698-82a6702b112d === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:34.876Z [DEBUG] TestAuthMethodReadCommand.server: Skipping self join check for node since the cluster is too small: node=Node-3f01282e-aea4-b9bd-eb7b-f2c4a3cf5ffd writer.go:29: 2021-01-29T19:31:34.881Z [DEBUG] TestAuthMethodReadCommand.http: Request finished: method=GET url=/v1/acl/auth-method/test-133571ee-81de-3546-bcf7-b768d6378ddb from=127.0.0.1:36874 latency=1.523667ms writer.go:29: 2021-01-29T19:31:34.884Z [INFO] TestAuthMethodReadCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:34.886Z [INFO] TestAuthMethodReadCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:34.887Z [DEBUG] TestAuthMethodReadCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:34.890Z [DEBUG] TestAuthMethodReadCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:34.893Z [DEBUG] TestAuthMethodReadCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:34.890Z [DEBUG] TestAuthMethodReadCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:34.893Z [DEBUG] TestAuthMethodReadCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:34.896Z [WARN] TestAuthMethodReadCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:34.896Z [DEBUG] TestAuthMethodReadCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:34.903Z [WARN] TestAuthMethodReadCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:34.925Z [INFO] TestAuthMethodReadCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:34.925Z [INFO] TestAuthMethodReadCommand: consul server down writer.go:29: 2021-01-29T19:31:34.929Z [INFO] TestAuthMethodReadCommand: shutdown complete writer.go:29: 2021-01-29T19:31:34.929Z [INFO] TestAuthMethodReadCommand: Stopping server: protocol=DNS address=127.0.0.1:18175 network=tcp writer.go:29: 2021-01-29T19:31:34.931Z [INFO] TestAuthMethodReadCommand: Stopping server: protocol=DNS address=127.0.0.1:18175 network=udp writer.go:29: 2021-01-29T19:31:34.932Z [INFO] TestAuthMethodReadCommand: Stopping server: protocol=HTTP address=127.0.0.1:18176 network=tcp === RUN TestAuthMethodReadCommand_JSON/read_by_name === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:34.964Z [DEBUG] TestAuthMethodReadCommand_JSON.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:41234 latency=12.193291ms writer.go:29: 2021-01-29T19:31:34.982Z [DEBUG] TestAuthMethodReadCommand_JSON.http: Request finished: method=GET url=/v1/acl/auth-method/test-7c35f44b-e6bf-6f86-6f6d-283f363672dc from=127.0.0.1:41236 latency=2.210506ms writer.go:29: 2021-01-29T19:31:34.987Z [INFO] TestAuthMethodReadCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:31:34.989Z [INFO] TestAuthMethodReadCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:31:34.991Z [DEBUG] TestAuthMethodReadCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:35.000Z [DEBUG] TestAuthMethodReadCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:34.993Z [ERROR] TestAuthMethodReadCommand_JSON.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:35.001Z [DEBUG] TestAuthMethodReadCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:35.008Z [DEBUG] TestAuthMethodReadCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:35.008Z [DEBUG] TestAuthMethodReadCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:35.015Z [WARN] TestAuthMethodReadCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:35.015Z [DEBUG] TestAuthMethodReadCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:35.020Z [WARN] TestAuthMethodReadCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:35.024Z [INFO] TestAuthMethodReadCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:35.026Z [INFO] TestAuthMethodReadCommand_JSON: consul server down writer.go:29: 2021-01-29T19:31:35.028Z [INFO] TestAuthMethodReadCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:31:35.030Z [INFO] TestAuthMethodReadCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:18169 network=tcp writer.go:29: 2021-01-29T19:31:35.032Z [INFO] TestAuthMethodReadCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:18169 network=udp writer.go:29: 2021-01-29T19:31:35.034Z [INFO] TestAuthMethodReadCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:18170 network=tcp === CONT TestAuthMethodReadCommand writer.go:29: 2021-01-29T19:31:35.433Z [INFO] TestAuthMethodReadCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:35.435Z [INFO] TestAuthMethodReadCommand: Endpoints down --- PASS: TestAuthMethodReadCommand (0.88s) --- PASS: TestAuthMethodReadCommand/name_required (0.01s) --- PASS: TestAuthMethodReadCommand/not_found (0.01s) --- PASS: TestAuthMethodReadCommand/read_by_name (0.04s) === CONT TestAuthMethodReadCommand_JSON writer.go:29: 2021-01-29T19:31:35.536Z [INFO] TestAuthMethodReadCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:35.538Z [INFO] TestAuthMethodReadCommand_JSON: Endpoints down --- PASS: TestAuthMethodReadCommand_JSON (0.98s) --- PASS: TestAuthMethodReadCommand_JSON/read_by_name (0.04s) PASS ok github.com/hashicorp/consul/command/acl/authmethod/read 1.234s === RUN TestAuthMethodUpdateCommand_noTabs === PAUSE TestAuthMethodUpdateCommand_noTabs === RUN TestAuthMethodUpdateCommand === PAUSE TestAuthMethodUpdateCommand === RUN TestAuthMethodUpdateCommand_JSON === PAUSE TestAuthMethodUpdateCommand_JSON === RUN TestAuthMethodUpdateCommand_noMerge === PAUSE TestAuthMethodUpdateCommand_noMerge === RUN TestAuthMethodUpdateCommand_k8s === PAUSE TestAuthMethodUpdateCommand_k8s === RUN TestAuthMethodUpdateCommand_k8s_noMerge === PAUSE TestAuthMethodUpdateCommand_k8s_noMerge === CONT TestAuthMethodUpdateCommand_noTabs --- PASS: TestAuthMethodUpdateCommand_noTabs (0.00s) === CONT TestAuthMethodUpdateCommand_JSON === CONT TestAuthMethodUpdateCommand_k8s_noMerge [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestAuthMethodUpdateCommand_noMerge === CONT TestAuthMethodUpdateCommand [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.049Z [WARN] TestAuthMethodUpdateCommand_noMerge: bootstrap = true: do not enable unless necessary === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.056Z [WARN] TestAuthMethodUpdateCommand: bootstrap = true: do not enable unless necessary === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.061Z [WARN] TestAuthMethodUpdateCommand_k8s_noMerge: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:39.063Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:39.071Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.050Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.tlsutil: Update: version=1 === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.072Z [DEBUG] TestAuthMethodUpdateCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:39.073Z [DEBUG] TestAuthMethodUpdateCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.075Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.077Z [WARN] TestAuthMethodUpdateCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:39.080Z [DEBUG] TestAuthMethodUpdateCommand_JSON.tlsutil: Update: version=1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.081Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cb851575-f3ce-2efc-e02d-59ebafe1d53f Address:127.0.0.1:12060}]" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.082Z [DEBUG] TestAuthMethodUpdateCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.084Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.raft: entering follower state: follower="Node at 127.0.0.1:12060 [Follower]" leader= writer.go:29: 2021-01-29T19:31:39.089Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.serf.wan: serf: EventMemberJoin: Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:39.091Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.serf.lan: serf: EventMemberJoin: Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f 127.0.0.1 writer.go:29: 2021-01-29T19:31:39.093Z [INFO] TestAuthMethodUpdateCommand_noMerge: Started DNS server: address=127.0.0.1:12055 network=udp === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.092Z [INFO] TestAuthMethodUpdateCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cc725c06-25c7-a2db-89fc-0a7a950e36a3 Address:127.0.0.1:12066}]" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.101Z [INFO] TestAuthMethodUpdateCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:34d6a66f-02c8-5291-1371-cbcd10c89dea Address:127.0.0.1:12048}]" writer.go:29: 2021-01-29T19:31:39.104Z [INFO] TestAuthMethodUpdateCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-34d6a66f-02c8-5291-1371-cbcd10c89dea.dc1 127.0.0.1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.101Z [INFO] TestAuthMethodUpdateCommand_noMerge: Started DNS server: address=127.0.0.1:12055 network=tcp === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.107Z [INFO] TestAuthMethodUpdateCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-34d6a66f-02c8-5291-1371-cbcd10c89dea 127.0.0.1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.094Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: Handled event for server in area: event=member-join server=Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f.dc1 area=wan === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.110Z [INFO] TestAuthMethodUpdateCommand_JSON: Started DNS server: address=127.0.0.1:12043 network=udp writer.go:29: 2021-01-29T19:31:39.112Z [INFO] TestAuthMethodUpdateCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:12048 [Follower]" leader= === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.115Z [INFO] TestAuthMethodUpdateCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:12066 [Follower]" leader= === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.096Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: Adding LAN server: server="Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f (Addr: tcp/127.0.0.1:12060) (DC: dc1)" === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.090Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0358cac9-a098-0d61-1416-89fd6c62f716 Address:127.0.0.1:12054}]" === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.117Z [INFO] TestAuthMethodUpdateCommand.server.serf.wan: serf: EventMemberJoin: Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3.dc1 127.0.0.1 === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.120Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.serf.wan: serf: EventMemberJoin: Node-0358cac9-a098-0d61-1416-89fd6c62f716.dc1 127.0.0.1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.127Z [INFO] TestAuthMethodUpdateCommand_noMerge: Started HTTP server: address=127.0.0.1:12056 network=tcp === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.142Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.raft: entering follower state: follower="Node at 127.0.0.1:12054 [Follower]" leader= === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.152Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Handled event for server in area: event=member-join server=Node-34d6a66f-02c8-5291-1371-cbcd10c89dea.dc1 area=wan === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.131Z [INFO] TestAuthMethodUpdateCommand_noMerge: started state syncer === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.172Z [WARN] TestAuthMethodUpdateCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:39.175Z [INFO] TestAuthMethodUpdateCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:12048 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:39.181Z [INFO] TestAuthMethodUpdateCommand_JSON: Started DNS server: address=127.0.0.1:12043 network=tcp === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.180Z [WARN] TestAuthMethodUpdateCommand_noMerge.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.183Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Adding LAN server: server="Node-34d6a66f-02c8-5291-1371-cbcd10c89dea (Addr: tcp/127.0.0.1:12048) (DC: dc1)" === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.186Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.raft: entering candidate state: node="Node at 127.0.0.1:12060 [Candidate]" term=2 === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.187Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.serf.lan: serf: EventMemberJoin: Node-0358cac9-a098-0d61-1416-89fd6c62f716 127.0.0.1 === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.187Z [INFO] TestAuthMethodUpdateCommand_JSON: Started HTTP server: address=127.0.0.1:12044 network=tcp === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.191Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.187Z [INFO] TestAuthMethodUpdateCommand.server.serf.lan: serf: EventMemberJoin: Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 127.0.0.1 === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.202Z [INFO] TestAuthMethodUpdateCommand_JSON: started state syncer === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.203Z [INFO] TestAuthMethodUpdateCommand: Started DNS server: address=127.0.0.1:12061 network=udp writer.go:29: 2021-01-29T19:31:39.191Z [WARN] TestAuthMethodUpdateCommand.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.192Z [DEBUG] TestAuthMethodUpdateCommand_JSON.server.raft: votes: needed=1 === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.205Z [INFO] TestAuthMethodUpdateCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:12066 [Candidate]" term=2 === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.205Z [DEBUG] TestAuthMethodUpdateCommand_JSON.server.raft: vote granted: from=34d6a66f-02c8-5291-1371-cbcd10c89dea term=2 tally=1 writer.go:29: 2021-01-29T19:31:39.207Z [INFO] TestAuthMethodUpdateCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:39.209Z [INFO] TestAuthMethodUpdateCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:12048 [Leader]" writer.go:29: 2021-01-29T19:31:39.211Z [INFO] TestAuthMethodUpdateCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:39.213Z [INFO] TestAuthMethodUpdateCommand_JSON.server: New leader elected: payload=Node-34d6a66f-02c8-5291-1371-cbcd10c89dea === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.214Z [INFO] TestAuthMethodUpdateCommand.server: Adding LAN server: server="Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 (Addr: tcp/127.0.0.1:12066) (DC: dc1)" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.213Z [ERROR] TestAuthMethodUpdateCommand_JSON.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.216Z [INFO] TestAuthMethodUpdateCommand.server: Handled event for server in area: event=member-join server=Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3.dc1 area=wan writer.go:29: 2021-01-29T19:31:39.218Z [INFO] TestAuthMethodUpdateCommand: Started DNS server: address=127.0.0.1:12061 network=tcp === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.218Z [WARN] TestAuthMethodUpdateCommand_k8s_noMerge.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:39.220Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.raft: entering candidate state: node="Node at 127.0.0.1:12054 [Candidate]" term=2 === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.221Z [INFO] TestAuthMethodUpdateCommand: Started HTTP server: address=127.0.0.1:12062 network=tcp writer.go:29: 2021-01-29T19:31:39.224Z [INFO] TestAuthMethodUpdateCommand: started state syncer === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.223Z [INFO] TestAuthMethodUpdateCommand_JSON.server: initializing acls === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.227Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.server.raft: votes: needed=1 === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.222Z [WARN] TestAuthMethodUpdateCommand_k8s_noMerge.server.raft: unable to get address for sever, using fallback address: id=0358cac9-a098-0d61-1416-89fd6c62f716 fallback=127.0.0.1:12054 error="Could not find address for server id 0358cac9-a098-0d61-1416-89fd6c62f716" === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.228Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.server.raft: vote granted: from=cb851575-f3ce-2efc-e02d-59ebafe1d53f term=2 tally=1 === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.229Z [DEBUG] TestAuthMethodUpdateCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:39.230Z [DEBUG] TestAuthMethodUpdateCommand.server.raft: vote granted: from=cc725c06-25c7-a2db-89fc-0a7a950e36a3 term=2 tally=1 writer.go:29: 2021-01-29T19:31:39.231Z [INFO] TestAuthMethodUpdateCommand.server.raft: election won: tally=1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.230Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.raft: election won: tally=1 === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.232Z [INFO] TestAuthMethodUpdateCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:12066 [Leader]" === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.233Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: Adding LAN server: server="Node-0358cac9-a098-0d61-1416-89fd6c62f716 (Addr: tcp/127.0.0.1:12054) (DC: dc1)" writer.go:29: 2021-01-29T19:31:39.234Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: Handled event for server in area: event=member-join server=Node-0358cac9-a098-0d61-1416-89fd6c62f716.dc1 area=wan === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.233Z [INFO] TestAuthMethodUpdateCommand.server: cluster leadership acquired === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.232Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.raft: entering leader state: leader="Node at 127.0.0.1:12060 [Leader]" === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.236Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: Started DNS server: address=127.0.0.1:12049 network=udp === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.236Z [INFO] TestAuthMethodUpdateCommand.server: New leader elected: payload=Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.236Z [INFO] TestAuthMethodUpdateCommand_JSON.server: initializing acls === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.238Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: cluster leadership acquired === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.238Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.server.raft: votes: needed=1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.238Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: New leader elected: payload=Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.240Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.server.raft: vote granted: from=0358cac9-a098-0d61-1416-89fd6c62f716 term=2 tally=1 writer.go:29: 2021-01-29T19:31:39.242Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:39.243Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.raft: entering leader state: leader="Node at 127.0.0.1:12054 [Leader]" writer.go:29: 2021-01-29T19:31:39.238Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: Started DNS server: address=127.0.0.1:12049 network=tcp writer.go:29: 2021-01-29T19:31:39.246Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:39.246Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: New leader elected: payload=Node-0358cac9-a098-0d61-1416-89fd6c62f716 writer.go:29: 2021-01-29T19:31:39.248Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: Started HTTP server: address=127.0.0.1:12050 network=tcp === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.245Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: initializing acls === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.248Z [INFO] TestAuthMethodUpdateCommand.server: initializing acls === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.250Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: started state syncer === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.251Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:39.253Z [WARN] TestAuthMethodUpdateCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:39.252Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Created ACL 'global-management' policy === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.254Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: Created ACL 'global-management' policy === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.255Z [INFO] TestAuthMethodUpdateCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:39.256Z [WARN] TestAuthMethodUpdateCommand.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.256Z [WARN] TestAuthMethodUpdateCommand_noMerge.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.251Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: initializing acls writer.go:29: 2021-01-29T19:31:39.251Z [ERROR] TestAuthMethodUpdateCommand_k8s_noMerge.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.255Z [WARN] TestAuthMethodUpdateCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.263Z [INFO] TestAuthMethodUpdateCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:39.266Z [INFO] TestAuthMethodUpdateCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:39.267Z [INFO] TestAuthMethodUpdateCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:39.268Z [INFO] TestAuthMethodUpdateCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:39.269Z [INFO] TestAuthMethodUpdateCommand.server.serf.lan: serf: EventMemberUpdate: Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.272Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.265Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: Bootstrapped ACL master token from configuration === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.273Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.275Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:39.277Z [WARN] TestAuthMethodUpdateCommand_k8s_noMerge.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.276Z [INFO] TestAuthMethodUpdateCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.266Z [INFO] TestAuthMethodUpdateCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:39.271Z [INFO] TestAuthMethodUpdateCommand.server.serf.wan: serf: EventMemberUpdate: Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3.dc1 writer.go:29: 2021-01-29T19:31:39.271Z [INFO] TestAuthMethodUpdateCommand.server: Updating LAN server: server="Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 (Addr: tcp/127.0.0.1:12066) (DC: dc1)" writer.go:29: 2021-01-29T19:31:39.281Z [WARN] TestAuthMethodUpdateCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:39.282Z [DEBUG] TestAuthMethodUpdateCommand.server: transitioning out of legacy ACL mode === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.283Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: Bootstrapped ACL master token from configuration === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.283Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: Created ACL anonymous token from configuration === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.279Z [INFO] TestAuthMethodUpdateCommand_JSON.leader: started routine: routine="acl token reaping" === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.283Z [INFO] TestAuthMethodUpdateCommand.server: Handled event for server in area: event=member-update server=Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3.dc1 area=wan === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.287Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: initializing acls === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.274Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Created ACL anonymous token from configuration === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.291Z [WARN] TestAuthMethodUpdateCommand_k8s_noMerge.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.292Z [INFO] TestAuthMethodUpdateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-34d6a66f-02c8-5291-1371-cbcd10c89dea === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.287Z [INFO] TestAuthMethodUpdateCommand.server.serf.lan: serf: EventMemberUpdate: Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.293Z [INFO] TestAuthMethodUpdateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-34d6a66f-02c8-5291-1371-cbcd10c89dea.dc1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.290Z [INFO] TestAuthMethodUpdateCommand_noMerge.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.291Z [DEBUG] TestAuthMethodUpdateCommand_JSON.server: transitioning out of legacy ACL mode === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.284Z [ERROR] TestAuthMethodUpdateCommand_noMerge.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.294Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Updating LAN server: server="Node-34d6a66f-02c8-5291-1371-cbcd10c89dea (Addr: tcp/127.0.0.1:12048) (DC: dc1)" === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.296Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:39.302Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.leader: started routine: routine="legacy ACL token upgrade" === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.295Z [INFO] TestAuthMethodUpdateCommand.server.serf.wan: serf: EventMemberUpdate: Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3.dc1 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.294Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: initializing acls === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.298Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-34d6a66f-02c8-5291-1371-cbcd10c89dea.dc1 area=wan === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.295Z [INFO] TestAuthMethodUpdateCommand_noMerge.leader: started routine: routine="acl token reaping" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.298Z [INFO] TestAuthMethodUpdateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-34d6a66f-02c8-5291-1371-cbcd10c89dea === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.294Z [INFO] TestAuthMethodUpdateCommand: Synced node info === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.311Z [INFO] TestAuthMethodUpdateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-34d6a66f-02c8-5291-1371-cbcd10c89dea.dc1 writer.go:29: 2021-01-29T19:31:39.313Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-34d6a66f-02c8-5291-1371-cbcd10c89dea.dc1 area=wan writer.go:29: 2021-01-29T19:31:39.314Z [INFO] TestAuthMethodUpdateCommand_JSON.server: Updating LAN server: server="Node-34d6a66f-02c8-5291-1371-cbcd10c89dea (Addr: tcp/127.0.0.1:12048) (DC: dc1)" === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.302Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:39.309Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.leader: started routine: routine="acl token reaping" === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.309Z [WARN] TestAuthMethodUpdateCommand_noMerge.server: Configuring a non-UUID master token is deprecated === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.294Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.320Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.server: transitioning out of legacy ACL mode === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.318Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.serf.lan: serf: EventMemberUpdate: Node-0358cac9-a098-0d61-1416-89fd6c62f716 === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.298Z [INFO] TestAuthMethodUpdateCommand.server: Updating LAN server: server="Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 (Addr: tcp/127.0.0.1:12066) (DC: dc1)" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.325Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.309Z [INFO] TestAuthMethodUpdateCommand.server: Handled event for server in area: event=member-update server=Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3.dc1 area=wan writer.go:29: 2021-01-29T19:31:39.318Z [DEBUG] TestAuthMethodUpdateCommand: Node info in sync === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.311Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.serf.lan: serf: EventMemberUpdate: Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f === RUN TestAuthMethodUpdateCommand/update_without_name === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.356Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.serf.wan: serf: EventMemberUpdate: Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f.dc1 writer.go:29: 2021-01-29T19:31:39.358Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.serf.lan: serf: EventMemberUpdate: Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f === RUN TestAuthMethodUpdateCommand/update_nonexistent_method === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.362Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.serf.wan: serf: EventMemberUpdate: Node-0358cac9-a098-0d61-1416-89fd6c62f716.dc1 writer.go:29: 2021-01-29T19:31:39.320Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:39.368Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.serf.lan: serf: EventMemberUpdate: Node-0358cac9-a098-0d61-1416-89fd6c62f716 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.360Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.serf.wan: serf: EventMemberUpdate: Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f.dc1 writer.go:29: 2021-01-29T19:31:39.361Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: Updating LAN server: server="Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f (Addr: tcp/127.0.0.1:12060) (DC: dc1)" === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.371Z [DEBUG] TestAuthMethodUpdateCommand.http: Request finished: method=GET url=/v1/acl/auth-method/test from=127.0.0.1:58054 latency=2.37639ms === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.367Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: Updating LAN server: server="Node-0358cac9-a098-0d61-1416-89fd6c62f716 (Addr: tcp/127.0.0.1:12054) (DC: dc1)" === RUN TestAuthMethodUpdateCommand/update_all_fields === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.372Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: Updating LAN server: server="Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f (Addr: tcp/127.0.0.1:12060) (DC: dc1)" writer.go:29: 2021-01-29T19:31:39.361Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: Handled event for server in area: event=member-update server=Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f.dc1 area=wan writer.go:29: 2021-01-29T19:31:39.379Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: Handled event for server in area: event=member-update server=Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f.dc1 area=wan === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.367Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: Handled event for server in area: event=member-update server=Node-0358cac9-a098-0d61-1416-89fd6c62f716.dc1 area=wan writer.go:29: 2021-01-29T19:31:39.375Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.serf.wan: serf: EventMemberUpdate: Node-0358cac9-a098-0d61-1416-89fd6c62f716.dc1 === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.393Z [INFO] TestAuthMethodUpdateCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.377Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: Updating LAN server: server="Node-0358cac9-a098-0d61-1416-89fd6c62f716 (Addr: tcp/127.0.0.1:12054) (DC: dc1)" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.395Z [INFO] TestAuthMethodUpdateCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.389Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.382Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.398Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: Handled event for server in area: event=member-update server=Node-0358cac9-a098-0d61-1416-89fd6c62f716.dc1 area=wan === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.397Z [DEBUG] TestAuthMethodUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-34d6a66f-02c8-5291-1371-cbcd10c89dea writer.go:29: 2021-01-29T19:31:39.400Z [INFO] TestAuthMethodUpdateCommand_JSON.server: member joined, marking health alive: member=Node-34d6a66f-02c8-5291-1371-cbcd10c89dea === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.403Z [DEBUG] TestAuthMethodUpdateCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:58056 latency=12.548899ms writer.go:29: 2021-01-29T19:31:39.406Z [INFO] TestAuthMethodUpdateCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:39.420Z [INFO] TestAuthMethodUpdateCommand.leader: started routine: routine="CA root pruning" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.424Z [DEBUG] TestAuthMethodUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-34d6a66f-02c8-5291-1371-cbcd10c89dea writer.go:29: 2021-01-29T19:31:39.425Z [DEBUG] TestAuthMethodUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-34d6a66f-02c8-5291-1371-cbcd10c89dea === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.425Z [DEBUG] TestAuthMethodUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 writer.go:29: 2021-01-29T19:31:39.428Z [INFO] TestAuthMethodUpdateCommand.server: member joined, marking health alive: member=Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.437Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:39.454Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.leader: started routine: routine="CA root pruning" === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.463Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:39.464Z [INFO] TestAuthMethodUpdateCommand_noMerge.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.465Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.460Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-0358cac9-a098-0d61-1416-89fd6c62f716 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.467Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: member joined, marking health alive: member=Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.468Z [DEBUG] TestAuthMethodUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.469Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: member joined, marking health alive: member=Node-0358cac9-a098-0d61-1416-89fd6c62f716 === RUN TestAuthMethodUpdateCommand_JSON/update_without_name === RUN TestAuthMethodUpdateCommand_JSON/update_all_fields === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.475Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f === RUN TestAuthMethodUpdateCommand_noMerge/update_without_name === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.471Z [DEBUG] TestAuthMethodUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-cc725c06-25c7-a2db-89fc-0a7a950e36a3 === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.481Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-0358cac9-a098-0d61-1416-89fd6c62f716 writer.go:29: 2021-01-29T19:31:39.482Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-0358cac9-a098-0d61-1416-89fd6c62f716 === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.480Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-cb851575-f3ce-2efc-e02d-59ebafe1d53f === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.474Z [DEBUG] TestAuthMethodUpdateCommand.http: Request finished: method=GET url=/v1/acl/auth-method/test-016c00b8-2e06-b7af-f10b-41bd4800c863 from=127.0.0.1:58058 latency=1.412774ms === RUN TestAuthMethodUpdateCommand_noMerge/update_nonexistent_method === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.487Z [DEBUG] TestAuthMethodUpdateCommand.http: Request finished: method=PUT url=/v1/acl/auth-method/test-016c00b8-2e06-b7af-f10b-41bd4800c863 from=127.0.0.1:58058 latency=1.457462ms === RUN TestAuthMethodUpdateCommand_k8s_noMerge/update_missing_k8s_host === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.494Z [DEBUG] TestAuthMethodUpdateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:37812 latency=6.77127ms === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.494Z [DEBUG] TestAuthMethodUpdateCommand.http: Request finished: method=GET url=/v1/acl/auth-method/test-016c00b8-2e06-b7af-f10b-41bd4800c863 from=127.0.0.1:58056 latency=4.658117ms === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.501Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/auth-method/test from=127.0.0.1:44020 latency=2.20003ms === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.500Z [INFO] TestAuthMethodUpdateCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:39.503Z [INFO] TestAuthMethodUpdateCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:39.504Z [DEBUG] TestAuthMethodUpdateCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:39.505Z [DEBUG] TestAuthMethodUpdateCommand.leader: stopping routine: routine="acl token reaping" === RUN TestAuthMethodUpdateCommand_noMerge/update_all_fields === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.506Z [DEBUG] TestAuthMethodUpdateCommand.leader: stopping routine: routine="CA root pruning" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.507Z [DEBUG] TestAuthMethodUpdateCommand_JSON.http: Request finished: method=GET url=/v1/acl/auth-method/test-d9e22bf5-f267-97e9-1a7e-b4b3724ea938 from=127.0.0.1:37816 latency=3.181562ms === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.518Z [WARN] TestAuthMethodUpdateCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.507Z [DEBUG] TestAuthMethodUpdateCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:39.507Z [DEBUG] TestAuthMethodUpdateCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:39.521Z [DEBUG] TestAuthMethodUpdateCommand.leader: stopped routine: routine="CA root pruning" === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.529Z [DEBUG] TestAuthMethodUpdateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/auth-method/test-d9e22bf5-f267-97e9-1a7e-b4b3724ea938 from=127.0.0.1:37816 latency=6.449546ms === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:39.530Z [WARN] TestAuthMethodUpdateCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.534Z [INFO] TestAuthMethodUpdateCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:39.534Z [INFO] TestAuthMethodUpdateCommand: consul server down writer.go:29: 2021-01-29T19:31:39.536Z [INFO] TestAuthMethodUpdateCommand: shutdown complete writer.go:29: 2021-01-29T19:31:39.542Z [INFO] TestAuthMethodUpdateCommand: Stopping server: protocol=DNS address=127.0.0.1:12061 network=tcp writer.go:29: 2021-01-29T19:31:39.543Z [INFO] TestAuthMethodUpdateCommand: Stopping server: protocol=DNS address=127.0.0.1:12061 network=udp writer.go:29: 2021-01-29T19:31:39.545Z [INFO] TestAuthMethodUpdateCommand: Stopping server: protocol=HTTP address=127.0.0.1:12062 network=tcp === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.533Z [DEBUG] TestAuthMethodUpdateCommand_JSON.http: Request finished: method=GET url=/v1/acl/auth-method/test-d9e22bf5-f267-97e9-1a7e-b4b3724ea938 from=127.0.0.1:37812 latency=146.416µs === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.556Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:40634 latency=13.951572ms === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.561Z [INFO] TestAuthMethodUpdateCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:31:39.564Z [INFO] TestAuthMethodUpdateCommand_JSON.server: shutting down server === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.565Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-633fb1a1-12ae-0538-b1d7-04707a19e924 from=127.0.0.1:40638 latency=1.510979ms === RUN TestAuthMethodUpdateCommand_k8s_noMerge/update_missing_k8s_ca_cert === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.566Z [DEBUG] TestAuthMethodUpdateCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:39.570Z [DEBUG] TestAuthMethodUpdateCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.571Z [DEBUG] TestAuthMethodUpdateCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:39.582Z [WARN] TestAuthMethodUpdateCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.571Z [DEBUG] TestAuthMethodUpdateCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:39.571Z [DEBUG] TestAuthMethodUpdateCommand_JSON.leader: stopped routine: routine="CA root pruning" === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.585Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:44026 latency=40.270827ms === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.587Z [DEBUG] TestAuthMethodUpdateCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:39.605Z [WARN] TestAuthMethodUpdateCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.615Z [INFO] TestAuthMethodUpdateCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:39.616Z [INFO] TestAuthMethodUpdateCommand_JSON: consul server down writer.go:29: 2021-01-29T19:31:39.617Z [INFO] TestAuthMethodUpdateCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:31:39.623Z [INFO] TestAuthMethodUpdateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:12043 network=tcp === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.620Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:40634 latency=50.189899ms === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.630Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/auth-method/test-4a54724d-fb99-4c6a-1a6f-0d7a80665384 from=127.0.0.1:44030 latency=9.555845ms === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:39.625Z [INFO] TestAuthMethodUpdateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:12043 network=udp writer.go:29: 2021-01-29T19:31:39.634Z [INFO] TestAuthMethodUpdateCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:12044 network=tcp === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.641Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-ef36fec8-a5b7-2882-e03e-82e4e936ae52 from=127.0.0.1:40642 latency=1.164148ms === RUN TestAuthMethodUpdateCommand_k8s_noMerge/update_missing_k8s_jwt === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.642Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/auth-method/test-4a54724d-fb99-4c6a-1a6f-0d7a80665384 from=127.0.0.1:44030 latency=9.830297ms === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.653Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:40634 latency=8.756798ms === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.655Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/auth-method/test-4a54724d-fb99-4c6a-1a6f-0d7a80665384 from=127.0.0.1:44026 latency=3.457641ms writer.go:29: 2021-01-29T19:31:39.666Z [INFO] TestAuthMethodUpdateCommand_noMerge: Requesting shutdown writer.go:29: 2021-01-29T19:31:39.669Z [INFO] TestAuthMethodUpdateCommand_noMerge.server: shutting down server writer.go:29: 2021-01-29T19:31:39.670Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:39.672Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:39.673Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.671Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-c1bbbe98-2a43-4ce9-eb54-6362b3de00ee from=127.0.0.1:40644 latency=2.073995ms === RUN TestAuthMethodUpdateCommand_k8s_noMerge/update_all_fields === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.678Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.687Z [WARN] TestAuthMethodUpdateCommand_noMerge.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.691Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.678Z [DEBUG] TestAuthMethodUpdateCommand_noMerge.leader: stopped routine: routine="acl token reaping" === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.696Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:40634 latency=5.854665ms === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.700Z [WARN] TestAuthMethodUpdateCommand_noMerge.server.serf.wan: serf: Shutdown without a Leave === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.708Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-bf1fb863-fed4-4933-38e1-1d5249945004 from=127.0.0.1:40646 latency=2.851138ms === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:39.711Z [INFO] TestAuthMethodUpdateCommand_noMerge.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:39.713Z [INFO] TestAuthMethodUpdateCommand_noMerge: consul server down writer.go:29: 2021-01-29T19:31:39.716Z [INFO] TestAuthMethodUpdateCommand_noMerge: shutdown complete writer.go:29: 2021-01-29T19:31:39.718Z [INFO] TestAuthMethodUpdateCommand_noMerge: Stopping server: protocol=DNS address=127.0.0.1:12055 network=tcp writer.go:29: 2021-01-29T19:31:39.720Z [INFO] TestAuthMethodUpdateCommand_noMerge: Stopping server: protocol=DNS address=127.0.0.1:12055 network=udp writer.go:29: 2021-01-29T19:31:39.722Z [INFO] TestAuthMethodUpdateCommand_noMerge: Stopping server: protocol=HTTP address=127.0.0.1:12056 network=tcp === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.721Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=PUT url=/v1/acl/auth-method/k8s-bf1fb863-fed4-4933-38e1-1d5249945004 from=127.0.0.1:40646 latency=10.39675ms writer.go:29: 2021-01-29T19:31:39.730Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-bf1fb863-fed4-4933-38e1-1d5249945004 from=127.0.0.1:40634 latency=122.127µs === RUN TestAuthMethodUpdateCommand_k8s_noMerge/update_all_fields_with_cert_file === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:39.753Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:40634 latency=3.078829ms writer.go:29: 2021-01-29T19:31:39.762Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-cf962df6-9025-49b1-3058-ac28f9032c15 from=127.0.0.1:40648 latency=2.548097ms writer.go:29: 2021-01-29T19:31:39.767Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=PUT url=/v1/acl/auth-method/k8s-cf962df6-9025-49b1-3058-ac28f9032c15 from=127.0.0.1:40648 latency=3.071738ms writer.go:29: 2021-01-29T19:31:39.772Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-cf962df6-9025-49b1-3058-ac28f9032c15 from=127.0.0.1:40634 latency=232.265µs writer.go:29: 2021-01-29T19:31:39.782Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: Requesting shutdown writer.go:29: 2021-01-29T19:31:39.783Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server: shutting down server writer.go:29: 2021-01-29T19:31:39.784Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:39.785Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:39.786Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.785Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:39.787Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:39.790Z [WARN] TestAuthMethodUpdateCommand_k8s_noMerge.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.791Z [DEBUG] TestAuthMethodUpdateCommand_k8s_noMerge.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:39.794Z [WARN] TestAuthMethodUpdateCommand_k8s_noMerge.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:39.802Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:39.805Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: consul server down writer.go:29: 2021-01-29T19:31:39.806Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: shutdown complete writer.go:29: 2021-01-29T19:31:39.808Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: Stopping server: protocol=DNS address=127.0.0.1:12049 network=tcp writer.go:29: 2021-01-29T19:31:39.809Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: Stopping server: protocol=DNS address=127.0.0.1:12049 network=udp writer.go:29: 2021-01-29T19:31:39.811Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: Stopping server: protocol=HTTP address=127.0.0.1:12050 network=tcp === CONT TestAuthMethodUpdateCommand writer.go:29: 2021-01-29T19:31:40.057Z [INFO] TestAuthMethodUpdateCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:40.059Z [INFO] TestAuthMethodUpdateCommand: Endpoints down --- PASS: TestAuthMethodUpdateCommand (1.06s) --- PASS: TestAuthMethodUpdateCommand/update_without_name (0.00s) --- PASS: TestAuthMethodUpdateCommand/update_nonexistent_method (0.01s) --- PASS: TestAuthMethodUpdateCommand/update_all_fields (0.12s) === CONT TestAuthMethodUpdateCommand_k8s writer.go:29: 2021-01-29T19:31:40.092Z [WARN] TestAuthMethodUpdateCommand_k8s: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:40.094Z [DEBUG] TestAuthMethodUpdateCommand_k8s.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:40.096Z [DEBUG] TestAuthMethodUpdateCommand_k8s.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:40.111Z [INFO] TestAuthMethodUpdateCommand_k8s.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d23f8ce2-14ec-ed80-89b6-4be18eebe60b Address:127.0.0.1:12072}]" writer.go:29: 2021-01-29T19:31:40.119Z [INFO] TestAuthMethodUpdateCommand_k8s.server.raft: entering follower state: follower="Node at 127.0.0.1:12072 [Follower]" leader= === CONT TestAuthMethodUpdateCommand_JSON writer.go:29: 2021-01-29T19:31:40.136Z [INFO] TestAuthMethodUpdateCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:40.138Z [INFO] TestAuthMethodUpdateCommand_JSON: Endpoints down --- PASS: TestAuthMethodUpdateCommand_JSON (1.14s) --- PASS: TestAuthMethodUpdateCommand_JSON/update_without_name (0.00s) --- PASS: TestAuthMethodUpdateCommand_JSON/update_all_fields (0.08s) === CONT TestAuthMethodUpdateCommand_k8s writer.go:29: 2021-01-29T19:31:40.131Z [INFO] TestAuthMethodUpdateCommand_k8s.server.serf.wan: serf: EventMemberJoin: Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:40.151Z [INFO] TestAuthMethodUpdateCommand_k8s.server.serf.lan: serf: EventMemberJoin: Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b 127.0.0.1 writer.go:29: 2021-01-29T19:31:40.159Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Adding LAN server: server="Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b (Addr: tcp/127.0.0.1:12072) (DC: dc1)" writer.go:29: 2021-01-29T19:31:40.162Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Handled event for server in area: event=member-join server=Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b.dc1 area=wan writer.go:29: 2021-01-29T19:31:40.168Z [INFO] TestAuthMethodUpdateCommand_k8s: Started DNS server: address=127.0.0.1:12067 network=tcp writer.go:29: 2021-01-29T19:31:40.169Z [INFO] TestAuthMethodUpdateCommand_k8s: Started DNS server: address=127.0.0.1:12067 network=udp writer.go:29: 2021-01-29T19:31:40.172Z [INFO] TestAuthMethodUpdateCommand_k8s: Started HTTP server: address=127.0.0.1:12068 network=tcp writer.go:29: 2021-01-29T19:31:40.176Z [INFO] TestAuthMethodUpdateCommand_k8s: started state syncer writer.go:29: 2021-01-29T19:31:40.190Z [WARN] TestAuthMethodUpdateCommand_k8s.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:40.193Z [INFO] TestAuthMethodUpdateCommand_k8s.server.raft: entering candidate state: node="Node at 127.0.0.1:12072 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:40.197Z [DEBUG] TestAuthMethodUpdateCommand_k8s.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:40.200Z [DEBUG] TestAuthMethodUpdateCommand_k8s.server.raft: vote granted: from=d23f8ce2-14ec-ed80-89b6-4be18eebe60b term=2 tally=1 writer.go:29: 2021-01-29T19:31:40.204Z [INFO] TestAuthMethodUpdateCommand_k8s.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:40.209Z [INFO] TestAuthMethodUpdateCommand_k8s.server.raft: entering leader state: leader="Node at 127.0.0.1:12072 [Leader]" writer.go:29: 2021-01-29T19:31:40.212Z [INFO] TestAuthMethodUpdateCommand_k8s.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:40.213Z [INFO] TestAuthMethodUpdateCommand_k8s.server: New leader elected: payload=Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b writer.go:29: 2021-01-29T19:31:40.214Z [INFO] TestAuthMethodUpdateCommand_k8s.server: initializing acls writer.go:29: 2021-01-29T19:31:40.215Z [INFO] TestAuthMethodUpdateCommand_k8s.server: initializing acls === CONT TestAuthMethodUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:40.224Z [INFO] TestAuthMethodUpdateCommand_noMerge: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:40.225Z [INFO] TestAuthMethodUpdateCommand_noMerge: Endpoints down --- PASS: TestAuthMethodUpdateCommand_noMerge (1.22s) --- PASS: TestAuthMethodUpdateCommand_noMerge/update_without_name (0.00s) --- PASS: TestAuthMethodUpdateCommand_noMerge/update_nonexistent_method (0.02s) --- PASS: TestAuthMethodUpdateCommand_noMerge/update_all_fields (0.16s) === CONT TestAuthMethodUpdateCommand_k8s writer.go:29: 2021-01-29T19:31:40.224Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:40.226Z [WARN] TestAuthMethodUpdateCommand_k8s.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:40.225Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:40.229Z [WARN] TestAuthMethodUpdateCommand_k8s.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:40.240Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:40.240Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:40.243Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:40.244Z [INFO] TestAuthMethodUpdateCommand_k8s.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:40.246Z [INFO] TestAuthMethodUpdateCommand_k8s.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:40.247Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:40.249Z [DEBUG] TestAuthMethodUpdateCommand_k8s.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:40.250Z [INFO] TestAuthMethodUpdateCommand_k8s.server.serf.lan: serf: EventMemberUpdate: Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b writer.go:29: 2021-01-29T19:31:40.252Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Updating LAN server: server="Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b (Addr: tcp/127.0.0.1:12072) (DC: dc1)" writer.go:29: 2021-01-29T19:31:40.252Z [INFO] TestAuthMethodUpdateCommand_k8s.server.serf.lan: serf: EventMemberUpdate: Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b writer.go:29: 2021-01-29T19:31:40.256Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Updating LAN server: server="Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b (Addr: tcp/127.0.0.1:12072) (DC: dc1)" writer.go:29: 2021-01-29T19:31:40.254Z [INFO] TestAuthMethodUpdateCommand_k8s.server.serf.wan: serf: EventMemberUpdate: Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b.dc1 writer.go:29: 2021-01-29T19:31:40.261Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Handled event for server in area: event=member-update server=Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b.dc1 area=wan writer.go:29: 2021-01-29T19:31:40.261Z [INFO] TestAuthMethodUpdateCommand_k8s.server.serf.wan: serf: EventMemberUpdate: Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b.dc1 writer.go:29: 2021-01-29T19:31:40.269Z [INFO] TestAuthMethodUpdateCommand_k8s.server: Handled event for server in area: event=member-update server=Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b.dc1 area=wan writer.go:29: 2021-01-29T19:31:40.263Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:40.279Z [INFO] TestAuthMethodUpdateCommand_k8s.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:40.280Z [INFO] TestAuthMethodUpdateCommand_k8s.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:40.281Z [DEBUG] TestAuthMethodUpdateCommand_k8s.server: Skipping self join check for node since the cluster is too small: node=Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b writer.go:29: 2021-01-29T19:31:40.282Z [INFO] TestAuthMethodUpdateCommand_k8s.server: member joined, marking health alive: member=Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b writer.go:29: 2021-01-29T19:31:40.283Z [DEBUG] TestAuthMethodUpdateCommand_k8s.server: Skipping self join check for node since the cluster is too small: node=Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b writer.go:29: 2021-01-29T19:31:40.284Z [DEBUG] TestAuthMethodUpdateCommand_k8s.server: Skipping self join check for node since the cluster is too small: node=Node-d23f8ce2-14ec-ed80-89b6-4be18eebe60b writer.go:29: 2021-01-29T19:31:40.325Z [DEBUG] TestAuthMethodUpdateCommand_k8s: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:40.325Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: Waiting for endpoints to shut down === CONT TestAuthMethodUpdateCommand_k8s writer.go:29: 2021-01-29T19:31:40.328Z [INFO] TestAuthMethodUpdateCommand_k8s: Synced node info === CONT TestAuthMethodUpdateCommand_k8s_noMerge writer.go:29: 2021-01-29T19:31:40.328Z [INFO] TestAuthMethodUpdateCommand_k8s_noMerge: Endpoints down --- PASS: TestAuthMethodUpdateCommand_k8s_noMerge (1.33s) --- PASS: TestAuthMethodUpdateCommand_k8s_noMerge/update_missing_k8s_host (0.07s) --- PASS: TestAuthMethodUpdateCommand_k8s_noMerge/update_missing_k8s_ca_cert (0.07s) --- PASS: TestAuthMethodUpdateCommand_k8s_noMerge/update_missing_k8s_jwt (0.04s) --- PASS: TestAuthMethodUpdateCommand_k8s_noMerge/update_all_fields (0.06s) --- PASS: TestAuthMethodUpdateCommand_k8s_noMerge/update_all_fields_with_cert_file (0.04s) === RUN TestAuthMethodUpdateCommand_k8s/update_all_fields === CONT TestAuthMethodUpdateCommand_k8s writer.go:29: 2021-01-29T19:31:40.575Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:33296 latency=7.483452ms writer.go:29: 2021-01-29T19:31:40.583Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-7f910163-6cf3-6f7a-8642-d18665e548bc from=127.0.0.1:33298 latency=2.425938ms writer.go:29: 2021-01-29T19:31:40.595Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method/k8s-7f910163-6cf3-6f7a-8642-d18665e548bc from=127.0.0.1:33298 latency=3.563879ms writer.go:29: 2021-01-29T19:31:40.599Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-7f910163-6cf3-6f7a-8642-d18665e548bc from=127.0.0.1:33296 latency=259.751µs === RUN TestAuthMethodUpdateCommand_k8s/update_all_fields_with_cert_file === CONT TestAuthMethodUpdateCommand_k8s writer.go:29: 2021-01-29T19:31:40.614Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:33296 latency=2.300715ms writer.go:29: 2021-01-29T19:31:40.636Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-33d5a4a3-f558-8ccd-1242-68421f972b60 from=127.0.0.1:33300 latency=1.916759ms writer.go:29: 2021-01-29T19:31:40.644Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method/k8s-33d5a4a3-f558-8ccd-1242-68421f972b60 from=127.0.0.1:33300 latency=3.050589ms writer.go:29: 2021-01-29T19:31:40.647Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-33d5a4a3-f558-8ccd-1242-68421f972b60 from=127.0.0.1:33296 latency=123.726µs === RUN TestAuthMethodUpdateCommand_k8s/update_all_fields_but_k8s_host === CONT TestAuthMethodUpdateCommand_k8s writer.go:29: 2021-01-29T19:31:40.656Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:33296 latency=1.895786ms writer.go:29: 2021-01-29T19:31:40.662Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-bb7378e6-97e9-fcc7-76f4-9c74396baebe from=127.0.0.1:33302 latency=1.648353ms writer.go:29: 2021-01-29T19:31:40.669Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method/k8s-bb7378e6-97e9-fcc7-76f4-9c74396baebe from=127.0.0.1:33302 latency=1.651276ms writer.go:29: 2021-01-29T19:31:40.674Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-bb7378e6-97e9-fcc7-76f4-9c74396baebe from=127.0.0.1:33296 latency=181.033µs === RUN TestAuthMethodUpdateCommand_k8s/update_all_fields_but_k8s_ca_cert === CONT TestAuthMethodUpdateCommand_k8s writer.go:29: 2021-01-29T19:31:40.689Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:33296 latency=2.230213ms writer.go:29: 2021-01-29T19:31:40.698Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-94afcdc2-e20c-211f-3c57-9d3603d0937d from=127.0.0.1:33304 latency=4.264955ms writer.go:29: 2021-01-29T19:31:40.710Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method/k8s-94afcdc2-e20c-211f-3c57-9d3603d0937d from=127.0.0.1:33304 latency=1.740439ms writer.go:29: 2021-01-29T19:31:40.713Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-94afcdc2-e20c-211f-3c57-9d3603d0937d from=127.0.0.1:33296 latency=227.561µs === RUN TestAuthMethodUpdateCommand_k8s/update_all_fields_but_k8s_jwt === CONT TestAuthMethodUpdateCommand_k8s writer.go:29: 2021-01-29T19:31:40.723Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:33296 latency=1.853487ms writer.go:29: 2021-01-29T19:31:40.731Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-9dc01e08-f3df-a05b-1eae-2bacc6c9a027 from=127.0.0.1:33306 latency=2.206083ms writer.go:29: 2021-01-29T19:31:40.735Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method/k8s-9dc01e08-f3df-a05b-1eae-2bacc6c9a027 from=127.0.0.1:33306 latency=1.274465ms writer.go:29: 2021-01-29T19:31:40.738Z [DEBUG] TestAuthMethodUpdateCommand_k8s.http: Request finished: method=GET url=/v1/acl/auth-method/k8s-9dc01e08-f3df-a05b-1eae-2bacc6c9a027 from=127.0.0.1:33296 latency=76.878µs writer.go:29: 2021-01-29T19:31:40.746Z [INFO] TestAuthMethodUpdateCommand_k8s: Requesting shutdown writer.go:29: 2021-01-29T19:31:40.747Z [INFO] TestAuthMethodUpdateCommand_k8s.server: shutting down server writer.go:29: 2021-01-29T19:31:40.749Z [DEBUG] TestAuthMethodUpdateCommand_k8s.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:40.751Z [DEBUG] TestAuthMethodUpdateCommand_k8s.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:40.753Z [DEBUG] TestAuthMethodUpdateCommand_k8s.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:40.751Z [DEBUG] TestAuthMethodUpdateCommand_k8s.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:40.753Z [DEBUG] TestAuthMethodUpdateCommand_k8s.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:40.755Z [DEBUG] TestAuthMethodUpdateCommand_k8s.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:40.755Z [WARN] TestAuthMethodUpdateCommand_k8s.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:40.765Z [WARN] TestAuthMethodUpdateCommand_k8s.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:40.768Z [INFO] TestAuthMethodUpdateCommand_k8s.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:40.768Z [INFO] TestAuthMethodUpdateCommand_k8s: consul server down writer.go:29: 2021-01-29T19:31:40.770Z [INFO] TestAuthMethodUpdateCommand_k8s: shutdown complete writer.go:29: 2021-01-29T19:31:40.770Z [INFO] TestAuthMethodUpdateCommand_k8s: Stopping server: protocol=DNS address=127.0.0.1:12067 network=tcp writer.go:29: 2021-01-29T19:31:40.771Z [INFO] TestAuthMethodUpdateCommand_k8s: Stopping server: protocol=DNS address=127.0.0.1:12067 network=udp writer.go:29: 2021-01-29T19:31:40.772Z [INFO] TestAuthMethodUpdateCommand_k8s: Stopping server: protocol=HTTP address=127.0.0.1:12068 network=tcp writer.go:29: 2021-01-29T19:31:41.274Z [INFO] TestAuthMethodUpdateCommand_k8s: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:41.275Z [INFO] TestAuthMethodUpdateCommand_k8s: Endpoints down --- PASS: TestAuthMethodUpdateCommand_k8s (1.22s) --- PASS: TestAuthMethodUpdateCommand_k8s/update_all_fields (0.04s) --- PASS: TestAuthMethodUpdateCommand_k8s/update_all_fields_with_cert_file (0.04s) --- PASS: TestAuthMethodUpdateCommand_k8s/update_all_fields_but_k8s_host (0.03s) --- PASS: TestAuthMethodUpdateCommand_k8s/update_all_fields_but_k8s_ca_cert (0.04s) --- PASS: TestAuthMethodUpdateCommand_k8s/update_all_fields_but_k8s_jwt (0.03s) PASS ok github.com/hashicorp/consul/command/acl/authmethod/update 2.419s ? github.com/hashicorp/consul/command/acl/bindingrule [no test files] === RUN TestBindingRuleCreateCommand_noTabs === PAUSE TestBindingRuleCreateCommand_noTabs === RUN TestBindingRuleCreateCommand === PAUSE TestBindingRuleCreateCommand === RUN TestBindingRuleCreateCommand_JSON === PAUSE TestBindingRuleCreateCommand_JSON === CONT TestBindingRuleCreateCommand_noTabs === CONT TestBindingRuleCreateCommand_JSON === CONT TestBindingRuleCreateCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestBindingRuleCreateCommand_noTabs (0.00s) === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.115Z [WARN] TestBindingRuleCreateCommand_JSON: bootstrap = true: do not enable unless necessary === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.116Z [WARN] TestBindingRuleCreateCommand: bootstrap = true: do not enable unless necessary === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.117Z [DEBUG] TestBindingRuleCreateCommand_JSON.tlsutil: Update: version=1 === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.117Z [DEBUG] TestBindingRuleCreateCommand.tlsutil: Update: version=1 === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.119Z [DEBUG] TestBindingRuleCreateCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.119Z [DEBUG] TestBindingRuleCreateCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.123Z [INFO] TestBindingRuleCreateCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f9ba69df-7a87-a262-72f9-8229dbd9d06e Address:127.0.0.1:25321}]" === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.124Z [INFO] TestBindingRuleCreateCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ed9c3672-b50f-d505-e9db-60e5fa2322e2 Address:127.0.0.1:25327}]" === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.126Z [INFO] TestBindingRuleCreateCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e.dc1 127.0.0.1 === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.126Z [INFO] TestBindingRuleCreateCommand.server.serf.wan: serf: EventMemberJoin: Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:41.133Z [INFO] TestBindingRuleCreateCommand.server.serf.lan: serf: EventMemberJoin: Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2 127.0.0.1 === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.132Z [INFO] TestBindingRuleCreateCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e 127.0.0.1 === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.142Z [INFO] TestBindingRuleCreateCommand: Started DNS server: address=127.0.0.1:25322 network=udp writer.go:29: 2021-01-29T19:31:41.147Z [INFO] TestBindingRuleCreateCommand.server: Adding LAN server: server="Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2 (Addr: tcp/127.0.0.1:25327) (DC: dc1)" === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.127Z [INFO] TestBindingRuleCreateCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:25321 [Follower]" leader= writer.go:29: 2021-01-29T19:31:41.148Z [INFO] TestBindingRuleCreateCommand_JSON: Started DNS server: address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:31:41.161Z [INFO] TestBindingRuleCreateCommand_JSON: Started DNS server: address=127.0.0.1:25316 network=tcp writer.go:29: 2021-01-29T19:31:41.156Z [INFO] TestBindingRuleCreateCommand_JSON.server: Handled event for server in area: event=member-join server=Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e.dc1 area=wan writer.go:29: 2021-01-29T19:31:41.158Z [INFO] TestBindingRuleCreateCommand_JSON.server: Adding LAN server: server="Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e (Addr: tcp/127.0.0.1:25321) (DC: dc1)" === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.128Z [INFO] TestBindingRuleCreateCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:25327 [Follower]" leader= === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.163Z [INFO] TestBindingRuleCreateCommand_JSON: Started HTTP server: address=127.0.0.1:25317 network=tcp === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.158Z [INFO] TestBindingRuleCreateCommand.server: Handled event for server in area: event=member-join server=Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2.dc1 area=wan === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.166Z [INFO] TestBindingRuleCreateCommand_JSON: started state syncer === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.159Z [INFO] TestBindingRuleCreateCommand: Started DNS server: address=127.0.0.1:25322 network=tcp writer.go:29: 2021-01-29T19:31:41.168Z [INFO] TestBindingRuleCreateCommand: Started HTTP server: address=127.0.0.1:25323 network=tcp writer.go:29: 2021-01-29T19:31:41.170Z [INFO] TestBindingRuleCreateCommand: started state syncer === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.202Z [WARN] TestBindingRuleCreateCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:41.204Z [INFO] TestBindingRuleCreateCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:25321 [Candidate]" term=2 === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.208Z [DEBUG] TestBindingRuleCreateCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.208Z [DEBUG] TestBindingRuleCreateCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:41.209Z [DEBUG] TestBindingRuleCreateCommand_JSON.server.raft: vote granted: from=f9ba69df-7a87-a262-72f9-8229dbd9d06e term=2 tally=1 writer.go:29: 2021-01-29T19:31:41.208Z [DEBUG] TestBindingRuleCreateCommand_JSON.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:41.211Z [INFO] TestBindingRuleCreateCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:41.214Z [INFO] TestBindingRuleCreateCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:25321 [Leader]" writer.go:29: 2021-01-29T19:31:41.215Z [INFO] TestBindingRuleCreateCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:41.217Z [INFO] TestBindingRuleCreateCommand_JSON.server: New leader elected: payload=Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.219Z [WARN] TestBindingRuleCreateCommand.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.217Z [INFO] TestBindingRuleCreateCommand_JSON.server: initializing acls === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.220Z [INFO] TestBindingRuleCreateCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:25327 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:41.223Z [DEBUG] TestBindingRuleCreateCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:41.225Z [DEBUG] TestBindingRuleCreateCommand.server.raft: vote granted: from=ed9c3672-b50f-d505-e9db-60e5fa2322e2 term=2 tally=1 === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.222Z [INFO] TestBindingRuleCreateCommand_JSON.server: Created ACL 'global-management' policy === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.227Z [INFO] TestBindingRuleCreateCommand.server.raft: election won: tally=1 === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.228Z [WARN] TestBindingRuleCreateCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:41.229Z [ERROR] TestBindingRuleCreateCommand_JSON.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.228Z [INFO] TestBindingRuleCreateCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:25327 [Leader]" writer.go:29: 2021-01-29T19:31:41.230Z [INFO] TestBindingRuleCreateCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:41.231Z [INFO] TestBindingRuleCreateCommand.server: New leader elected: payload=Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2 === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.230Z [INFO] TestBindingRuleCreateCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.233Z [INFO] TestBindingRuleCreateCommand.server: initializing acls === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.234Z [INFO] TestBindingRuleCreateCommand_JSON.server: Created ACL anonymous token from configuration === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.237Z [INFO] TestBindingRuleCreateCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:41.238Z [WARN] TestBindingRuleCreateCommand.server: Configuring a non-UUID master token is deprecated === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.237Z [INFO] TestBindingRuleCreateCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.239Z [INFO] TestBindingRuleCreateCommand.server: Bootstrapped ACL master token from configuration === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.240Z [INFO] TestBindingRuleCreateCommand_JSON.leader: started routine: routine="acl token reaping" === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.241Z [INFO] TestBindingRuleCreateCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:41.242Z [INFO] TestBindingRuleCreateCommand.leader: started routine: routine="legacy ACL token upgrade" === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.242Z [INFO] TestBindingRuleCreateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.243Z [INFO] TestBindingRuleCreateCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:41.245Z [INFO] TestBindingRuleCreateCommand.server.serf.lan: serf: EventMemberUpdate: Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2 === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.245Z [INFO] TestBindingRuleCreateCommand_JSON.server: Updating LAN server: server="Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e (Addr: tcp/127.0.0.1:25321) (DC: dc1)" === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.246Z [INFO] TestBindingRuleCreateCommand.server.serf.wan: serf: EventMemberUpdate: Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2.dc1 === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.245Z [INFO] TestBindingRuleCreateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e.dc1 === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.247Z [INFO] TestBindingRuleCreateCommand.server: Updating LAN server: server="Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2 (Addr: tcp/127.0.0.1:25327) (DC: dc1)" === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.249Z [INFO] TestBindingRuleCreateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e.dc1 area=wan === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.247Z [INFO] TestBindingRuleCreateCommand.server: Handled event for server in area: event=member-update server=Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2.dc1 area=wan === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.251Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.258Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.280Z [INFO] TestBindingRuleCreateCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:41.282Z [INFO] TestBindingRuleCreateCommand_JSON.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.285Z [DEBUG] TestBindingRuleCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e writer.go:29: 2021-01-29T19:31:41.288Z [INFO] TestBindingRuleCreateCommand_JSON.server: member joined, marking health alive: member=Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e writer.go:29: 2021-01-29T19:31:41.292Z [DEBUG] TestBindingRuleCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-f9ba69df-7a87-a262-72f9-8229dbd9d06e === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.297Z [INFO] TestBindingRuleCreateCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:41.298Z [INFO] TestBindingRuleCreateCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.299Z [DEBUG] TestBindingRuleCreateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2 writer.go:29: 2021-01-29T19:31:41.300Z [INFO] TestBindingRuleCreateCommand.server: member joined, marking health alive: member=Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2 writer.go:29: 2021-01-29T19:31:41.307Z [DEBUG] TestBindingRuleCreateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-ed9c3672-b50f-d505-e9db-60e5fa2322e2 === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.355Z [DEBUG] TestBindingRuleCreateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:53804 latency=5.148548ms === RUN TestBindingRuleCreateCommand_JSON/create_it_with_no_selector === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.384Z [DEBUG] TestBindingRuleCreateCommand: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.389Z [DEBUG] TestBindingRuleCreateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53806 latency=25.825204ms === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.390Z [INFO] TestBindingRuleCreateCommand: Synced node info writer.go:29: 2021-01-29T19:31:41.392Z [DEBUG] TestBindingRuleCreateCommand: Node info in sync === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.404Z [INFO] TestBindingRuleCreateCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:31:41.405Z [INFO] TestBindingRuleCreateCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:31:41.406Z [DEBUG] TestBindingRuleCreateCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:41.407Z [DEBUG] TestBindingRuleCreateCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:41.408Z [DEBUG] TestBindingRuleCreateCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.409Z [WARN] TestBindingRuleCreateCommand_JSON.server.serf.lan: serf: Shutdown without a Leave === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.409Z [DEBUG] TestBindingRuleCreateCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:33030 latency=3.608911ms === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.408Z [DEBUG] TestBindingRuleCreateCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" === RUN TestBindingRuleCreateCommand/method_is_required === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.408Z [DEBUG] TestBindingRuleCreateCommand_JSON.leader: stopped routine: routine="acl token reaping" === RUN TestBindingRuleCreateCommand/bind_type_required === RUN TestBindingRuleCreateCommand/bind_name_required === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.409Z [DEBUG] TestBindingRuleCreateCommand_JSON.leader: stopped routine: routine="CA root pruning" === RUN TestBindingRuleCreateCommand/must_use_roughly_valid_selector === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.413Z [WARN] TestBindingRuleCreateCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:41.434Z [INFO] TestBindingRuleCreateCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:41.436Z [INFO] TestBindingRuleCreateCommand_JSON: consul server down writer.go:29: 2021-01-29T19:31:41.438Z [INFO] TestBindingRuleCreateCommand_JSON: shutdown complete === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.440Z [ERROR] TestBindingRuleCreateCommand.http: Request error: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:33032 error="invalid Binding Rule: Selector is invalid: 1:4 (3): no match found, expected: "!=", ".", "==", "[", [ \t\r\n] or [a-zA-Z0-9_]" === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.439Z [INFO] TestBindingRuleCreateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:25316 network=tcp === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.441Z [DEBUG] TestBindingRuleCreateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:33032 latency=12.507868ms === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.441Z [INFO] TestBindingRuleCreateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:25316 network=udp === RUN TestBindingRuleCreateCommand/create_it_with_no_selector === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.443Z [INFO] TestBindingRuleCreateCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:25317 network=tcp === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.450Z [DEBUG] TestBindingRuleCreateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:33034 latency=2.04403ms === RUN TestBindingRuleCreateCommand/create_it_with_a_match_selector === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.466Z [DEBUG] TestBindingRuleCreateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:33036 latency=5.838237ms === RUN TestBindingRuleCreateCommand/create_it_with_type_role === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:41.495Z [DEBUG] TestBindingRuleCreateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:33038 latency=6.694736ms writer.go:29: 2021-01-29T19:31:41.499Z [INFO] TestBindingRuleCreateCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:41.500Z [INFO] TestBindingRuleCreateCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:41.500Z [DEBUG] TestBindingRuleCreateCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:41.501Z [DEBUG] TestBindingRuleCreateCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:41.502Z [DEBUG] TestBindingRuleCreateCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.503Z [WARN] TestBindingRuleCreateCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:41.501Z [DEBUG] TestBindingRuleCreateCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:41.502Z [DEBUG] TestBindingRuleCreateCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:41.503Z [DEBUG] TestBindingRuleCreateCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:41.505Z [WARN] TestBindingRuleCreateCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:41.511Z [INFO] TestBindingRuleCreateCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:41.511Z [INFO] TestBindingRuleCreateCommand: consul server down writer.go:29: 2021-01-29T19:31:41.513Z [INFO] TestBindingRuleCreateCommand: shutdown complete writer.go:29: 2021-01-29T19:31:41.514Z [INFO] TestBindingRuleCreateCommand: Stopping server: protocol=DNS address=127.0.0.1:25322 network=tcp writer.go:29: 2021-01-29T19:31:41.514Z [INFO] TestBindingRuleCreateCommand: Stopping server: protocol=DNS address=127.0.0.1:25322 network=udp writer.go:29: 2021-01-29T19:31:41.515Z [INFO] TestBindingRuleCreateCommand: Stopping server: protocol=HTTP address=127.0.0.1:25323 network=tcp === CONT TestBindingRuleCreateCommand_JSON writer.go:29: 2021-01-29T19:31:41.944Z [INFO] TestBindingRuleCreateCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:41.946Z [INFO] TestBindingRuleCreateCommand_JSON: Endpoints down --- PASS: TestBindingRuleCreateCommand_JSON (0.86s) --- PASS: TestBindingRuleCreateCommand_JSON/create_it_with_no_selector (0.05s) === CONT TestBindingRuleCreateCommand writer.go:29: 2021-01-29T19:31:42.020Z [INFO] TestBindingRuleCreateCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:42.022Z [INFO] TestBindingRuleCreateCommand: Endpoints down --- PASS: TestBindingRuleCreateCommand (0.94s) --- PASS: TestBindingRuleCreateCommand/method_is_required (0.00s) --- PASS: TestBindingRuleCreateCommand/bind_type_required (0.00s) --- PASS: TestBindingRuleCreateCommand/bind_name_required (0.00s) --- PASS: TestBindingRuleCreateCommand/must_use_roughly_valid_selector (0.02s) --- PASS: TestBindingRuleCreateCommand/create_it_with_no_selector (0.01s) --- PASS: TestBindingRuleCreateCommand/create_it_with_a_match_selector (0.02s) --- PASS: TestBindingRuleCreateCommand/create_it_with_type_role (0.03s) PASS ok github.com/hashicorp/consul/command/acl/bindingrule/create 1.044s === RUN TestBindingRuleDeleteCommand_noTabs === PAUSE TestBindingRuleDeleteCommand_noTabs === RUN TestBindingRuleDeleteCommand === PAUSE TestBindingRuleDeleteCommand === CONT TestBindingRuleDeleteCommand_noTabs === CONT TestBindingRuleDeleteCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestBindingRuleDeleteCommand_noTabs (0.01s) === CONT TestBindingRuleDeleteCommand writer.go:29: 2021-01-29T19:31:42.685Z [WARN] TestBindingRuleDeleteCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:42.687Z [DEBUG] TestBindingRuleDeleteCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:42.689Z [DEBUG] TestBindingRuleDeleteCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:42.695Z [INFO] TestBindingRuleDeleteCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:38514117-e587-c555-05ce-e569b681440d Address:127.0.0.1:21237}]" writer.go:29: 2021-01-29T19:31:42.699Z [INFO] TestBindingRuleDeleteCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:21237 [Follower]" leader= writer.go:29: 2021-01-29T19:31:42.700Z [INFO] TestBindingRuleDeleteCommand.server.serf.wan: serf: EventMemberJoin: Node-38514117-e587-c555-05ce-e569b681440d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:42.713Z [INFO] TestBindingRuleDeleteCommand.server.serf.lan: serf: EventMemberJoin: Node-38514117-e587-c555-05ce-e569b681440d 127.0.0.1 writer.go:29: 2021-01-29T19:31:42.717Z [INFO] TestBindingRuleDeleteCommand.server: Adding LAN server: server="Node-38514117-e587-c555-05ce-e569b681440d (Addr: tcp/127.0.0.1:21237) (DC: dc1)" writer.go:29: 2021-01-29T19:31:42.717Z [INFO] TestBindingRuleDeleteCommand: Started DNS server: address=127.0.0.1:21232 network=udp writer.go:29: 2021-01-29T19:31:42.718Z [INFO] TestBindingRuleDeleteCommand.server: Handled event for server in area: event=member-join server=Node-38514117-e587-c555-05ce-e569b681440d.dc1 area=wan writer.go:29: 2021-01-29T19:31:42.720Z [INFO] TestBindingRuleDeleteCommand: Started DNS server: address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:31:42.722Z [INFO] TestBindingRuleDeleteCommand: Started HTTP server: address=127.0.0.1:21233 network=tcp writer.go:29: 2021-01-29T19:31:42.723Z [INFO] TestBindingRuleDeleteCommand: started state syncer writer.go:29: 2021-01-29T19:31:42.768Z [WARN] TestBindingRuleDeleteCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:42.772Z [INFO] TestBindingRuleDeleteCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:21237 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:42.770Z [DEBUG] TestBindingRuleDeleteCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:42.775Z [DEBUG] TestBindingRuleDeleteCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:42.779Z [DEBUG] TestBindingRuleDeleteCommand.server.raft: vote granted: from=38514117-e587-c555-05ce-e569b681440d term=2 tally=1 writer.go:29: 2021-01-29T19:31:42.781Z [INFO] TestBindingRuleDeleteCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:42.784Z [INFO] TestBindingRuleDeleteCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:21237 [Leader]" writer.go:29: 2021-01-29T19:31:42.786Z [INFO] TestBindingRuleDeleteCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:42.787Z [INFO] TestBindingRuleDeleteCommand.server: New leader elected: payload=Node-38514117-e587-c555-05ce-e569b681440d writer.go:29: 2021-01-29T19:31:42.787Z [INFO] TestBindingRuleDeleteCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:42.812Z [INFO] TestBindingRuleDeleteCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:42.814Z [WARN] TestBindingRuleDeleteCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:42.818Z [INFO] TestBindingRuleDeleteCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:42.821Z [INFO] TestBindingRuleDeleteCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:42.826Z [INFO] TestBindingRuleDeleteCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:42.828Z [INFO] TestBindingRuleDeleteCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:42.832Z [INFO] TestBindingRuleDeleteCommand.server.serf.lan: serf: EventMemberUpdate: Node-38514117-e587-c555-05ce-e569b681440d writer.go:29: 2021-01-29T19:31:42.835Z [INFO] TestBindingRuleDeleteCommand.server: Updating LAN server: server="Node-38514117-e587-c555-05ce-e569b681440d (Addr: tcp/127.0.0.1:21237) (DC: dc1)" writer.go:29: 2021-01-29T19:31:42.835Z [INFO] TestBindingRuleDeleteCommand.server.serf.wan: serf: EventMemberUpdate: Node-38514117-e587-c555-05ce-e569b681440d.dc1 writer.go:29: 2021-01-29T19:31:42.839Z [INFO] TestBindingRuleDeleteCommand.server: Handled event for server in area: event=member-update server=Node-38514117-e587-c555-05ce-e569b681440d.dc1 area=wan writer.go:29: 2021-01-29T19:31:42.840Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:42.934Z [INFO] TestBindingRuleDeleteCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:42.935Z [INFO] TestBindingRuleDeleteCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:42.938Z [DEBUG] TestBindingRuleDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-38514117-e587-c555-05ce-e569b681440d writer.go:29: 2021-01-29T19:31:42.941Z [INFO] TestBindingRuleDeleteCommand.server: member joined, marking health alive: member=Node-38514117-e587-c555-05ce-e569b681440d writer.go:29: 2021-01-29T19:31:42.956Z [DEBUG] TestBindingRuleDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-38514117-e587-c555-05ce-e569b681440d writer.go:29: 2021-01-29T19:31:43.159Z [DEBUG] TestBindingRuleDeleteCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:43.184Z [INFO] TestBindingRuleDeleteCommand: Synced node info writer.go:29: 2021-01-29T19:31:43.186Z [DEBUG] TestBindingRuleDeleteCommand: Node info in sync writer.go:29: 2021-01-29T19:31:43.185Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:41814 latency=33.340938ms === RUN TestBindingRuleDeleteCommand/id_required === RUN TestBindingRuleDeleteCommand/delete_works === CONT TestBindingRuleDeleteCommand writer.go:29: 2021-01-29T19:31:43.207Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:41814 latency=9.497818ms writer.go:29: 2021-01-29T19:31:43.227Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=DELETE url=/v1/acl/binding-rule/8f6b480b-d8c9-a36c-948f-eed59cd23e67 from=127.0.0.1:41816 latency=2.792939ms writer.go:29: 2021-01-29T19:31:43.232Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/8f6b480b-d8c9-a36c-948f-eed59cd23e67 from=127.0.0.1:41814 latency=75.498µs === RUN TestBindingRuleDeleteCommand/delete_works_via_prefixes === CONT TestBindingRuleDeleteCommand writer.go:29: 2021-01-29T19:31:43.238Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:41814 latency=1.027359ms writer.go:29: 2021-01-29T19:31:43.243Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:41818 latency=217.964µs writer.go:29: 2021-01-29T19:31:43.247Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=DELETE url=/v1/acl/binding-rule/665d6210-f126-f0a9-fca8-50e8de510830 from=127.0.0.1:41818 latency=2.284905ms writer.go:29: 2021-01-29T19:31:43.252Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/665d6210-f126-f0a9-fca8-50e8de510830 from=127.0.0.1:41814 latency=55.355µs === RUN TestBindingRuleDeleteCommand/delete_fails_when_prefix_matches_more_than_one_rule === CONT TestBindingRuleDeleteCommand writer.go:29: 2021-01-29T19:31:43.257Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:41814 latency=180.816µs writer.go:29: 2021-01-29T19:31:43.270Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:41814 latency=10.613839ms writer.go:29: 2021-01-29T19:31:43.274Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:41814 latency=128.274µs writer.go:29: 2021-01-29T19:31:43.277Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:41814 latency=786.247µs writer.go:29: 2021-01-29T19:31:43.279Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:41814 latency=70.239µs writer.go:29: 2021-01-29T19:31:43.282Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:41814 latency=659.932µs writer.go:29: 2021-01-29T19:31:43.284Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:41814 latency=73.026µs writer.go:29: 2021-01-29T19:31:43.288Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:41814 latency=848.364µs writer.go:29: 2021-01-29T19:31:43.290Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:41814 latency=164.625µs writer.go:29: 2021-01-29T19:31:43.293Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:41814 latency=819.722µs writer.go:29: 2021-01-29T19:31:43.295Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:41814 latency=281.981µs writer.go:29: 2021-01-29T19:31:43.313Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:41814 latency=13.951133ms writer.go:29: 2021-01-29T19:31:43.323Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:41814 latency=1.57128ms writer.go:29: 2021-01-29T19:31:43.345Z [DEBUG] TestBindingRuleDeleteCommand.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:41820 latency=229.669µs writer.go:29: 2021-01-29T19:31:43.347Z [INFO] TestBindingRuleDeleteCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:43.349Z [INFO] TestBindingRuleDeleteCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:43.350Z [DEBUG] TestBindingRuleDeleteCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:43.351Z [DEBUG] TestBindingRuleDeleteCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:43.352Z [DEBUG] TestBindingRuleDeleteCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:43.352Z [DEBUG] TestBindingRuleDeleteCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:43.352Z [DEBUG] TestBindingRuleDeleteCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:43.354Z [DEBUG] TestBindingRuleDeleteCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:43.354Z [WARN] TestBindingRuleDeleteCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:43.358Z [WARN] TestBindingRuleDeleteCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:43.361Z [INFO] TestBindingRuleDeleteCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:43.361Z [INFO] TestBindingRuleDeleteCommand: consul server down writer.go:29: 2021-01-29T19:31:43.363Z [INFO] TestBindingRuleDeleteCommand: shutdown complete writer.go:29: 2021-01-29T19:31:43.364Z [INFO] TestBindingRuleDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:31:43.365Z [INFO] TestBindingRuleDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:21232 network=udp writer.go:29: 2021-01-29T19:31:43.367Z [INFO] TestBindingRuleDeleteCommand: Stopping server: protocol=HTTP address=127.0.0.1:21233 network=tcp writer.go:29: 2021-01-29T19:31:43.869Z [INFO] TestBindingRuleDeleteCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:43.870Z [INFO] TestBindingRuleDeleteCommand: Endpoints down --- PASS: TestBindingRuleDeleteCommand (1.26s) --- PASS: TestBindingRuleDeleteCommand/id_required (0.00s) --- PASS: TestBindingRuleDeleteCommand/delete_works (0.04s) --- PASS: TestBindingRuleDeleteCommand/delete_works_via_prefixes (0.02s) --- PASS: TestBindingRuleDeleteCommand/delete_fails_when_prefix_matches_more_than_one_rule (0.09s) PASS ok github.com/hashicorp/consul/command/acl/bindingrule/delete 1.559s === RUN TestBindingRuleListCommand_noTabs === PAUSE TestBindingRuleListCommand_noTabs === RUN TestBindingRuleListCommand === PAUSE TestBindingRuleListCommand === CONT TestBindingRuleListCommand_noTabs --- PASS: TestBindingRuleListCommand_noTabs (0.00s) === CONT TestBindingRuleListCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:31:48.153Z [WARN] TestBindingRuleListCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:48.155Z [DEBUG] TestBindingRuleListCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:48.158Z [DEBUG] TestBindingRuleListCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:48.164Z [INFO] TestBindingRuleListCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:837a17fd-0efb-62d8-3bd4-753466288e35 Address:127.0.0.1:25321}]" writer.go:29: 2021-01-29T19:31:48.167Z [INFO] TestBindingRuleListCommand.server.serf.wan: serf: EventMemberJoin: Node-837a17fd-0efb-62d8-3bd4-753466288e35.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:48.167Z [INFO] TestBindingRuleListCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:25321 [Follower]" leader= writer.go:29: 2021-01-29T19:31:48.171Z [INFO] TestBindingRuleListCommand.server.serf.lan: serf: EventMemberJoin: Node-837a17fd-0efb-62d8-3bd4-753466288e35 127.0.0.1 writer.go:29: 2021-01-29T19:31:48.175Z [INFO] TestBindingRuleListCommand.server: Adding LAN server: server="Node-837a17fd-0efb-62d8-3bd4-753466288e35 (Addr: tcp/127.0.0.1:25321) (DC: dc1)" writer.go:29: 2021-01-29T19:31:48.176Z [INFO] TestBindingRuleListCommand.server: Handled event for server in area: event=member-join server=Node-837a17fd-0efb-62d8-3bd4-753466288e35.dc1 area=wan writer.go:29: 2021-01-29T19:31:48.179Z [INFO] TestBindingRuleListCommand: Started DNS server: address=127.0.0.1:25316 network=tcp writer.go:29: 2021-01-29T19:31:48.180Z [INFO] TestBindingRuleListCommand: Started DNS server: address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:31:48.181Z [INFO] TestBindingRuleListCommand: Started HTTP server: address=127.0.0.1:25317 network=tcp writer.go:29: 2021-01-29T19:31:48.182Z [INFO] TestBindingRuleListCommand: started state syncer writer.go:29: 2021-01-29T19:31:48.228Z [DEBUG] TestBindingRuleListCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:48.234Z [WARN] TestBindingRuleListCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:48.235Z [INFO] TestBindingRuleListCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:25321 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:48.238Z [DEBUG] TestBindingRuleListCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:48.240Z [DEBUG] TestBindingRuleListCommand.server.raft: vote granted: from=837a17fd-0efb-62d8-3bd4-753466288e35 term=2 tally=1 writer.go:29: 2021-01-29T19:31:48.242Z [INFO] TestBindingRuleListCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:48.243Z [INFO] TestBindingRuleListCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:25321 [Leader]" writer.go:29: 2021-01-29T19:31:48.247Z [INFO] TestBindingRuleListCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:48.248Z [INFO] TestBindingRuleListCommand.server: New leader elected: payload=Node-837a17fd-0efb-62d8-3bd4-753466288e35 writer.go:29: 2021-01-29T19:31:48.249Z [INFO] TestBindingRuleListCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:48.253Z [INFO] TestBindingRuleListCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:48.255Z [WARN] TestBindingRuleListCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:48.257Z [INFO] TestBindingRuleListCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:48.260Z [INFO] TestBindingRuleListCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:48.261Z [INFO] TestBindingRuleListCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:48.263Z [INFO] TestBindingRuleListCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:48.264Z [INFO] TestBindingRuleListCommand.server.serf.lan: serf: EventMemberUpdate: Node-837a17fd-0efb-62d8-3bd4-753466288e35 writer.go:29: 2021-01-29T19:31:48.267Z [INFO] TestBindingRuleListCommand.server.serf.wan: serf: EventMemberUpdate: Node-837a17fd-0efb-62d8-3bd4-753466288e35.dc1 writer.go:29: 2021-01-29T19:31:48.269Z [INFO] TestBindingRuleListCommand.server: Updating LAN server: server="Node-837a17fd-0efb-62d8-3bd4-753466288e35 (Addr: tcp/127.0.0.1:25321) (DC: dc1)" writer.go:29: 2021-01-29T19:31:48.269Z [INFO] TestBindingRuleListCommand.server: Handled event for server in area: event=member-update server=Node-837a17fd-0efb-62d8-3bd4-753466288e35.dc1 area=wan writer.go:29: 2021-01-29T19:31:48.271Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:48.302Z [INFO] TestBindingRuleListCommand: Synced node info writer.go:29: 2021-01-29T19:31:48.373Z [INFO] TestBindingRuleListCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:48.375Z [INFO] TestBindingRuleListCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:48.377Z [DEBUG] TestBindingRuleListCommand.server: Skipping self join check for node since the cluster is too small: node=Node-837a17fd-0efb-62d8-3bd4-753466288e35 writer.go:29: 2021-01-29T19:31:48.379Z [INFO] TestBindingRuleListCommand.server: member joined, marking health alive: member=Node-837a17fd-0efb-62d8-3bd4-753466288e35 writer.go:29: 2021-01-29T19:31:48.382Z [DEBUG] TestBindingRuleListCommand.server: Skipping self join check for node since the cluster is too small: node=Node-837a17fd-0efb-62d8-3bd4-753466288e35 writer.go:29: 2021-01-29T19:31:48.404Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:53826 latency=8.781497ms writer.go:29: 2021-01-29T19:31:48.426Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:53826 latency=12.508419ms writer.go:29: 2021-01-29T19:31:48.433Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53826 latency=4.98206ms writer.go:29: 2021-01-29T19:31:48.436Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53826 latency=597.423µs writer.go:29: 2021-01-29T19:31:48.438Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53826 latency=562.481µs writer.go:29: 2021-01-29T19:31:48.440Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53826 latency=907.299µs writer.go:29: 2021-01-29T19:31:48.443Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53826 latency=493.573µs writer.go:29: 2021-01-29T19:31:48.446Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53826 latency=1.343985ms writer.go:29: 2021-01-29T19:31:48.449Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53826 latency=855.78µs writer.go:29: 2021-01-29T19:31:48.451Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53826 latency=723.323µs writer.go:29: 2021-01-29T19:31:48.454Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53826 latency=840.296µs writer.go:29: 2021-01-29T19:31:48.457Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:53826 latency=708.712µs === RUN TestBindingRuleListCommand/normal === CONT TestBindingRuleListCommand writer.go:29: 2021-01-29T19:31:48.463Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:53828 latency=121.687µs === RUN TestBindingRuleListCommand/filter_by_method_1 === CONT TestBindingRuleListCommand writer.go:29: 2021-01-29T19:31:48.477Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test-1 from=127.0.0.1:53830 latency=93.087µs === RUN TestBindingRuleListCommand/filter_by_method_2 === CONT TestBindingRuleListCommand writer.go:29: 2021-01-29T19:31:48.492Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test-2 from=127.0.0.1:53832 latency=2.438695ms === RUN TestBindingRuleListCommand/normal_json_formatted === CONT TestBindingRuleListCommand writer.go:29: 2021-01-29T19:31:48.507Z [DEBUG] TestBindingRuleListCommand.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:53834 latency=101.498µs writer.go:29: 2021-01-29T19:31:48.524Z [INFO] TestBindingRuleListCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:48.526Z [INFO] TestBindingRuleListCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:48.527Z [DEBUG] TestBindingRuleListCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:48.528Z [DEBUG] TestBindingRuleListCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:48.530Z [DEBUG] TestBindingRuleListCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:48.528Z [DEBUG] TestBindingRuleListCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:48.530Z [DEBUG] TestBindingRuleListCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:48.532Z [WARN] TestBindingRuleListCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:48.532Z [DEBUG] TestBindingRuleListCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:48.540Z [WARN] TestBindingRuleListCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:48.544Z [INFO] TestBindingRuleListCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:48.544Z [INFO] TestBindingRuleListCommand: consul server down writer.go:29: 2021-01-29T19:31:48.546Z [INFO] TestBindingRuleListCommand: shutdown complete writer.go:29: 2021-01-29T19:31:48.548Z [INFO] TestBindingRuleListCommand: Stopping server: protocol=DNS address=127.0.0.1:25316 network=tcp writer.go:29: 2021-01-29T19:31:48.549Z [INFO] TestBindingRuleListCommand: Stopping server: protocol=DNS address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:31:48.550Z [INFO] TestBindingRuleListCommand: Stopping server: protocol=HTTP address=127.0.0.1:25317 network=tcp writer.go:29: 2021-01-29T19:31:49.051Z [INFO] TestBindingRuleListCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:49.053Z [INFO] TestBindingRuleListCommand: Endpoints down --- PASS: TestBindingRuleListCommand (0.96s) --- PASS: TestBindingRuleListCommand/normal (0.01s) --- PASS: TestBindingRuleListCommand/filter_by_method_1 (0.01s) --- PASS: TestBindingRuleListCommand/filter_by_method_2 (0.02s) --- PASS: TestBindingRuleListCommand/normal_json_formatted (0.02s) PASS ok github.com/hashicorp/consul/command/acl/bindingrule/list 1.172s === RUN TestBindingRuleReadCommand_noTabs === PAUSE TestBindingRuleReadCommand_noTabs === RUN TestBindingRuleReadCommand === PAUSE TestBindingRuleReadCommand === CONT TestBindingRuleReadCommand_noTabs === CONT TestBindingRuleReadCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestBindingRuleReadCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestBindingRuleReadCommand writer.go:29: 2021-01-29T19:31:49.570Z [WARN] TestBindingRuleReadCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:49.571Z [DEBUG] TestBindingRuleReadCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:49.573Z [DEBUG] TestBindingRuleReadCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:49.579Z [INFO] TestBindingRuleReadCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fcea8600-2f5c-0505-484c-088fda679108 Address:127.0.0.1:24300}]" writer.go:29: 2021-01-29T19:31:49.580Z [INFO] TestBindingRuleReadCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:24300 [Follower]" leader= writer.go:29: 2021-01-29T19:31:49.584Z [INFO] TestBindingRuleReadCommand.server.serf.wan: serf: EventMemberJoin: Node-fcea8600-2f5c-0505-484c-088fda679108.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:49.586Z [INFO] TestBindingRuleReadCommand.server.serf.lan: serf: EventMemberJoin: Node-fcea8600-2f5c-0505-484c-088fda679108 127.0.0.1 writer.go:29: 2021-01-29T19:31:49.588Z [INFO] TestBindingRuleReadCommand: Started DNS server: address=127.0.0.1:24295 network=udp writer.go:29: 2021-01-29T19:31:49.589Z [INFO] TestBindingRuleReadCommand.server: Adding LAN server: server="Node-fcea8600-2f5c-0505-484c-088fda679108 (Addr: tcp/127.0.0.1:24300) (DC: dc1)" writer.go:29: 2021-01-29T19:31:49.589Z [INFO] TestBindingRuleReadCommand.server: Handled event for server in area: event=member-join server=Node-fcea8600-2f5c-0505-484c-088fda679108.dc1 area=wan writer.go:29: 2021-01-29T19:31:49.590Z [INFO] TestBindingRuleReadCommand: Started DNS server: address=127.0.0.1:24295 network=tcp writer.go:29: 2021-01-29T19:31:49.592Z [INFO] TestBindingRuleReadCommand: Started HTTP server: address=127.0.0.1:24296 network=tcp writer.go:29: 2021-01-29T19:31:49.593Z [INFO] TestBindingRuleReadCommand: started state syncer writer.go:29: 2021-01-29T19:31:49.640Z [DEBUG] TestBindingRuleReadCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:49.642Z [WARN] TestBindingRuleReadCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:49.643Z [INFO] TestBindingRuleReadCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:24300 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:49.645Z [DEBUG] TestBindingRuleReadCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:49.646Z [DEBUG] TestBindingRuleReadCommand.server.raft: vote granted: from=fcea8600-2f5c-0505-484c-088fda679108 term=2 tally=1 writer.go:29: 2021-01-29T19:31:49.647Z [INFO] TestBindingRuleReadCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:49.648Z [INFO] TestBindingRuleReadCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:24300 [Leader]" writer.go:29: 2021-01-29T19:31:49.649Z [INFO] TestBindingRuleReadCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:49.650Z [INFO] TestBindingRuleReadCommand.server: New leader elected: payload=Node-fcea8600-2f5c-0505-484c-088fda679108 writer.go:29: 2021-01-29T19:31:49.660Z [INFO] TestBindingRuleReadCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:49.665Z [INFO] TestBindingRuleReadCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:49.667Z [WARN] TestBindingRuleReadCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:49.670Z [INFO] TestBindingRuleReadCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:49.673Z [INFO] TestBindingRuleReadCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:49.674Z [INFO] TestBindingRuleReadCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:49.676Z [INFO] TestBindingRuleReadCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:49.678Z [INFO] TestBindingRuleReadCommand.server.serf.lan: serf: EventMemberUpdate: Node-fcea8600-2f5c-0505-484c-088fda679108 writer.go:29: 2021-01-29T19:31:49.681Z [INFO] TestBindingRuleReadCommand.server.serf.wan: serf: EventMemberUpdate: Node-fcea8600-2f5c-0505-484c-088fda679108.dc1 writer.go:29: 2021-01-29T19:31:49.681Z [INFO] TestBindingRuleReadCommand.server: Updating LAN server: server="Node-fcea8600-2f5c-0505-484c-088fda679108 (Addr: tcp/127.0.0.1:24300) (DC: dc1)" writer.go:29: 2021-01-29T19:31:49.683Z [INFO] TestBindingRuleReadCommand.server: Handled event for server in area: event=member-update server=Node-fcea8600-2f5c-0505-484c-088fda679108.dc1 area=wan writer.go:29: 2021-01-29T19:31:49.685Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:49.726Z [INFO] TestBindingRuleReadCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:49.727Z [INFO] TestBindingRuleReadCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:49.730Z [DEBUG] TestBindingRuleReadCommand.server: Skipping self join check for node since the cluster is too small: node=Node-fcea8600-2f5c-0505-484c-088fda679108 writer.go:29: 2021-01-29T19:31:49.732Z [INFO] TestBindingRuleReadCommand.server: member joined, marking health alive: member=Node-fcea8600-2f5c-0505-484c-088fda679108 writer.go:29: 2021-01-29T19:31:49.740Z [DEBUG] TestBindingRuleReadCommand.server: Skipping self join check for node since the cluster is too small: node=Node-fcea8600-2f5c-0505-484c-088fda679108 writer.go:29: 2021-01-29T19:31:49.792Z [DEBUG] TestBindingRuleReadCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:49.795Z [INFO] TestBindingRuleReadCommand: Synced node info writer.go:29: 2021-01-29T19:31:49.797Z [DEBUG] TestBindingRuleReadCommand: Node info in sync writer.go:29: 2021-01-29T19:31:49.996Z [DEBUG] TestBindingRuleReadCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:37724 latency=11.564879ms === RUN TestBindingRuleReadCommand/id_required === RUN TestBindingRuleReadCommand/read_by_id_not_found === CONT TestBindingRuleReadCommand writer.go:29: 2021-01-29T19:31:50.046Z [DEBUG] TestBindingRuleReadCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/b7b5f45c-381f-1f42-d5ce-235c8d20f893 from=127.0.0.1:37726 latency=2.439849ms === RUN TestBindingRuleReadCommand/read_by_id === CONT TestBindingRuleReadCommand writer.go:29: 2021-01-29T19:31:50.083Z [DEBUG] TestBindingRuleReadCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:37724 latency=21.490908ms writer.go:29: 2021-01-29T19:31:50.093Z [DEBUG] TestBindingRuleReadCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/43a092c4-9d67-4779-50a9-68f7069811f0 from=127.0.0.1:37728 latency=3.21825ms === RUN TestBindingRuleReadCommand/read_by_id_prefix === CONT TestBindingRuleReadCommand writer.go:29: 2021-01-29T19:31:50.105Z [DEBUG] TestBindingRuleReadCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:37724 latency=2.212033ms writer.go:29: 2021-01-29T19:31:50.126Z [DEBUG] TestBindingRuleReadCommand.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:37730 latency=2.787529ms writer.go:29: 2021-01-29T19:31:50.131Z [DEBUG] TestBindingRuleReadCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/3b4652bf-2148-2191-d547-bba1bc3a1cf0 from=127.0.0.1:37730 latency=413.251µs === RUN TestBindingRuleReadCommand/read_by_id_json_formatted === CONT TestBindingRuleReadCommand writer.go:29: 2021-01-29T19:31:50.142Z [DEBUG] TestBindingRuleReadCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:37724 latency=1.780166ms writer.go:29: 2021-01-29T19:31:50.165Z [DEBUG] TestBindingRuleReadCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/4059eb98-3b89-655c-9c26-7970424145e9 from=127.0.0.1:37732 latency=2.912361ms writer.go:29: 2021-01-29T19:31:50.171Z [INFO] TestBindingRuleReadCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:50.174Z [INFO] TestBindingRuleReadCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:50.176Z [DEBUG] TestBindingRuleReadCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:50.177Z [DEBUG] TestBindingRuleReadCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:50.179Z [DEBUG] TestBindingRuleReadCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:50.178Z [DEBUG] TestBindingRuleReadCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:50.179Z [DEBUG] TestBindingRuleReadCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:50.181Z [WARN] TestBindingRuleReadCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:50.181Z [DEBUG] TestBindingRuleReadCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:50.188Z [WARN] TestBindingRuleReadCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:50.193Z [INFO] TestBindingRuleReadCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:50.195Z [INFO] TestBindingRuleReadCommand: consul server down writer.go:29: 2021-01-29T19:31:50.196Z [INFO] TestBindingRuleReadCommand: shutdown complete writer.go:29: 2021-01-29T19:31:50.197Z [INFO] TestBindingRuleReadCommand: Stopping server: protocol=DNS address=127.0.0.1:24295 network=tcp writer.go:29: 2021-01-29T19:31:50.199Z [INFO] TestBindingRuleReadCommand: Stopping server: protocol=DNS address=127.0.0.1:24295 network=udp writer.go:29: 2021-01-29T19:31:50.200Z [INFO] TestBindingRuleReadCommand: Stopping server: protocol=HTTP address=127.0.0.1:24296 network=tcp writer.go:29: 2021-01-29T19:31:50.701Z [INFO] TestBindingRuleReadCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:50.703Z [INFO] TestBindingRuleReadCommand: Endpoints down --- PASS: TestBindingRuleReadCommand (1.19s) --- PASS: TestBindingRuleReadCommand/id_required (0.00s) --- PASS: TestBindingRuleReadCommand/read_by_id_not_found (0.05s) --- PASS: TestBindingRuleReadCommand/read_by_id (0.04s) --- PASS: TestBindingRuleReadCommand/read_by_id_prefix (0.04s) --- PASS: TestBindingRuleReadCommand/read_by_id_json_formatted (0.03s) PASS ok github.com/hashicorp/consul/command/acl/bindingrule/read 1.294s === RUN TestBindingRuleUpdateCommand_noTabs === PAUSE TestBindingRuleUpdateCommand_noTabs === RUN TestBindingRuleUpdateCommand === PAUSE TestBindingRuleUpdateCommand === RUN TestBindingRuleUpdateCommand_noMerge === PAUSE TestBindingRuleUpdateCommand_noMerge === CONT TestBindingRuleUpdateCommand_noTabs --- PASS: TestBindingRuleUpdateCommand_noTabs (0.00s) === CONT TestBindingRuleUpdateCommand_noMerge [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestBindingRuleUpdateCommand [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:31:51.104Z [WARN] TestBindingRuleUpdateCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:51.105Z [DEBUG] TestBindingRuleUpdateCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:51.106Z [DEBUG] TestBindingRuleUpdateCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:51.112Z [INFO] TestBindingRuleUpdateCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6a51ed67-a879-0e32-fb99-131b78abd1c0 Address:127.0.0.1:19201}]" writer.go:29: 2021-01-29T19:31:51.114Z [INFO] TestBindingRuleUpdateCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:19201 [Follower]" leader= === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.118Z [WARN] TestBindingRuleUpdateCommand_noMerge: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:51.119Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.tlsutil: Update: version=1 === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.116Z [INFO] TestBindingRuleUpdateCommand.server.serf.wan: serf: EventMemberJoin: Node-6a51ed67-a879-0e32-fb99-131b78abd1c0.dc1 127.0.0.1 === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.120Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.123Z [INFO] TestBindingRuleUpdateCommand.server.serf.lan: serf: EventMemberJoin: Node-6a51ed67-a879-0e32-fb99-131b78abd1c0 127.0.0.1 writer.go:29: 2021-01-29T19:31:51.126Z [INFO] TestBindingRuleUpdateCommand.server: Adding LAN server: server="Node-6a51ed67-a879-0e32-fb99-131b78abd1c0 (Addr: tcp/127.0.0.1:19201) (DC: dc1)" writer.go:29: 2021-01-29T19:31:51.127Z [INFO] TestBindingRuleUpdateCommand: Started DNS server: address=127.0.0.1:19196 network=udp writer.go:29: 2021-01-29T19:31:51.127Z [INFO] TestBindingRuleUpdateCommand.server: Handled event for server in area: event=member-join server=Node-6a51ed67-a879-0e32-fb99-131b78abd1c0.dc1 area=wan writer.go:29: 2021-01-29T19:31:51.130Z [INFO] TestBindingRuleUpdateCommand: Started DNS server: address=127.0.0.1:19196 network=tcp === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.152Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7f099f42-6329-7ab2-60a0-721e18d02010 Address:127.0.0.1:19195}]" === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.156Z [INFO] TestBindingRuleUpdateCommand: Started HTTP server: address=127.0.0.1:19197 network=tcp writer.go:29: 2021-01-29T19:31:51.158Z [INFO] TestBindingRuleUpdateCommand: started state syncer === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.155Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.raft: entering follower state: follower="Node at 127.0.0.1:19195 [Follower]" leader= writer.go:29: 2021-01-29T19:31:51.155Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.serf.wan: serf: EventMemberJoin: Node-7f099f42-6329-7ab2-60a0-721e18d02010.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:51.165Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.serf.lan: serf: EventMemberJoin: Node-7f099f42-6329-7ab2-60a0-721e18d02010 127.0.0.1 writer.go:29: 2021-01-29T19:31:51.168Z [INFO] TestBindingRuleUpdateCommand_noMerge: Started DNS server: address=127.0.0.1:19190 network=udp writer.go:29: 2021-01-29T19:31:51.169Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Adding LAN server: server="Node-7f099f42-6329-7ab2-60a0-721e18d02010 (Addr: tcp/127.0.0.1:19195) (DC: dc1)" writer.go:29: 2021-01-29T19:31:51.172Z [INFO] TestBindingRuleUpdateCommand_noMerge: Started DNS server: address=127.0.0.1:19190 network=tcp writer.go:29: 2021-01-29T19:31:51.169Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Handled event for server in area: event=member-join server=Node-7f099f42-6329-7ab2-60a0-721e18d02010.dc1 area=wan === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.178Z [DEBUG] TestBindingRuleUpdateCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.181Z [INFO] TestBindingRuleUpdateCommand_noMerge: Started HTTP server: address=127.0.0.1:19191 network=tcp writer.go:29: 2021-01-29T19:31:51.182Z [INFO] TestBindingRuleUpdateCommand_noMerge: started state syncer === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.186Z [WARN] TestBindingRuleUpdateCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:51.187Z [INFO] TestBindingRuleUpdateCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:19201 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:51.189Z [DEBUG] TestBindingRuleUpdateCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:51.190Z [DEBUG] TestBindingRuleUpdateCommand.server.raft: vote granted: from=6a51ed67-a879-0e32-fb99-131b78abd1c0 term=2 tally=1 writer.go:29: 2021-01-29T19:31:51.191Z [INFO] TestBindingRuleUpdateCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:51.192Z [INFO] TestBindingRuleUpdateCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:19201 [Leader]" writer.go:29: 2021-01-29T19:31:51.193Z [INFO] TestBindingRuleUpdateCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:51.194Z [INFO] TestBindingRuleUpdateCommand.server: New leader elected: payload=Node-6a51ed67-a879-0e32-fb99-131b78abd1c0 writer.go:29: 2021-01-29T19:31:51.196Z [INFO] TestBindingRuleUpdateCommand.server: initializing acls === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.203Z [WARN] TestBindingRuleUpdateCommand_noMerge.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:51.206Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.raft: entering candidate state: node="Node at 127.0.0.1:19195 [Candidate]" term=2 === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.201Z [INFO] TestBindingRuleUpdateCommand.server: Created ACL 'global-management' policy === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.209Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:51.211Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.server.raft: vote granted: from=7f099f42-6329-7ab2-60a0-721e18d02010 term=2 tally=1 writer.go:29: 2021-01-29T19:31:51.214Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.raft: election won: tally=1 === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.209Z [WARN] TestBindingRuleUpdateCommand.server: Configuring a non-UUID master token is deprecated === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.216Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.raft: entering leader state: leader="Node at 127.0.0.1:19195 [Leader]" writer.go:29: 2021-01-29T19:31:51.221Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: cluster leadership acquired === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.222Z [INFO] TestBindingRuleUpdateCommand.server: Bootstrapped ACL master token from configuration === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.223Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: initializing acls === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.224Z [INFO] TestBindingRuleUpdateCommand.server: Created ACL anonymous token from configuration === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.221Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: New leader elected: payload=Node-7f099f42-6329-7ab2-60a0-721e18d02010 === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.226Z [INFO] TestBindingRuleUpdateCommand.leader: started routine: routine="legacy ACL token upgrade" === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.222Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: initializing acls === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.227Z [INFO] TestBindingRuleUpdateCommand.leader: started routine: routine="acl token reaping" === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.231Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Created ACL 'global-management' policy === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.229Z [INFO] TestBindingRuleUpdateCommand.server.serf.lan: serf: EventMemberUpdate: Node-6a51ed67-a879-0e32-fb99-131b78abd1c0 === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.233Z [WARN] TestBindingRuleUpdateCommand_noMerge.server: Configuring a non-UUID master token is deprecated === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.234Z [INFO] TestBindingRuleUpdateCommand.server: Updating LAN server: server="Node-6a51ed67-a879-0e32-fb99-131b78abd1c0 (Addr: tcp/127.0.0.1:19201) (DC: dc1)" writer.go:29: 2021-01-29T19:31:51.234Z [INFO] TestBindingRuleUpdateCommand.server.serf.wan: serf: EventMemberUpdate: Node-6a51ed67-a879-0e32-fb99-131b78abd1c0.dc1 === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.231Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Created ACL 'global-management' policy === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.236Z [INFO] TestBindingRuleUpdateCommand.server: Handled event for server in area: event=member-update server=Node-6a51ed67-a879-0e32-fb99-131b78abd1c0.dc1 area=wan === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.237Z [WARN] TestBindingRuleUpdateCommand_noMerge.server: Configuring a non-UUID master token is deprecated === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.239Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.242Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:51.243Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:51.246Z [INFO] TestBindingRuleUpdateCommand_noMerge.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:51.244Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:51.247Z [INFO] TestBindingRuleUpdateCommand_noMerge.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:51.250Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:51.257Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.serf.lan: serf: EventMemberUpdate: Node-7f099f42-6329-7ab2-60a0-721e18d02010 writer.go:29: 2021-01-29T19:31:51.266Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Updating LAN server: server="Node-7f099f42-6329-7ab2-60a0-721e18d02010 (Addr: tcp/127.0.0.1:19195) (DC: dc1)" writer.go:29: 2021-01-29T19:31:51.266Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.serf.wan: serf: EventMemberUpdate: Node-7f099f42-6329-7ab2-60a0-721e18d02010.dc1 writer.go:29: 2021-01-29T19:31:51.266Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.serf.lan: serf: EventMemberUpdate: Node-7f099f42-6329-7ab2-60a0-721e18d02010 writer.go:29: 2021-01-29T19:31:51.273Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.serf.wan: serf: EventMemberUpdate: Node-7f099f42-6329-7ab2-60a0-721e18d02010.dc1 writer.go:29: 2021-01-29T19:31:51.270Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Handled event for server in area: event=member-update server=Node-7f099f42-6329-7ab2-60a0-721e18d02010.dc1 area=wan writer.go:29: 2021-01-29T19:31:51.277Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Handled event for server in area: event=member-update server=Node-7f099f42-6329-7ab2-60a0-721e18d02010.dc1 area=wan writer.go:29: 2021-01-29T19:31:51.271Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:51.276Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: Updating LAN server: server="Node-7f099f42-6329-7ab2-60a0-721e18d02010 (Addr: tcp/127.0.0.1:19195) (DC: dc1)" writer.go:29: 2021-01-29T19:31:51.342Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:51.344Z [INFO] TestBindingRuleUpdateCommand_noMerge.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:51.345Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-7f099f42-6329-7ab2-60a0-721e18d02010 writer.go:29: 2021-01-29T19:31:51.347Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: member joined, marking health alive: member=Node-7f099f42-6329-7ab2-60a0-721e18d02010 writer.go:29: 2021-01-29T19:31:51.350Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-7f099f42-6329-7ab2-60a0-721e18d02010 writer.go:29: 2021-01-29T19:31:51.352Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-7f099f42-6329-7ab2-60a0-721e18d02010 === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.353Z [INFO] TestBindingRuleUpdateCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:51.355Z [INFO] TestBindingRuleUpdateCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:51.357Z [DEBUG] TestBindingRuleUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-6a51ed67-a879-0e32-fb99-131b78abd1c0 writer.go:29: 2021-01-29T19:31:51.358Z [INFO] TestBindingRuleUpdateCommand.server: member joined, marking health alive: member=Node-6a51ed67-a879-0e32-fb99-131b78abd1c0 writer.go:29: 2021-01-29T19:31:51.395Z [DEBUG] TestBindingRuleUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-6a51ed67-a879-0e32-fb99-131b78abd1c0 === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.452Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:59402 latency=52.714907ms === RUN TestBindingRuleUpdateCommand_noMerge/rule_id_required === RUN TestBindingRuleUpdateCommand_noMerge/rule_id_partial_matches_nothing === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.464Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:56360 latency=4.363825ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.484Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:59406 latency=17.590587ms === RUN TestBindingRuleUpdateCommand/rule_id_required === RUN TestBindingRuleUpdateCommand/rule_id_partial_matches_nothing === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.523Z [DEBUG] TestBindingRuleUpdateCommand_noMerge: Skipping remote check since it is managed automatically: check=serfHealth === RUN TestBindingRuleUpdateCommand_noMerge/rule_id_exact_match_doesn't_exist === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.532Z [INFO] TestBindingRuleUpdateCommand_noMerge: Synced node info writer.go:29: 2021-01-29T19:31:51.533Z [DEBUG] TestBindingRuleUpdateCommand_noMerge: Node info in sync === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.532Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:56364 latency=1.973529ms writer.go:29: 2021-01-29T19:31:51.533Z [DEBUG] TestBindingRuleUpdateCommand: Skipping remote check since it is managed automatically: check=serfHealth === RUN TestBindingRuleUpdateCommand/rule_id_exact_match_doesn't_exist === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.541Z [INFO] TestBindingRuleUpdateCommand: Synced node info === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.543Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/46662220-b746-14e4-aee0-088a7dcc7719 from=127.0.0.1:59410 latency=5.332047ms === RUN TestBindingRuleUpdateCommand_noMerge/rule_id_partial_matches_multiple === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.547Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/06cc94c9-fe22-cb91-1e9d-b0f52d697266 from=127.0.0.1:56368 latency=2.304115ms === RUN TestBindingRuleUpdateCommand/rule_id_partial_matches_multiple === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.554Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:59402 latency=297.765µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.556Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:56360 latency=2.013119ms writer.go:29: 2021-01-29T19:31:51.563Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56360 latency=5.238658ms writer.go:29: 2021-01-29T19:31:51.566Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:56360 latency=397.964µs === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.569Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59402 latency=13.07174ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.570Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56360 latency=854.188µs writer.go:29: 2021-01-29T19:31:51.573Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:56360 latency=133.189µs === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.573Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:59402 latency=415.734µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.576Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56360 latency=1.03343ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.577Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59402 latency=1.131033ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.581Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:56360 latency=1.831175ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.580Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:59402 latency=141.019µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.585Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56360 latency=1.662682ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.586Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59402 latency=1.432129ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.587Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:56360 latency=73.246µs === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.588Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:59402 latency=69.07µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.591Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56360 latency=1.960995ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.591Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59402 latency=1.053804ms writer.go:29: 2021-01-29T19:31:51.593Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:59402 latency=79.919µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.593Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:56360 latency=227.337µs writer.go:29: 2021-01-29T19:31:51.596Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56360 latency=1.072412ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.598Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59402 latency=3.11199ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.599Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:56360 latency=193.919µs === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.601Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:59402 latency=95.626µs writer.go:29: 2021-01-29T19:31:51.604Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59402 latency=695.578µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.605Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:56370 latency=97.271µs === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.606Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:59402 latency=95.853µs === RUN TestBindingRuleUpdateCommand/must_use_roughly_valid_selector === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.611Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56360 latency=895.45µs === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.612Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59402 latency=2.91441ms writer.go:29: 2021-01-29T19:31:51.616Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:59402 latency=237.882µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.621Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/a8a6d1c0-bcaa-8cc6-711c-e3bd5735ec60 from=127.0.0.1:56372 latency=1.669994ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.623Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59402 latency=1.873886ms writer.go:29: 2021-01-29T19:31:51.639Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:59402 latency=242.931µs writer.go:29: 2021-01-29T19:31:51.646Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:59418 latency=93.047µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.646Z [ERROR] TestBindingRuleUpdateCommand.http: Request error: method=PUT url=/v1/acl/binding-rule/a8a6d1c0-bcaa-8cc6-711c-e3bd5735ec60 from=127.0.0.1:56372 error="invalid Binding Rule: Selector is invalid: 1:4 (3): no match found, expected: "!=", ".", "==", "[", [ \t\r\n] or [a-zA-Z0-9_]" === RUN TestBindingRuleUpdateCommand_noMerge/must_use_roughly_valid_selector === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.648Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule/a8a6d1c0-bcaa-8cc6-711c-e3bd5735ec60 from=127.0.0.1:56372 latency=10.851814ms === RUN TestBindingRuleUpdateCommand/update_all_fields === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.665Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59402 latency=12.049919ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.669Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56360 latency=11.623586ms writer.go:29: 2021-01-29T19:31:51.678Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/6b3197fd-192b-13aa-9053-84cf39e29138 from=127.0.0.1:56378 latency=1.25326ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.680Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/30531324-5382-b452-e901-41e980225462 from=127.0.0.1:59420 latency=3.636577ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.687Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule/6b3197fd-192b-13aa-9053-84cf39e29138 from=127.0.0.1:56378 latency=7.128179ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.690Z [ERROR] TestBindingRuleUpdateCommand_noMerge.http: Request error: method=PUT url=/v1/acl/binding-rule/30531324-5382-b452-e901-41e980225462 from=127.0.0.1:59420 error="invalid Binding Rule: Selector is invalid: 1:4 (3): no match found, expected: "!=", ".", "==", "[", [ \t\r\n] or [a-zA-Z0-9_]" === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.690Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/6b3197fd-192b-13aa-9053-84cf39e29138 from=127.0.0.1:56360 latency=148.448µs === RUN TestBindingRuleUpdateCommand/update_all_fields_-_partial === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.692Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule/30531324-5382-b452-e901-41e980225462 from=127.0.0.1:59420 latency=8.733181ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.696Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:56360 latency=107.181µs === RUN TestBindingRuleUpdateCommand_noMerge/update_all_fields === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.703Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59402 latency=1.637825ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.700Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=DELETE url=/v1/acl/binding-rule/084e4195-b376-135c-d2b2-a9744ccdb876 from=127.0.0.1:56360 latency=637.017µs === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.711Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/d55883e9-8917-0619-0dca-6147a8fdc93c from=127.0.0.1:59424 latency=1.692713ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.724Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=DELETE url=/v1/acl/binding-rule/0dcc7237-eebc-8e6f-1945-eefe20318e7d from=127.0.0.1:56382 latency=8.323756ms writer.go:29: 2021-01-29T19:31:51.738Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=DELETE url=/v1/acl/binding-rule/361456ec-11a4-32d7-a941-cfca77b1387b from=127.0.0.1:56384 latency=2.750417ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.743Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule/d55883e9-8917-0619-0dca-6147a8fdc93c from=127.0.0.1:59424 latency=23.01607ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.748Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=DELETE url=/v1/acl/binding-rule/5a4aa09e-883a-a2a3-59bb-476db0ff0591 from=127.0.0.1:56386 latency=2.239677ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.748Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/d55883e9-8917-0619-0dca-6147a8fdc93c from=127.0.0.1:59402 latency=126.435µs === RUN TestBindingRuleUpdateCommand_noMerge/update_all_fields_-_partial === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.760Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=DELETE url=/v1/acl/binding-rule/6b3197fd-192b-13aa-9053-84cf39e29138 from=127.0.0.1:56388 latency=4.162607ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.770Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules?authmethod=test from=127.0.0.1:59402 latency=262.389µs writer.go:29: 2021-01-29T19:31:51.774Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=DELETE url=/v1/acl/binding-rule/1ef16173-3cd8-2a48-1970-7a7ef88eb432 from=127.0.0.1:59402 latency=762.672µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.772Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=DELETE url=/v1/acl/binding-rule/7fd5fccd-d161-6eaf-7d66-aefba388a902 from=127.0.0.1:56390 latency=4.400582ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.786Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=DELETE url=/v1/acl/binding-rule/2014143a-19d1-f907-818b-c68c17a2cd77 from=127.0.0.1:59436 latency=3.122383ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.791Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=DELETE url=/v1/acl/binding-rule/a8a6d1c0-bcaa-8cc6-711c-e3bd5735ec60 from=127.0.0.1:56394 latency=4.95827ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.797Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=DELETE url=/v1/acl/binding-rule/30531324-5382-b452-e901-41e980225462 from=127.0.0.1:59440 latency=5.93405ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.806Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=DELETE url=/v1/acl/binding-rule/cfa01c1b-e0da-f854-0345-c8c5d1fed460 from=127.0.0.1:56398 latency=7.781739ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.808Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=DELETE url=/v1/acl/binding-rule/77ce92c5-a1ec-8209-2886-4b063b2e321e from=127.0.0.1:59444 latency=5.20003ms writer.go:29: 2021-01-29T19:31:51.817Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=DELETE url=/v1/acl/binding-rule/98f97165-095e-24b5-46d2-d4b31f99ca04 from=127.0.0.1:59446 latency=2.897051ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.832Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56404 latency=12.39952ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.845Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=DELETE url=/v1/acl/binding-rule/d55883e9-8917-0619-0dca-6147a8fdc93c from=127.0.0.1:59450 latency=20.026887ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.846Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:56408 latency=6.433494ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.856Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=DELETE url=/v1/acl/binding-rule/d684afe1-0366-5459-9c20-8e710d641a86 from=127.0.0.1:59454 latency=1.889672ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.852Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/d040c695-9f64-08ca-4937-06d2c9d48bd4 from=127.0.0.1:56408 latency=668.961µs === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.863Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=DELETE url=/v1/acl/binding-rule/d969ae84-1be2-4208-6a3a-df90a869171c from=127.0.0.1:59456 latency=2.164865ms writer.go:29: 2021-01-29T19:31:51.882Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=DELETE url=/v1/acl/binding-rule/ebe1ffe6-3043-bd79-6c74-a33a71b391ff from=127.0.0.1:59458 latency=6.322311ms writer.go:29: 2021-01-29T19:31:51.904Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=DELETE url=/v1/acl/binding-rule/fe60fddf-048e-5dd2-2ee4-d513b48591d9 from=127.0.0.1:59460 latency=17.803696ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.904Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule/d040c695-9f64-08ca-4937-06d2c9d48bd4 from=127.0.0.1:56408 latency=43.38833ms writer.go:29: 2021-01-29T19:31:51.909Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/d040c695-9f64-08ca-4937-06d2c9d48bd4 from=127.0.0.1:56404 latency=136.232µs === RUN TestBindingRuleUpdateCommand/update_all_fields_but_description === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.922Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56404 latency=1.556302ms writer.go:29: 2021-01-29T19:31:51.929Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/de60515e-1160-581d-940a-ac7c86f3339d from=127.0.0.1:56420 latency=1.34861ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.931Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59462 latency=21.377902ms writer.go:29: 2021-01-29T19:31:51.937Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rules from=127.0.0.1:59466 latency=218.221µs writer.go:29: 2021-01-29T19:31:51.941Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/b673d27f-a354-5c93-b0ae-dc70edbdbcce from=127.0.0.1:59466 latency=1.716358ms === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.960Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule/de60515e-1160-581d-940a-ac7c86f3339d from=127.0.0.1:56420 latency=29.404109ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.963Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule/b673d27f-a354-5c93-b0ae-dc70edbdbcce from=127.0.0.1:59466 latency=19.06054ms writer.go:29: 2021-01-29T19:31:51.966Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/b673d27f-a354-5c93-b0ae-dc70edbdbcce from=127.0.0.1:59462 latency=153.051µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.962Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/de60515e-1160-581d-940a-ac7c86f3339d from=127.0.0.1:56404 latency=176.437µs === RUN TestBindingRuleUpdateCommand/update_all_fields_but_bind_name === RUN TestBindingRuleUpdateCommand_noMerge/update_all_fields_but_description === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.978Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59462 latency=961.567µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:51.978Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56404 latency=1.693547ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:51.985Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/bac8594a-7910-309c-7eb9-0af14dde440f from=127.0.0.1:59468 latency=1.100981ms writer.go:29: 2021-01-29T19:31:51.995Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule/bac8594a-7910-309c-7eb9-0af14dde440f from=127.0.0.1:59468 latency=9.023429ms writer.go:29: 2021-01-29T19:31:51.999Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/bac8594a-7910-309c-7eb9-0af14dde440f from=127.0.0.1:59462 latency=87.965µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.005Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/71f43ade-13ce-a08b-e2a8-70b81053c988 from=127.0.0.1:56426 latency=4.031163ms === RUN TestBindingRuleUpdateCommand_noMerge/missing_bind_name === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.027Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule/71f43ade-13ce-a08b-e2a8-70b81053c988 from=127.0.0.1:56426 latency=18.1272ms writer.go:29: 2021-01-29T19:31:52.030Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/71f43ade-13ce-a08b-e2a8-70b81053c988 from=127.0.0.1:56404 latency=82.809µs === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:52.032Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59462 latency=1.020422ms === RUN TestBindingRuleUpdateCommand/update_all_fields_but_must_exist === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.052Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56404 latency=13.033667ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:52.084Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/ff74f945-0547-a198-7a89-01837c3931e6 from=127.0.0.1:59472 latency=1.127947ms === RUN TestBindingRuleUpdateCommand_noMerge/update_all_fields_but_selector === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.088Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/f6d0460d-6fef-712b-aae8-25ec5c2a2ec8 from=127.0.0.1:56430 latency=1.078227ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:52.089Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:59462 latency=959.499µs === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.094Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule/f6d0460d-6fef-712b-aae8-25ec5c2a2ec8 from=127.0.0.1:56430 latency=5.119962ms writer.go:29: 2021-01-29T19:31:52.097Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/f6d0460d-6fef-712b-aae8-25ec5c2a2ec8 from=127.0.0.1:56404 latency=143.621µs === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:52.103Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/572241af-17c2-63f9-3447-b322b65e3570 from=127.0.0.1:59476 latency=6.617287ms === RUN TestBindingRuleUpdateCommand/update_all_fields_but_selector === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:52.109Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/binding-rule/572241af-17c2-63f9-3447-b322b65e3570 from=127.0.0.1:59476 latency=1.868966ms writer.go:29: 2021-01-29T19:31:52.113Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/binding-rule/572241af-17c2-63f9-3447-b322b65e3570 from=127.0.0.1:59462 latency=92.263µs writer.go:29: 2021-01-29T19:31:52.121Z [INFO] TestBindingRuleUpdateCommand_noMerge: Requesting shutdown writer.go:29: 2021-01-29T19:31:52.122Z [INFO] TestBindingRuleUpdateCommand_noMerge.server: shutting down server === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.124Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56404 latency=14.9601ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:52.123Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:52.127Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:52.129Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:52.130Z [WARN] TestBindingRuleUpdateCommand_noMerge.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:52.131Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:52.132Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:52.133Z [DEBUG] TestBindingRuleUpdateCommand_noMerge.leader: stopped routine: routine="acl token reaping" === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.146Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/886c6205-cbe2-d4ac-1a18-adc1cc84af0e from=127.0.0.1:56434 latency=1.948194ms === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:52.149Z [WARN] TestBindingRuleUpdateCommand_noMerge.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:52.151Z [INFO] TestBindingRuleUpdateCommand_noMerge.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:52.153Z [INFO] TestBindingRuleUpdateCommand_noMerge: consul server down writer.go:29: 2021-01-29T19:31:52.154Z [INFO] TestBindingRuleUpdateCommand_noMerge: shutdown complete writer.go:29: 2021-01-29T19:31:52.155Z [INFO] TestBindingRuleUpdateCommand_noMerge: Stopping server: protocol=DNS address=127.0.0.1:19190 network=tcp writer.go:29: 2021-01-29T19:31:52.155Z [INFO] TestBindingRuleUpdateCommand_noMerge: Stopping server: protocol=DNS address=127.0.0.1:19190 network=udp writer.go:29: 2021-01-29T19:31:52.156Z [INFO] TestBindingRuleUpdateCommand_noMerge: Stopping server: protocol=HTTP address=127.0.0.1:19191 network=tcp === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.198Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule/886c6205-cbe2-d4ac-1a18-adc1cc84af0e from=127.0.0.1:56434 latency=48.404407ms writer.go:29: 2021-01-29T19:31:52.200Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/886c6205-cbe2-d4ac-1a18-adc1cc84af0e from=127.0.0.1:56404 latency=113.502µs === RUN TestBindingRuleUpdateCommand/update_all_fields_clear_selector === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.223Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56404 latency=15.381797ms writer.go:29: 2021-01-29T19:31:52.233Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/3da0c36f-1167-b263-4be4-70ac2cb4e290 from=127.0.0.1:56436 latency=2.686961ms writer.go:29: 2021-01-29T19:31:52.236Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule/3da0c36f-1167-b263-4be4-70ac2cb4e290 from=127.0.0.1:56436 latency=1.097736ms writer.go:29: 2021-01-29T19:31:52.240Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/3da0c36f-1167-b263-4be4-70ac2cb4e290 from=127.0.0.1:56404 latency=127.019µs === RUN TestBindingRuleUpdateCommand/update_all_fields_json_formatted === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.253Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:56404 latency=1.150898ms writer.go:29: 2021-01-29T19:31:52.261Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/71a52989-5a5b-05c3-bd69-86baf7c71264 from=127.0.0.1:56438 latency=2.44227ms writer.go:29: 2021-01-29T19:31:52.282Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule/71a52989-5a5b-05c3-bd69-86baf7c71264 from=127.0.0.1:56438 latency=18.075403ms writer.go:29: 2021-01-29T19:31:52.284Z [DEBUG] TestBindingRuleUpdateCommand.http: Request finished: method=GET url=/v1/acl/binding-rule/71a52989-5a5b-05c3-bd69-86baf7c71264 from=127.0.0.1:56404 latency=185.497µs writer.go:29: 2021-01-29T19:31:52.288Z [INFO] TestBindingRuleUpdateCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:52.289Z [INFO] TestBindingRuleUpdateCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:52.290Z [DEBUG] TestBindingRuleUpdateCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:52.291Z [DEBUG] TestBindingRuleUpdateCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:52.292Z [DEBUG] TestBindingRuleUpdateCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:52.293Z [WARN] TestBindingRuleUpdateCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:52.291Z [DEBUG] TestBindingRuleUpdateCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:52.292Z [DEBUG] TestBindingRuleUpdateCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:52.293Z [DEBUG] TestBindingRuleUpdateCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:52.309Z [WARN] TestBindingRuleUpdateCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:52.312Z [INFO] TestBindingRuleUpdateCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:52.314Z [INFO] TestBindingRuleUpdateCommand: consul server down writer.go:29: 2021-01-29T19:31:52.316Z [INFO] TestBindingRuleUpdateCommand: shutdown complete writer.go:29: 2021-01-29T19:31:52.317Z [INFO] TestBindingRuleUpdateCommand: Stopping server: protocol=DNS address=127.0.0.1:19196 network=tcp writer.go:29: 2021-01-29T19:31:52.317Z [INFO] TestBindingRuleUpdateCommand: Stopping server: protocol=DNS address=127.0.0.1:19196 network=udp writer.go:29: 2021-01-29T19:31:52.318Z [INFO] TestBindingRuleUpdateCommand: Stopping server: protocol=HTTP address=127.0.0.1:19197 network=tcp === CONT TestBindingRuleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:31:52.657Z [INFO] TestBindingRuleUpdateCommand_noMerge: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:52.658Z [INFO] TestBindingRuleUpdateCommand_noMerge: Endpoints down --- PASS: TestBindingRuleUpdateCommand_noMerge (1.63s) --- PASS: TestBindingRuleUpdateCommand_noMerge/rule_id_required (0.00s) --- PASS: TestBindingRuleUpdateCommand_noMerge/rule_id_partial_matches_nothing (0.07s) --- PASS: TestBindingRuleUpdateCommand_noMerge/rule_id_exact_match_doesn't_exist (0.02s) --- PASS: TestBindingRuleUpdateCommand_noMerge/rule_id_partial_matches_multiple (0.10s) --- PASS: TestBindingRuleUpdateCommand_noMerge/must_use_roughly_valid_selector (0.05s) --- PASS: TestBindingRuleUpdateCommand_noMerge/update_all_fields (0.06s) --- PASS: TestBindingRuleUpdateCommand_noMerge/update_all_fields_-_partial (0.22s) --- PASS: TestBindingRuleUpdateCommand_noMerge/update_all_fields_but_description (0.04s) --- PASS: TestBindingRuleUpdateCommand_noMerge/missing_bind_name (0.06s) --- PASS: TestBindingRuleUpdateCommand_noMerge/update_all_fields_but_selector (0.03s) === CONT TestBindingRuleUpdateCommand writer.go:29: 2021-01-29T19:31:52.819Z [INFO] TestBindingRuleUpdateCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:52.821Z [INFO] TestBindingRuleUpdateCommand: Endpoints down --- PASS: TestBindingRuleUpdateCommand (1.79s) --- PASS: TestBindingRuleUpdateCommand/rule_id_required (0.00s) --- PASS: TestBindingRuleUpdateCommand/rule_id_partial_matches_nothing (0.01s) --- PASS: TestBindingRuleUpdateCommand/rule_id_exact_match_doesn't_exist (0.01s) --- PASS: TestBindingRuleUpdateCommand/rule_id_partial_matches_multiple (0.06s) --- PASS: TestBindingRuleUpdateCommand/must_use_roughly_valid_selector (0.05s) --- PASS: TestBindingRuleUpdateCommand/update_all_fields (0.04s) --- PASS: TestBindingRuleUpdateCommand/update_all_fields_-_partial (0.22s) --- PASS: TestBindingRuleUpdateCommand/update_all_fields_but_description (0.06s) --- PASS: TestBindingRuleUpdateCommand/update_all_fields_but_bind_name (0.06s) --- PASS: TestBindingRuleUpdateCommand/update_all_fields_but_must_exist (0.07s) --- PASS: TestBindingRuleUpdateCommand/update_all_fields_but_selector (0.10s) --- PASS: TestBindingRuleUpdateCommand/update_all_fields_clear_selector (0.04s) --- PASS: TestBindingRuleUpdateCommand/update_all_fields_json_formatted (0.04s) PASS ok github.com/hashicorp/consul/command/acl/bindingrule/update 1.970s === RUN TestBootstrapCommand_noTabs === PAUSE TestBootstrapCommand_noTabs === RUN TestBootstrapCommand_Pretty === PAUSE TestBootstrapCommand_Pretty === RUN TestBootstrapCommand_JSON === PAUSE TestBootstrapCommand_JSON === CONT TestBootstrapCommand_noTabs === CONT TestBootstrapCommand_Pretty --- PASS: TestBootstrapCommand_noTabs (0.00s) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestBootstrapCommand_JSON [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:31:55.106Z [WARN] TestBootstrapCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:55.107Z [DEBUG] TestBootstrapCommand_JSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:55.111Z [DEBUG] TestBootstrapCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.118Z [WARN] TestBootstrapCommand_Pretty: bootstrap = true: do not enable unless necessary === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.120Z [INFO] TestBootstrapCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:78dabddb-5cbd-7a23-8cc3-40a8839c69a8 Address:127.0.0.1:13075}]" writer.go:29: 2021-01-29T19:31:55.122Z [INFO] TestBootstrapCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:13075 [Follower]" leader= writer.go:29: 2021-01-29T19:31:55.125Z [INFO] TestBootstrapCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:55.130Z [INFO] TestBootstrapCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 127.0.0.1 === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.120Z [DEBUG] TestBootstrapCommand_Pretty.tlsutil: Update: version=1 === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.133Z [INFO] TestBootstrapCommand_JSON: Started DNS server: address=127.0.0.1:13070 network=udp === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.137Z [DEBUG] TestBootstrapCommand_Pretty.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.138Z [INFO] TestBootstrapCommand_JSON: Started DNS server: address=127.0.0.1:13070 network=tcp writer.go:29: 2021-01-29T19:31:55.135Z [INFO] TestBootstrapCommand_JSON.server: Adding LAN server: server="Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 (Addr: tcp/127.0.0.1:13075) (DC: dc1)" writer.go:29: 2021-01-29T19:31:55.135Z [INFO] TestBootstrapCommand_JSON.server: Handled event for server in area: event=member-join server=Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8.dc1 area=wan === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.143Z [INFO] TestBootstrapCommand_Pretty.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:db187671-efec-c544-5b44-96d837c2a853 Address:127.0.0.1:13069}]" === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.143Z [INFO] TestBootstrapCommand_JSON: Started HTTP server: address=127.0.0.1:13071 network=tcp writer.go:29: 2021-01-29T19:31:55.146Z [INFO] TestBootstrapCommand_JSON: started state syncer === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.144Z [INFO] TestBootstrapCommand_Pretty.server.raft: entering follower state: follower="Node at 127.0.0.1:13069 [Follower]" leader= writer.go:29: 2021-01-29T19:31:55.145Z [INFO] TestBootstrapCommand_Pretty.server.serf.wan: serf: EventMemberJoin: Node-db187671-efec-c544-5b44-96d837c2a853.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:55.153Z [INFO] TestBootstrapCommand_Pretty.server.serf.lan: serf: EventMemberJoin: Node-db187671-efec-c544-5b44-96d837c2a853 127.0.0.1 writer.go:29: 2021-01-29T19:31:55.156Z [INFO] TestBootstrapCommand_Pretty.server: Handled event for server in area: event=member-join server=Node-db187671-efec-c544-5b44-96d837c2a853.dc1 area=wan writer.go:29: 2021-01-29T19:31:55.157Z [INFO] TestBootstrapCommand_Pretty.server: Adding LAN server: server="Node-db187671-efec-c544-5b44-96d837c2a853 (Addr: tcp/127.0.0.1:13069) (DC: dc1)" writer.go:29: 2021-01-29T19:31:55.157Z [INFO] TestBootstrapCommand_Pretty: Started DNS server: address=127.0.0.1:13064 network=tcp writer.go:29: 2021-01-29T19:31:55.159Z [INFO] TestBootstrapCommand_Pretty: Started DNS server: address=127.0.0.1:13064 network=udp writer.go:29: 2021-01-29T19:31:55.160Z [INFO] TestBootstrapCommand_Pretty: Started HTTP server: address=127.0.0.1:13065 network=tcp writer.go:29: 2021-01-29T19:31:55.161Z [INFO] TestBootstrapCommand_Pretty: started state syncer === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.164Z [WARN] TestBootstrapCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:55.165Z [INFO] TestBootstrapCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:13075 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:55.167Z [DEBUG] TestBootstrapCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:55.168Z [DEBUG] TestBootstrapCommand_JSON.server.raft: vote granted: from=78dabddb-5cbd-7a23-8cc3-40a8839c69a8 term=2 tally=1 writer.go:29: 2021-01-29T19:31:55.170Z [INFO] TestBootstrapCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:55.171Z [INFO] TestBootstrapCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:13075 [Leader]" writer.go:29: 2021-01-29T19:31:55.173Z [INFO] TestBootstrapCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:55.174Z [INFO] TestBootstrapCommand_JSON.server: New leader elected: payload=Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 writer.go:29: 2021-01-29T19:31:55.178Z [INFO] TestBootstrapCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:31:55.181Z [INFO] TestBootstrapCommand_JSON.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:55.185Z [INFO] TestBootstrapCommand_JSON.server: Created ACL anonymous token from configuration === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.186Z [WARN] TestBootstrapCommand_Pretty.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:55.187Z [INFO] TestBootstrapCommand_Pretty.server.raft: entering candidate state: node="Node at 127.0.0.1:13069 [Candidate]" term=2 === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.186Z [INFO] TestBootstrapCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.188Z [DEBUG] TestBootstrapCommand_Pretty.server.raft: votes: needed=1 === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.186Z [INFO] TestBootstrapCommand_JSON.server: initializing acls === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.189Z [DEBUG] TestBootstrapCommand_Pretty.server.raft: vote granted: from=db187671-efec-c544-5b44-96d837c2a853 term=2 tally=1 === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.189Z [INFO] TestBootstrapCommand_JSON.leader: started routine: routine="acl token reaping" === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.190Z [INFO] TestBootstrapCommand_Pretty.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:55.191Z [INFO] TestBootstrapCommand_Pretty.server.raft: entering leader state: leader="Node at 127.0.0.1:13069 [Leader]" === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.191Z [DEBUG] TestBootstrapCommand_JSON.server: transitioning out of legacy ACL mode === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.191Z [INFO] TestBootstrapCommand_Pretty.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:55.192Z [INFO] TestBootstrapCommand_Pretty.server: New leader elected: payload=Node-db187671-efec-c544-5b44-96d837c2a853 === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.192Z [INFO] TestBootstrapCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 writer.go:29: 2021-01-29T19:31:55.195Z [INFO] TestBootstrapCommand_JSON.server: Updating LAN server: server="Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 (Addr: tcp/127.0.0.1:13075) (DC: dc1)" === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.194Z [INFO] TestBootstrapCommand_Pretty.server: initializing acls === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.197Z [INFO] TestBootstrapCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8.dc1 writer.go:29: 2021-01-29T19:31:55.197Z [INFO] TestBootstrapCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 writer.go:29: 2021-01-29T19:31:55.199Z [INFO] TestBootstrapCommand_JSON.server: Handled event for server in area: event=member-update server=Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8.dc1 area=wan writer.go:29: 2021-01-29T19:31:55.201Z [INFO] TestBootstrapCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8.dc1 === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.199Z [INFO] TestBootstrapCommand_Pretty.server: Created ACL 'global-management' policy === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.201Z [INFO] TestBootstrapCommand_JSON.server: Updating LAN server: server="Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 (Addr: tcp/127.0.0.1:13075) (DC: dc1)" writer.go:29: 2021-01-29T19:31:55.203Z [INFO] TestBootstrapCommand_JSON.server: Handled event for server in area: event=member-update server=Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8.dc1 area=wan === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.206Z [INFO] TestBootstrapCommand_Pretty.server: initializing acls writer.go:29: 2021-01-29T19:31:55.208Z [INFO] TestBootstrapCommand_Pretty.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:55.206Z [INFO] TestBootstrapCommand_Pretty.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:55.209Z [INFO] TestBootstrapCommand_Pretty.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:55.211Z [DEBUG] TestBootstrapCommand_Pretty.server: transitioning out of legacy ACL mode === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.206Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.211Z [INFO] TestBootstrapCommand_Pretty.server.serf.lan: serf: EventMemberUpdate: Node-db187671-efec-c544-5b44-96d837c2a853 writer.go:29: 2021-01-29T19:31:55.221Z [INFO] TestBootstrapCommand_Pretty.server: Updating LAN server: server="Node-db187671-efec-c544-5b44-96d837c2a853 (Addr: tcp/127.0.0.1:13069) (DC: dc1)" writer.go:29: 2021-01-29T19:31:55.221Z [INFO] TestBootstrapCommand_Pretty.server.serf.lan: serf: EventMemberUpdate: Node-db187671-efec-c544-5b44-96d837c2a853 writer.go:29: 2021-01-29T19:31:55.221Z [INFO] TestBootstrapCommand_Pretty.server.serf.wan: serf: EventMemberUpdate: Node-db187671-efec-c544-5b44-96d837c2a853.dc1 writer.go:29: 2021-01-29T19:31:55.223Z [INFO] TestBootstrapCommand_Pretty.server: Updating LAN server: server="Node-db187671-efec-c544-5b44-96d837c2a853 (Addr: tcp/127.0.0.1:13069) (DC: dc1)" writer.go:29: 2021-01-29T19:31:55.225Z [INFO] TestBootstrapCommand_Pretty.server: Handled event for server in area: event=member-update server=Node-db187671-efec-c544-5b44-96d837c2a853.dc1 area=wan writer.go:29: 2021-01-29T19:31:55.226Z [INFO] TestBootstrapCommand_Pretty.server.serf.wan: serf: EventMemberUpdate: Node-db187671-efec-c544-5b44-96d837c2a853.dc1 writer.go:29: 2021-01-29T19:31:55.227Z [INFO] TestBootstrapCommand_Pretty.server: Handled event for server in area: event=member-update server=Node-db187671-efec-c544-5b44-96d837c2a853.dc1 area=wan writer.go:29: 2021-01-29T19:31:55.226Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.272Z [INFO] TestBootstrapCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:55.273Z [INFO] TestBootstrapCommand_JSON.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:55.276Z [DEBUG] TestBootstrapCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.276Z [INFO] TestBootstrapCommand_Pretty.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.278Z [INFO] TestBootstrapCommand_JSON.server: member joined, marking health alive: member=Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 writer.go:29: 2021-01-29T19:31:55.281Z [DEBUG] TestBootstrapCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 writer.go:29: 2021-01-29T19:31:55.282Z [DEBUG] TestBootstrapCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-78dabddb-5cbd-7a23-8cc3-40a8839c69a8 === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.279Z [INFO] TestBootstrapCommand_Pretty.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:55.287Z [DEBUG] TestBootstrapCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-db187671-efec-c544-5b44-96d837c2a853 writer.go:29: 2021-01-29T19:31:55.293Z [INFO] TestBootstrapCommand_Pretty.server: member joined, marking health alive: member=Node-db187671-efec-c544-5b44-96d837c2a853 writer.go:29: 2021-01-29T19:31:55.302Z [DEBUG] TestBootstrapCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-db187671-efec-c544-5b44-96d837c2a853 writer.go:29: 2021-01-29T19:31:55.312Z [DEBUG] TestBootstrapCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-db187671-efec-c544-5b44-96d837c2a853 writer.go:29: 2021-01-29T19:31:55.361Z [WARN] TestBootstrapCommand_Pretty.server.acl: failed to remove bootstrap file: error="remove /tmp/TestBootstrapCommand_Pretty-agent052073785/acl-bootstrap-reset: no such file or directory" writer.go:29: 2021-01-29T19:31:55.382Z [INFO] TestBootstrapCommand_Pretty.server.acl: ACL bootstrap completed writer.go:29: 2021-01-29T19:31:55.423Z [DEBUG] TestBootstrapCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/bootstrap from=127.0.0.1:58696 latency=61.890187ms === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.421Z [WARN] TestBootstrapCommand_JSON.server.acl: failed to remove bootstrap file: error="remove /tmp/TestBootstrapCommand_JSON-agent767529939/acl-bootstrap-reset: no such file or directory" === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.462Z [INFO] TestBootstrapCommand_Pretty: Requesting shutdown writer.go:29: 2021-01-29T19:31:55.464Z [INFO] TestBootstrapCommand_Pretty.server: shutting down server writer.go:29: 2021-01-29T19:31:55.466Z [DEBUG] TestBootstrapCommand_Pretty.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:55.469Z [DEBUG] TestBootstrapCommand_Pretty.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:55.471Z [DEBUG] TestBootstrapCommand_Pretty.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.483Z [INFO] TestBootstrapCommand_JSON.server.acl: ACL bootstrap completed writer.go:29: 2021-01-29T19:31:55.484Z [DEBUG] TestBootstrapCommand_JSON.http: Request finished: method=PUT url=/v1/acl/bootstrap from=127.0.0.1:39378 latency=62.919672ms writer.go:29: 2021-01-29T19:31:55.487Z [INFO] TestBootstrapCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:31:55.488Z [INFO] TestBootstrapCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:31:55.489Z [DEBUG] TestBootstrapCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:55.490Z [DEBUG] TestBootstrapCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:55.491Z [DEBUG] TestBootstrapCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.469Z [ERROR] TestBootstrapCommand_Pretty.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:31:55.469Z [DEBUG] TestBootstrapCommand_Pretty.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:55.476Z [DEBUG] TestBootstrapCommand_Pretty.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:55.488Z [WARN] TestBootstrapCommand_Pretty.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:55.491Z [DEBUG] TestBootstrapCommand_Pretty.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.491Z [WARN] TestBootstrapCommand_JSON.server.serf.lan: serf: Shutdown without a Leave === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.501Z [WARN] TestBootstrapCommand_Pretty.server.serf.wan: serf: Shutdown without a Leave === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.491Z [ERROR] TestBootstrapCommand_JSON.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.504Z [INFO] TestBootstrapCommand_Pretty.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:55.504Z [INFO] TestBootstrapCommand_Pretty: consul server down === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.491Z [DEBUG] TestBootstrapCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:55.491Z [DEBUG] TestBootstrapCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:55.491Z [DEBUG] TestBootstrapCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:55.504Z [WARN] TestBootstrapCommand_JSON.server.serf.wan: serf: Shutdown without a Leave === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:55.508Z [INFO] TestBootstrapCommand_Pretty: shutdown complete writer.go:29: 2021-01-29T19:31:55.514Z [INFO] TestBootstrapCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:13064 network=tcp writer.go:29: 2021-01-29T19:31:55.516Z [INFO] TestBootstrapCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:13064 network=udp writer.go:29: 2021-01-29T19:31:55.518Z [INFO] TestBootstrapCommand_Pretty: Stopping server: protocol=HTTP address=127.0.0.1:13065 network=tcp === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:55.514Z [INFO] TestBootstrapCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:55.514Z [INFO] TestBootstrapCommand_JSON: consul server down writer.go:29: 2021-01-29T19:31:55.536Z [INFO] TestBootstrapCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:31:55.537Z [INFO] TestBootstrapCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:13070 network=tcp writer.go:29: 2021-01-29T19:31:55.538Z [INFO] TestBootstrapCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:13070 network=udp writer.go:29: 2021-01-29T19:31:55.540Z [INFO] TestBootstrapCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:13071 network=tcp === CONT TestBootstrapCommand_Pretty writer.go:29: 2021-01-29T19:31:56.019Z [INFO] TestBootstrapCommand_Pretty: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:56.020Z [INFO] TestBootstrapCommand_Pretty: Endpoints down --- PASS: TestBootstrapCommand_Pretty (0.96s) === CONT TestBootstrapCommand_JSON writer.go:29: 2021-01-29T19:31:56.041Z [INFO] TestBootstrapCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:56.042Z [INFO] TestBootstrapCommand_JSON: Endpoints down --- PASS: TestBootstrapCommand_JSON (0.98s) PASS ok github.com/hashicorp/consul/command/acl/bootstrap 1.155s ? github.com/hashicorp/consul/command/acl/policy [no test files] === RUN TestPolicyCreateCommand_noTabs === PAUSE TestPolicyCreateCommand_noTabs === RUN TestPolicyCreateCommand === PAUSE TestPolicyCreateCommand === RUN TestPolicyCreateCommand_JSON === PAUSE TestPolicyCreateCommand_JSON === CONT TestPolicyCreateCommand_noTabs === CONT TestPolicyCreateCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestPolicyCreateCommand_JSON [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestPolicyCreateCommand_noTabs (0.01s) === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.297Z [WARN] TestPolicyCreateCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:57.298Z [DEBUG] TestPolicyCreateCommand_JSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:57.300Z [DEBUG] TestPolicyCreateCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:57.314Z [INFO] TestPolicyCreateCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2 Address:127.0.0.1:27369}]" writer.go:29: 2021-01-29T19:31:57.316Z [INFO] TestPolicyCreateCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:27369 [Follower]" leader= writer.go:29: 2021-01-29T19:31:57.317Z [INFO] TestPolicyCreateCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:57.320Z [INFO] TestPolicyCreateCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2 127.0.0.1 writer.go:29: 2021-01-29T19:31:57.322Z [INFO] TestPolicyCreateCommand_JSON.server: Adding LAN server: server="Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2 (Addr: tcp/127.0.0.1:27369) (DC: dc1)" writer.go:29: 2021-01-29T19:31:57.323Z [INFO] TestPolicyCreateCommand_JSON: Started DNS server: address=127.0.0.1:27364 network=udp writer.go:29: 2021-01-29T19:31:57.324Z [INFO] TestPolicyCreateCommand_JSON: Started DNS server: address=127.0.0.1:27364 network=tcp writer.go:29: 2021-01-29T19:31:57.323Z [INFO] TestPolicyCreateCommand_JSON.server: Handled event for server in area: event=member-join server=Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2.dc1 area=wan writer.go:29: 2021-01-29T19:31:57.326Z [INFO] TestPolicyCreateCommand_JSON: Started HTTP server: address=127.0.0.1:27365 network=tcp writer.go:29: 2021-01-29T19:31:57.327Z [INFO] TestPolicyCreateCommand_JSON: started state syncer === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.334Z [WARN] TestPolicyCreateCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:57.337Z [DEBUG] TestPolicyCreateCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:57.342Z [DEBUG] TestPolicyCreateCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.377Z [WARN] TestPolicyCreateCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:57.378Z [INFO] TestPolicyCreateCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:27369 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:57.381Z [DEBUG] TestPolicyCreateCommand_JSON.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:57.383Z [DEBUG] TestPolicyCreateCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:57.386Z [DEBUG] TestPolicyCreateCommand_JSON.server.raft: vote granted: from=e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2 term=2 tally=1 writer.go:29: 2021-01-29T19:31:57.387Z [INFO] TestPolicyCreateCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:57.388Z [INFO] TestPolicyCreateCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:27369 [Leader]" === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.389Z [INFO] TestPolicyCreateCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:407854f1-093b-bfe9-6474-a2d52872919c Address:127.0.0.1:27363}]" === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.389Z [INFO] TestPolicyCreateCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:57.393Z [INFO] TestPolicyCreateCommand_JSON.server: New leader elected: payload=Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2 === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.393Z [INFO] TestPolicyCreateCommand.server.serf.wan: serf: EventMemberJoin: Node-407854f1-093b-bfe9-6474-a2d52872919c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:57.394Z [INFO] TestPolicyCreateCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:27363 [Follower]" leader= writer.go:29: 2021-01-29T19:31:57.400Z [INFO] TestPolicyCreateCommand.server.serf.lan: serf: EventMemberJoin: Node-407854f1-093b-bfe9-6474-a2d52872919c 127.0.0.1 === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.402Z [INFO] TestPolicyCreateCommand_JSON.server: initializing acls === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.423Z [INFO] TestPolicyCreateCommand.server: Handled event for server in area: event=member-join server=Node-407854f1-093b-bfe9-6474-a2d52872919c.dc1 area=wan === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.426Z [INFO] TestPolicyCreateCommand_JSON.server: Created ACL 'global-management' policy === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.424Z [INFO] TestPolicyCreateCommand: Started DNS server: address=127.0.0.1:27358 network=udp === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.427Z [WARN] TestPolicyCreateCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:57.429Z [INFO] TestPolicyCreateCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.427Z [INFO] TestPolicyCreateCommand: Started DNS server: address=127.0.0.1:27358 network=tcp === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.431Z [INFO] TestPolicyCreateCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:57.432Z [INFO] TestPolicyCreateCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:57.433Z [INFO] TestPolicyCreateCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:57.434Z [INFO] TestPolicyCreateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2 writer.go:29: 2021-01-29T19:31:57.435Z [INFO] TestPolicyCreateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2.dc1 === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.425Z [INFO] TestPolicyCreateCommand.server: Adding LAN server: server="Node-407854f1-093b-bfe9-6474-a2d52872919c (Addr: tcp/127.0.0.1:27363) (DC: dc1)" writer.go:29: 2021-01-29T19:31:57.438Z [INFO] TestPolicyCreateCommand: Started HTTP server: address=127.0.0.1:27359 network=tcp writer.go:29: 2021-01-29T19:31:57.439Z [INFO] TestPolicyCreateCommand: started state syncer === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.437Z [INFO] TestPolicyCreateCommand_JSON.server: Updating LAN server: server="Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2 (Addr: tcp/127.0.0.1:27369) (DC: dc1)" writer.go:29: 2021-01-29T19:31:57.437Z [INFO] TestPolicyCreateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2.dc1 area=wan writer.go:29: 2021-01-29T19:31:57.441Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.452Z [WARN] TestPolicyCreateCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:57.453Z [INFO] TestPolicyCreateCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:27363 [Candidate]" term=2 === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.450Z [INFO] TestPolicyCreateCommand_JSON: Synced node info === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.455Z [DEBUG] TestPolicyCreateCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:57.456Z [DEBUG] TestPolicyCreateCommand.server.raft: vote granted: from=407854f1-093b-bfe9-6474-a2d52872919c term=2 tally=1 === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.454Z [DEBUG] TestPolicyCreateCommand_JSON: Node info in sync === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.457Z [INFO] TestPolicyCreateCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:57.457Z [INFO] TestPolicyCreateCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:27363 [Leader]" writer.go:29: 2021-01-29T19:31:57.458Z [INFO] TestPolicyCreateCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:57.459Z [INFO] TestPolicyCreateCommand.server: New leader elected: payload=Node-407854f1-093b-bfe9-6474-a2d52872919c writer.go:29: 2021-01-29T19:31:57.459Z [INFO] TestPolicyCreateCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:57.473Z [INFO] TestPolicyCreateCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:57.477Z [INFO] TestPolicyCreateCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:57.478Z [WARN] TestPolicyCreateCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:57.480Z [INFO] TestPolicyCreateCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:57.483Z [WARN] TestPolicyCreateCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:57.481Z [INFO] TestPolicyCreateCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:57.488Z [INFO] TestPolicyCreateCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:57.489Z [INFO] TestPolicyCreateCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:57.488Z [INFO] TestPolicyCreateCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:57.490Z [INFO] TestPolicyCreateCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:57.493Z [INFO] TestPolicyCreateCommand.server.serf.lan: serf: EventMemberUpdate: Node-407854f1-093b-bfe9-6474-a2d52872919c writer.go:29: 2021-01-29T19:31:57.495Z [DEBUG] TestPolicyCreateCommand.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:31:57.496Z [INFO] TestPolicyCreateCommand.server.serf.lan: serf: EventMemberUpdate: Node-407854f1-093b-bfe9-6474-a2d52872919c writer.go:29: 2021-01-29T19:31:57.495Z [INFO] TestPolicyCreateCommand.server.serf.wan: serf: EventMemberUpdate: Node-407854f1-093b-bfe9-6474-a2d52872919c.dc1 writer.go:29: 2021-01-29T19:31:57.495Z [INFO] TestPolicyCreateCommand.server: Updating LAN server: server="Node-407854f1-093b-bfe9-6474-a2d52872919c (Addr: tcp/127.0.0.1:27363) (DC: dc1)" writer.go:29: 2021-01-29T19:31:57.501Z [INFO] TestPolicyCreateCommand.server: Updating LAN server: server="Node-407854f1-093b-bfe9-6474-a2d52872919c (Addr: tcp/127.0.0.1:27363) (DC: dc1)" writer.go:29: 2021-01-29T19:31:57.499Z [INFO] TestPolicyCreateCommand.server: Handled event for server in area: event=member-update server=Node-407854f1-093b-bfe9-6474-a2d52872919c.dc1 area=wan writer.go:29: 2021-01-29T19:31:57.500Z [INFO] TestPolicyCreateCommand.server.serf.wan: serf: EventMemberUpdate: Node-407854f1-093b-bfe9-6474-a2d52872919c.dc1 writer.go:29: 2021-01-29T19:31:57.502Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:57.506Z [INFO] TestPolicyCreateCommand.server: Handled event for server in area: event=member-update server=Node-407854f1-093b-bfe9-6474-a2d52872919c.dc1 area=wan writer.go:29: 2021-01-29T19:31:57.523Z [INFO] TestPolicyCreateCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.523Z [INFO] TestPolicyCreateCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.524Z [INFO] TestPolicyCreateCommand.leader: started routine: routine="CA root pruning" === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.524Z [INFO] TestPolicyCreateCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.525Z [DEBUG] TestPolicyCreateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-407854f1-093b-bfe9-6474-a2d52872919c writer.go:29: 2021-01-29T19:31:57.526Z [INFO] TestPolicyCreateCommand.server: member joined, marking health alive: member=Node-407854f1-093b-bfe9-6474-a2d52872919c === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.526Z [DEBUG] TestPolicyCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2 === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.529Z [DEBUG] TestPolicyCreateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-407854f1-093b-bfe9-6474-a2d52872919c === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.528Z [INFO] TestPolicyCreateCommand_JSON.server: member joined, marking health alive: member=Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2 === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.530Z [DEBUG] TestPolicyCreateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-407854f1-093b-bfe9-6474-a2d52872919c === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.531Z [DEBUG] TestPolicyCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-e9ce7dae-b2b7-1545-adcb-d58a9ec3f4d2 === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:57.534Z [DEBUG] TestPolicyCreateCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:57.537Z [INFO] TestPolicyCreateCommand: Synced node info writer.go:29: 2021-01-29T19:31:57.553Z [DEBUG] TestPolicyCreateCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:33192 latency=4.452825ms writer.go:29: 2021-01-29T19:31:57.557Z [INFO] TestPolicyCreateCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:57.558Z [INFO] TestPolicyCreateCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:57.558Z [DEBUG] TestPolicyCreateCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:57.559Z [DEBUG] TestPolicyCreateCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:57.560Z [DEBUG] TestPolicyCreateCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:57.561Z [WARN] TestPolicyCreateCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:57.559Z [DEBUG] TestPolicyCreateCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:57.560Z [DEBUG] TestPolicyCreateCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:57.561Z [DEBUG] TestPolicyCreateCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:57.563Z [WARN] TestPolicyCreateCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:57.569Z [INFO] TestPolicyCreateCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:57.569Z [INFO] TestPolicyCreateCommand: consul server down writer.go:29: 2021-01-29T19:31:57.571Z [INFO] TestPolicyCreateCommand: shutdown complete writer.go:29: 2021-01-29T19:31:57.572Z [INFO] TestPolicyCreateCommand: Stopping server: protocol=DNS address=127.0.0.1:27358 network=tcp writer.go:29: 2021-01-29T19:31:57.573Z [INFO] TestPolicyCreateCommand: Stopping server: protocol=DNS address=127.0.0.1:27358 network=udp writer.go:29: 2021-01-29T19:31:57.574Z [INFO] TestPolicyCreateCommand: Stopping server: protocol=HTTP address=127.0.0.1:27359 network=tcp === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:57.672Z [DEBUG] TestPolicyCreateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:53222 latency=8.692938ms writer.go:29: 2021-01-29T19:31:57.685Z [INFO] TestPolicyCreateCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:31:57.686Z [INFO] TestPolicyCreateCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:31:57.688Z [DEBUG] TestPolicyCreateCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:57.689Z [DEBUG] TestPolicyCreateCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:57.691Z [DEBUG] TestPolicyCreateCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:57.689Z [DEBUG] TestPolicyCreateCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:57.691Z [DEBUG] TestPolicyCreateCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:57.693Z [DEBUG] TestPolicyCreateCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:57.693Z [WARN] TestPolicyCreateCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:57.703Z [WARN] TestPolicyCreateCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:57.706Z [INFO] TestPolicyCreateCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:57.706Z [INFO] TestPolicyCreateCommand_JSON: consul server down writer.go:29: 2021-01-29T19:31:57.708Z [INFO] TestPolicyCreateCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:31:57.708Z [INFO] TestPolicyCreateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:27364 network=tcp writer.go:29: 2021-01-29T19:31:57.709Z [INFO] TestPolicyCreateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:27364 network=udp writer.go:29: 2021-01-29T19:31:57.710Z [INFO] TestPolicyCreateCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:27365 network=tcp === CONT TestPolicyCreateCommand writer.go:29: 2021-01-29T19:31:58.074Z [INFO] TestPolicyCreateCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:58.077Z [INFO] TestPolicyCreateCommand: Endpoints down --- PASS: TestPolicyCreateCommand (0.84s) === CONT TestPolicyCreateCommand_JSON writer.go:29: 2021-01-29T19:31:58.211Z [INFO] TestPolicyCreateCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:31:58.213Z [INFO] TestPolicyCreateCommand_JSON: Endpoints down --- PASS: TestPolicyCreateCommand_JSON (0.97s) PASS ok github.com/hashicorp/consul/command/acl/policy/create 1.150s === RUN TestPolicyDeleteCommand_noTabs === PAUSE TestPolicyDeleteCommand_noTabs === RUN TestPolicyDeleteCommand === PAUSE TestPolicyDeleteCommand === CONT TestPolicyDeleteCommand_noTabs === CONT TestPolicyDeleteCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestPolicyDeleteCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestPolicyDeleteCommand writer.go:29: 2021-01-29T19:31:59.418Z [WARN] TestPolicyDeleteCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:31:59.432Z [DEBUG] TestPolicyDeleteCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:31:59.435Z [DEBUG] TestPolicyDeleteCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:31:59.442Z [INFO] TestPolicyDeleteCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b428a14f-a52c-c871-e809-28e71ffbd849 Address:127.0.0.1:16132}]" writer.go:29: 2021-01-29T19:31:59.444Z [INFO] TestPolicyDeleteCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:16132 [Follower]" leader= writer.go:29: 2021-01-29T19:31:59.447Z [INFO] TestPolicyDeleteCommand.server.serf.wan: serf: EventMemberJoin: Node-b428a14f-a52c-c871-e809-28e71ffbd849.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:31:59.452Z [INFO] TestPolicyDeleteCommand.server.serf.lan: serf: EventMemberJoin: Node-b428a14f-a52c-c871-e809-28e71ffbd849 127.0.0.1 writer.go:29: 2021-01-29T19:31:59.463Z [INFO] TestPolicyDeleteCommand.server: Adding LAN server: server="Node-b428a14f-a52c-c871-e809-28e71ffbd849 (Addr: tcp/127.0.0.1:16132) (DC: dc1)" writer.go:29: 2021-01-29T19:31:59.465Z [INFO] TestPolicyDeleteCommand: Started DNS server: address=127.0.0.1:16127 network=udp writer.go:29: 2021-01-29T19:31:59.466Z [INFO] TestPolicyDeleteCommand: Started DNS server: address=127.0.0.1:16127 network=tcp writer.go:29: 2021-01-29T19:31:59.467Z [INFO] TestPolicyDeleteCommand: Started HTTP server: address=127.0.0.1:16128 network=tcp writer.go:29: 2021-01-29T19:31:59.468Z [INFO] TestPolicyDeleteCommand: started state syncer writer.go:29: 2021-01-29T19:31:59.468Z [INFO] TestPolicyDeleteCommand.server: Handled event for server in area: event=member-join server=Node-b428a14f-a52c-c871-e809-28e71ffbd849.dc1 area=wan writer.go:29: 2021-01-29T19:31:59.511Z [WARN] TestPolicyDeleteCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:31:59.512Z [INFO] TestPolicyDeleteCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:16132 [Candidate]" term=2 writer.go:29: 2021-01-29T19:31:59.515Z [DEBUG] TestPolicyDeleteCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:31:59.515Z [DEBUG] TestPolicyDeleteCommand.server.raft: vote granted: from=b428a14f-a52c-c871-e809-28e71ffbd849 term=2 tally=1 writer.go:29: 2021-01-29T19:31:59.517Z [INFO] TestPolicyDeleteCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:31:59.515Z [DEBUG] TestPolicyDeleteCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:31:59.518Z [INFO] TestPolicyDeleteCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:16132 [Leader]" writer.go:29: 2021-01-29T19:31:59.520Z [INFO] TestPolicyDeleteCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:31:59.520Z [INFO] TestPolicyDeleteCommand.server: New leader elected: payload=Node-b428a14f-a52c-c871-e809-28e71ffbd849 writer.go:29: 2021-01-29T19:31:59.522Z [INFO] TestPolicyDeleteCommand.server: initializing acls writer.go:29: 2021-01-29T19:31:59.525Z [INFO] TestPolicyDeleteCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:31:59.525Z [WARN] TestPolicyDeleteCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:31:59.527Z [INFO] TestPolicyDeleteCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:31:59.529Z [INFO] TestPolicyDeleteCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:31:59.531Z [INFO] TestPolicyDeleteCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:59.532Z [INFO] TestPolicyDeleteCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:59.533Z [INFO] TestPolicyDeleteCommand.server.serf.lan: serf: EventMemberUpdate: Node-b428a14f-a52c-c871-e809-28e71ffbd849 writer.go:29: 2021-01-29T19:31:59.535Z [INFO] TestPolicyDeleteCommand.server.serf.wan: serf: EventMemberUpdate: Node-b428a14f-a52c-c871-e809-28e71ffbd849.dc1 writer.go:29: 2021-01-29T19:31:59.535Z [INFO] TestPolicyDeleteCommand.server: Updating LAN server: server="Node-b428a14f-a52c-c871-e809-28e71ffbd849 (Addr: tcp/127.0.0.1:16132) (DC: dc1)" writer.go:29: 2021-01-29T19:31:59.537Z [INFO] TestPolicyDeleteCommand.server: Handled event for server in area: event=member-update server=Node-b428a14f-a52c-c871-e809-28e71ffbd849.dc1 area=wan writer.go:29: 2021-01-29T19:31:59.541Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:31:59.577Z [INFO] TestPolicyDeleteCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:31:59.578Z [INFO] TestPolicyDeleteCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:59.579Z [DEBUG] TestPolicyDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-b428a14f-a52c-c871-e809-28e71ffbd849 writer.go:29: 2021-01-29T19:31:59.580Z [INFO] TestPolicyDeleteCommand.server: member joined, marking health alive: member=Node-b428a14f-a52c-c871-e809-28e71ffbd849 writer.go:29: 2021-01-29T19:31:59.582Z [DEBUG] TestPolicyDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-b428a14f-a52c-c871-e809-28e71ffbd849 writer.go:29: 2021-01-29T19:31:59.630Z [DEBUG] TestPolicyDeleteCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:59.633Z [INFO] TestPolicyDeleteCommand: Synced node info writer.go:29: 2021-01-29T19:31:59.690Z [DEBUG] TestPolicyDeleteCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:31:59.692Z [DEBUG] TestPolicyDeleteCommand: Node info in sync writer.go:29: 2021-01-29T19:31:59.695Z [DEBUG] TestPolicyDeleteCommand: Node info in sync writer.go:29: 2021-01-29T19:31:59.869Z [DEBUG] TestPolicyDeleteCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:46730 latency=9.989568ms writer.go:29: 2021-01-29T19:31:59.886Z [DEBUG] TestPolicyDeleteCommand.http: Request finished: method=DELETE url=/v1/acl/policy/909f17b1-4ee6-c7c2-d61b-2b555be958d6 from=127.0.0.1:46732 latency=5.559249ms writer.go:29: 2021-01-29T19:31:59.890Z [ERROR] TestPolicyDeleteCommand.http: Request error: method=GET url=/v1/acl/policy/909f17b1-4ee6-c7c2-d61b-2b555be958d6 from=127.0.0.1:46730 error="ACL not found" writer.go:29: 2021-01-29T19:31:59.896Z [DEBUG] TestPolicyDeleteCommand.http: Request finished: method=GET url=/v1/acl/policy/909f17b1-4ee6-c7c2-d61b-2b555be958d6 from=127.0.0.1:46730 latency=5.751159ms writer.go:29: 2021-01-29T19:31:59.899Z [INFO] TestPolicyDeleteCommand: Requesting shutdown writer.go:29: 2021-01-29T19:31:59.900Z [INFO] TestPolicyDeleteCommand.server: shutting down server writer.go:29: 2021-01-29T19:31:59.902Z [DEBUG] TestPolicyDeleteCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:59.903Z [DEBUG] TestPolicyDeleteCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:59.904Z [DEBUG] TestPolicyDeleteCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:59.906Z [WARN] TestPolicyDeleteCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:59.906Z [DEBUG] TestPolicyDeleteCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:31:59.905Z [DEBUG] TestPolicyDeleteCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:31:59.905Z [DEBUG] TestPolicyDeleteCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:31:59.907Z [WARN] TestPolicyDeleteCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:31:59.916Z [INFO] TestPolicyDeleteCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:31:59.916Z [INFO] TestPolicyDeleteCommand: consul server down writer.go:29: 2021-01-29T19:31:59.918Z [INFO] TestPolicyDeleteCommand: shutdown complete writer.go:29: 2021-01-29T19:31:59.919Z [INFO] TestPolicyDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:16127 network=tcp writer.go:29: 2021-01-29T19:31:59.920Z [INFO] TestPolicyDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:16127 network=udp writer.go:29: 2021-01-29T19:31:59.921Z [INFO] TestPolicyDeleteCommand: Stopping server: protocol=HTTP address=127.0.0.1:16128 network=tcp writer.go:29: 2021-01-29T19:32:00.421Z [INFO] TestPolicyDeleteCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:00.423Z [INFO] TestPolicyDeleteCommand: Endpoints down --- PASS: TestPolicyDeleteCommand (1.06s) PASS ok github.com/hashicorp/consul/command/acl/policy/delete 1.199s === RUN TestPolicyListCommand_noTabs === PAUSE TestPolicyListCommand_noTabs === RUN TestPolicyListCommand === PAUSE TestPolicyListCommand === RUN TestPolicyListCommand_JSON === PAUSE TestPolicyListCommand_JSON === CONT TestPolicyListCommand_noTabs === CONT TestPolicyListCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestPolicyListCommand_noTabs (0.00s) === CONT TestPolicyListCommand_JSON [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:32:02.748Z [WARN] TestPolicyListCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:02.751Z [DEBUG] TestPolicyListCommand_JSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:02.755Z [DEBUG] TestPolicyListCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.760Z [WARN] TestPolicyListCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:02.761Z [DEBUG] TestPolicyListCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:02.762Z [DEBUG] TestPolicyListCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.764Z [INFO] TestPolicyListCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9089df41-b4db-89e0-e746-756b2e1a7721 Address:127.0.0.1:20222}]" writer.go:29: 2021-01-29T19:32:02.767Z [INFO] TestPolicyListCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-9089df41-b4db-89e0-e746-756b2e1a7721.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:02.768Z [INFO] TestPolicyListCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:20222 [Follower]" leader= writer.go:29: 2021-01-29T19:32:02.770Z [INFO] TestPolicyListCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-9089df41-b4db-89e0-e746-756b2e1a7721 127.0.0.1 writer.go:29: 2021-01-29T19:32:02.774Z [INFO] TestPolicyListCommand_JSON: Started DNS server: address=127.0.0.1:20217 network=udp writer.go:29: 2021-01-29T19:32:02.774Z [INFO] TestPolicyListCommand_JSON.server: Handled event for server in area: event=member-join server=Node-9089df41-b4db-89e0-e746-756b2e1a7721.dc1 area=wan writer.go:29: 2021-01-29T19:32:02.775Z [INFO] TestPolicyListCommand_JSON.server: Adding LAN server: server="Node-9089df41-b4db-89e0-e746-756b2e1a7721 (Addr: tcp/127.0.0.1:20222) (DC: dc1)" writer.go:29: 2021-01-29T19:32:02.777Z [INFO] TestPolicyListCommand_JSON: Started DNS server: address=127.0.0.1:20217 network=tcp writer.go:29: 2021-01-29T19:32:02.780Z [INFO] TestPolicyListCommand_JSON: Started HTTP server: address=127.0.0.1:20218 network=tcp === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.781Z [INFO] TestPolicyListCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a260f3ce-f03e-4594-9e45-1abe834ca549 Address:127.0.0.1:20216}]" === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.781Z [INFO] TestPolicyListCommand_JSON: started state syncer === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.783Z [INFO] TestPolicyListCommand.server.serf.wan: serf: EventMemberJoin: Node-a260f3ce-f03e-4594-9e45-1abe834ca549.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:02.784Z [INFO] TestPolicyListCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:20216 [Follower]" leader= writer.go:29: 2021-01-29T19:32:02.795Z [INFO] TestPolicyListCommand.server.serf.lan: serf: EventMemberJoin: Node-a260f3ce-f03e-4594-9e45-1abe834ca549 127.0.0.1 writer.go:29: 2021-01-29T19:32:02.810Z [INFO] TestPolicyListCommand: Started DNS server: address=127.0.0.1:20211 network=udp writer.go:29: 2021-01-29T19:32:02.823Z [INFO] TestPolicyListCommand.server: Adding LAN server: server="Node-a260f3ce-f03e-4594-9e45-1abe834ca549 (Addr: tcp/127.0.0.1:20216) (DC: dc1)" writer.go:29: 2021-01-29T19:32:02.824Z [INFO] TestPolicyListCommand.server: Handled event for server in area: event=member-join server=Node-a260f3ce-f03e-4594-9e45-1abe834ca549.dc1 area=wan writer.go:29: 2021-01-29T19:32:02.825Z [INFO] TestPolicyListCommand: Started DNS server: address=127.0.0.1:20211 network=tcp === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.829Z [WARN] TestPolicyListCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:02.834Z [INFO] TestPolicyListCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:20222 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:02.829Z [DEBUG] TestPolicyListCommand_JSON.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.834Z [INFO] TestPolicyListCommand: Started HTTP server: address=127.0.0.1:20212 network=tcp writer.go:29: 2021-01-29T19:32:02.859Z [INFO] TestPolicyListCommand: started state syncer === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.863Z [DEBUG] TestPolicyListCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:02.865Z [DEBUG] TestPolicyListCommand_JSON.server.raft: vote granted: from=9089df41-b4db-89e0-e746-756b2e1a7721 term=2 tally=1 writer.go:29: 2021-01-29T19:32:02.865Z [INFO] TestPolicyListCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:02.866Z [INFO] TestPolicyListCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:20222 [Leader]" writer.go:29: 2021-01-29T19:32:02.867Z [INFO] TestPolicyListCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:02.869Z [INFO] TestPolicyListCommand_JSON.server: New leader elected: payload=Node-9089df41-b4db-89e0-e746-756b2e1a7721 === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.858Z [WARN] TestPolicyListCommand.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.869Z [INFO] TestPolicyListCommand_JSON.server: initializing acls === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.873Z [INFO] TestPolicyListCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:20216 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:02.875Z [DEBUG] TestPolicyListCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:02.884Z [DEBUG] TestPolicyListCommand.server.raft: votes: needed=1 === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.886Z [INFO] TestPolicyListCommand_JSON.server: Created ACL 'global-management' policy === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.891Z [DEBUG] TestPolicyListCommand.server.raft: vote granted: from=a260f3ce-f03e-4594-9e45-1abe834ca549 term=2 tally=1 === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.892Z [WARN] TestPolicyListCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.896Z [INFO] TestPolicyListCommand.server.raft: election won: tally=1 === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.899Z [INFO] TestPolicyListCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.901Z [INFO] TestPolicyListCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:20216 [Leader]" writer.go:29: 2021-01-29T19:32:02.907Z [INFO] TestPolicyListCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:02.907Z [INFO] TestPolicyListCommand.server: New leader elected: payload=Node-a260f3ce-f03e-4594-9e45-1abe834ca549 === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.905Z [INFO] TestPolicyListCommand_JSON.server: Created ACL anonymous token from configuration === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.908Z [INFO] TestPolicyListCommand.server: initializing acls === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.910Z [INFO] TestPolicyListCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.913Z [INFO] TestPolicyListCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:02.915Z [WARN] TestPolicyListCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:02.917Z [INFO] TestPolicyListCommand.server: Bootstrapped ACL master token from configuration === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.915Z [INFO] TestPolicyListCommand_JSON.leader: started routine: routine="acl token reaping" === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.923Z [INFO] TestPolicyListCommand.server: Created ACL anonymous token from configuration === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.925Z [INFO] TestPolicyListCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-9089df41-b4db-89e0-e746-756b2e1a7721 === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.930Z [INFO] TestPolicyListCommand.leader: started routine: routine="legacy ACL token upgrade" === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.934Z [INFO] TestPolicyListCommand_JSON.server: Updating LAN server: server="Node-9089df41-b4db-89e0-e746-756b2e1a7721 (Addr: tcp/127.0.0.1:20222) (DC: dc1)" writer.go:29: 2021-01-29T19:32:02.935Z [INFO] TestPolicyListCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-9089df41-b4db-89e0-e746-756b2e1a7721.dc1 writer.go:29: 2021-01-29T19:32:02.936Z [INFO] TestPolicyListCommand_JSON.server: Handled event for server in area: event=member-update server=Node-9089df41-b4db-89e0-e746-756b2e1a7721.dc1 area=wan === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.935Z [INFO] TestPolicyListCommand.leader: started routine: routine="acl token reaping" === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:02.940Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:02.941Z [INFO] TestPolicyListCommand.server.serf.lan: serf: EventMemberUpdate: Node-a260f3ce-f03e-4594-9e45-1abe834ca549 writer.go:29: 2021-01-29T19:32:02.945Z [INFO] TestPolicyListCommand.server: Updating LAN server: server="Node-a260f3ce-f03e-4594-9e45-1abe834ca549 (Addr: tcp/127.0.0.1:20216) (DC: dc1)" writer.go:29: 2021-01-29T19:32:02.945Z [INFO] TestPolicyListCommand.server.serf.wan: serf: EventMemberUpdate: Node-a260f3ce-f03e-4594-9e45-1abe834ca549.dc1 writer.go:29: 2021-01-29T19:32:02.948Z [INFO] TestPolicyListCommand.server: Handled event for server in area: event=member-update server=Node-a260f3ce-f03e-4594-9e45-1abe834ca549.dc1 area=wan writer.go:29: 2021-01-29T19:32:02.949Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:02.962Z [INFO] TestPolicyListCommand: Synced node info writer.go:29: 2021-01-29T19:32:03.010Z [INFO] TestPolicyListCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.011Z [INFO] TestPolicyListCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.011Z [INFO] TestPolicyListCommand.leader: started routine: routine="CA root pruning" === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.011Z [INFO] TestPolicyListCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.011Z [DEBUG] TestPolicyListCommand.server: Skipping self join check for node since the cluster is too small: node=Node-a260f3ce-f03e-4594-9e45-1abe834ca549 === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.012Z [DEBUG] TestPolicyListCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-9089df41-b4db-89e0-e746-756b2e1a7721 === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.012Z [INFO] TestPolicyListCommand.server: member joined, marking health alive: member=Node-a260f3ce-f03e-4594-9e45-1abe834ca549 === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.013Z [INFO] TestPolicyListCommand_JSON.server: member joined, marking health alive: member=Node-9089df41-b4db-89e0-e746-756b2e1a7721 === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.014Z [DEBUG] TestPolicyListCommand.server: Skipping self join check for node since the cluster is too small: node=Node-a260f3ce-f03e-4594-9e45-1abe834ca549 === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.015Z [DEBUG] TestPolicyListCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-9089df41-b4db-89e0-e746-756b2e1a7721 writer.go:29: 2021-01-29T19:32:03.020Z [DEBUG] TestPolicyListCommand_JSON: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:03.022Z [INFO] TestPolicyListCommand_JSON: Synced node info writer.go:29: 2021-01-29T19:32:03.023Z [DEBUG] TestPolicyListCommand_JSON: Node info in sync === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.027Z [DEBUG] TestPolicyListCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:43514 latency=6.060055ms writer.go:29: 2021-01-29T19:32:03.032Z [DEBUG] TestPolicyListCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:43514 latency=546.015µs writer.go:29: 2021-01-29T19:32:03.041Z [DEBUG] TestPolicyListCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:43514 latency=835.588µs writer.go:29: 2021-01-29T19:32:03.044Z [DEBUG] TestPolicyListCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:43514 latency=859.546µs writer.go:29: 2021-01-29T19:32:03.082Z [DEBUG] TestPolicyListCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:43514 latency=13.03054ms === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.087Z [DEBUG] TestPolicyListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:50652 latency=13.581851ms writer.go:29: 2021-01-29T19:32:03.090Z [DEBUG] TestPolicyListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:50652 latency=511.831µs writer.go:29: 2021-01-29T19:32:03.092Z [DEBUG] TestPolicyListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:50652 latency=830.03µs === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.095Z [DEBUG] TestPolicyListCommand.http: Request finished: method=GET url=/v1/acl/policies from=127.0.0.1:43518 latency=2.531171ms === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.095Z [DEBUG] TestPolicyListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:50652 latency=665.379µs writer.go:29: 2021-01-29T19:32:03.098Z [DEBUG] TestPolicyListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:50652 latency=792.114µs writer.go:29: 2021-01-29T19:32:03.105Z [DEBUG] TestPolicyListCommand_JSON.http: Request finished: method=GET url=/v1/acl/policies from=127.0.0.1:50656 latency=1.573814ms === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.112Z [INFO] TestPolicyListCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:03.114Z [INFO] TestPolicyListCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:03.115Z [DEBUG] TestPolicyListCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:03.116Z [DEBUG] TestPolicyListCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:03.118Z [DEBUG] TestPolicyListCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:03.116Z [DEBUG] TestPolicyListCommand.leader: stopped routine: routine="acl token reaping" === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.113Z [INFO] TestPolicyListCommand_JSON: Requesting shutdown === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.118Z [DEBUG] TestPolicyListCommand.leader: stopped routine: routine="CA root pruning" === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.122Z [INFO] TestPolicyListCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:03.122Z [DEBUG] TestPolicyListCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:03.123Z [DEBUG] TestPolicyListCommand_JSON.leader: stopping routine: routine="acl token reaping" === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.119Z [WARN] TestPolicyListCommand.server.serf.lan: serf: Shutdown without a Leave === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.124Z [DEBUG] TestPolicyListCommand_JSON.leader: stopping routine: routine="CA root pruning" === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.119Z [DEBUG] TestPolicyListCommand.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.125Z [WARN] TestPolicyListCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:03.126Z [DEBUG] TestPolicyListCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:03.127Z [WARN] TestPolicyListCommand_JSON.server.serf.wan: serf: Shutdown without a Leave === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.128Z [WARN] TestPolicyListCommand.server.serf.wan: serf: Shutdown without a Leave === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.127Z [DEBUG] TestPolicyListCommand_JSON.leader: stopped routine: routine="CA root pruning" === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.130Z [INFO] TestPolicyListCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:03.130Z [INFO] TestPolicyListCommand: consul server down === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.127Z [DEBUG] TestPolicyListCommand_JSON.leader: stopped routine: routine="acl token reaping" === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.132Z [INFO] TestPolicyListCommand: shutdown complete writer.go:29: 2021-01-29T19:32:03.132Z [INFO] TestPolicyListCommand: Stopping server: protocol=DNS address=127.0.0.1:20211 network=tcp writer.go:29: 2021-01-29T19:32:03.133Z [INFO] TestPolicyListCommand: Stopping server: protocol=DNS address=127.0.0.1:20211 network=udp writer.go:29: 2021-01-29T19:32:03.134Z [INFO] TestPolicyListCommand: Stopping server: protocol=HTTP address=127.0.0.1:20212 network=tcp === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.131Z [INFO] TestPolicyListCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:03.131Z [INFO] TestPolicyListCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:03.144Z [INFO] TestPolicyListCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:32:03.145Z [INFO] TestPolicyListCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:20217 network=tcp writer.go:29: 2021-01-29T19:32:03.146Z [INFO] TestPolicyListCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:20217 network=udp writer.go:29: 2021-01-29T19:32:03.146Z [INFO] TestPolicyListCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:20218 network=tcp === CONT TestPolicyListCommand writer.go:29: 2021-01-29T19:32:03.635Z [INFO] TestPolicyListCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:03.636Z [INFO] TestPolicyListCommand: Endpoints down --- PASS: TestPolicyListCommand (0.96s) === CONT TestPolicyListCommand_JSON writer.go:29: 2021-01-29T19:32:03.647Z [INFO] TestPolicyListCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:03.648Z [INFO] TestPolicyListCommand_JSON: Endpoints down --- PASS: TestPolicyListCommand_JSON (0.97s) PASS ok github.com/hashicorp/consul/command/acl/policy/list 1.151s === RUN TestPolicyReadCommand_noTabs === PAUSE TestPolicyReadCommand_noTabs === RUN TestPolicyReadCommand === PAUSE TestPolicyReadCommand === RUN TestPolicyReadCommand_JSON === PAUSE TestPolicyReadCommand_JSON === CONT TestPolicyReadCommand_noTabs === CONT TestPolicyReadCommand_JSON === CONT TestPolicyReadCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestPolicyReadCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:04.985Z [WARN] TestPolicyReadCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:04.988Z [DEBUG] TestPolicyReadCommand_JSON.tlsutil: Update: version=1 === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:04.993Z [WARN] TestPolicyReadCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:04.994Z [DEBUG] TestPolicyReadCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:04.995Z [DEBUG] TestPolicyReadCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:04.994Z [DEBUG] TestPolicyReadCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:05.001Z [INFO] TestPolicyReadCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1f2436aa-dee8-20aa-9280-296e2b3b0180 Address:127.0.0.1:26348}]" writer.go:29: 2021-01-29T19:32:05.004Z [INFO] TestPolicyReadCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:26348 [Follower]" leader= === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.001Z [INFO] TestPolicyReadCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3737ca17-c9cb-f8eb-1cfc-c39ffc724aea Address:127.0.0.1:26342}]" === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.005Z [INFO] TestPolicyReadCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-1f2436aa-dee8-20aa-9280-296e2b3b0180.dc1 127.0.0.1 === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.010Z [INFO] TestPolicyReadCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:26342 [Follower]" leader= === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.014Z [INFO] TestPolicyReadCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 127.0.0.1 writer.go:29: 2021-01-29T19:32:05.022Z [INFO] TestPolicyReadCommand_JSON.server: Adding LAN server: server="Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 (Addr: tcp/127.0.0.1:26348) (DC: dc1)" writer.go:29: 2021-01-29T19:32:05.030Z [INFO] TestPolicyReadCommand_JSON.server: Handled event for server in area: event=member-join server=Node-1f2436aa-dee8-20aa-9280-296e2b3b0180.dc1 area=wan writer.go:29: 2021-01-29T19:32:05.033Z [INFO] TestPolicyReadCommand_JSON: Started DNS server: address=127.0.0.1:26343 network=udp === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.025Z [INFO] TestPolicyReadCommand.server.serf.wan: serf: EventMemberJoin: Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea.dc1 127.0.0.1 === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.035Z [INFO] TestPolicyReadCommand_JSON: Started DNS server: address=127.0.0.1:26343 network=tcp writer.go:29: 2021-01-29T19:32:05.044Z [INFO] TestPolicyReadCommand_JSON: Started HTTP server: address=127.0.0.1:26344 network=tcp writer.go:29: 2021-01-29T19:32:05.046Z [INFO] TestPolicyReadCommand_JSON: started state syncer writer.go:29: 2021-01-29T19:32:05.047Z [WARN] TestPolicyReadCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:05.050Z [INFO] TestPolicyReadCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:26348 [Candidate]" term=2 === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.043Z [INFO] TestPolicyReadCommand.server.serf.lan: serf: EventMemberJoin: Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea 127.0.0.1 === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.052Z [DEBUG] TestPolicyReadCommand_JSON.server.raft: votes: needed=1 === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.053Z [INFO] TestPolicyReadCommand.server: Handled event for server in area: event=member-join server=Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea.dc1 area=wan === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.054Z [DEBUG] TestPolicyReadCommand_JSON.server.raft: vote granted: from=1f2436aa-dee8-20aa-9280-296e2b3b0180 term=2 tally=1 writer.go:29: 2021-01-29T19:32:05.056Z [INFO] TestPolicyReadCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:05.058Z [INFO] TestPolicyReadCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:26348 [Leader]" writer.go:29: 2021-01-29T19:32:05.060Z [INFO] TestPolicyReadCommand_JSON.server: cluster leadership acquired === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.058Z [INFO] TestPolicyReadCommand: Started DNS server: address=127.0.0.1:26337 network=udp === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.064Z [INFO] TestPolicyReadCommand_JSON.server: New leader elected: payload=Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.064Z [INFO] TestPolicyReadCommand: Started DNS server: address=127.0.0.1:26337 network=tcp writer.go:29: 2021-01-29T19:32:05.061Z [INFO] TestPolicyReadCommand.server: Adding LAN server: server="Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea (Addr: tcp/127.0.0.1:26342) (DC: dc1)" writer.go:29: 2021-01-29T19:32:05.066Z [INFO] TestPolicyReadCommand: Started HTTP server: address=127.0.0.1:26338 network=tcp writer.go:29: 2021-01-29T19:32:05.069Z [INFO] TestPolicyReadCommand: started state syncer === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.071Z [INFO] TestPolicyReadCommand_JSON.server: initializing acls === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.074Z [WARN] TestPolicyReadCommand.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.076Z [INFO] TestPolicyReadCommand_JSON.server: Created ACL 'global-management' policy === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.076Z [INFO] TestPolicyReadCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:26342 [Candidate]" term=2 === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.077Z [WARN] TestPolicyReadCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.079Z [DEBUG] TestPolicyReadCommand.server.raft: votes: needed=1 === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.080Z [INFO] TestPolicyReadCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:32:05.082Z [WARN] TestPolicyReadCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:05.081Z [INFO] TestPolicyReadCommand_JSON.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:05.083Z [INFO] TestPolicyReadCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:05.084Z [INFO] TestPolicyReadCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:05.086Z [INFO] TestPolicyReadCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:05.087Z [INFO] TestPolicyReadCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:05.088Z [DEBUG] TestPolicyReadCommand_JSON.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:32:05.089Z [INFO] TestPolicyReadCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 writer.go:29: 2021-01-29T19:32:05.090Z [INFO] TestPolicyReadCommand_JSON.server: Updating LAN server: server="Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 (Addr: tcp/127.0.0.1:26348) (DC: dc1)" === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.081Z [DEBUG] TestPolicyReadCommand.server.raft: vote granted: from=3737ca17-c9cb-f8eb-1cfc-c39ffc724aea term=2 tally=1 writer.go:29: 2021-01-29T19:32:05.095Z [INFO] TestPolicyReadCommand.server.raft: election won: tally=1 === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.090Z [INFO] TestPolicyReadCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.099Z [INFO] TestPolicyReadCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:26342 [Leader]" === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.091Z [INFO] TestPolicyReadCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-1f2436aa-dee8-20aa-9280-296e2b3b0180.dc1 === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.104Z [INFO] TestPolicyReadCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:05.104Z [INFO] TestPolicyReadCommand.server: New leader elected: payload=Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea writer.go:29: 2021-01-29T19:32:05.104Z [INFO] TestPolicyReadCommand.server: initializing acls === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.100Z [INFO] TestPolicyReadCommand_JSON.server: Updating LAN server: server="Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 (Addr: tcp/127.0.0.1:26348) (DC: dc1)" === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.108Z [INFO] TestPolicyReadCommand.server: initializing acls writer.go:29: 2021-01-29T19:32:05.113Z [INFO] TestPolicyReadCommand.server: Created ACL 'global-management' policy === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.105Z [INFO] TestPolicyReadCommand_JSON.server: Handled event for server in area: event=member-update server=Node-1f2436aa-dee8-20aa-9280-296e2b3b0180.dc1 area=wan === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.113Z [WARN] TestPolicyReadCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:05.113Z [INFO] TestPolicyReadCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:05.116Z [WARN] TestPolicyReadCommand.server: Configuring a non-UUID master token is deprecated === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.105Z [INFO] TestPolicyReadCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-1f2436aa-dee8-20aa-9280-296e2b3b0180.dc1 === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.120Z [INFO] TestPolicyReadCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:05.120Z [INFO] TestPolicyReadCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:05.122Z [INFO] TestPolicyReadCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:05.123Z [INFO] TestPolicyReadCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:05.121Z [INFO] TestPolicyReadCommand.server: Created ACL anonymous token from configuration === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.114Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.124Z [DEBUG] TestPolicyReadCommand.server: transitioning out of legacy ACL mode === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.119Z [INFO] TestPolicyReadCommand_JSON.server: Handled event for server in area: event=member-update server=Node-1f2436aa-dee8-20aa-9280-296e2b3b0180.dc1 area=wan === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.126Z [INFO] TestPolicyReadCommand.server.serf.lan: serf: EventMemberUpdate: Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea writer.go:29: 2021-01-29T19:32:05.128Z [INFO] TestPolicyReadCommand.server: Updating LAN server: server="Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea (Addr: tcp/127.0.0.1:26342) (DC: dc1)" writer.go:29: 2021-01-29T19:32:05.128Z [INFO] TestPolicyReadCommand.server.serf.wan: serf: EventMemberUpdate: Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea.dc1 writer.go:29: 2021-01-29T19:32:05.128Z [INFO] TestPolicyReadCommand.server.serf.lan: serf: EventMemberUpdate: Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea writer.go:29: 2021-01-29T19:32:05.136Z [INFO] TestPolicyReadCommand.server: Updating LAN server: server="Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea (Addr: tcp/127.0.0.1:26342) (DC: dc1)" writer.go:29: 2021-01-29T19:32:05.137Z [INFO] TestPolicyReadCommand.server: Handled event for server in area: event=member-update server=Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea.dc1 area=wan writer.go:29: 2021-01-29T19:32:05.137Z [INFO] TestPolicyReadCommand.server.serf.wan: serf: EventMemberUpdate: Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea.dc1 writer.go:29: 2021-01-29T19:32:05.140Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:05.148Z [INFO] TestPolicyReadCommand.server: Handled event for server in area: event=member-update server=Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea.dc1 area=wan === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.178Z [INFO] TestPolicyReadCommand_JSON: Synced node info === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.180Z [INFO] TestPolicyReadCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:05.181Z [INFO] TestPolicyReadCommand.leader: started routine: routine="CA root pruning" === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.180Z [DEBUG] TestPolicyReadCommand_JSON: Node info in sync === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.182Z [DEBUG] TestPolicyReadCommand.server: Skipping self join check for node since the cluster is too small: node=Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea writer.go:29: 2021-01-29T19:32:05.183Z [INFO] TestPolicyReadCommand.server: member joined, marking health alive: member=Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.181Z [INFO] TestPolicyReadCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.184Z [DEBUG] TestPolicyReadCommand.server: Skipping self join check for node since the cluster is too small: node=Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea writer.go:29: 2021-01-29T19:32:05.185Z [DEBUG] TestPolicyReadCommand.server: Skipping self join check for node since the cluster is too small: node=Node-3737ca17-c9cb-f8eb-1cfc-c39ffc724aea === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.184Z [INFO] TestPolicyReadCommand_JSON.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.186Z [DEBUG] TestPolicyReadCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 writer.go:29: 2021-01-29T19:32:05.188Z [INFO] TestPolicyReadCommand_JSON.server: member joined, marking health alive: member=Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 writer.go:29: 2021-01-29T19:32:05.190Z [DEBUG] TestPolicyReadCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 writer.go:29: 2021-01-29T19:32:05.191Z [DEBUG] TestPolicyReadCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-1f2436aa-dee8-20aa-9280-296e2b3b0180 writer.go:29: 2021-01-29T19:32:05.202Z [DEBUG] TestPolicyReadCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:37056 latency=6.393194ms writer.go:29: 2021-01-29T19:32:05.212Z [DEBUG] TestPolicyReadCommand_JSON.http: Request finished: method=GET url=/v1/acl/policy/40ad8cfc-d136-3245-c28b-98d2aa03f98e from=127.0.0.1:37058 latency=4.601434ms writer.go:29: 2021-01-29T19:32:05.222Z [INFO] TestPolicyReadCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:32:05.226Z [INFO] TestPolicyReadCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:05.230Z [DEBUG] TestPolicyReadCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:05.234Z [DEBUG] TestPolicyReadCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:05.238Z [DEBUG] TestPolicyReadCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.240Z [WARN] TestPolicyReadCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:05.235Z [DEBUG] TestPolicyReadCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:05.245Z [DEBUG] TestPolicyReadCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:05.246Z [DEBUG] TestPolicyReadCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.254Z [WARN] TestPolicyReadCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:05.259Z [INFO] TestPolicyReadCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:05.259Z [INFO] TestPolicyReadCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:05.263Z [INFO] TestPolicyReadCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:32:05.264Z [INFO] TestPolicyReadCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:26343 network=tcp writer.go:29: 2021-01-29T19:32:05.265Z [INFO] TestPolicyReadCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:26343 network=udp writer.go:29: 2021-01-29T19:32:05.266Z [INFO] TestPolicyReadCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:26344 network=tcp === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.390Z [DEBUG] TestPolicyReadCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:56692 latency=6.666869ms writer.go:29: 2021-01-29T19:32:05.397Z [DEBUG] TestPolicyReadCommand.http: Request finished: method=GET url=/v1/acl/policy/ef6ad1d1-63d5-5f5f-9da9-f9bfc6f2f18d from=127.0.0.1:56694 latency=1.334483ms writer.go:29: 2021-01-29T19:32:05.400Z [INFO] TestPolicyReadCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:05.401Z [INFO] TestPolicyReadCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:05.402Z [DEBUG] TestPolicyReadCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:05.403Z [DEBUG] TestPolicyReadCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:05.404Z [DEBUG] TestPolicyReadCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.402Z [ERROR] TestPolicyReadCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:05.403Z [DEBUG] TestPolicyReadCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:05.404Z [DEBUG] TestPolicyReadCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:05.405Z [WARN] TestPolicyReadCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:05.405Z [DEBUG] TestPolicyReadCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:05.410Z [WARN] TestPolicyReadCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:05.413Z [INFO] TestPolicyReadCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:05.413Z [INFO] TestPolicyReadCommand: consul server down writer.go:29: 2021-01-29T19:32:05.415Z [INFO] TestPolicyReadCommand: shutdown complete writer.go:29: 2021-01-29T19:32:05.416Z [INFO] TestPolicyReadCommand: Stopping server: protocol=DNS address=127.0.0.1:26337 network=tcp writer.go:29: 2021-01-29T19:32:05.417Z [INFO] TestPolicyReadCommand: Stopping server: protocol=DNS address=127.0.0.1:26337 network=udp writer.go:29: 2021-01-29T19:32:05.418Z [INFO] TestPolicyReadCommand: Stopping server: protocol=HTTP address=127.0.0.1:26338 network=tcp === CONT TestPolicyReadCommand_JSON writer.go:29: 2021-01-29T19:32:05.768Z [INFO] TestPolicyReadCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:05.769Z [INFO] TestPolicyReadCommand_JSON: Endpoints down --- PASS: TestPolicyReadCommand_JSON (0.83s) === CONT TestPolicyReadCommand writer.go:29: 2021-01-29T19:32:05.919Z [INFO] TestPolicyReadCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:05.920Z [INFO] TestPolicyReadCommand: Endpoints down --- PASS: TestPolicyReadCommand (0.98s) PASS ok github.com/hashicorp/consul/command/acl/policy/read 1.276s === RUN TestPolicyUpdateCommand_noTabs === PAUSE TestPolicyUpdateCommand_noTabs === RUN TestPolicyUpdateCommand === PAUSE TestPolicyUpdateCommand === RUN TestPolicyUpdateCommand_JSON === PAUSE TestPolicyUpdateCommand_JSON === CONT TestPolicyUpdateCommand_noTabs === CONT TestPolicyUpdateCommand_JSON === CONT TestPolicyUpdateCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestPolicyUpdateCommand_noTabs (0.01s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:06.969Z [WARN] TestPolicyUpdateCommand: bootstrap = true: do not enable unless necessary === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:06.969Z [WARN] TestPolicyUpdateCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:06.970Z [DEBUG] TestPolicyUpdateCommand_JSON.tlsutil: Update: version=1 === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:06.970Z [DEBUG] TestPolicyUpdateCommand.tlsutil: Update: version=1 === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:06.971Z [DEBUG] TestPolicyUpdateCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:06.973Z [DEBUG] TestPolicyUpdateCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:06.976Z [INFO] TestPolicyUpdateCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f895c207-6511-915a-4149-6343f1108bbf Address:127.0.0.1:27369}]" writer.go:29: 2021-01-29T19:32:06.978Z [INFO] TestPolicyUpdateCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:27369 [Follower]" leader= writer.go:29: 2021-01-29T19:32:06.980Z [INFO] TestPolicyUpdateCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-f895c207-6511-915a-4149-6343f1108bbf.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:06.992Z [INFO] TestPolicyUpdateCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-f895c207-6511-915a-4149-6343f1108bbf 127.0.0.1 === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:06.993Z [INFO] TestPolicyUpdateCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56 Address:127.0.0.1:27363}]" === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:06.994Z [INFO] TestPolicyUpdateCommand_JSON.server: Adding LAN server: server="Node-f895c207-6511-915a-4149-6343f1108bbf (Addr: tcp/127.0.0.1:27369) (DC: dc1)" writer.go:29: 2021-01-29T19:32:06.994Z [INFO] TestPolicyUpdateCommand_JSON: Started DNS server: address=127.0.0.1:27364 network=udp writer.go:29: 2021-01-29T19:32:06.994Z [INFO] TestPolicyUpdateCommand_JSON.server: Handled event for server in area: event=member-join server=Node-f895c207-6511-915a-4149-6343f1108bbf.dc1 area=wan === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:06.995Z [INFO] TestPolicyUpdateCommand.server.serf.wan: serf: EventMemberJoin: Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56.dc1 127.0.0.1 === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:06.995Z [INFO] TestPolicyUpdateCommand_JSON: Started DNS server: address=127.0.0.1:27364 network=tcp === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:06.997Z [INFO] TestPolicyUpdateCommand.server.serf.lan: serf: EventMemberJoin: Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56 127.0.0.1 writer.go:29: 2021-01-29T19:32:06.996Z [INFO] TestPolicyUpdateCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:27363 [Follower]" leader= writer.go:29: 2021-01-29T19:32:07.003Z [INFO] TestPolicyUpdateCommand: Started DNS server: address=127.0.0.1:27358 network=udp writer.go:29: 2021-01-29T19:32:07.005Z [INFO] TestPolicyUpdateCommand.server: Adding LAN server: server="Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56 (Addr: tcp/127.0.0.1:27363) (DC: dc1)" === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:07.001Z [INFO] TestPolicyUpdateCommand_JSON: Started HTTP server: address=127.0.0.1:27365 network=tcp writer.go:29: 2021-01-29T19:32:07.009Z [INFO] TestPolicyUpdateCommand_JSON: started state syncer === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:07.008Z [INFO] TestPolicyUpdateCommand.server: Handled event for server in area: event=member-join server=Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56.dc1 area=wan writer.go:29: 2021-01-29T19:32:07.011Z [INFO] TestPolicyUpdateCommand: Started DNS server: address=127.0.0.1:27358 network=tcp writer.go:29: 2021-01-29T19:32:07.020Z [INFO] TestPolicyUpdateCommand: Started HTTP server: address=127.0.0.1:27359 network=tcp writer.go:29: 2021-01-29T19:32:07.023Z [INFO] TestPolicyUpdateCommand: started state syncer === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:07.026Z [WARN] TestPolicyUpdateCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:07.027Z [INFO] TestPolicyUpdateCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:27369 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:07.030Z [DEBUG] TestPolicyUpdateCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:07.031Z [DEBUG] TestPolicyUpdateCommand_JSON.server.raft: vote granted: from=f895c207-6511-915a-4149-6343f1108bbf term=2 tally=1 writer.go:29: 2021-01-29T19:32:07.032Z [INFO] TestPolicyUpdateCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:07.034Z [INFO] TestPolicyUpdateCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:27369 [Leader]" writer.go:29: 2021-01-29T19:32:07.036Z [ERROR] TestPolicyUpdateCommand_JSON.anti_entropy: failed to sync remote state: error="ACL not found" writer.go:29: 2021-01-29T19:32:07.037Z [INFO] TestPolicyUpdateCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:07.038Z [INFO] TestPolicyUpdateCommand_JSON.server: New leader elected: payload=Node-f895c207-6511-915a-4149-6343f1108bbf writer.go:29: 2021-01-29T19:32:07.038Z [INFO] TestPolicyUpdateCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:32:07.044Z [INFO] TestPolicyUpdateCommand_JSON.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:07.045Z [WARN] TestPolicyUpdateCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:07.045Z [INFO] TestPolicyUpdateCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:32:07.048Z [WARN] TestPolicyUpdateCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:07.047Z [INFO] TestPolicyUpdateCommand_JSON.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:07.050Z [INFO] TestPolicyUpdateCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:07.051Z [INFO] TestPolicyUpdateCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:07.050Z [INFO] TestPolicyUpdateCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:07.052Z [DEBUG] TestPolicyUpdateCommand_JSON.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:32:07.052Z [INFO] TestPolicyUpdateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-f895c207-6511-915a-4149-6343f1108bbf writer.go:29: 2021-01-29T19:32:07.055Z [INFO] TestPolicyUpdateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-f895c207-6511-915a-4149-6343f1108bbf.dc1 writer.go:29: 2021-01-29T19:32:07.057Z [INFO] TestPolicyUpdateCommand_JSON.server: Updating LAN server: server="Node-f895c207-6511-915a-4149-6343f1108bbf (Addr: tcp/127.0.0.1:27369) (DC: dc1)" writer.go:29: 2021-01-29T19:32:07.057Z [INFO] TestPolicyUpdateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-f895c207-6511-915a-4149-6343f1108bbf === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:07.059Z [DEBUG] TestPolicyUpdateCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:07.060Z [INFO] TestPolicyUpdateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-f895c207-6511-915a-4149-6343f1108bbf.dc1 writer.go:29: 2021-01-29T19:32:07.058Z [INFO] TestPolicyUpdateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-f895c207-6511-915a-4149-6343f1108bbf.dc1 area=wan writer.go:29: 2021-01-29T19:32:07.059Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:07.061Z [INFO] TestPolicyUpdateCommand_JSON.server: Updating LAN server: server="Node-f895c207-6511-915a-4149-6343f1108bbf (Addr: tcp/127.0.0.1:27369) (DC: dc1)" writer.go:29: 2021-01-29T19:32:07.062Z [INFO] TestPolicyUpdateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-f895c207-6511-915a-4149-6343f1108bbf.dc1 area=wan === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:07.063Z [WARN] TestPolicyUpdateCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:07.064Z [INFO] TestPolicyUpdateCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:27363 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:07.068Z [DEBUG] TestPolicyUpdateCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:07.070Z [DEBUG] TestPolicyUpdateCommand.server.raft: vote granted: from=f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56 term=2 tally=1 writer.go:29: 2021-01-29T19:32:07.072Z [INFO] TestPolicyUpdateCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:07.074Z [INFO] TestPolicyUpdateCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:27363 [Leader]" writer.go:29: 2021-01-29T19:32:07.075Z [INFO] TestPolicyUpdateCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:07.076Z [INFO] TestPolicyUpdateCommand.server: New leader elected: payload=Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56 writer.go:29: 2021-01-29T19:32:07.092Z [INFO] TestPolicyUpdateCommand.server: initializing acls writer.go:29: 2021-01-29T19:32:07.103Z [INFO] TestPolicyUpdateCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:07.103Z [WARN] TestPolicyUpdateCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:07.105Z [INFO] TestPolicyUpdateCommand.server: Bootstrapped ACL master token from configuration === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:07.106Z [INFO] TestPolicyUpdateCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:07.107Z [INFO] TestPolicyUpdateCommand.server: Created ACL anonymous token from configuration === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:07.107Z [INFO] TestPolicyUpdateCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:07.107Z [INFO] TestPolicyUpdateCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:07.108Z [INFO] TestPolicyUpdateCommand.leader: started routine: routine="acl token reaping" === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:07.108Z [DEBUG] TestPolicyUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-f895c207-6511-915a-4149-6343f1108bbf === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:07.109Z [INFO] TestPolicyUpdateCommand.server.serf.lan: serf: EventMemberUpdate: Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56 === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:07.109Z [INFO] TestPolicyUpdateCommand_JSON.server: member joined, marking health alive: member=Node-f895c207-6511-915a-4149-6343f1108bbf === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:07.110Z [INFO] TestPolicyUpdateCommand.server: Updating LAN server: server="Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56 (Addr: tcp/127.0.0.1:27363) (DC: dc1)" writer.go:29: 2021-01-29T19:32:07.110Z [INFO] TestPolicyUpdateCommand.server.serf.wan: serf: EventMemberUpdate: Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56.dc1 === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:07.112Z [DEBUG] TestPolicyUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-f895c207-6511-915a-4149-6343f1108bbf writer.go:29: 2021-01-29T19:32:07.117Z [DEBUG] TestPolicyUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-f895c207-6511-915a-4149-6343f1108bbf === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:07.116Z [INFO] TestPolicyUpdateCommand.server: Handled event for server in area: event=member-update server=Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56.dc1 area=wan writer.go:29: 2021-01-29T19:32:07.123Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:07.129Z [INFO] TestPolicyUpdateCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:07.130Z [INFO] TestPolicyUpdateCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:07.132Z [DEBUG] TestPolicyUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56 writer.go:29: 2021-01-29T19:32:07.133Z [INFO] TestPolicyUpdateCommand.server: member joined, marking health alive: member=Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56 writer.go:29: 2021-01-29T19:32:07.138Z [DEBUG] TestPolicyUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-f30ebf6c-eba2-bbc8-c59e-314ef7fa4d56 writer.go:29: 2021-01-29T19:32:07.157Z [DEBUG] TestPolicyUpdateCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:33224 latency=4.48169ms writer.go:29: 2021-01-29T19:32:07.166Z [DEBUG] TestPolicyUpdateCommand.http: Request finished: method=GET url=/v1/acl/policy/5cd77eb1-dade-5fde-36c0-6f5a10206e90 from=127.0.0.1:33226 latency=2.620105ms writer.go:29: 2021-01-29T19:32:07.174Z [DEBUG] TestPolicyUpdateCommand.http: Request finished: method=PUT url=/v1/acl/policy/5cd77eb1-dade-5fde-36c0-6f5a10206e90 from=127.0.0.1:33226 latency=1.700638ms writer.go:29: 2021-01-29T19:32:07.177Z [INFO] TestPolicyUpdateCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:07.178Z [INFO] TestPolicyUpdateCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:07.179Z [DEBUG] TestPolicyUpdateCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:07.181Z [DEBUG] TestPolicyUpdateCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:07.183Z [DEBUG] TestPolicyUpdateCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:07.185Z [WARN] TestPolicyUpdateCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:07.187Z [DEBUG] TestPolicyUpdateCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:07.187Z [ERROR] TestPolicyUpdateCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:07.187Z [DEBUG] TestPolicyUpdateCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:07.187Z [DEBUG] TestPolicyUpdateCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:07.190Z [WARN] TestPolicyUpdateCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:07.200Z [INFO] TestPolicyUpdateCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:07.200Z [INFO] TestPolicyUpdateCommand: consul server down writer.go:29: 2021-01-29T19:32:07.202Z [INFO] TestPolicyUpdateCommand: shutdown complete writer.go:29: 2021-01-29T19:32:07.203Z [INFO] TestPolicyUpdateCommand: Stopping server: protocol=DNS address=127.0.0.1:27358 network=tcp writer.go:29: 2021-01-29T19:32:07.204Z [INFO] TestPolicyUpdateCommand: Stopping server: protocol=DNS address=127.0.0.1:27358 network=udp writer.go:29: 2021-01-29T19:32:07.204Z [INFO] TestPolicyUpdateCommand: Stopping server: protocol=HTTP address=127.0.0.1:27359 network=tcp === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:07.295Z [DEBUG] TestPolicyUpdateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:53256 latency=4.522169ms writer.go:29: 2021-01-29T19:32:07.301Z [DEBUG] TestPolicyUpdateCommand_JSON.http: Request finished: method=GET url=/v1/acl/policy/ea11f0fd-fcbd-b75f-1c1b-68abfe0411de from=127.0.0.1:53258 latency=1.349553ms writer.go:29: 2021-01-29T19:32:07.304Z [DEBUG] TestPolicyUpdateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy/ea11f0fd-fcbd-b75f-1c1b-68abfe0411de from=127.0.0.1:53258 latency=2.259812ms writer.go:29: 2021-01-29T19:32:07.308Z [INFO] TestPolicyUpdateCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:32:07.309Z [INFO] TestPolicyUpdateCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:07.310Z [DEBUG] TestPolicyUpdateCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:07.312Z [DEBUG] TestPolicyUpdateCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:07.313Z [DEBUG] TestPolicyUpdateCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:07.314Z [WARN] TestPolicyUpdateCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:07.315Z [DEBUG] TestPolicyUpdateCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:07.317Z [DEBUG] TestPolicyUpdateCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:07.318Z [DEBUG] TestPolicyUpdateCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:07.320Z [WARN] TestPolicyUpdateCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:07.323Z [INFO] TestPolicyUpdateCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:07.325Z [INFO] TestPolicyUpdateCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:07.327Z [INFO] TestPolicyUpdateCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:32:07.329Z [INFO] TestPolicyUpdateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:27364 network=tcp writer.go:29: 2021-01-29T19:32:07.330Z [INFO] TestPolicyUpdateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:27364 network=udp writer.go:29: 2021-01-29T19:32:07.332Z [INFO] TestPolicyUpdateCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:27365 network=tcp === CONT TestPolicyUpdateCommand writer.go:29: 2021-01-29T19:32:07.705Z [INFO] TestPolicyUpdateCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:07.707Z [INFO] TestPolicyUpdateCommand: Endpoints down --- PASS: TestPolicyUpdateCommand (0.79s) === CONT TestPolicyUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:07.834Z [INFO] TestPolicyUpdateCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:07.835Z [INFO] TestPolicyUpdateCommand_JSON: Endpoints down --- PASS: TestPolicyUpdateCommand_JSON (0.92s) PASS ok github.com/hashicorp/consul/command/acl/policy/update 1.105s ? github.com/hashicorp/consul/command/acl/role [no test files] === RUN TestRoleCreateCommand_noTabs === PAUSE TestRoleCreateCommand_noTabs === RUN TestRoleCreateCommand_Pretty === PAUSE TestRoleCreateCommand_Pretty === RUN TestRoleCreateCommand_JSON === PAUSE TestRoleCreateCommand_JSON === CONT TestRoleCreateCommand_noTabs === CONT TestRoleCreateCommand_JSON === CONT TestRoleCreateCommand_Pretty [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestRoleCreateCommand_noTabs (0.02s) === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:09.974Z [WARN] TestRoleCreateCommand_Pretty: bootstrap = true: do not enable unless necessary === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:09.975Z [WARN] TestRoleCreateCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:09.992Z [DEBUG] TestRoleCreateCommand_JSON.tlsutil: Update: version=1 === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:09.992Z [DEBUG] TestRoleCreateCommand_Pretty.tlsutil: Update: version=1 === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:09.994Z [DEBUG] TestRoleCreateCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:09.994Z [DEBUG] TestRoleCreateCommand_Pretty.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:09.999Z [INFO] TestRoleCreateCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a Address:127.0.0.1:12048}]" === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.004Z [INFO] TestRoleCreateCommand_Pretty.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4e6270ea-a1ca-727f-71ad-677cef400673 Address:127.0.0.1:12054}]" writer.go:29: 2021-01-29T19:32:10.006Z [INFO] TestRoleCreateCommand_Pretty.server.raft: entering follower state: follower="Node at 127.0.0.1:12054 [Follower]" leader= writer.go:29: 2021-01-29T19:32:10.007Z [INFO] TestRoleCreateCommand_Pretty.server.serf.wan: serf: EventMemberJoin: Node-4e6270ea-a1ca-727f-71ad-677cef400673.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:10.012Z [INFO] TestRoleCreateCommand_Pretty.server.serf.lan: serf: EventMemberJoin: Node-4e6270ea-a1ca-727f-71ad-677cef400673 127.0.0.1 === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.000Z [INFO] TestRoleCreateCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:12048 [Follower]" leader= writer.go:29: 2021-01-29T19:32:10.003Z [INFO] TestRoleCreateCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a.dc1 127.0.0.1 === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.020Z [INFO] TestRoleCreateCommand_Pretty.server: Handled event for server in area: event=member-join server=Node-4e6270ea-a1ca-727f-71ad-677cef400673.dc1 area=wan writer.go:29: 2021-01-29T19:32:10.020Z [INFO] TestRoleCreateCommand_Pretty.server: Adding LAN server: server="Node-4e6270ea-a1ca-727f-71ad-677cef400673 (Addr: tcp/127.0.0.1:12054) (DC: dc1)" writer.go:29: 2021-01-29T19:32:10.021Z [INFO] TestRoleCreateCommand_Pretty: Started DNS server: address=127.0.0.1:12049 network=udp writer.go:29: 2021-01-29T19:32:10.027Z [INFO] TestRoleCreateCommand_Pretty: Started DNS server: address=127.0.0.1:12049 network=tcp writer.go:29: 2021-01-29T19:32:10.030Z [INFO] TestRoleCreateCommand_Pretty: Started HTTP server: address=127.0.0.1:12050 network=tcp writer.go:29: 2021-01-29T19:32:10.031Z [INFO] TestRoleCreateCommand_Pretty: started state syncer === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.025Z [INFO] TestRoleCreateCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a 127.0.0.1 writer.go:29: 2021-01-29T19:32:10.041Z [INFO] TestRoleCreateCommand_JSON.server: Handled event for server in area: event=member-join server=Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a.dc1 area=wan writer.go:29: 2021-01-29T19:32:10.041Z [INFO] TestRoleCreateCommand_JSON: Started DNS server: address=127.0.0.1:12043 network=udp writer.go:29: 2021-01-29T19:32:10.041Z [INFO] TestRoleCreateCommand_JSON.server: Adding LAN server: server="Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a (Addr: tcp/127.0.0.1:12048) (DC: dc1)" writer.go:29: 2021-01-29T19:32:10.045Z [INFO] TestRoleCreateCommand_JSON: Started DNS server: address=127.0.0.1:12043 network=tcp writer.go:29: 2021-01-29T19:32:10.047Z [INFO] TestRoleCreateCommand_JSON: Started HTTP server: address=127.0.0.1:12044 network=tcp writer.go:29: 2021-01-29T19:32:10.048Z [INFO] TestRoleCreateCommand_JSON: started state syncer writer.go:29: 2021-01-29T19:32:10.065Z [WARN] TestRoleCreateCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:10.066Z [INFO] TestRoleCreateCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:12048 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:10.068Z [DEBUG] TestRoleCreateCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:10.069Z [DEBUG] TestRoleCreateCommand_JSON.server.raft: vote granted: from=8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a term=2 tally=1 writer.go:29: 2021-01-29T19:32:10.070Z [INFO] TestRoleCreateCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:10.071Z [INFO] TestRoleCreateCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:12048 [Leader]" === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.069Z [WARN] TestRoleCreateCommand_Pretty.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.072Z [INFO] TestRoleCreateCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:10.074Z [INFO] TestRoleCreateCommand_JSON.server: New leader elected: payload=Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.072Z [INFO] TestRoleCreateCommand_Pretty.server.raft: entering candidate state: node="Node at 127.0.0.1:12054 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:10.070Z [DEBUG] TestRoleCreateCommand_Pretty.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.075Z [INFO] TestRoleCreateCommand_JSON.server: initializing acls === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.077Z [DEBUG] TestRoleCreateCommand_Pretty.server.raft: votes: needed=1 === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.078Z [INFO] TestRoleCreateCommand_JSON.server: Created ACL 'global-management' policy === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.078Z [DEBUG] TestRoleCreateCommand_Pretty.server.raft: vote granted: from=4e6270ea-a1ca-727f-71ad-677cef400673 term=2 tally=1 === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.079Z [WARN] TestRoleCreateCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.079Z [INFO] TestRoleCreateCommand_Pretty.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:10.080Z [INFO] TestRoleCreateCommand_Pretty.server.raft: entering leader state: leader="Node at 127.0.0.1:12054 [Leader]" === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.082Z [INFO] TestRoleCreateCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.081Z [INFO] TestRoleCreateCommand_Pretty.server: cluster leadership acquired === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.084Z [INFO] TestRoleCreateCommand_JSON.server: Created ACL anonymous token from configuration === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.081Z [INFO] TestRoleCreateCommand_Pretty.server: New leader elected: payload=Node-4e6270ea-a1ca-727f-71ad-677cef400673 === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.085Z [INFO] TestRoleCreateCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:10.086Z [INFO] TestRoleCreateCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:10.087Z [INFO] TestRoleCreateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a writer.go:29: 2021-01-29T19:32:10.089Z [INFO] TestRoleCreateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a.dc1 writer.go:29: 2021-01-29T19:32:10.090Z [INFO] TestRoleCreateCommand_JSON.server: Updating LAN server: server="Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a (Addr: tcp/127.0.0.1:12048) (DC: dc1)" writer.go:29: 2021-01-29T19:32:10.090Z [INFO] TestRoleCreateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a.dc1 area=wan writer.go:29: 2021-01-29T19:32:10.093Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.082Z [INFO] TestRoleCreateCommand_Pretty.server: initializing acls writer.go:29: 2021-01-29T19:32:10.106Z [INFO] TestRoleCreateCommand_Pretty.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:10.109Z [WARN] TestRoleCreateCommand_Pretty.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:10.112Z [INFO] TestRoleCreateCommand_Pretty.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:10.115Z [INFO] TestRoleCreateCommand_Pretty.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:10.117Z [INFO] TestRoleCreateCommand_Pretty.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:10.119Z [INFO] TestRoleCreateCommand_Pretty.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:10.121Z [INFO] TestRoleCreateCommand_Pretty.server.serf.lan: serf: EventMemberUpdate: Node-4e6270ea-a1ca-727f-71ad-677cef400673 writer.go:29: 2021-01-29T19:32:10.124Z [INFO] TestRoleCreateCommand_Pretty.server: Updating LAN server: server="Node-4e6270ea-a1ca-727f-71ad-677cef400673 (Addr: tcp/127.0.0.1:12054) (DC: dc1)" writer.go:29: 2021-01-29T19:32:10.124Z [INFO] TestRoleCreateCommand_Pretty.server.serf.wan: serf: EventMemberUpdate: Node-4e6270ea-a1ca-727f-71ad-677cef400673.dc1 === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.123Z [INFO] TestRoleCreateCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:10.126Z [INFO] TestRoleCreateCommand_JSON: Synced node info === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.126Z [INFO] TestRoleCreateCommand_Pretty.server: Handled event for server in area: event=member-update server=Node-4e6270ea-a1ca-727f-71ad-677cef400673.dc1 area=wan === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.127Z [DEBUG] TestRoleCreateCommand_JSON: Node info in sync writer.go:29: 2021-01-29T19:32:10.127Z [INFO] TestRoleCreateCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.127Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.129Z [DEBUG] TestRoleCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a writer.go:29: 2021-01-29T19:32:10.130Z [INFO] TestRoleCreateCommand_JSON.server: member joined, marking health alive: member=Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a writer.go:29: 2021-01-29T19:32:10.132Z [DEBUG] TestRoleCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-8f6dc4bb-9a36-3c2b-c0d2-1d79faf9a05a === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.134Z [INFO] TestRoleCreateCommand_Pretty.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:10.135Z [INFO] TestRoleCreateCommand_Pretty.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:10.136Z [DEBUG] TestRoleCreateCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-4e6270ea-a1ca-727f-71ad-677cef400673 writer.go:29: 2021-01-29T19:32:10.137Z [INFO] TestRoleCreateCommand_Pretty.server: member joined, marking health alive: member=Node-4e6270ea-a1ca-727f-71ad-677cef400673 writer.go:29: 2021-01-29T19:32:10.139Z [DEBUG] TestRoleCreateCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-4e6270ea-a1ca-727f-71ad-677cef400673 writer.go:29: 2021-01-29T19:32:10.174Z [DEBUG] TestRoleCreateCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:40830 latency=9.284444ms writer.go:29: 2021-01-29T19:32:10.188Z [DEBUG] TestRoleCreateCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:40832 latency=4.520591ms writer.go:29: 2021-01-29T19:32:10.207Z [DEBUG] TestRoleCreateCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:40834 latency=3.989571ms writer.go:29: 2021-01-29T19:32:10.214Z [DEBUG] TestRoleCreateCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:40836 latency=2.587623ms writer.go:29: 2021-01-29T19:32:10.225Z [DEBUG] TestRoleCreateCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:40838 latency=5.548703ms writer.go:29: 2021-01-29T19:32:10.235Z [DEBUG] TestRoleCreateCommand_Pretty: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:10.238Z [INFO] TestRoleCreateCommand_Pretty: Synced node info writer.go:29: 2021-01-29T19:32:10.239Z [INFO] TestRoleCreateCommand_Pretty: Requesting shutdown writer.go:29: 2021-01-29T19:32:10.242Z [INFO] TestRoleCreateCommand_Pretty.server: shutting down server writer.go:29: 2021-01-29T19:32:10.243Z [DEBUG] TestRoleCreateCommand_Pretty.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:10.244Z [DEBUG] TestRoleCreateCommand_Pretty.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:10.245Z [DEBUG] TestRoleCreateCommand_Pretty.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:10.246Z [WARN] TestRoleCreateCommand_Pretty.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:10.244Z [DEBUG] TestRoleCreateCommand_Pretty.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:10.245Z [DEBUG] TestRoleCreateCommand_Pretty.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:10.246Z [DEBUG] TestRoleCreateCommand_Pretty.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:10.257Z [WARN] TestRoleCreateCommand_Pretty.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:10.260Z [INFO] TestRoleCreateCommand_Pretty.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:10.261Z [INFO] TestRoleCreateCommand_Pretty: consul server down writer.go:29: 2021-01-29T19:32:10.262Z [INFO] TestRoleCreateCommand_Pretty: shutdown complete writer.go:29: 2021-01-29T19:32:10.263Z [INFO] TestRoleCreateCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:12049 network=tcp writer.go:29: 2021-01-29T19:32:10.264Z [INFO] TestRoleCreateCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:12049 network=udp writer.go:29: 2021-01-29T19:32:10.266Z [INFO] TestRoleCreateCommand_Pretty: Stopping server: protocol=HTTP address=127.0.0.1:12050 network=tcp === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:10.493Z [DEBUG] TestRoleCreateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:38024 latency=4.353767ms writer.go:29: 2021-01-29T19:32:10.497Z [DEBUG] TestRoleCreateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:38026 latency=1.573167ms writer.go:29: 2021-01-29T19:32:10.499Z [INFO] TestRoleCreateCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:32:10.500Z [INFO] TestRoleCreateCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:10.501Z [DEBUG] TestRoleCreateCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:10.502Z [DEBUG] TestRoleCreateCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:10.503Z [DEBUG] TestRoleCreateCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:10.502Z [DEBUG] TestRoleCreateCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:10.504Z [DEBUG] TestRoleCreateCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:10.504Z [WARN] TestRoleCreateCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:10.504Z [DEBUG] TestRoleCreateCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:10.509Z [WARN] TestRoleCreateCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:10.510Z [INFO] TestRoleCreateCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:10.512Z [INFO] TestRoleCreateCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:10.513Z [INFO] TestRoleCreateCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:32:10.513Z [INFO] TestRoleCreateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:12043 network=tcp writer.go:29: 2021-01-29T19:32:10.514Z [INFO] TestRoleCreateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:12043 network=udp writer.go:29: 2021-01-29T19:32:10.516Z [INFO] TestRoleCreateCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:12044 network=tcp === CONT TestRoleCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:10.769Z [INFO] TestRoleCreateCommand_Pretty: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:10.771Z [INFO] TestRoleCreateCommand_Pretty: Endpoints down --- PASS: TestRoleCreateCommand_Pretty (0.84s) === CONT TestRoleCreateCommand_JSON writer.go:29: 2021-01-29T19:32:11.018Z [INFO] TestRoleCreateCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:11.020Z [INFO] TestRoleCreateCommand_JSON: Endpoints down --- PASS: TestRoleCreateCommand_JSON (1.09s) PASS ok github.com/hashicorp/consul/command/acl/role/create 1.257s === RUN TestRoleDeleteCommand_noTabs === PAUSE TestRoleDeleteCommand_noTabs === RUN TestRoleDeleteCommand === PAUSE TestRoleDeleteCommand === CONT TestRoleDeleteCommand_noTabs === CONT TestRoleDeleteCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestRoleDeleteCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestRoleDeleteCommand writer.go:29: 2021-01-29T19:32:12.362Z [WARN] TestRoleDeleteCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:12.364Z [DEBUG] TestRoleDeleteCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:12.368Z [DEBUG] TestRoleDeleteCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:12.382Z [INFO] TestRoleDeleteCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a Address:127.0.0.1:14090}]" writer.go:29: 2021-01-29T19:32:12.385Z [INFO] TestRoleDeleteCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:14090 [Follower]" leader= writer.go:29: 2021-01-29T19:32:12.386Z [INFO] TestRoleDeleteCommand.server.serf.wan: serf: EventMemberJoin: Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:12.391Z [INFO] TestRoleDeleteCommand.server.serf.lan: serf: EventMemberJoin: Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a 127.0.0.1 writer.go:29: 2021-01-29T19:32:12.401Z [INFO] TestRoleDeleteCommand.server: Adding LAN server: server="Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a (Addr: tcp/127.0.0.1:14090) (DC: dc1)" writer.go:29: 2021-01-29T19:32:12.401Z [INFO] TestRoleDeleteCommand.server: Handled event for server in area: event=member-join server=Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a.dc1 area=wan writer.go:29: 2021-01-29T19:32:12.403Z [INFO] TestRoleDeleteCommand: Started DNS server: address=127.0.0.1:14085 network=tcp writer.go:29: 2021-01-29T19:32:12.404Z [INFO] TestRoleDeleteCommand: Started DNS server: address=127.0.0.1:14085 network=udp writer.go:29: 2021-01-29T19:32:12.407Z [INFO] TestRoleDeleteCommand: Started HTTP server: address=127.0.0.1:14086 network=tcp writer.go:29: 2021-01-29T19:32:12.409Z [INFO] TestRoleDeleteCommand: started state syncer writer.go:29: 2021-01-29T19:32:12.450Z [WARN] TestRoleDeleteCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:12.453Z [INFO] TestRoleDeleteCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:14090 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:12.452Z [DEBUG] TestRoleDeleteCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:12.456Z [DEBUG] TestRoleDeleteCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:12.457Z [DEBUG] TestRoleDeleteCommand.server.raft: vote granted: from=f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a term=2 tally=1 writer.go:29: 2021-01-29T19:32:12.458Z [INFO] TestRoleDeleteCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:12.459Z [INFO] TestRoleDeleteCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:14090 [Leader]" writer.go:29: 2021-01-29T19:32:12.460Z [INFO] TestRoleDeleteCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:12.461Z [INFO] TestRoleDeleteCommand.server: initializing acls writer.go:29: 2021-01-29T19:32:12.461Z [INFO] TestRoleDeleteCommand.server: New leader elected: payload=Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a writer.go:29: 2021-01-29T19:32:12.468Z [INFO] TestRoleDeleteCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:12.469Z [WARN] TestRoleDeleteCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:12.472Z [INFO] TestRoleDeleteCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:12.475Z [INFO] TestRoleDeleteCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:12.477Z [INFO] TestRoleDeleteCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:12.478Z [INFO] TestRoleDeleteCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:12.480Z [INFO] TestRoleDeleteCommand.server.serf.lan: serf: EventMemberUpdate: Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a writer.go:29: 2021-01-29T19:32:12.483Z [INFO] TestRoleDeleteCommand.server.serf.wan: serf: EventMemberUpdate: Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a.dc1 writer.go:29: 2021-01-29T19:32:12.483Z [INFO] TestRoleDeleteCommand.server: Updating LAN server: server="Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a (Addr: tcp/127.0.0.1:14090) (DC: dc1)" writer.go:29: 2021-01-29T19:32:12.485Z [INFO] TestRoleDeleteCommand.server: Handled event for server in area: event=member-update server=Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a.dc1 area=wan writer.go:29: 2021-01-29T19:32:12.487Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:12.516Z [INFO] TestRoleDeleteCommand: Synced node info writer.go:29: 2021-01-29T19:32:12.519Z [INFO] TestRoleDeleteCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:12.520Z [INFO] TestRoleDeleteCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:12.522Z [DEBUG] TestRoleDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a writer.go:29: 2021-01-29T19:32:12.523Z [INFO] TestRoleDeleteCommand.server: member joined, marking health alive: member=Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a writer.go:29: 2021-01-29T19:32:12.526Z [DEBUG] TestRoleDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-f19a19bd-ddf5-6e6d-9347-74ddb6fdd90a === RUN TestRoleDeleteCommand/id_or_name_required === RUN TestRoleDeleteCommand/delete_works === CONT TestRoleDeleteCommand writer.go:29: 2021-01-29T19:32:12.652Z [DEBUG] TestRoleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:51156 latency=21.578756ms writer.go:29: 2021-01-29T19:32:12.685Z [DEBUG] TestRoleDeleteCommand.http: Request finished: method=DELETE url=/v1/acl/role/f0c6d62b-7260-6c74-caa1-d3ba00e6eed3 from=127.0.0.1:51158 latency=4.989879ms writer.go:29: 2021-01-29T19:32:12.696Z [DEBUG] TestRoleDeleteCommand.http: Request finished: method=GET url=/v1/acl/role/f0c6d62b-7260-6c74-caa1-d3ba00e6eed3 from=127.0.0.1:51156 latency=89.266µs === RUN TestRoleDeleteCommand/delete_works_via_prefixes === CONT TestRoleDeleteCommand writer.go:29: 2021-01-29T19:32:12.703Z [DEBUG] TestRoleDeleteCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:51156 latency=987.165µs writer.go:29: 2021-01-29T19:32:12.711Z [DEBUG] TestRoleDeleteCommand.http: Request finished: method=GET url=/v1/acl/roles from=127.0.0.1:51160 latency=376.164µs writer.go:29: 2021-01-29T19:32:12.718Z [DEBUG] TestRoleDeleteCommand.http: Request finished: method=DELETE url=/v1/acl/role/a5529858-8d40-3b4e-316b-d9a3033ced55 from=127.0.0.1:51160 latency=3.85352ms writer.go:29: 2021-01-29T19:32:12.725Z [DEBUG] TestRoleDeleteCommand.http: Request finished: method=GET url=/v1/acl/role/a5529858-8d40-3b4e-316b-d9a3033ced55 from=127.0.0.1:51156 latency=72.252µs writer.go:29: 2021-01-29T19:32:12.729Z [INFO] TestRoleDeleteCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:12.732Z [INFO] TestRoleDeleteCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:12.734Z [DEBUG] TestRoleDeleteCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:12.736Z [DEBUG] TestRoleDeleteCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:12.738Z [DEBUG] TestRoleDeleteCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:12.736Z [DEBUG] TestRoleDeleteCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:12.738Z [DEBUG] TestRoleDeleteCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:12.740Z [WARN] TestRoleDeleteCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:12.740Z [DEBUG] TestRoleDeleteCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:12.747Z [WARN] TestRoleDeleteCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:12.752Z [INFO] TestRoleDeleteCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:12.753Z [INFO] TestRoleDeleteCommand: consul server down writer.go:29: 2021-01-29T19:32:12.756Z [INFO] TestRoleDeleteCommand: shutdown complete writer.go:29: 2021-01-29T19:32:12.758Z [INFO] TestRoleDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:14085 network=tcp writer.go:29: 2021-01-29T19:32:12.759Z [INFO] TestRoleDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:14085 network=udp writer.go:29: 2021-01-29T19:32:12.761Z [INFO] TestRoleDeleteCommand: Stopping server: protocol=HTTP address=127.0.0.1:14086 network=tcp writer.go:29: 2021-01-29T19:32:13.262Z [INFO] TestRoleDeleteCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:13.266Z [INFO] TestRoleDeleteCommand: Endpoints down --- PASS: TestRoleDeleteCommand (0.96s) --- PASS: TestRoleDeleteCommand/id_or_name_required (0.00s) --- PASS: TestRoleDeleteCommand/delete_works (0.08s) --- PASS: TestRoleDeleteCommand/delete_works_via_prefixes (0.03s) PASS ok github.com/hashicorp/consul/command/acl/role/delete 1.255s === RUN TestRoleListCommand_noTabs === PAUSE TestRoleListCommand_noTabs === RUN TestRoleListCommand === PAUSE TestRoleListCommand === RUN TestRoleListCommand_JSON === PAUSE TestRoleListCommand_JSON === CONT TestRoleListCommand_noTabs === CONT TestRoleListCommand_JSON === CONT TestRoleListCommand --- PASS: TestRoleListCommand_noTabs (0.00s) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:13.924Z [WARN] TestRoleListCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:13.925Z [DEBUG] TestRoleListCommand_JSON.tlsutil: Update: version=1 === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:13.926Z [WARN] TestRoleListCommand: bootstrap = true: do not enable unless necessary === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:13.926Z [DEBUG] TestRoleListCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:13.927Z [DEBUG] TestRoleListCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:13.929Z [DEBUG] TestRoleListCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:13.930Z [INFO] TestRoleListCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:34ccbb62-b778-c21b-9747-85694df99c0c Address:127.0.0.1:21237}]" writer.go:29: 2021-01-29T19:32:13.931Z [INFO] TestRoleListCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-34ccbb62-b778-c21b-9747-85694df99c0c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:13.931Z [INFO] TestRoleListCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:21237 [Follower]" leader= writer.go:29: 2021-01-29T19:32:13.933Z [INFO] TestRoleListCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-34ccbb62-b778-c21b-9747-85694df99c0c 127.0.0.1 writer.go:29: 2021-01-29T19:32:13.941Z [INFO] TestRoleListCommand_JSON.server: Adding LAN server: server="Node-34ccbb62-b778-c21b-9747-85694df99c0c (Addr: tcp/127.0.0.1:21237) (DC: dc1)" === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:13.942Z [INFO] TestRoleListCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2cf41022-246d-4c8e-00c1-0026fa4bcf5d Address:127.0.0.1:21243}]" === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:13.944Z [INFO] TestRoleListCommand_JSON.server: Handled event for server in area: event=member-join server=Node-34ccbb62-b778-c21b-9747-85694df99c0c.dc1 area=wan writer.go:29: 2021-01-29T19:32:13.945Z [INFO] TestRoleListCommand_JSON: Started DNS server: address=127.0.0.1:21232 network=udp === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:13.947Z [INFO] TestRoleListCommand.server.serf.wan: serf: EventMemberJoin: Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d.dc1 127.0.0.1 === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:13.948Z [INFO] TestRoleListCommand_JSON: Started DNS server: address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:32:13.953Z [INFO] TestRoleListCommand_JSON: Started HTTP server: address=127.0.0.1:21233 network=tcp === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:13.947Z [INFO] TestRoleListCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:21243 [Follower]" leader= === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:13.954Z [INFO] TestRoleListCommand_JSON: started state syncer === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:13.952Z [INFO] TestRoleListCommand.server.serf.lan: serf: EventMemberJoin: Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d 127.0.0.1 writer.go:29: 2021-01-29T19:32:13.974Z [INFO] TestRoleListCommand.server: Handled event for server in area: event=member-join server=Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d.dc1 area=wan writer.go:29: 2021-01-29T19:32:13.976Z [INFO] TestRoleListCommand: Started DNS server: address=127.0.0.1:21238 network=udp writer.go:29: 2021-01-29T19:32:13.978Z [INFO] TestRoleListCommand: Started DNS server: address=127.0.0.1:21238 network=tcp writer.go:29: 2021-01-29T19:32:13.978Z [INFO] TestRoleListCommand.server: Adding LAN server: server="Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d (Addr: tcp/127.0.0.1:21243) (DC: dc1)" writer.go:29: 2021-01-29T19:32:13.979Z [INFO] TestRoleListCommand: Started HTTP server: address=127.0.0.1:21239 network=tcp writer.go:29: 2021-01-29T19:32:13.982Z [INFO] TestRoleListCommand: started state syncer === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:13.995Z [DEBUG] TestRoleListCommand_JSON.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:13.997Z [WARN] TestRoleListCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:13.998Z [INFO] TestRoleListCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:21237 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:14.001Z [DEBUG] TestRoleListCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:14.002Z [DEBUG] TestRoleListCommand_JSON.server.raft: vote granted: from=34ccbb62-b778-c21b-9747-85694df99c0c term=2 tally=1 writer.go:29: 2021-01-29T19:32:14.004Z [INFO] TestRoleListCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:14.006Z [INFO] TestRoleListCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:21237 [Leader]" writer.go:29: 2021-01-29T19:32:14.008Z [INFO] TestRoleListCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:14.009Z [INFO] TestRoleListCommand_JSON.server: New leader elected: payload=Node-34ccbb62-b778-c21b-9747-85694df99c0c writer.go:29: 2021-01-29T19:32:14.010Z [INFO] TestRoleListCommand_JSON.server: initializing acls === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.011Z [DEBUG] TestRoleListCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:14.013Z [WARN] TestRoleListCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:14.014Z [INFO] TestRoleListCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:21243 [Candidate]" term=2 === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.015Z [INFO] TestRoleListCommand_JSON.server: Created ACL 'global-management' policy === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.017Z [DEBUG] TestRoleListCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:14.019Z [DEBUG] TestRoleListCommand.server.raft: vote granted: from=2cf41022-246d-4c8e-00c1-0026fa4bcf5d term=2 tally=1 === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.019Z [WARN] TestRoleListCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.020Z [INFO] TestRoleListCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:14.022Z [INFO] TestRoleListCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:21243 [Leader]" === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.022Z [INFO] TestRoleListCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.024Z [INFO] TestRoleListCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:14.025Z [INFO] TestRoleListCommand.server: New leader elected: payload=Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.026Z [INFO] TestRoleListCommand_JSON.server: Created ACL anonymous token from configuration === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.025Z [INFO] TestRoleListCommand.server: initializing acls === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.027Z [INFO] TestRoleListCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.030Z [INFO] TestRoleListCommand.server: Created ACL 'global-management' policy === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.030Z [INFO] TestRoleListCommand_JSON.leader: started routine: routine="acl token reaping" === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.032Z [WARN] TestRoleListCommand.server: Configuring a non-UUID master token is deprecated === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.032Z [INFO] TestRoleListCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-34ccbb62-b778-c21b-9747-85694df99c0c === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.034Z [INFO] TestRoleListCommand.server: Bootstrapped ACL master token from configuration === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.034Z [INFO] TestRoleListCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-34ccbb62-b778-c21b-9747-85694df99c0c.dc1 === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.036Z [INFO] TestRoleListCommand.server: Created ACL anonymous token from configuration === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.034Z [INFO] TestRoleListCommand_JSON.server: Updating LAN server: server="Node-34ccbb62-b778-c21b-9747-85694df99c0c (Addr: tcp/127.0.0.1:21237) (DC: dc1)" === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.038Z [INFO] TestRoleListCommand.leader: started routine: routine="legacy ACL token upgrade" === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.036Z [INFO] TestRoleListCommand_JSON.server: Handled event for server in area: event=member-update server=Node-34ccbb62-b778-c21b-9747-85694df99c0c.dc1 area=wan === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.039Z [INFO] TestRoleListCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:14.040Z [INFO] TestRoleListCommand.server.serf.lan: serf: EventMemberUpdate: Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.038Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.041Z [INFO] TestRoleListCommand.server.serf.wan: serf: EventMemberUpdate: Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d.dc1 writer.go:29: 2021-01-29T19:32:14.041Z [INFO] TestRoleListCommand.server: Updating LAN server: server="Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d (Addr: tcp/127.0.0.1:21243) (DC: dc1)" === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.043Z [INFO] TestRoleListCommand_JSON: Synced node info === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.042Z [INFO] TestRoleListCommand.server: Handled event for server in area: event=member-update server=Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d.dc1 area=wan writer.go:29: 2021-01-29T19:32:14.043Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:14.079Z [INFO] TestRoleListCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.081Z [INFO] TestRoleListCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:14.083Z [INFO] TestRoleListCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.082Z [INFO] TestRoleListCommand.leader: started routine: routine="CA root pruning" === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.085Z [DEBUG] TestRoleListCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-34ccbb62-b778-c21b-9747-85694df99c0c === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.085Z [DEBUG] TestRoleListCommand.server: Skipping self join check for node since the cluster is too small: node=Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.087Z [INFO] TestRoleListCommand_JSON.server: member joined, marking health alive: member=Node-34ccbb62-b778-c21b-9747-85694df99c0c === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.088Z [INFO] TestRoleListCommand.server: member joined, marking health alive: member=Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.094Z [DEBUG] TestRoleListCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-34ccbb62-b778-c21b-9747-85694df99c0c === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.094Z [DEBUG] TestRoleListCommand.server: Skipping self join check for node since the cluster is too small: node=Node-2cf41022-246d-4c8e-00c1-0026fa4bcf5d === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.144Z [DEBUG] TestRoleListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:41988 latency=4.900486ms writer.go:29: 2021-01-29T19:32:14.150Z [DEBUG] TestRoleListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:41988 latency=2.94609ms === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.170Z [DEBUG] TestRoleListCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:14.179Z [INFO] TestRoleListCommand: Synced node info === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.189Z [DEBUG] TestRoleListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:41988 latency=31.733206ms writer.go:29: 2021-01-29T19:32:14.201Z [DEBUG] TestRoleListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:41988 latency=2.416333ms writer.go:29: 2021-01-29T19:32:14.206Z [DEBUG] TestRoleListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:41988 latency=1.134016ms writer.go:29: 2021-01-29T19:32:14.212Z [DEBUG] TestRoleListCommand_JSON.http: Request finished: method=GET url=/v1/acl/roles from=127.0.0.1:41990 latency=203.677µs writer.go:29: 2021-01-29T19:32:14.224Z [INFO] TestRoleListCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:32:14.225Z [INFO] TestRoleListCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:14.229Z [DEBUG] TestRoleListCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:14.230Z [DEBUG] TestRoleListCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:14.231Z [DEBUG] TestRoleListCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:14.232Z [WARN] TestRoleListCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:14.230Z [DEBUG] TestRoleListCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:14.231Z [DEBUG] TestRoleListCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:14.232Z [DEBUG] TestRoleListCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:14.234Z [WARN] TestRoleListCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:14.239Z [INFO] TestRoleListCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:14.239Z [INFO] TestRoleListCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:14.241Z [INFO] TestRoleListCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:32:14.242Z [INFO] TestRoleListCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:32:14.244Z [INFO] TestRoleListCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:21232 network=udp writer.go:29: 2021-01-29T19:32:14.245Z [INFO] TestRoleListCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:21233 network=tcp === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.327Z [DEBUG] TestRoleListCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:47568 latency=3.787287ms writer.go:29: 2021-01-29T19:32:14.331Z [DEBUG] TestRoleListCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:47568 latency=1.109745ms writer.go:29: 2021-01-29T19:32:14.337Z [DEBUG] TestRoleListCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:47568 latency=826.304µs writer.go:29: 2021-01-29T19:32:14.340Z [DEBUG] TestRoleListCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:47568 latency=693.459µs writer.go:29: 2021-01-29T19:32:14.343Z [DEBUG] TestRoleListCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:47568 latency=596.329µs writer.go:29: 2021-01-29T19:32:14.348Z [DEBUG] TestRoleListCommand.http: Request finished: method=GET url=/v1/acl/roles from=127.0.0.1:47570 latency=129.45µs writer.go:29: 2021-01-29T19:32:14.370Z [INFO] TestRoleListCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:14.371Z [INFO] TestRoleListCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:14.373Z [DEBUG] TestRoleListCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:14.375Z [DEBUG] TestRoleListCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:14.376Z [DEBUG] TestRoleListCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:14.375Z [DEBUG] TestRoleListCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:14.376Z [DEBUG] TestRoleListCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:14.378Z [WARN] TestRoleListCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:14.378Z [DEBUG] TestRoleListCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:14.385Z [WARN] TestRoleListCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:14.389Z [INFO] TestRoleListCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:14.389Z [INFO] TestRoleListCommand: consul server down writer.go:29: 2021-01-29T19:32:14.391Z [INFO] TestRoleListCommand: shutdown complete writer.go:29: 2021-01-29T19:32:14.393Z [INFO] TestRoleListCommand: Stopping server: protocol=DNS address=127.0.0.1:21238 network=tcp writer.go:29: 2021-01-29T19:32:14.395Z [INFO] TestRoleListCommand: Stopping server: protocol=DNS address=127.0.0.1:21238 network=udp writer.go:29: 2021-01-29T19:32:14.397Z [INFO] TestRoleListCommand: Stopping server: protocol=HTTP address=127.0.0.1:21239 network=tcp === CONT TestRoleListCommand_JSON writer.go:29: 2021-01-29T19:32:14.746Z [INFO] TestRoleListCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:14.748Z [INFO] TestRoleListCommand_JSON: Endpoints down --- PASS: TestRoleListCommand_JSON (0.86s) === CONT TestRoleListCommand writer.go:29: 2021-01-29T19:32:14.899Z [INFO] TestRoleListCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:14.900Z [INFO] TestRoleListCommand: Endpoints down --- PASS: TestRoleListCommand (1.01s) PASS ok github.com/hashicorp/consul/command/acl/role/list 1.137s === RUN TestRoleReadCommand_noTabs === PAUSE TestRoleReadCommand_noTabs === RUN TestRoleReadCommand === PAUSE TestRoleReadCommand === RUN TestRoleReadCommand_JSON === PAUSE TestRoleReadCommand_JSON === CONT TestRoleReadCommand_noTabs === CONT TestRoleReadCommand_JSON === CONT TestRoleReadCommand --- PASS: TestRoleReadCommand_noTabs (0.00s) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.542Z [WARN] TestRoleReadCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:17.544Z [DEBUG] TestRoleReadCommand_JSON.tlsutil: Update: version=1 === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.547Z [WARN] TestRoleReadCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:17.550Z [DEBUG] TestRoleReadCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:17.555Z [DEBUG] TestRoleReadCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.547Z [DEBUG] TestRoleReadCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.569Z [INFO] TestRoleReadCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b693acc6-5758-204c-8bfe-fd6c77b8ab62 Address:127.0.0.1:31453}]" writer.go:29: 2021-01-29T19:32:17.570Z [INFO] TestRoleReadCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:31453 [Follower]" leader= === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.573Z [INFO] TestRoleReadCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3e568802-d10b-3fbf-43c8-ebae90d3fec6 Address:127.0.0.1:31447}]" writer.go:29: 2021-01-29T19:32:17.575Z [INFO] TestRoleReadCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:31447 [Follower]" leader= writer.go:29: 2021-01-29T19:32:17.576Z [INFO] TestRoleReadCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6.dc1 127.0.0.1 === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.575Z [INFO] TestRoleReadCommand.server.serf.wan: serf: EventMemberJoin: Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62.dc1 127.0.0.1 === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.579Z [INFO] TestRoleReadCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6 127.0.0.1 writer.go:29: 2021-01-29T19:32:17.582Z [INFO] TestRoleReadCommand_JSON.server: Handled event for server in area: event=member-join server=Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6.dc1 area=wan writer.go:29: 2021-01-29T19:32:17.582Z [INFO] TestRoleReadCommand_JSON.server: Adding LAN server: server="Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6 (Addr: tcp/127.0.0.1:31447) (DC: dc1)" writer.go:29: 2021-01-29T19:32:17.582Z [INFO] TestRoleReadCommand_JSON: Started DNS server: address=127.0.0.1:31442 network=udp writer.go:29: 2021-01-29T19:32:17.585Z [INFO] TestRoleReadCommand_JSON: Started DNS server: address=127.0.0.1:31442 network=tcp === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.585Z [INFO] TestRoleReadCommand.server.serf.lan: serf: EventMemberJoin: Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62 127.0.0.1 === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.587Z [INFO] TestRoleReadCommand_JSON: Started HTTP server: address=127.0.0.1:31443 network=tcp writer.go:29: 2021-01-29T19:32:17.589Z [INFO] TestRoleReadCommand_JSON: started state syncer === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.587Z [INFO] TestRoleReadCommand: Started DNS server: address=127.0.0.1:31448 network=udp writer.go:29: 2021-01-29T19:32:17.593Z [INFO] TestRoleReadCommand: Started DNS server: address=127.0.0.1:31448 network=tcp writer.go:29: 2021-01-29T19:32:17.596Z [INFO] TestRoleReadCommand: Started HTTP server: address=127.0.0.1:31449 network=tcp writer.go:29: 2021-01-29T19:32:17.597Z [INFO] TestRoleReadCommand: started state syncer writer.go:29: 2021-01-29T19:32:17.588Z [INFO] TestRoleReadCommand.server: Adding LAN server: server="Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62 (Addr: tcp/127.0.0.1:31453) (DC: dc1)" writer.go:29: 2021-01-29T19:32:17.588Z [INFO] TestRoleReadCommand.server: Handled event for server in area: event=member-join server=Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62.dc1 area=wan === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.633Z [DEBUG] TestRoleReadCommand_JSON.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.635Z [WARN] TestRoleReadCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:17.636Z [INFO] TestRoleReadCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:31453 [Candidate]" term=2 === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.637Z [WARN] TestRoleReadCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.638Z [DEBUG] TestRoleReadCommand.server.raft: votes: needed=1 === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.638Z [INFO] TestRoleReadCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:31447 [Candidate]" term=2 === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.639Z [DEBUG] TestRoleReadCommand.server.raft: vote granted: from=b693acc6-5758-204c-8bfe-fd6c77b8ab62 term=2 tally=1 writer.go:29: 2021-01-29T19:32:17.640Z [INFO] TestRoleReadCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:17.639Z [DEBUG] TestRoleReadCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:32:17.641Z [INFO] TestRoleReadCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:31453 [Leader]" writer.go:29: 2021-01-29T19:32:17.643Z [INFO] TestRoleReadCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:17.645Z [INFO] TestRoleReadCommand.server: New leader elected: payload=Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62 === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.640Z [DEBUG] TestRoleReadCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:17.650Z [DEBUG] TestRoleReadCommand_JSON.server.raft: vote granted: from=3e568802-d10b-3fbf-43c8-ebae90d3fec6 term=2 tally=1 writer.go:29: 2021-01-29T19:32:17.652Z [INFO] TestRoleReadCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:17.653Z [INFO] TestRoleReadCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:31447 [Leader]" writer.go:29: 2021-01-29T19:32:17.655Z [INFO] TestRoleReadCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:17.656Z [INFO] TestRoleReadCommand_JSON.server: New leader elected: payload=Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6 === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.662Z [INFO] TestRoleReadCommand.server: initializing acls writer.go:29: 2021-01-29T19:32:17.664Z [INFO] TestRoleReadCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:17.665Z [WARN] TestRoleReadCommand.server: Configuring a non-UUID master token is deprecated === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.659Z [INFO] TestRoleReadCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:32:17.669Z [INFO] TestRoleReadCommand_JSON.server: Created ACL 'global-management' policy === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.667Z [INFO] TestRoleReadCommand.server: Bootstrapped ACL master token from configuration === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.670Z [WARN] TestRoleReadCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:17.672Z [INFO] TestRoleReadCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.672Z [INFO] TestRoleReadCommand.server: Created ACL anonymous token from configuration === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.673Z [INFO] TestRoleReadCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:17.674Z [INFO] TestRoleReadCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:17.675Z [INFO] TestRoleReadCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:17.677Z [INFO] TestRoleReadCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6 writer.go:29: 2021-01-29T19:32:17.678Z [INFO] TestRoleReadCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6.dc1 === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.674Z [INFO] TestRoleReadCommand.leader: started routine: routine="legacy ACL token upgrade" === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.680Z [INFO] TestRoleReadCommand_JSON.server: Updating LAN server: server="Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6 (Addr: tcp/127.0.0.1:31447) (DC: dc1)" writer.go:29: 2021-01-29T19:32:17.680Z [INFO] TestRoleReadCommand_JSON.server: Handled event for server in area: event=member-update server=Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6.dc1 area=wan === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.680Z [INFO] TestRoleReadCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:17.682Z [INFO] TestRoleReadCommand.server.serf.lan: serf: EventMemberUpdate: Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62 writer.go:29: 2021-01-29T19:32:17.684Z [INFO] TestRoleReadCommand: Synced node info writer.go:29: 2021-01-29T19:32:17.685Z [INFO] TestRoleReadCommand.server: Updating LAN server: server="Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62 (Addr: tcp/127.0.0.1:31453) (DC: dc1)" writer.go:29: 2021-01-29T19:32:17.685Z [INFO] TestRoleReadCommand.server.serf.wan: serf: EventMemberUpdate: Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62.dc1 writer.go:29: 2021-01-29T19:32:17.688Z [INFO] TestRoleReadCommand.server: Handled event for server in area: event=member-update server=Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62.dc1 area=wan === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.691Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === RUN TestRoleReadCommand/id_or_name_required === RUN TestRoleReadCommand/read_by_id_not_found === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.702Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:17.707Z [DEBUG] TestRoleReadCommand.http: Request finished: method=GET url=/v1/acl/role/f539afc3-7c83-bdf5-ff78-1ff8b006783c from=127.0.0.1:39484 latency=2.25601ms === RUN TestRoleReadCommand/read_by_name_not_found === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.728Z [DEBUG] TestRoleReadCommand.http: Request finished: method=GET url=/v1/acl/role/name/blah from=127.0.0.1:39486 latency=2.193698ms === RUN TestRoleReadCommand/read_by_id === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.744Z [DEBUG] TestRoleReadCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:39488 latency=2.916559ms writer.go:29: 2021-01-29T19:32:17.755Z [DEBUG] TestRoleReadCommand.http: Request finished: method=GET url=/v1/acl/role/a20102ab-099e-d62c-d02e-e26e7cc1583f from=127.0.0.1:39490 latency=3.527568ms writer.go:29: 2021-01-29T19:32:17.764Z [INFO] TestRoleReadCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.765Z [INFO] TestRoleReadCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.765Z [INFO] TestRoleReadCommand.leader: started routine: routine="CA root pruning" === RUN TestRoleReadCommand/read_by_id_prefix === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.766Z [INFO] TestRoleReadCommand_JSON.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:17.767Z [DEBUG] TestRoleReadCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6 writer.go:29: 2021-01-29T19:32:17.768Z [INFO] TestRoleReadCommand_JSON.server: member joined, marking health alive: member=Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6 === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.766Z [DEBUG] TestRoleReadCommand.server: Skipping self join check for node since the cluster is too small: node=Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62 writer.go:29: 2021-01-29T19:32:17.778Z [INFO] TestRoleReadCommand.server: member joined, marking health alive: member=Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62 writer.go:29: 2021-01-29T19:32:17.770Z [DEBUG] TestRoleReadCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:39488 latency=2.7967ms === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.779Z [DEBUG] TestRoleReadCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-3e568802-d10b-3fbf-43c8-ebae90d3fec6 === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.782Z [DEBUG] TestRoleReadCommand.server: Skipping self join check for node since the cluster is too small: node=Node-b693acc6-5758-204c-8bfe-fd6c77b8ab62 writer.go:29: 2021-01-29T19:32:17.823Z [DEBUG] TestRoleReadCommand.http: Request finished: method=GET url=/v1/acl/roles from=127.0.0.1:39492 latency=35.824217ms === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.831Z [DEBUG] TestRoleReadCommand_JSON: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.834Z [DEBUG] TestRoleReadCommand.http: Request finished: method=GET url=/v1/acl/role/a9439aea-8592-17a8-51b4-4bc58fa3d968 from=127.0.0.1:39492 latency=2.107012ms === RUN TestRoleReadCommand/read_by_name === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.836Z [INFO] TestRoleReadCommand_JSON: Synced node info writer.go:29: 2021-01-29T19:32:17.845Z [DEBUG] TestRoleReadCommand_JSON: Node info in sync === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.854Z [DEBUG] TestRoleReadCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:39488 latency=2.488362ms === RUN TestRoleReadCommand_JSON/read_by_id === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.860Z [DEBUG] TestRoleReadCommand_JSON.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:56678 latency=1.943193ms writer.go:29: 2021-01-29T19:32:17.867Z [DEBUG] TestRoleReadCommand_JSON.http: Request finished: method=GET url=/v1/acl/role/067553b9-3d47-e8f2-97a5-ce5aa96fc753 from=127.0.0.1:56682 latency=1.762857ms === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.869Z [DEBUG] TestRoleReadCommand.http: Request finished: method=GET url=/v1/acl/role/name/test-role-by-name from=127.0.0.1:39494 latency=8.257118ms === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.872Z [INFO] TestRoleReadCommand_JSON: Requesting shutdown === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.872Z [INFO] TestRoleReadCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:17.874Z [INFO] TestRoleReadCommand.server: shutting down server === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.874Z [INFO] TestRoleReadCommand_JSON.server: shutting down server === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.874Z [DEBUG] TestRoleReadCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:17.875Z [DEBUG] TestRoleReadCommand.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.875Z [DEBUG] TestRoleReadCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.876Z [DEBUG] TestRoleReadCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:17.877Z [WARN] TestRoleReadCommand.server.serf.lan: serf: Shutdown without a Leave === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.876Z [DEBUG] TestRoleReadCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:17.879Z [DEBUG] TestRoleReadCommand_JSON.leader: stopping routine: routine="CA root pruning" === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.875Z [DEBUG] TestRoleReadCommand.leader: stopped routine: routine="CA root pruning" === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.877Z [DEBUG] TestRoleReadCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.876Z [DEBUG] TestRoleReadCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:17.877Z [DEBUG] TestRoleReadCommand.leader: stopped routine: routine="acl token reaping" === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.879Z [DEBUG] TestRoleReadCommand_JSON.leader: stopped routine: routine="acl token reaping" === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:17.879Z [WARN] TestRoleReadCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:17.894Z [INFO] TestRoleReadCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:17.895Z [INFO] TestRoleReadCommand: consul server down writer.go:29: 2021-01-29T19:32:17.896Z [INFO] TestRoleReadCommand: shutdown complete writer.go:29: 2021-01-29T19:32:17.897Z [INFO] TestRoleReadCommand: Stopping server: protocol=DNS address=127.0.0.1:31448 network=tcp writer.go:29: 2021-01-29T19:32:17.898Z [INFO] TestRoleReadCommand: Stopping server: protocol=DNS address=127.0.0.1:31448 network=udp writer.go:29: 2021-01-29T19:32:17.899Z [INFO] TestRoleReadCommand: Stopping server: protocol=HTTP address=127.0.0.1:31449 network=tcp === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:17.893Z [WARN] TestRoleReadCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:17.893Z [DEBUG] TestRoleReadCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:17.921Z [WARN] TestRoleReadCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:17.925Z [INFO] TestRoleReadCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:17.925Z [INFO] TestRoleReadCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:17.926Z [INFO] TestRoleReadCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:32:17.927Z [INFO] TestRoleReadCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:31442 network=tcp writer.go:29: 2021-01-29T19:32:17.928Z [INFO] TestRoleReadCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:31442 network=udp writer.go:29: 2021-01-29T19:32:17.929Z [INFO] TestRoleReadCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:31443 network=tcp === CONT TestRoleReadCommand writer.go:29: 2021-01-29T19:32:18.400Z [INFO] TestRoleReadCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:18.401Z [INFO] TestRoleReadCommand: Endpoints down --- PASS: TestRoleReadCommand (0.94s) --- PASS: TestRoleReadCommand/id_or_name_required (0.00s) --- PASS: TestRoleReadCommand/read_by_id_not_found (0.01s) --- PASS: TestRoleReadCommand/read_by_name_not_found (0.02s) --- PASS: TestRoleReadCommand/read_by_id (0.03s) --- PASS: TestRoleReadCommand/read_by_id_prefix (0.07s) --- PASS: TestRoleReadCommand/read_by_name (0.03s) === CONT TestRoleReadCommand_JSON writer.go:29: 2021-01-29T19:32:18.430Z [INFO] TestRoleReadCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:18.431Z [INFO] TestRoleReadCommand_JSON: Endpoints down --- PASS: TestRoleReadCommand_JSON (0.97s) --- PASS: TestRoleReadCommand_JSON/read_by_id (0.02s) PASS ok github.com/hashicorp/consul/command/acl/role/read 1.123s === RUN TestRoleUpdateCommand_noTabs === PAUSE TestRoleUpdateCommand_noTabs === RUN TestRoleUpdateCommand === PAUSE TestRoleUpdateCommand === RUN TestRoleUpdateCommand_JSON === PAUSE TestRoleUpdateCommand_JSON === RUN TestRoleUpdateCommand_noMerge === PAUSE TestRoleUpdateCommand_noMerge === CONT TestRoleUpdateCommand_noTabs === CONT TestRoleUpdateCommand_noMerge === CONT TestRoleUpdateCommand_JSON [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestRoleUpdateCommand --- PASS: TestRoleUpdateCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.525Z [WARN] TestRoleUpdateCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:20.527Z [DEBUG] TestRoleUpdateCommand_JSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:20.540Z [DEBUG] TestRoleUpdateCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.544Z [WARN] TestRoleUpdateCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:20.545Z [DEBUG] TestRoleUpdateCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:20.547Z [DEBUG] TestRoleUpdateCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:20.552Z [INFO] TestRoleUpdateCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ee964517-99f6-a5db-cba7-e8359aaa4437 Address:127.0.0.1:19207}]" writer.go:29: 2021-01-29T19:32:20.554Z [INFO] TestRoleUpdateCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:19207 [Follower]" leader= writer.go:29: 2021-01-29T19:32:20.557Z [INFO] TestRoleUpdateCommand.server.serf.wan: serf: EventMemberJoin: Node-ee964517-99f6-a5db-cba7-e8359aaa4437.dc1 127.0.0.1 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.559Z [WARN] TestRoleUpdateCommand_noMerge: bootstrap = true: do not enable unless necessary === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.559Z [INFO] TestRoleUpdateCommand.server.serf.lan: serf: EventMemberJoin: Node-ee964517-99f6-a5db-cba7-e8359aaa4437 127.0.0.1 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.560Z [DEBUG] TestRoleUpdateCommand_noMerge.tlsutil: Update: version=1 === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.563Z [INFO] TestRoleUpdateCommand: Started DNS server: address=127.0.0.1:19202 network=tcp === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.563Z [DEBUG] TestRoleUpdateCommand_noMerge.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.564Z [INFO] TestRoleUpdateCommand: Started DNS server: address=127.0.0.1:19202 network=udp writer.go:29: 2021-01-29T19:32:20.566Z [INFO] TestRoleUpdateCommand.server: Handled event for server in area: event=member-join server=Node-ee964517-99f6-a5db-cba7-e8359aaa4437.dc1 area=wan === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.568Z [INFO] TestRoleUpdateCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:89a6bdf0-4606-4854-8f54-5cb52eb5a08f Address:127.0.0.1:19195}]" === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.566Z [INFO] TestRoleUpdateCommand: Started HTTP server: address=127.0.0.1:19203 network=tcp === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.571Z [INFO] TestRoleUpdateCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f.dc1 127.0.0.1 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.574Z [INFO] TestRoleUpdateCommand_noMerge.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2f48b1e1-1210-1e2f-151d-7da676919304 Address:127.0.0.1:19201}]" writer.go:29: 2021-01-29T19:32:20.584Z [INFO] TestRoleUpdateCommand_noMerge.server.serf.wan: serf: EventMemberJoin: Node-2f48b1e1-1210-1e2f-151d-7da676919304.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:20.593Z [INFO] TestRoleUpdateCommand_noMerge.server.raft: entering follower state: follower="Node at 127.0.0.1:19201 [Follower]" leader= === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.569Z [INFO] TestRoleUpdateCommand: started state syncer writer.go:29: 2021-01-29T19:32:20.567Z [INFO] TestRoleUpdateCommand.server: Adding LAN server: server="Node-ee964517-99f6-a5db-cba7-e8359aaa4437 (Addr: tcp/127.0.0.1:19207) (DC: dc1)" === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.571Z [INFO] TestRoleUpdateCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:19195 [Follower]" leader= === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.605Z [INFO] TestRoleUpdateCommand_noMerge.server.serf.lan: serf: EventMemberJoin: Node-2f48b1e1-1210-1e2f-151d-7da676919304 127.0.0.1 === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.605Z [WARN] TestRoleUpdateCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:20.611Z [INFO] TestRoleUpdateCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:19207 [Candidate]" term=2 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.632Z [INFO] TestRoleUpdateCommand_noMerge.server: Handled event for server in area: event=member-join server=Node-2f48b1e1-1210-1e2f-151d-7da676919304.dc1 area=wan === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.632Z [DEBUG] TestRoleUpdateCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.632Z [INFO] TestRoleUpdateCommand_noMerge.server: Adding LAN server: server="Node-2f48b1e1-1210-1e2f-151d-7da676919304 (Addr: tcp/127.0.0.1:19201) (DC: dc1)" writer.go:29: 2021-01-29T19:32:20.633Z [INFO] TestRoleUpdateCommand_noMerge: Started DNS server: address=127.0.0.1:19196 network=udp === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.634Z [INFO] TestRoleUpdateCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f 127.0.0.1 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.634Z [INFO] TestRoleUpdateCommand_noMerge: Started DNS server: address=127.0.0.1:19196 network=tcp writer.go:29: 2021-01-29T19:32:20.636Z [INFO] TestRoleUpdateCommand_noMerge: Started HTTP server: address=127.0.0.1:19197 network=tcp === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.636Z [INFO] TestRoleUpdateCommand_JSON: Started DNS server: address=127.0.0.1:19190 network=udp === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.637Z [INFO] TestRoleUpdateCommand_noMerge: started state syncer === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.638Z [INFO] TestRoleUpdateCommand_JSON.server: Adding LAN server: server="Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f (Addr: tcp/127.0.0.1:19195) (DC: dc1)" === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.634Z [WARN] TestRoleUpdateCommand_noMerge.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.639Z [INFO] TestRoleUpdateCommand_JSON.server: Handled event for server in area: event=member-join server=Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f.dc1 area=wan === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.639Z [INFO] TestRoleUpdateCommand_noMerge.server.raft: entering candidate state: node="Node at 127.0.0.1:19201 [Candidate]" term=2 === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.642Z [INFO] TestRoleUpdateCommand_JSON: Started DNS server: address=127.0.0.1:19190 network=tcp === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.643Z [DEBUG] TestRoleUpdateCommand.server.raft: votes: needed=1 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.644Z [DEBUG] TestRoleUpdateCommand_noMerge.server.raft: votes: needed=1 === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.645Z [DEBUG] TestRoleUpdateCommand.server.raft: vote granted: from=ee964517-99f6-a5db-cba7-e8359aaa4437 term=2 tally=1 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.646Z [DEBUG] TestRoleUpdateCommand_noMerge.server.raft: vote granted: from=2f48b1e1-1210-1e2f-151d-7da676919304 term=2 tally=1 === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.646Z [INFO] TestRoleUpdateCommand_JSON: Started HTTP server: address=127.0.0.1:19191 network=tcp === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.647Z [INFO] TestRoleUpdateCommand.server.raft: election won: tally=1 === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.648Z [INFO] TestRoleUpdateCommand_JSON: started state syncer === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.648Z [INFO] TestRoleUpdateCommand_noMerge.server.raft: election won: tally=1 === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.649Z [INFO] TestRoleUpdateCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:19207 [Leader]" === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.650Z [INFO] TestRoleUpdateCommand_noMerge.server.raft: entering leader state: leader="Node at 127.0.0.1:19201 [Leader]" === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.651Z [INFO] TestRoleUpdateCommand.server: cluster leadership acquired === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.654Z [INFO] TestRoleUpdateCommand_noMerge.server: cluster leadership acquired === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.653Z [INFO] TestRoleUpdateCommand.server: New leader elected: payload=Node-ee964517-99f6-a5db-cba7-e8359aaa4437 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.656Z [INFO] TestRoleUpdateCommand_noMerge.server: New leader elected: payload=Node-2f48b1e1-1210-1e2f-151d-7da676919304 === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.656Z [INFO] TestRoleUpdateCommand.server: initializing acls === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.658Z [INFO] TestRoleUpdateCommand_noMerge.server: initializing acls === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.662Z [INFO] TestRoleUpdateCommand.server: Created ACL 'global-management' policy === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.664Z [INFO] TestRoleUpdateCommand_noMerge.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:20.665Z [WARN] TestRoleUpdateCommand_noMerge.server: Configuring a non-UUID master token is deprecated === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.664Z [WARN] TestRoleUpdateCommand.server: Configuring a non-UUID master token is deprecated === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.668Z [INFO] TestRoleUpdateCommand_noMerge.server: Bootstrapped ACL master token from configuration === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.668Z [INFO] TestRoleUpdateCommand.server: Bootstrapped ACL master token from configuration === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.671Z [INFO] TestRoleUpdateCommand_noMerge.server: Created ACL anonymous token from configuration === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.673Z [WARN] TestRoleUpdateCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.671Z [INFO] TestRoleUpdateCommand.server: Created ACL anonymous token from configuration === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.674Z [INFO] TestRoleUpdateCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:19195 [Candidate]" term=2 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.672Z [INFO] TestRoleUpdateCommand_noMerge.leader: started routine: routine="legacy ACL token upgrade" === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.675Z [DEBUG] TestRoleUpdateCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:20.676Z [DEBUG] TestRoleUpdateCommand_JSON.server.raft: vote granted: from=89a6bdf0-4606-4854-8f54-5cb52eb5a08f term=2 tally=1 === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.674Z [INFO] TestRoleUpdateCommand.leader: started routine: routine="legacy ACL token upgrade" === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.677Z [INFO] TestRoleUpdateCommand_JSON.server.raft: election won: tally=1 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.675Z [INFO] TestRoleUpdateCommand_noMerge.leader: started routine: routine="acl token reaping" === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.677Z [INFO] TestRoleUpdateCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:19195 [Leader]" writer.go:29: 2021-01-29T19:32:20.678Z [INFO] TestRoleUpdateCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:20.680Z [INFO] TestRoleUpdateCommand_JSON.server: New leader elected: payload=Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.677Z [INFO] TestRoleUpdateCommand.leader: started routine: routine="acl token reaping" === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.682Z [INFO] TestRoleUpdateCommand_JSON.server: initializing acls === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.678Z [INFO] TestRoleUpdateCommand_noMerge.server.serf.lan: serf: EventMemberUpdate: Node-2f48b1e1-1210-1e2f-151d-7da676919304 writer.go:29: 2021-01-29T19:32:20.682Z [INFO] TestRoleUpdateCommand_noMerge: Synced node info === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.687Z [INFO] TestRoleUpdateCommand_JSON.server: Created ACL 'global-management' policy === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.685Z [INFO] TestRoleUpdateCommand_noMerge.server.serf.wan: serf: EventMemberUpdate: Node-2f48b1e1-1210-1e2f-151d-7da676919304.dc1 writer.go:29: 2021-01-29T19:32:20.685Z [INFO] TestRoleUpdateCommand_noMerge.server: Updating LAN server: server="Node-2f48b1e1-1210-1e2f-151d-7da676919304 (Addr: tcp/127.0.0.1:19201) (DC: dc1)" === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.690Z [WARN] TestRoleUpdateCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.692Z [INFO] TestRoleUpdateCommand_noMerge.server: Handled event for server in area: event=member-update server=Node-2f48b1e1-1210-1e2f-151d-7da676919304.dc1 area=wan === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.688Z [ERROR] TestRoleUpdateCommand_JSON.anti_entropy: failed to sync remote state: error="ACL not found" === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.694Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.692Z [INFO] TestRoleUpdateCommand_JSON.server: initializing acls === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.683Z [INFO] TestRoleUpdateCommand.server.serf.lan: serf: EventMemberUpdate: Node-ee964517-99f6-a5db-cba7-e8359aaa4437 === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.697Z [WARN] TestRoleUpdateCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.698Z [INFO] TestRoleUpdateCommand.server.serf.wan: serf: EventMemberUpdate: Node-ee964517-99f6-a5db-cba7-e8359aaa4437.dc1 writer.go:29: 2021-01-29T19:32:20.700Z [INFO] TestRoleUpdateCommand.server: Updating LAN server: server="Node-ee964517-99f6-a5db-cba7-e8359aaa4437 (Addr: tcp/127.0.0.1:19207) (DC: dc1)" writer.go:29: 2021-01-29T19:32:20.702Z [INFO] TestRoleUpdateCommand.server: Handled event for server in area: event=member-update server=Node-ee964517-99f6-a5db-cba7-e8359aaa4437.dc1 area=wan writer.go:29: 2021-01-29T19:32:20.702Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.695Z [INFO] TestRoleUpdateCommand_JSON.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:20.702Z [INFO] TestRoleUpdateCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:20.707Z [INFO] TestRoleUpdateCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:20.709Z [INFO] TestRoleUpdateCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:20.710Z [DEBUG] TestRoleUpdateCommand_JSON.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:32:20.711Z [INFO] TestRoleUpdateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f writer.go:29: 2021-01-29T19:32:20.714Z [INFO] TestRoleUpdateCommand_JSON.server: Updating LAN server: server="Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f (Addr: tcp/127.0.0.1:19195) (DC: dc1)" writer.go:29: 2021-01-29T19:32:20.714Z [INFO] TestRoleUpdateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f.dc1 writer.go:29: 2021-01-29T19:32:20.715Z [INFO] TestRoleUpdateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f writer.go:29: 2021-01-29T19:32:20.722Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:20.720Z [INFO] TestRoleUpdateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f.dc1 area=wan writer.go:29: 2021-01-29T19:32:20.722Z [INFO] TestRoleUpdateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f.dc1 writer.go:29: 2021-01-29T19:32:20.722Z [INFO] TestRoleUpdateCommand_JSON.server: Updating LAN server: server="Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f (Addr: tcp/127.0.0.1:19195) (DC: dc1)" writer.go:29: 2021-01-29T19:32:20.738Z [INFO] TestRoleUpdateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f.dc1 area=wan writer.go:29: 2021-01-29T19:32:20.760Z [INFO] TestRoleUpdateCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:20.761Z [INFO] TestRoleUpdateCommand_JSON.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:20.762Z [DEBUG] TestRoleUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f writer.go:29: 2021-01-29T19:32:20.762Z [INFO] TestRoleUpdateCommand_JSON.server: member joined, marking health alive: member=Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.762Z [INFO] TestRoleUpdateCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:20.764Z [INFO] TestRoleUpdateCommand.leader: started routine: routine="CA root pruning" === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.766Z [DEBUG] TestRoleUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.765Z [DEBUG] TestRoleUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-ee964517-99f6-a5db-cba7-e8359aaa4437 === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.767Z [DEBUG] TestRoleUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-89a6bdf0-4606-4854-8f54-5cb52eb5a08f === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.767Z [INFO] TestRoleUpdateCommand.server: member joined, marking health alive: member=Node-ee964517-99f6-a5db-cba7-e8359aaa4437 === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.770Z [INFO] TestRoleUpdateCommand_noMerge.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:20.771Z [INFO] TestRoleUpdateCommand_noMerge.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:20.772Z [DEBUG] TestRoleUpdateCommand_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-2f48b1e1-1210-1e2f-151d-7da676919304 writer.go:29: 2021-01-29T19:32:20.773Z [INFO] TestRoleUpdateCommand_noMerge.server: member joined, marking health alive: member=Node-2f48b1e1-1210-1e2f-151d-7da676919304 writer.go:29: 2021-01-29T19:32:20.775Z [DEBUG] TestRoleUpdateCommand_noMerge.server: Skipping self join check for node since the cluster is too small: node=Node-2f48b1e1-1210-1e2f-151d-7da676919304 === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.773Z [DEBUG] TestRoleUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-ee964517-99f6-a5db-cba7-e8359aaa4437 writer.go:29: 2021-01-29T19:32:20.833Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:32920 latency=35.291177ms === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.842Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:56916 latency=16.099323ms === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.848Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:32920 latency=5.951248ms === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.840Z [DEBUG] TestRoleUpdateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:59964 latency=9.599549ms === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.881Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:56916 latency=3.912451ms writer.go:29: 2021-01-29T19:32:20.885Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:56916 latency=950.145µs writer.go:29: 2021-01-29T19:32:20.885Z [DEBUG] TestRoleUpdateCommand_noMerge: Skipping remote check since it is managed automatically: check=serfHealth === RUN TestRoleUpdateCommand_noMerge/update_a_role_that_does_not_exist === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.888Z [DEBUG] TestRoleUpdateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:59964 latency=3.083285ms === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.888Z [DEBUG] TestRoleUpdateCommand_noMerge: Node info in sync === RUN TestRoleUpdateCommand_JSON/update_a_role_that_does_not_exist === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.891Z [DEBUG] TestRoleUpdateCommand_noMerge: Node info in sync === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.898Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:32920 latency=4.565531ms === RUN TestRoleUpdateCommand/update_a_role_that_does_not_exist === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.905Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/role/07f9f6c6-bb0a-c38b-c143-bf9bba91d671 from=127.0.0.1:56936 latency=2.61243ms === RUN TestRoleUpdateCommand_noMerge/update_with_policy_by_name === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.919Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=GET url=/v1/acl/role/a4f1fcec-73a4-4e20-5a48-d512f43c0b0c from=127.0.0.1:32946 latency=2.766436ms === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.920Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:56916 latency=1.183302ms === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.921Z [DEBUG] TestRoleUpdateCommand_JSON.http: Request finished: method=GET url=/v1/acl/role/106dcafd-cae9-72d6-d27e-cdb9a0e40adf from=127.0.0.1:59978 latency=17.293983ms === RUN TestRoleUpdateCommand_JSON/update_with_policy_by_name === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.935Z [DEBUG] TestRoleUpdateCommand_JSON.http: Request finished: method=GET url=/v1/acl/role/c0c721eb-1943-1d53-b330-b1aae508f260 from=127.0.0.1:59990 latency=2.025254ms === RUN TestRoleUpdateCommand/update_with_policy_by_name === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.938Z [DEBUG] TestRoleUpdateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/role/c0c721eb-1943-1d53-b330-b1aae508f260 from=127.0.0.1:59990 latency=863.827µs writer.go:29: 2021-01-29T19:32:20.944Z [INFO] TestRoleUpdateCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:32:20.946Z [INFO] TestRoleUpdateCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:20.948Z [DEBUG] TestRoleUpdateCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:20.949Z [DEBUG] TestRoleUpdateCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.944Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/role/4553a05f-39d4-a828-7632-217c29c4d124 from=127.0.0.1:56944 latency=8.11984ms === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.951Z [DEBUG] TestRoleUpdateCommand_JSON.leader: stopping routine: routine="acl token reaping" === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.953Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/role/4553a05f-39d4-a828-7632-217c29c4d124 from=127.0.0.1:56944 latency=797.031µs === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.949Z [DEBUG] TestRoleUpdateCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:20.951Z [DEBUG] TestRoleUpdateCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.955Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/role/4553a05f-39d4-a828-7632-217c29c4d124 from=127.0.0.1:56916 latency=69.823µs === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.953Z [WARN] TestRoleUpdateCommand_JSON.server.serf.lan: serf: Shutdown without a Leave === RUN TestRoleUpdateCommand_noMerge/update_with_policy_by_id === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.953Z [DEBUG] TestRoleUpdateCommand_JSON.leader: stopped routine: routine="acl token reaping" === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.959Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=GET url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32958 latency=2.095053ms === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.963Z [WARN] TestRoleUpdateCommand_JSON.server.serf.wan: serf: Shutdown without a Leave === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.967Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:56916 latency=5.543894ms === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.967Z [INFO] TestRoleUpdateCommand_JSON.server.router.manager: shutting down === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.968Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32958 latency=1.214612ms === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.969Z [INFO] TestRoleUpdateCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:20.971Z [INFO] TestRoleUpdateCommand_JSON: shutdown complete === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.975Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/role/08f4cad9-4e8b-93e5-99da-b172e789f0d7 from=127.0.0.1:56954 latency=1.124155ms === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.975Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=GET url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32920 latency=113.009µs === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.974Z [INFO] TestRoleUpdateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:19190 network=tcp === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.978Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/role/08f4cad9-4e8b-93e5-99da-b172e789f0d7 from=127.0.0.1:56954 latency=1.095542ms === RUN TestRoleUpdateCommand/update_with_policy_by_id === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.981Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/role/08f4cad9-4e8b-93e5-99da-b172e789f0d7 from=127.0.0.1:56916 latency=96.398µs === RUN TestRoleUpdateCommand_noMerge/update_with_service_identity === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.977Z [INFO] TestRoleUpdateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:19190 network=udp === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:20.991Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:56916 latency=2.251971ms === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:20.989Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=GET url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32966 latency=2.089889ms writer.go:29: 2021-01-29T19:32:20.997Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32966 latency=1.023428ms === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:20.990Z [INFO] TestRoleUpdateCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:19191 network=tcp === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:21.012Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=GET url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32920 latency=356.053µs === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:21.014Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/role/7e7a5040-6ae1-ea01-2228-045bb2cd7be1 from=127.0.0.1:56964 latency=4.414541ms === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:21.018Z [DEBUG] TestRoleUpdateCommand: Skipping remote check since it is managed automatically: check=serfHealth === RUN TestRoleUpdateCommand/update_with_service_identity === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:21.030Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/role/7e7a5040-6ae1-ea01-2228-045bb2cd7be1 from=127.0.0.1:56964 latency=7.164581ms === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:21.033Z [INFO] TestRoleUpdateCommand: Synced node info === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:21.041Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/role/7e7a5040-6ae1-ea01-2228-045bb2cd7be1 from=127.0.0.1:56916 latency=4.889296ms === RUN TestRoleUpdateCommand_noMerge/update_with_service_identity_scoped_to_2_DCs === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:21.065Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=GET url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32982 latency=4.666091ms writer.go:29: 2021-01-29T19:32:21.073Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32982 latency=1.757443ms writer.go:29: 2021-01-29T19:32:21.079Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=GET url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32920 latency=155.242µs === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:21.085Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/role from=127.0.0.1:56916 latency=15.429769ms === RUN TestRoleUpdateCommand/update_with_service_identity_scoped_to_2_DCs === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:21.091Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/role/85dc9443-e4bc-6524-0de4-671f102119ae from=127.0.0.1:56986 latency=1.276238ms writer.go:29: 2021-01-29T19:32:21.096Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=PUT url=/v1/acl/role/85dc9443-e4bc-6524-0de4-671f102119ae from=127.0.0.1:56986 latency=1.669975ms writer.go:29: 2021-01-29T19:32:21.098Z [DEBUG] TestRoleUpdateCommand_noMerge.http: Request finished: method=GET url=/v1/acl/role/85dc9443-e4bc-6524-0de4-671f102119ae from=127.0.0.1:56916 latency=87.805µs writer.go:29: 2021-01-29T19:32:21.102Z [INFO] TestRoleUpdateCommand_noMerge: Requesting shutdown === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:21.101Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=GET url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32996 latency=4.725788ms === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:21.104Z [INFO] TestRoleUpdateCommand_noMerge.server: shutting down server writer.go:29: 2021-01-29T19:32:21.113Z [DEBUG] TestRoleUpdateCommand_noMerge.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:21.114Z [DEBUG] TestRoleUpdateCommand_noMerge.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:21.114Z [DEBUG] TestRoleUpdateCommand_noMerge.leader: stopping routine: routine="CA root pruning" === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:21.123Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=PUT url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32996 latency=14.663673ms === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:21.124Z [WARN] TestRoleUpdateCommand_noMerge.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:21.122Z [DEBUG] TestRoleUpdateCommand_noMerge.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:21.122Z [DEBUG] TestRoleUpdateCommand_noMerge.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:21.127Z [DEBUG] TestRoleUpdateCommand_noMerge.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:21.130Z [WARN] TestRoleUpdateCommand_noMerge.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:21.134Z [INFO] TestRoleUpdateCommand_noMerge.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:21.137Z [INFO] TestRoleUpdateCommand_noMerge: consul server down writer.go:29: 2021-01-29T19:32:21.139Z [INFO] TestRoleUpdateCommand_noMerge: shutdown complete writer.go:29: 2021-01-29T19:32:21.140Z [INFO] TestRoleUpdateCommand_noMerge: Stopping server: protocol=DNS address=127.0.0.1:19196 network=tcp writer.go:29: 2021-01-29T19:32:21.142Z [INFO] TestRoleUpdateCommand_noMerge: Stopping server: protocol=DNS address=127.0.0.1:19196 network=udp === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:21.146Z [DEBUG] TestRoleUpdateCommand.http: Request finished: method=GET url=/v1/acl/role/28a16101-b8cf-133a-8797-a20915abecdf from=127.0.0.1:32920 latency=9.575878ms === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:21.144Z [INFO] TestRoleUpdateCommand_noMerge: Stopping server: protocol=HTTP address=127.0.0.1:19197 network=tcp === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:21.150Z [INFO] TestRoleUpdateCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:21.153Z [INFO] TestRoleUpdateCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:21.155Z [DEBUG] TestRoleUpdateCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:21.157Z [DEBUG] TestRoleUpdateCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:21.158Z [DEBUG] TestRoleUpdateCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:21.157Z [DEBUG] TestRoleUpdateCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:21.159Z [DEBUG] TestRoleUpdateCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:21.160Z [DEBUG] TestRoleUpdateCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:21.160Z [WARN] TestRoleUpdateCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:21.170Z [WARN] TestRoleUpdateCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:21.173Z [INFO] TestRoleUpdateCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:21.173Z [INFO] TestRoleUpdateCommand: consul server down writer.go:29: 2021-01-29T19:32:21.176Z [INFO] TestRoleUpdateCommand: shutdown complete writer.go:29: 2021-01-29T19:32:21.178Z [INFO] TestRoleUpdateCommand: Stopping server: protocol=DNS address=127.0.0.1:19202 network=tcp writer.go:29: 2021-01-29T19:32:21.180Z [INFO] TestRoleUpdateCommand: Stopping server: protocol=DNS address=127.0.0.1:19202 network=udp writer.go:29: 2021-01-29T19:32:21.181Z [INFO] TestRoleUpdateCommand: Stopping server: protocol=HTTP address=127.0.0.1:19203 network=tcp === CONT TestRoleUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:21.508Z [INFO] TestRoleUpdateCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:21.510Z [INFO] TestRoleUpdateCommand_JSON: Endpoints down --- PASS: TestRoleUpdateCommand_JSON (1.03s) --- PASS: TestRoleUpdateCommand_JSON/update_a_role_that_does_not_exist (0.03s) --- PASS: TestRoleUpdateCommand_JSON/update_with_policy_by_name (0.02s) === CONT TestRoleUpdateCommand_noMerge writer.go:29: 2021-01-29T19:32:21.653Z [INFO] TestRoleUpdateCommand_noMerge: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:21.654Z [INFO] TestRoleUpdateCommand_noMerge: Endpoints down --- PASS: TestRoleUpdateCommand_noMerge (1.17s) --- PASS: TestRoleUpdateCommand_noMerge/update_a_role_that_does_not_exist (0.02s) --- PASS: TestRoleUpdateCommand_noMerge/update_with_policy_by_name (0.05s) --- PASS: TestRoleUpdateCommand_noMerge/update_with_policy_by_id (0.03s) --- PASS: TestRoleUpdateCommand_noMerge/update_with_service_identity (0.07s) --- PASS: TestRoleUpdateCommand_noMerge/update_with_service_identity_scoped_to_2_DCs (0.04s) === CONT TestRoleUpdateCommand writer.go:29: 2021-01-29T19:32:21.685Z [INFO] TestRoleUpdateCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:21.686Z [INFO] TestRoleUpdateCommand: Endpoints down --- PASS: TestRoleUpdateCommand (1.20s) --- PASS: TestRoleUpdateCommand/update_a_role_that_does_not_exist (0.03s) --- PASS: TestRoleUpdateCommand/update_with_policy_by_name (0.04s) --- PASS: TestRoleUpdateCommand/update_with_policy_by_id (0.05s) --- PASS: TestRoleUpdateCommand/update_with_service_identity (0.06s) --- PASS: TestRoleUpdateCommand/update_with_service_identity_scoped_to_2_DCs (0.06s) PASS ok github.com/hashicorp/consul/command/acl/role/update 1.424s === RUN TestRulesTranslateCommand_noTabs === PAUSE TestRulesTranslateCommand_noTabs === RUN TestRulesTranslateCommand === PAUSE TestRulesTranslateCommand === CONT TestRulesTranslateCommand_noTabs === CONT TestRulesTranslateCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestRulesTranslateCommand_noTabs (0.01s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestRulesTranslateCommand writer.go:29: 2021-01-29T19:32:22.167Z [WARN] TestRulesTranslateCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:22.169Z [DEBUG] TestRulesTranslateCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:22.185Z [DEBUG] TestRulesTranslateCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:22.205Z [INFO] TestRulesTranslateCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:86ed2529-2ad2-33dc-dc38-9330d79329fb Address:127.0.0.1:15111}]" writer.go:29: 2021-01-29T19:32:22.207Z [INFO] TestRulesTranslateCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:15111 [Follower]" leader= writer.go:29: 2021-01-29T19:32:22.221Z [INFO] TestRulesTranslateCommand.server.serf.wan: serf: EventMemberJoin: Node-86ed2529-2ad2-33dc-dc38-9330d79329fb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:22.250Z [INFO] TestRulesTranslateCommand.server.serf.lan: serf: EventMemberJoin: Node-86ed2529-2ad2-33dc-dc38-9330d79329fb 127.0.0.1 writer.go:29: 2021-01-29T19:32:22.259Z [INFO] TestRulesTranslateCommand.server: Handled event for server in area: event=member-join server=Node-86ed2529-2ad2-33dc-dc38-9330d79329fb.dc1 area=wan writer.go:29: 2021-01-29T19:32:22.260Z [INFO] TestRulesTranslateCommand.server: Adding LAN server: server="Node-86ed2529-2ad2-33dc-dc38-9330d79329fb (Addr: tcp/127.0.0.1:15111) (DC: dc1)" writer.go:29: 2021-01-29T19:32:22.266Z [WARN] TestRulesTranslateCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:22.267Z [INFO] TestRulesTranslateCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:15111 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:22.270Z [INFO] TestRulesTranslateCommand: Started DNS server: address=127.0.0.1:15106 network=tcp writer.go:29: 2021-01-29T19:32:22.271Z [INFO] TestRulesTranslateCommand: Started DNS server: address=127.0.0.1:15106 network=udp writer.go:29: 2021-01-29T19:32:22.271Z [DEBUG] TestRulesTranslateCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:22.275Z [DEBUG] TestRulesTranslateCommand.server.raft: vote granted: from=86ed2529-2ad2-33dc-dc38-9330d79329fb term=2 tally=1 writer.go:29: 2021-01-29T19:32:22.276Z [INFO] TestRulesTranslateCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:22.274Z [INFO] TestRulesTranslateCommand: Started HTTP server: address=127.0.0.1:15107 network=tcp writer.go:29: 2021-01-29T19:32:22.279Z [INFO] TestRulesTranslateCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:15111 [Leader]" writer.go:29: 2021-01-29T19:32:22.282Z [INFO] TestRulesTranslateCommand: started state syncer writer.go:29: 2021-01-29T19:32:22.284Z [ERROR] TestRulesTranslateCommand.anti_entropy: failed to sync remote state: error="ACL not found" writer.go:29: 2021-01-29T19:32:22.289Z [INFO] TestRulesTranslateCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:22.292Z [INFO] TestRulesTranslateCommand.server: New leader elected: payload=Node-86ed2529-2ad2-33dc-dc38-9330d79329fb writer.go:29: 2021-01-29T19:32:22.292Z [INFO] TestRulesTranslateCommand.server: initializing acls writer.go:29: 2021-01-29T19:32:22.302Z [INFO] TestRulesTranslateCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:22.313Z [WARN] TestRulesTranslateCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:22.309Z [INFO] TestRulesTranslateCommand.server: initializing acls writer.go:29: 2021-01-29T19:32:22.325Z [WARN] TestRulesTranslateCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:22.320Z [INFO] TestRulesTranslateCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:22.330Z [INFO] TestRulesTranslateCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:22.331Z [INFO] TestRulesTranslateCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:22.332Z [INFO] TestRulesTranslateCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:22.333Z [DEBUG] TestRulesTranslateCommand.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:32:22.332Z [INFO] TestRulesTranslateCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:22.334Z [INFO] TestRulesTranslateCommand.server.serf.lan: serf: EventMemberUpdate: Node-86ed2529-2ad2-33dc-dc38-9330d79329fb writer.go:29: 2021-01-29T19:32:22.336Z [INFO] TestRulesTranslateCommand.server.serf.wan: serf: EventMemberUpdate: Node-86ed2529-2ad2-33dc-dc38-9330d79329fb.dc1 writer.go:29: 2021-01-29T19:32:22.338Z [INFO] TestRulesTranslateCommand.server: Handled event for server in area: event=member-update server=Node-86ed2529-2ad2-33dc-dc38-9330d79329fb.dc1 area=wan writer.go:29: 2021-01-29T19:32:22.339Z [INFO] TestRulesTranslateCommand.server: Updating LAN server: server="Node-86ed2529-2ad2-33dc-dc38-9330d79329fb (Addr: tcp/127.0.0.1:15111) (DC: dc1)" writer.go:29: 2021-01-29T19:32:22.339Z [INFO] TestRulesTranslateCommand.server.serf.lan: serf: EventMemberUpdate: Node-86ed2529-2ad2-33dc-dc38-9330d79329fb writer.go:29: 2021-01-29T19:32:22.344Z [INFO] TestRulesTranslateCommand.server.serf.wan: serf: EventMemberUpdate: Node-86ed2529-2ad2-33dc-dc38-9330d79329fb.dc1 writer.go:29: 2021-01-29T19:32:22.346Z [INFO] TestRulesTranslateCommand.server: Updating LAN server: server="Node-86ed2529-2ad2-33dc-dc38-9330d79329fb (Addr: tcp/127.0.0.1:15111) (DC: dc1)" writer.go:29: 2021-01-29T19:32:22.346Z [INFO] TestRulesTranslateCommand.server: Handled event for server in area: event=member-update server=Node-86ed2529-2ad2-33dc-dc38-9330d79329fb.dc1 area=wan writer.go:29: 2021-01-29T19:32:22.352Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:22.492Z [INFO] TestRulesTranslateCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:22.493Z [INFO] TestRulesTranslateCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:22.495Z [DEBUG] TestRulesTranslateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-86ed2529-2ad2-33dc-dc38-9330d79329fb writer.go:29: 2021-01-29T19:32:22.496Z [INFO] TestRulesTranslateCommand.server: member joined, marking health alive: member=Node-86ed2529-2ad2-33dc-dc38-9330d79329fb writer.go:29: 2021-01-29T19:32:22.504Z [DEBUG] TestRulesTranslateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-86ed2529-2ad2-33dc-dc38-9330d79329fb writer.go:29: 2021-01-29T19:32:22.506Z [DEBUG] TestRulesTranslateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-86ed2529-2ad2-33dc-dc38-9330d79329fb === RUN TestRulesTranslateCommand/file === RUN TestRulesTranslateCommand/stdin === RUN TestRulesTranslateCommand/arg === RUN TestRulesTranslateCommand/exclusive-options === CONT TestRulesTranslateCommand writer.go:29: 2021-01-29T19:32:22.556Z [INFO] TestRulesTranslateCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:22.558Z [INFO] TestRulesTranslateCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:22.559Z [DEBUG] TestRulesTranslateCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:22.559Z [DEBUG] TestRulesTranslateCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:22.560Z [DEBUG] TestRulesTranslateCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:22.561Z [WARN] TestRulesTranslateCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:22.563Z [DEBUG] TestRulesTranslateCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:22.563Z [DEBUG] TestRulesTranslateCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:22.564Z [DEBUG] TestRulesTranslateCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:22.566Z [WARN] TestRulesTranslateCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:22.568Z [INFO] TestRulesTranslateCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:22.570Z [INFO] TestRulesTranslateCommand: consul server down writer.go:29: 2021-01-29T19:32:22.572Z [INFO] TestRulesTranslateCommand: shutdown complete writer.go:29: 2021-01-29T19:32:22.591Z [INFO] TestRulesTranslateCommand: Stopping server: protocol=DNS address=127.0.0.1:15106 network=tcp writer.go:29: 2021-01-29T19:32:22.593Z [INFO] TestRulesTranslateCommand: Stopping server: protocol=DNS address=127.0.0.1:15106 network=udp writer.go:29: 2021-01-29T19:32:22.595Z [INFO] TestRulesTranslateCommand: Stopping server: protocol=HTTP address=127.0.0.1:15107 network=tcp writer.go:29: 2021-01-29T19:32:23.097Z [INFO] TestRulesTranslateCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:23.098Z [INFO] TestRulesTranslateCommand: Endpoints down --- PASS: TestRulesTranslateCommand (0.98s) --- PASS: TestRulesTranslateCommand/file (0.01s) --- PASS: TestRulesTranslateCommand/stdin (0.00s) --- PASS: TestRulesTranslateCommand/arg (0.02s) --- PASS: TestRulesTranslateCommand/exclusive-options (0.00s) PASS ok github.com/hashicorp/consul/command/acl/rules 1.159s ? github.com/hashicorp/consul/command/acl/token [no test files] === RUN TestTokenCloneCommand_noTabs === PAUSE TestTokenCloneCommand_noTabs === RUN TestTokenCloneCommand_Pretty === PAUSE TestTokenCloneCommand_Pretty === RUN TestTokenCloneCommand_JSON === PAUSE TestTokenCloneCommand_JSON === CONT TestTokenCloneCommand_noTabs === CONT TestTokenCloneCommand_JSON [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestTokenCloneCommand_Pretty --- PASS: TestTokenCloneCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.031Z [WARN] TestTokenCloneCommand_Pretty: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:26.035Z [DEBUG] TestTokenCloneCommand_Pretty.tlsutil: Update: version=1 === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.040Z [WARN] TestTokenCloneCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:26.042Z [DEBUG] TestTokenCloneCommand_JSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:26.043Z [DEBUG] TestTokenCloneCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.041Z [DEBUG] TestTokenCloneCommand_Pretty.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.048Z [INFO] TestTokenCloneCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 Address:127.0.0.1:11027}]" writer.go:29: 2021-01-29T19:32:26.050Z [INFO] TestTokenCloneCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:11027 [Follower]" leader= === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.053Z [INFO] TestTokenCloneCommand_Pretty.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:da180c94-a2fb-f710-f6d3-69607f1fe8df Address:127.0.0.1:11033}]" === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.054Z [INFO] TestTokenCloneCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1.dc1 127.0.0.1 === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.055Z [INFO] TestTokenCloneCommand_Pretty.server.raft: entering follower state: follower="Node at 127.0.0.1:11033 [Follower]" leader= writer.go:29: 2021-01-29T19:32:26.056Z [INFO] TestTokenCloneCommand_Pretty.server.serf.wan: serf: EventMemberJoin: Node-da180c94-a2fb-f710-f6d3-69607f1fe8df.dc1 127.0.0.1 === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.059Z [INFO] TestTokenCloneCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 127.0.0.1 writer.go:29: 2021-01-29T19:32:26.061Z [INFO] TestTokenCloneCommand_JSON.server: Adding LAN server: server="Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 (Addr: tcp/127.0.0.1:11027) (DC: dc1)" === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.062Z [INFO] TestTokenCloneCommand_Pretty.server.serf.lan: serf: EventMemberJoin: Node-da180c94-a2fb-f710-f6d3-69607f1fe8df 127.0.0.1 === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.062Z [INFO] TestTokenCloneCommand_JSON.server: Handled event for server in area: event=member-join server=Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1.dc1 area=wan === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.066Z [INFO] TestTokenCloneCommand_Pretty: Started DNS server: address=127.0.0.1:11028 network=udp === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.065Z [INFO] TestTokenCloneCommand_JSON: Started DNS server: address=127.0.0.1:11022 network=tcp === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.068Z [INFO] TestTokenCloneCommand_Pretty.server: Adding LAN server: server="Node-da180c94-a2fb-f710-f6d3-69607f1fe8df (Addr: tcp/127.0.0.1:11033) (DC: dc1)" === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.069Z [INFO] TestTokenCloneCommand_JSON: Started DNS server: address=127.0.0.1:11022 network=udp === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.068Z [INFO] TestTokenCloneCommand_Pretty.server: Handled event for server in area: event=member-join server=Node-da180c94-a2fb-f710-f6d3-69607f1fe8df.dc1 area=wan writer.go:29: 2021-01-29T19:32:26.069Z [INFO] TestTokenCloneCommand_Pretty: Started DNS server: address=127.0.0.1:11028 network=tcp === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.072Z [INFO] TestTokenCloneCommand_JSON: Started HTTP server: address=127.0.0.1:11023 network=tcp === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.073Z [INFO] TestTokenCloneCommand_Pretty: Started HTTP server: address=127.0.0.1:11029 network=tcp writer.go:29: 2021-01-29T19:32:26.074Z [INFO] TestTokenCloneCommand_Pretty: started state syncer === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.074Z [INFO] TestTokenCloneCommand_JSON: started state syncer writer.go:29: 2021-01-29T19:32:26.092Z [WARN] TestTokenCloneCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:26.094Z [INFO] TestTokenCloneCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:11027 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:26.096Z [DEBUG] TestTokenCloneCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:26.098Z [DEBUG] TestTokenCloneCommand_JSON.server.raft: vote granted: from=aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 term=2 tally=1 writer.go:29: 2021-01-29T19:32:26.099Z [INFO] TestTokenCloneCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:26.101Z [INFO] TestTokenCloneCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:11027 [Leader]" writer.go:29: 2021-01-29T19:32:26.102Z [INFO] TestTokenCloneCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:26.104Z [INFO] TestTokenCloneCommand_JSON.server: New leader elected: payload=Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 writer.go:29: 2021-01-29T19:32:26.105Z [INFO] TestTokenCloneCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:32:26.112Z [INFO] TestTokenCloneCommand_JSON.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:26.113Z [WARN] TestTokenCloneCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:26.115Z [INFO] TestTokenCloneCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:32:26.117Z [WARN] TestTokenCloneCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:26.116Z [INFO] TestTokenCloneCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.118Z [WARN] TestTokenCloneCommand_Pretty.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.121Z [INFO] TestTokenCloneCommand_JSON.server: Created ACL anonymous token from configuration === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.119Z [INFO] TestTokenCloneCommand_Pretty.server.raft: entering candidate state: node="Node at 127.0.0.1:11033 [Candidate]" term=2 === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.122Z [INFO] TestTokenCloneCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.120Z [DEBUG] TestTokenCloneCommand_Pretty.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.123Z [INFO] TestTokenCloneCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:26.124Z [INFO] TestTokenCloneCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 writer.go:29: 2021-01-29T19:32:26.123Z [INFO] TestTokenCloneCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:26.127Z [DEBUG] TestTokenCloneCommand_JSON.server: transitioning out of legacy ACL mode === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.126Z [DEBUG] TestTokenCloneCommand_Pretty.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:26.128Z [DEBUG] TestTokenCloneCommand_Pretty.server.raft: vote granted: from=da180c94-a2fb-f710-f6d3-69607f1fe8df term=2 tally=1 writer.go:29: 2021-01-29T19:32:26.130Z [INFO] TestTokenCloneCommand_Pretty.server.raft: election won: tally=1 === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.128Z [INFO] TestTokenCloneCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.131Z [INFO] TestTokenCloneCommand_Pretty.server.raft: entering leader state: leader="Node at 127.0.0.1:11033 [Leader]" === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.126Z [INFO] TestTokenCloneCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1.dc1 === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.133Z [INFO] TestTokenCloneCommand_Pretty.server: cluster leadership acquired === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.126Z [INFO] TestTokenCloneCommand_JSON.server: Updating LAN server: server="Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 (Addr: tcp/127.0.0.1:11027) (DC: dc1)" === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.135Z [INFO] TestTokenCloneCommand_Pretty.server: New leader elected: payload=Node-da180c94-a2fb-f710-f6d3-69607f1fe8df === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.135Z [INFO] TestTokenCloneCommand_JSON.server: Updating LAN server: server="Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 (Addr: tcp/127.0.0.1:11027) (DC: dc1)" writer.go:29: 2021-01-29T19:32:26.137Z [INFO] TestTokenCloneCommand_JSON.server: Handled event for server in area: event=member-update server=Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1.dc1 area=wan writer.go:29: 2021-01-29T19:32:26.138Z [INFO] TestTokenCloneCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1.dc1 writer.go:29: 2021-01-29T19:32:26.139Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:26.151Z [INFO] TestTokenCloneCommand_JSON.server: Handled event for server in area: event=member-update server=Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1.dc1 area=wan === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.152Z [INFO] TestTokenCloneCommand_Pretty.server: initializing acls writer.go:29: 2021-01-29T19:32:26.157Z [INFO] TestTokenCloneCommand_Pretty.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:26.159Z [WARN] TestTokenCloneCommand_Pretty.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:26.161Z [INFO] TestTokenCloneCommand_Pretty.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:26.164Z [INFO] TestTokenCloneCommand_Pretty.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:26.166Z [INFO] TestTokenCloneCommand_Pretty.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:26.171Z [INFO] TestTokenCloneCommand_Pretty.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:26.173Z [INFO] TestTokenCloneCommand_Pretty.server.serf.lan: serf: EventMemberUpdate: Node-da180c94-a2fb-f710-f6d3-69607f1fe8df writer.go:29: 2021-01-29T19:32:26.176Z [INFO] TestTokenCloneCommand_Pretty.server: Updating LAN server: server="Node-da180c94-a2fb-f710-f6d3-69607f1fe8df (Addr: tcp/127.0.0.1:11033) (DC: dc1)" writer.go:29: 2021-01-29T19:32:26.176Z [INFO] TestTokenCloneCommand_Pretty.server.serf.wan: serf: EventMemberUpdate: Node-da180c94-a2fb-f710-f6d3-69607f1fe8df.dc1 === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.176Z [INFO] TestTokenCloneCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.178Z [INFO] TestTokenCloneCommand_Pretty.server: Handled event for server in area: event=member-update server=Node-da180c94-a2fb-f710-f6d3-69607f1fe8df.dc1 area=wan === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.178Z [INFO] TestTokenCloneCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.179Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.180Z [DEBUG] TestTokenCloneCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 writer.go:29: 2021-01-29T19:32:26.181Z [INFO] TestTokenCloneCommand_JSON.server: member joined, marking health alive: member=Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.183Z [INFO] TestTokenCloneCommand_Pretty.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:26.184Z [INFO] TestTokenCloneCommand_Pretty.leader: started routine: routine="CA root pruning" === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.184Z [DEBUG] TestTokenCloneCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.185Z [DEBUG] TestTokenCloneCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-da180c94-a2fb-f710-f6d3-69607f1fe8df writer.go:29: 2021-01-29T19:32:26.186Z [INFO] TestTokenCloneCommand_Pretty.server: member joined, marking health alive: member=Node-da180c94-a2fb-f710-f6d3-69607f1fe8df === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.185Z [DEBUG] TestTokenCloneCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-aa46b9d5-10b9-1b63-d0bb-fe3ffeff73d1 === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.193Z [DEBUG] TestTokenCloneCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-da180c94-a2fb-f710-f6d3-69607f1fe8df writer.go:29: 2021-01-29T19:32:26.389Z [DEBUG] TestTokenCloneCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:38936 latency=8.087983ms writer.go:29: 2021-01-29T19:32:26.397Z [DEBUG] TestTokenCloneCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:38936 latency=2.820925ms === RUN TestTokenCloneCommand_Pretty/Description === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.404Z [DEBUG] TestTokenCloneCommand_Pretty: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:26.406Z [INFO] TestTokenCloneCommand_Pretty: Synced node info writer.go:29: 2021-01-29T19:32:26.407Z [DEBUG] TestTokenCloneCommand_Pretty: Node info in sync writer.go:29: 2021-01-29T19:32:26.407Z [DEBUG] TestTokenCloneCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token/8c33f200-67dc-cbea-48d0-425d3beb0cd4/clone from=127.0.0.1:38938 latency=3.118438ms === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.425Z [DEBUG] TestTokenCloneCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:54228 latency=2.553678ms writer.go:29: 2021-01-29T19:32:26.429Z [DEBUG] TestTokenCloneCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:54228 latency=1.317155ms === RUN TestTokenCloneCommand_JSON/Description === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.437Z [DEBUG] TestTokenCloneCommand_JSON: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.446Z [DEBUG] TestTokenCloneCommand_Pretty.http: Request finished: method=GET url=/v1/acl/token/71aeb5aa-e1b8-cbd8-e844-a5ca39ff2f0f from=127.0.0.1:38936 latency=17.719675ms === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.455Z [INFO] TestTokenCloneCommand_JSON: Synced node info writer.go:29: 2021-01-29T19:32:26.458Z [DEBUG] TestTokenCloneCommand_JSON: Node info in sync === RUN TestTokenCloneCommand_Pretty/Without_Description === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.479Z [DEBUG] TestTokenCloneCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token/c2eeae89-90af-43a9-0ee8-b1ae26829343/clone from=127.0.0.1:54230 latency=15.170202ms === RUN TestTokenCloneCommand_JSON/Without_Description === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.507Z [DEBUG] TestTokenCloneCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token/c2eeae89-90af-43a9-0ee8-b1ae26829343/clone from=127.0.0.1:54234 latency=2.605231ms writer.go:29: 2021-01-29T19:32:26.511Z [INFO] TestTokenCloneCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:32:26.512Z [INFO] TestTokenCloneCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:26.513Z [DEBUG] TestTokenCloneCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:26.515Z [DEBUG] TestTokenCloneCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:26.517Z [DEBUG] TestTokenCloneCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:26.516Z [DEBUG] TestTokenCloneCommand_JSON.leader: stopped routine: routine="CA root pruning" === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.522Z [DEBUG] TestTokenCloneCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token/8c33f200-67dc-cbea-48d0-425d3beb0cd4/clone from=127.0.0.1:38944 latency=44.829211ms === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.519Z [DEBUG] TestTokenCloneCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:26.520Z [WARN] TestTokenCloneCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:26.520Z [DEBUG] TestTokenCloneCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:26.528Z [WARN] TestTokenCloneCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:26.531Z [INFO] TestTokenCloneCommand_JSON: consul server down === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.529Z [DEBUG] TestTokenCloneCommand_Pretty.http: Request finished: method=GET url=/v1/acl/token/324750a6-e993-0629-44d0-f7a6a3eb8a74 from=127.0.0.1:38936 latency=459.823µs === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:26.533Z [INFO] TestTokenCloneCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:32:26.531Z [INFO] TestTokenCloneCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:26.535Z [INFO] TestTokenCloneCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:11022 network=tcp writer.go:29: 2021-01-29T19:32:26.539Z [INFO] TestTokenCloneCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:11022 network=udp writer.go:29: 2021-01-29T19:32:26.541Z [INFO] TestTokenCloneCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:11023 network=tcp === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:26.544Z [INFO] TestTokenCloneCommand_Pretty: Requesting shutdown writer.go:29: 2021-01-29T19:32:26.546Z [INFO] TestTokenCloneCommand_Pretty.server: shutting down server writer.go:29: 2021-01-29T19:32:26.548Z [DEBUG] TestTokenCloneCommand_Pretty.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:26.551Z [DEBUG] TestTokenCloneCommand_Pretty.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:26.552Z [DEBUG] TestTokenCloneCommand_Pretty.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:26.551Z [DEBUG] TestTokenCloneCommand_Pretty.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:26.553Z [DEBUG] TestTokenCloneCommand_Pretty.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:26.555Z [DEBUG] TestTokenCloneCommand_Pretty.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:26.555Z [WARN] TestTokenCloneCommand_Pretty.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:26.575Z [WARN] TestTokenCloneCommand_Pretty.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:26.583Z [INFO] TestTokenCloneCommand_Pretty.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:26.588Z [INFO] TestTokenCloneCommand_Pretty: consul server down writer.go:29: 2021-01-29T19:32:26.588Z [INFO] TestTokenCloneCommand_Pretty: shutdown complete writer.go:29: 2021-01-29T19:32:26.589Z [INFO] TestTokenCloneCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:11028 network=tcp writer.go:29: 2021-01-29T19:32:26.590Z [INFO] TestTokenCloneCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:11028 network=udp writer.go:29: 2021-01-29T19:32:26.591Z [INFO] TestTokenCloneCommand_Pretty: Stopping server: protocol=HTTP address=127.0.0.1:11029 network=tcp === CONT TestTokenCloneCommand_JSON writer.go:29: 2021-01-29T19:32:27.046Z [INFO] TestTokenCloneCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:27.048Z [INFO] TestTokenCloneCommand_JSON: Endpoints down --- PASS: TestTokenCloneCommand_JSON (1.07s) --- PASS: TestTokenCloneCommand_JSON/Description (0.04s) --- PASS: TestTokenCloneCommand_JSON/Without_Description (0.02s) === CONT TestTokenCloneCommand_Pretty writer.go:29: 2021-01-29T19:32:27.092Z [INFO] TestTokenCloneCommand_Pretty: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:27.093Z [INFO] TestTokenCloneCommand_Pretty: Endpoints down --- PASS: TestTokenCloneCommand_Pretty (1.11s) --- PASS: TestTokenCloneCommand_Pretty/Description (0.06s) --- PASS: TestTokenCloneCommand_Pretty/Without_Description (0.08s) PASS ok github.com/hashicorp/consul/command/acl/token/clone 1.255s === RUN TestTokenCreateCommand_noTabs === PAUSE TestTokenCreateCommand_noTabs === RUN TestTokenCreateCommand_Pretty === PAUSE TestTokenCreateCommand_Pretty === RUN TestTokenCreateCommand_JSON === PAUSE TestTokenCreateCommand_JSON === CONT TestTokenCreateCommand_noTabs === CONT TestTokenCreateCommand_Pretty [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestTokenCreateCommand_noTabs (0.00s) === CONT TestTokenCreateCommand_JSON [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.769Z [WARN] TestTokenCreateCommand_Pretty: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:28.770Z [DEBUG] TestTokenCreateCommand_Pretty.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:28.771Z [DEBUG] TestTokenCreateCommand_Pretty.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:28.776Z [INFO] TestTokenCreateCommand_Pretty.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c289919f-c28d-fbf3-5796-d5d40a528af7 Address:127.0.0.1:20216}]" writer.go:29: 2021-01-29T19:32:28.778Z [INFO] TestTokenCreateCommand_Pretty.server.raft: entering follower state: follower="Node at 127.0.0.1:20216 [Follower]" leader= writer.go:29: 2021-01-29T19:32:28.779Z [INFO] TestTokenCreateCommand_Pretty.server.serf.wan: serf: EventMemberJoin: Node-c289919f-c28d-fbf3-5796-d5d40a528af7.dc1 127.0.0.1 === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.784Z [WARN] TestTokenCreateCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:28.785Z [DEBUG] TestTokenCreateCommand_JSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:28.786Z [DEBUG] TestTokenCreateCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:28.789Z [INFO] TestTokenCreateCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:12568595-d6c9-ca87-1ba4-c5c2ffe62910 Address:127.0.0.1:20222}]" === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.788Z [INFO] TestTokenCreateCommand_Pretty.server.serf.lan: serf: EventMemberJoin: Node-c289919f-c28d-fbf3-5796-d5d40a528af7 127.0.0.1 writer.go:29: 2021-01-29T19:32:28.795Z [INFO] TestTokenCreateCommand_Pretty: Started DNS server: address=127.0.0.1:20211 network=udp writer.go:29: 2021-01-29T19:32:28.798Z [INFO] TestTokenCreateCommand_Pretty.server: Adding LAN server: server="Node-c289919f-c28d-fbf3-5796-d5d40a528af7 (Addr: tcp/127.0.0.1:20216) (DC: dc1)" writer.go:29: 2021-01-29T19:32:28.799Z [INFO] TestTokenCreateCommand_Pretty.server: Handled event for server in area: event=member-join server=Node-c289919f-c28d-fbf3-5796-d5d40a528af7.dc1 area=wan === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.802Z [INFO] TestTokenCreateCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910.dc1 127.0.0.1 === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.804Z [INFO] TestTokenCreateCommand_Pretty: Started DNS server: address=127.0.0.1:20211 network=tcp writer.go:29: 2021-01-29T19:32:28.807Z [INFO] TestTokenCreateCommand_Pretty: Started HTTP server: address=127.0.0.1:20212 network=tcp === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.805Z [INFO] TestTokenCreateCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910 127.0.0.1 === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.807Z [INFO] TestTokenCreateCommand_Pretty: started state syncer === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.811Z [INFO] TestTokenCreateCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:20222 [Follower]" leader= writer.go:29: 2021-01-29T19:32:28.811Z [INFO] TestTokenCreateCommand_JSON: Started DNS server: address=127.0.0.1:20217 network=udp writer.go:29: 2021-01-29T19:32:28.817Z [INFO] TestTokenCreateCommand_JSON.server: Handled event for server in area: event=member-join server=Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910.dc1 area=wan writer.go:29: 2021-01-29T19:32:28.822Z [INFO] TestTokenCreateCommand_JSON.server: Adding LAN server: server="Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910 (Addr: tcp/127.0.0.1:20222) (DC: dc1)" writer.go:29: 2021-01-29T19:32:28.824Z [INFO] TestTokenCreateCommand_JSON: Started DNS server: address=127.0.0.1:20217 network=tcp writer.go:29: 2021-01-29T19:32:28.825Z [INFO] TestTokenCreateCommand_JSON: Started HTTP server: address=127.0.0.1:20218 network=tcp writer.go:29: 2021-01-29T19:32:28.826Z [INFO] TestTokenCreateCommand_JSON: started state syncer === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.844Z [WARN] TestTokenCreateCommand_Pretty.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:28.845Z [INFO] TestTokenCreateCommand_Pretty.server.raft: entering candidate state: node="Node at 127.0.0.1:20216 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:28.847Z [DEBUG] TestTokenCreateCommand_Pretty.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:28.848Z [DEBUG] TestTokenCreateCommand_Pretty.server.raft: vote granted: from=c289919f-c28d-fbf3-5796-d5d40a528af7 term=2 tally=1 writer.go:29: 2021-01-29T19:32:28.849Z [INFO] TestTokenCreateCommand_Pretty.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:28.850Z [INFO] TestTokenCreateCommand_Pretty.server.raft: entering leader state: leader="Node at 127.0.0.1:20216 [Leader]" writer.go:29: 2021-01-29T19:32:28.851Z [INFO] TestTokenCreateCommand_Pretty.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:28.852Z [INFO] TestTokenCreateCommand_Pretty.server: initializing acls writer.go:29: 2021-01-29T19:32:28.854Z [INFO] TestTokenCreateCommand_Pretty.server: New leader elected: payload=Node-c289919f-c28d-fbf3-5796-d5d40a528af7 writer.go:29: 2021-01-29T19:32:28.857Z [INFO] TestTokenCreateCommand_Pretty.server: initializing acls writer.go:29: 2021-01-29T19:32:28.858Z [WARN] TestTokenCreateCommand_Pretty.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:28.857Z [INFO] TestTokenCreateCommand_Pretty.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:28.859Z [WARN] TestTokenCreateCommand_Pretty.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:28.860Z [INFO] TestTokenCreateCommand_Pretty.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:28.861Z [INFO] TestTokenCreateCommand_Pretty.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:28.862Z [INFO] TestTokenCreateCommand_Pretty.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:28.862Z [INFO] TestTokenCreateCommand_Pretty.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:28.863Z [INFO] TestTokenCreateCommand_Pretty.leader: started routine: routine="acl token reaping" === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.864Z [WARN] TestTokenCreateCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.865Z [DEBUG] TestTokenCreateCommand_Pretty.server: transitioning out of legacy ACL mode === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.865Z [INFO] TestTokenCreateCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:20222 [Candidate]" term=2 === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.865Z [INFO] TestTokenCreateCommand_Pretty.server.serf.lan: serf: EventMemberUpdate: Node-c289919f-c28d-fbf3-5796-d5d40a528af7 writer.go:29: 2021-01-29T19:32:28.867Z [INFO] TestTokenCreateCommand_Pretty.server: Updating LAN server: server="Node-c289919f-c28d-fbf3-5796-d5d40a528af7 (Addr: tcp/127.0.0.1:20216) (DC: dc1)" writer.go:29: 2021-01-29T19:32:28.867Z [INFO] TestTokenCreateCommand_Pretty.server.serf.wan: serf: EventMemberUpdate: Node-c289919f-c28d-fbf3-5796-d5d40a528af7.dc1 writer.go:29: 2021-01-29T19:32:28.867Z [INFO] TestTokenCreateCommand_Pretty.server.serf.lan: serf: EventMemberUpdate: Node-c289919f-c28d-fbf3-5796-d5d40a528af7 === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.870Z [DEBUG] TestTokenCreateCommand_JSON.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.870Z [INFO] TestTokenCreateCommand_Pretty.server.serf.wan: serf: EventMemberUpdate: Node-c289919f-c28d-fbf3-5796-d5d40a528af7.dc1 writer.go:29: 2021-01-29T19:32:28.869Z [INFO] TestTokenCreateCommand_Pretty.server: Handled event for server in area: event=member-update server=Node-c289919f-c28d-fbf3-5796-d5d40a528af7.dc1 area=wan === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.871Z [DEBUG] TestTokenCreateCommand_JSON.server.raft: votes: needed=1 === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.872Z [INFO] TestTokenCreateCommand_Pretty.server: Handled event for server in area: event=member-update server=Node-c289919f-c28d-fbf3-5796-d5d40a528af7.dc1 area=wan === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.872Z [DEBUG] TestTokenCreateCommand_JSON.server.raft: vote granted: from=12568595-d6c9-ca87-1ba4-c5c2ffe62910 term=2 tally=1 === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.871Z [INFO] TestTokenCreateCommand_Pretty.server: Updating LAN server: server="Node-c289919f-c28d-fbf3-5796-d5d40a528af7 (Addr: tcp/127.0.0.1:20216) (DC: dc1)" === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.873Z [INFO] TestTokenCreateCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:28.874Z [INFO] TestTokenCreateCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:20222 [Leader]" writer.go:29: 2021-01-29T19:32:28.875Z [INFO] TestTokenCreateCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:28.876Z [INFO] TestTokenCreateCommand_JSON.server: New leader elected: payload=Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910 === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.878Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.886Z [INFO] TestTokenCreateCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:32:28.889Z [INFO] TestTokenCreateCommand_JSON.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:28.890Z [WARN] TestTokenCreateCommand_JSON.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:28.901Z [INFO] TestTokenCreateCommand_JSON.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:28.904Z [INFO] TestTokenCreateCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:28.905Z [INFO] TestTokenCreateCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:28.906Z [INFO] TestTokenCreateCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:28.907Z [INFO] TestTokenCreateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910 writer.go:29: 2021-01-29T19:32:28.909Z [INFO] TestTokenCreateCommand_JSON.server: Updating LAN server: server="Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910 (Addr: tcp/127.0.0.1:20222) (DC: dc1)" writer.go:29: 2021-01-29T19:32:28.909Z [INFO] TestTokenCreateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910.dc1 writer.go:29: 2021-01-29T19:32:28.911Z [INFO] TestTokenCreateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910.dc1 area=wan writer.go:29: 2021-01-29T19:32:28.915Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.917Z [INFO] TestTokenCreateCommand_Pretty.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:28.918Z [INFO] TestTokenCreateCommand_Pretty.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:28.919Z [DEBUG] TestTokenCreateCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-c289919f-c28d-fbf3-5796-d5d40a528af7 writer.go:29: 2021-01-29T19:32:28.920Z [INFO] TestTokenCreateCommand_Pretty.server: member joined, marking health alive: member=Node-c289919f-c28d-fbf3-5796-d5d40a528af7 writer.go:29: 2021-01-29T19:32:28.922Z [DEBUG] TestTokenCreateCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-c289919f-c28d-fbf3-5796-d5d40a528af7 === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.922Z [INFO] TestTokenCreateCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.923Z [DEBUG] TestTokenCreateCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-c289919f-c28d-fbf3-5796-d5d40a528af7 === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.923Z [INFO] TestTokenCreateCommand_JSON.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:28.924Z [DEBUG] TestTokenCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910 writer.go:29: 2021-01-29T19:32:28.925Z [INFO] TestTokenCreateCommand_JSON.server: member joined, marking health alive: member=Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910 writer.go:29: 2021-01-29T19:32:28.927Z [DEBUG] TestTokenCreateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-12568595-d6c9-ca87-1ba4-c5c2ffe62910 writer.go:29: 2021-01-29T19:32:28.979Z [DEBUG] TestTokenCreateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:51632 latency=3.768577ms writer.go:29: 2021-01-29T19:32:28.994Z [DEBUG] TestTokenCreateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:51634 latency=6.635585ms === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:28.996Z [DEBUG] TestTokenCreateCommand_Pretty: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:28.998Z [INFO] TestTokenCreateCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:32:28.999Z [INFO] TestTokenCreateCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:29.000Z [DEBUG] TestTokenCreateCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:29.001Z [DEBUG] TestTokenCreateCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:29.002Z [DEBUG] TestTokenCreateCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:29.003Z [WARN] TestTokenCreateCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:29.003Z [ERROR] TestTokenCreateCommand_JSON.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:29.003Z [DEBUG] TestTokenCreateCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:29.005Z [INFO] TestTokenCreateCommand_Pretty: Synced node info === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:29.003Z [DEBUG] TestTokenCreateCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:29.003Z [DEBUG] TestTokenCreateCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:29.011Z [WARN] TestTokenCreateCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:29.017Z [INFO] TestTokenCreateCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:29.017Z [INFO] TestTokenCreateCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:29.019Z [INFO] TestTokenCreateCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:32:29.020Z [INFO] TestTokenCreateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:20217 network=tcp writer.go:29: 2021-01-29T19:32:29.020Z [INFO] TestTokenCreateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:20217 network=udp writer.go:29: 2021-01-29T19:32:29.021Z [INFO] TestTokenCreateCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:20218 network=tcp === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:29.226Z [DEBUG] TestTokenCreateCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:44500 latency=2.451483ms writer.go:29: 2021-01-29T19:32:29.233Z [DEBUG] TestTokenCreateCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:44502 latency=2.289621ms writer.go:29: 2021-01-29T19:32:29.242Z [DEBUG] TestTokenCreateCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:44504 latency=4.191396ms writer.go:29: 2021-01-29T19:32:29.249Z [DEBUG] TestTokenCreateCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:44506 latency=2.062643ms writer.go:29: 2021-01-29T19:32:29.256Z [DEBUG] TestTokenCreateCommand_Pretty.http: Request finished: method=GET url=/v1/acl/token/3d852bb8-5153-4388-a3ca-8ca78661889f from=127.0.0.1:44508 latency=1.462308ms writer.go:29: 2021-01-29T19:32:29.261Z [INFO] TestTokenCreateCommand_Pretty: Requesting shutdown writer.go:29: 2021-01-29T19:32:29.263Z [INFO] TestTokenCreateCommand_Pretty.server: shutting down server writer.go:29: 2021-01-29T19:32:29.264Z [DEBUG] TestTokenCreateCommand_Pretty.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:29.266Z [DEBUG] TestTokenCreateCommand_Pretty.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:29.267Z [DEBUG] TestTokenCreateCommand_Pretty.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:29.266Z [DEBUG] TestTokenCreateCommand_Pretty.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:29.267Z [DEBUG] TestTokenCreateCommand_Pretty.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:29.269Z [DEBUG] TestTokenCreateCommand_Pretty.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:29.269Z [WARN] TestTokenCreateCommand_Pretty.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:29.276Z [WARN] TestTokenCreateCommand_Pretty.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:29.279Z [INFO] TestTokenCreateCommand_Pretty.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:29.279Z [INFO] TestTokenCreateCommand_Pretty: consul server down writer.go:29: 2021-01-29T19:32:29.282Z [INFO] TestTokenCreateCommand_Pretty: shutdown complete writer.go:29: 2021-01-29T19:32:29.283Z [INFO] TestTokenCreateCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:20211 network=tcp writer.go:29: 2021-01-29T19:32:29.284Z [INFO] TestTokenCreateCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:20211 network=udp writer.go:29: 2021-01-29T19:32:29.285Z [INFO] TestTokenCreateCommand_Pretty: Stopping server: protocol=HTTP address=127.0.0.1:20212 network=tcp === CONT TestTokenCreateCommand_JSON writer.go:29: 2021-01-29T19:32:29.522Z [INFO] TestTokenCreateCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:29.524Z [INFO] TestTokenCreateCommand_JSON: Endpoints down --- PASS: TestTokenCreateCommand_JSON (0.78s) === CONT TestTokenCreateCommand_Pretty writer.go:29: 2021-01-29T19:32:29.787Z [INFO] TestTokenCreateCommand_Pretty: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:29.788Z [INFO] TestTokenCreateCommand_Pretty: Endpoints down --- PASS: TestTokenCreateCommand_Pretty (1.05s) PASS ok github.com/hashicorp/consul/command/acl/token/create 1.252s === RUN TestTokenDeleteCommand_noTabs === PAUSE TestTokenDeleteCommand_noTabs === RUN TestTokenDeleteCommand === PAUSE TestTokenDeleteCommand === CONT TestTokenDeleteCommand_noTabs === CONT TestTokenDeleteCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestTokenDeleteCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestTokenDeleteCommand writer.go:29: 2021-01-29T19:32:29.864Z [WARN] TestTokenDeleteCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:29.865Z [DEBUG] TestTokenDeleteCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:29.867Z [DEBUG] TestTokenDeleteCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:29.870Z [INFO] TestTokenDeleteCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4e233977-5a93-fe6e-e915-c35d04f9ab70 Address:127.0.0.1:11027}]" writer.go:29: 2021-01-29T19:32:29.872Z [INFO] TestTokenDeleteCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:11027 [Follower]" leader= writer.go:29: 2021-01-29T19:32:29.879Z [INFO] TestTokenDeleteCommand.server.serf.wan: serf: EventMemberJoin: Node-4e233977-5a93-fe6e-e915-c35d04f9ab70.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:29.887Z [INFO] TestTokenDeleteCommand.server.serf.lan: serf: EventMemberJoin: Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 127.0.0.1 writer.go:29: 2021-01-29T19:32:29.889Z [INFO] TestTokenDeleteCommand.server: Handled event for server in area: event=member-join server=Node-4e233977-5a93-fe6e-e915-c35d04f9ab70.dc1 area=wan writer.go:29: 2021-01-29T19:32:29.890Z [INFO] TestTokenDeleteCommand: Started DNS server: address=127.0.0.1:11022 network=udp writer.go:29: 2021-01-29T19:32:29.892Z [INFO] TestTokenDeleteCommand: Started DNS server: address=127.0.0.1:11022 network=tcp writer.go:29: 2021-01-29T19:32:29.891Z [INFO] TestTokenDeleteCommand.server: Adding LAN server: server="Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 (Addr: tcp/127.0.0.1:11027) (DC: dc1)" writer.go:29: 2021-01-29T19:32:29.894Z [INFO] TestTokenDeleteCommand: Started HTTP server: address=127.0.0.1:11023 network=tcp writer.go:29: 2021-01-29T19:32:29.895Z [INFO] TestTokenDeleteCommand: started state syncer writer.go:29: 2021-01-29T19:32:29.922Z [WARN] TestTokenDeleteCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:29.923Z [INFO] TestTokenDeleteCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:11027 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:29.925Z [DEBUG] TestTokenDeleteCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:29.926Z [DEBUG] TestTokenDeleteCommand.server.raft: vote granted: from=4e233977-5a93-fe6e-e915-c35d04f9ab70 term=2 tally=1 writer.go:29: 2021-01-29T19:32:29.927Z [INFO] TestTokenDeleteCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:29.928Z [INFO] TestTokenDeleteCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:11027 [Leader]" writer.go:29: 2021-01-29T19:32:29.930Z [INFO] TestTokenDeleteCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:29.930Z [INFO] TestTokenDeleteCommand.server: New leader elected: payload=Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 writer.go:29: 2021-01-29T19:32:29.931Z [INFO] TestTokenDeleteCommand.server: initializing acls writer.go:29: 2021-01-29T19:32:29.935Z [INFO] TestTokenDeleteCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:29.936Z [WARN] TestTokenDeleteCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:29.938Z [INFO] TestTokenDeleteCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:29.940Z [INFO] TestTokenDeleteCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:29.941Z [INFO] TestTokenDeleteCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:29.940Z [INFO] TestTokenDeleteCommand.server: initializing acls writer.go:29: 2021-01-29T19:32:29.944Z [WARN] TestTokenDeleteCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:29.942Z [INFO] TestTokenDeleteCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:29.946Z [INFO] TestTokenDeleteCommand.server.serf.lan: serf: EventMemberUpdate: Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 writer.go:29: 2021-01-29T19:32:29.948Z [INFO] TestTokenDeleteCommand.server.serf.wan: serf: EventMemberUpdate: Node-4e233977-5a93-fe6e-e915-c35d04f9ab70.dc1 writer.go:29: 2021-01-29T19:32:29.949Z [DEBUG] TestTokenDeleteCommand.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:32:29.949Z [INFO] TestTokenDeleteCommand.server: Updating LAN server: server="Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 (Addr: tcp/127.0.0.1:11027) (DC: dc1)" writer.go:29: 2021-01-29T19:32:29.949Z [INFO] TestTokenDeleteCommand.server: Handled event for server in area: event=member-update server=Node-4e233977-5a93-fe6e-e915-c35d04f9ab70.dc1 area=wan writer.go:29: 2021-01-29T19:32:29.951Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:29.951Z [INFO] TestTokenDeleteCommand.server.serf.lan: serf: EventMemberUpdate: Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 writer.go:29: 2021-01-29T19:32:29.967Z [INFO] TestTokenDeleteCommand.server.serf.wan: serf: EventMemberUpdate: Node-4e233977-5a93-fe6e-e915-c35d04f9ab70.dc1 writer.go:29: 2021-01-29T19:32:29.968Z [INFO] TestTokenDeleteCommand.server: Handled event for server in area: event=member-update server=Node-4e233977-5a93-fe6e-e915-c35d04f9ab70.dc1 area=wan writer.go:29: 2021-01-29T19:32:29.969Z [INFO] TestTokenDeleteCommand.server: Updating LAN server: server="Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 (Addr: tcp/127.0.0.1:11027) (DC: dc1)" writer.go:29: 2021-01-29T19:32:30.001Z [INFO] TestTokenDeleteCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:30.002Z [INFO] TestTokenDeleteCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:30.003Z [DEBUG] TestTokenDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 writer.go:29: 2021-01-29T19:32:30.004Z [INFO] TestTokenDeleteCommand.server: member joined, marking health alive: member=Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 writer.go:29: 2021-01-29T19:32:30.007Z [DEBUG] TestTokenDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 writer.go:29: 2021-01-29T19:32:30.008Z [DEBUG] TestTokenDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-4e233977-5a93-fe6e-e915-c35d04f9ab70 writer.go:29: 2021-01-29T19:32:30.016Z [DEBUG] TestTokenDeleteCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:30.020Z [INFO] TestTokenDeleteCommand: Synced node info writer.go:29: 2021-01-29T19:32:30.023Z [DEBUG] TestTokenDeleteCommand: Node info in sync writer.go:29: 2021-01-29T19:32:30.243Z [DEBUG] TestTokenDeleteCommand.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:54250 latency=11.613989ms writer.go:29: 2021-01-29T19:32:30.251Z [DEBUG] TestTokenDeleteCommand.http: Request finished: method=DELETE url=/v1/acl/token/a8aa8da4-8a42-5af9-c8c0-1142b01baec6 from=127.0.0.1:54252 latency=3.680179ms writer.go:29: 2021-01-29T19:32:30.256Z [ERROR] TestTokenDeleteCommand.http: Request error: method=GET url=/v1/acl/token/a8aa8da4-8a42-5af9-c8c0-1142b01baec6 from=127.0.0.1:54250 error="ACL not found" writer.go:29: 2021-01-29T19:32:30.258Z [DEBUG] TestTokenDeleteCommand.http: Request finished: method=GET url=/v1/acl/token/a8aa8da4-8a42-5af9-c8c0-1142b01baec6 from=127.0.0.1:54250 latency=1.970117ms writer.go:29: 2021-01-29T19:32:30.268Z [INFO] TestTokenDeleteCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:30.270Z [INFO] TestTokenDeleteCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:30.271Z [DEBUG] TestTokenDeleteCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:30.272Z [DEBUG] TestTokenDeleteCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:30.273Z [DEBUG] TestTokenDeleteCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:30.274Z [WARN] TestTokenDeleteCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:30.272Z [DEBUG] TestTokenDeleteCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:30.273Z [DEBUG] TestTokenDeleteCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:30.274Z [DEBUG] TestTokenDeleteCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:30.276Z [WARN] TestTokenDeleteCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:30.279Z [INFO] TestTokenDeleteCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:30.279Z [INFO] TestTokenDeleteCommand: consul server down writer.go:29: 2021-01-29T19:32:30.281Z [INFO] TestTokenDeleteCommand: shutdown complete writer.go:29: 2021-01-29T19:32:30.282Z [INFO] TestTokenDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:11022 network=tcp writer.go:29: 2021-01-29T19:32:30.283Z [INFO] TestTokenDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:11022 network=udp writer.go:29: 2021-01-29T19:32:30.284Z [INFO] TestTokenDeleteCommand: Stopping server: protocol=HTTP address=127.0.0.1:11023 network=tcp writer.go:29: 2021-01-29T19:32:30.785Z [INFO] TestTokenDeleteCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:30.786Z [INFO] TestTokenDeleteCommand: Endpoints down --- PASS: TestTokenDeleteCommand (0.96s) PASS ok github.com/hashicorp/consul/command/acl/token/delete 1.094s === RUN TestTokenListCommand_noTabs === PAUSE TestTokenListCommand_noTabs === RUN TestTokenListCommand_Pretty === PAUSE TestTokenListCommand_Pretty === RUN TestTokenListCommand_JSON === PAUSE TestTokenListCommand_JSON === CONT TestTokenListCommand_noTabs === CONT TestTokenListCommand_JSON === CONT TestTokenListCommand_Pretty [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestTokenListCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.443Z [WARN] TestTokenListCommand_Pretty: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:33.444Z [DEBUG] TestTokenListCommand_Pretty.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:33.446Z [DEBUG] TestTokenListCommand_Pretty.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.445Z [WARN] TestTokenListCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:33.448Z [DEBUG] TestTokenListCommand_JSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:33.450Z [DEBUG] TestTokenListCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.453Z [INFO] TestTokenListCommand_Pretty.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e2ecca89-c793-ba61-f2e1-1955b9ed5090 Address:127.0.0.1:22264}]" writer.go:29: 2021-01-29T19:32:33.454Z [INFO] TestTokenListCommand_Pretty.server.raft: entering follower state: follower="Node at 127.0.0.1:22264 [Follower]" leader= writer.go:29: 2021-01-29T19:32:33.456Z [INFO] TestTokenListCommand_Pretty.server.serf.wan: serf: EventMemberJoin: Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090.dc1 127.0.0.1 === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.457Z [INFO] TestTokenListCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3fc23864-d826-a0f9-8eec-ff08c2e06751 Address:127.0.0.1:22258}]" writer.go:29: 2021-01-29T19:32:33.459Z [INFO] TestTokenListCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:22258 [Follower]" leader= writer.go:29: 2021-01-29T19:32:33.459Z [INFO] TestTokenListCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-3fc23864-d826-a0f9-8eec-ff08c2e06751.dc1 127.0.0.1 === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.461Z [INFO] TestTokenListCommand_Pretty.server.serf.lan: serf: EventMemberJoin: Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090 127.0.0.1 writer.go:29: 2021-01-29T19:32:33.467Z [INFO] TestTokenListCommand_Pretty.server: Adding LAN server: server="Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090 (Addr: tcp/127.0.0.1:22264) (DC: dc1)" writer.go:29: 2021-01-29T19:32:33.468Z [INFO] TestTokenListCommand_Pretty.server: Handled event for server in area: event=member-join server=Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090.dc1 area=wan === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.468Z [INFO] TestTokenListCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-3fc23864-d826-a0f9-8eec-ff08c2e06751 127.0.0.1 writer.go:29: 2021-01-29T19:32:33.477Z [INFO] TestTokenListCommand_JSON: Started DNS server: address=127.0.0.1:22253 network=udp === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.477Z [INFO] TestTokenListCommand_Pretty: Started DNS server: address=127.0.0.1:22259 network=tcp writer.go:29: 2021-01-29T19:32:33.478Z [INFO] TestTokenListCommand_Pretty: Started DNS server: address=127.0.0.1:22259 network=udp === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.480Z [INFO] TestTokenListCommand_JSON.server: Adding LAN server: server="Node-3fc23864-d826-a0f9-8eec-ff08c2e06751 (Addr: tcp/127.0.0.1:22258) (DC: dc1)" writer.go:29: 2021-01-29T19:32:33.482Z [INFO] TestTokenListCommand_JSON.server: Handled event for server in area: event=member-join server=Node-3fc23864-d826-a0f9-8eec-ff08c2e06751.dc1 area=wan writer.go:29: 2021-01-29T19:32:33.494Z [INFO] TestTokenListCommand_JSON: Started DNS server: address=127.0.0.1:22253 network=tcp writer.go:29: 2021-01-29T19:32:33.499Z [INFO] TestTokenListCommand_JSON: Started HTTP server: address=127.0.0.1:22254 network=tcp writer.go:29: 2021-01-29T19:32:33.500Z [INFO] TestTokenListCommand_JSON: started state syncer === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.499Z [INFO] TestTokenListCommand_Pretty: Started HTTP server: address=127.0.0.1:22260 network=tcp writer.go:29: 2021-01-29T19:32:33.502Z [INFO] TestTokenListCommand_Pretty: started state syncer === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.508Z [WARN] TestTokenListCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:33.510Z [INFO] TestTokenListCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:22258 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:33.512Z [DEBUG] TestTokenListCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:33.513Z [DEBUG] TestTokenListCommand_JSON.server.raft: vote granted: from=3fc23864-d826-a0f9-8eec-ff08c2e06751 term=2 tally=1 writer.go:29: 2021-01-29T19:32:33.515Z [INFO] TestTokenListCommand_JSON.server.raft: election won: tally=1 === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.517Z [WARN] TestTokenListCommand_Pretty.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.517Z [INFO] TestTokenListCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:22258 [Leader]" === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.518Z [INFO] TestTokenListCommand_Pretty.server.raft: entering candidate state: node="Node at 127.0.0.1:22264 [Candidate]" term=2 === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.519Z [INFO] TestTokenListCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:33.519Z [INFO] TestTokenListCommand_JSON.server: New leader elected: payload=Node-3fc23864-d826-a0f9-8eec-ff08c2e06751 writer.go:29: 2021-01-29T19:32:33.519Z [INFO] TestTokenListCommand_JSON.server: initializing acls === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.521Z [DEBUG] TestTokenListCommand_Pretty.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.523Z [INFO] TestTokenListCommand_JSON.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:33.523Z [WARN] TestTokenListCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.525Z [DEBUG] TestTokenListCommand_Pretty.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:33.526Z [DEBUG] TestTokenListCommand_Pretty.server.raft: vote granted: from=e2ecca89-c793-ba61-f2e1-1955b9ed5090 term=2 tally=1 writer.go:29: 2021-01-29T19:32:33.527Z [INFO] TestTokenListCommand_Pretty.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:33.528Z [INFO] TestTokenListCommand_Pretty.server.raft: entering leader state: leader="Node at 127.0.0.1:22264 [Leader]" === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.530Z [INFO] TestTokenListCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.529Z [INFO] TestTokenListCommand_Pretty.server: cluster leadership acquired === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.532Z [INFO] TestTokenListCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:33.533Z [INFO] TestTokenListCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:33.534Z [INFO] TestTokenListCommand_JSON.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:33.535Z [INFO] TestTokenListCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-3fc23864-d826-a0f9-8eec-ff08c2e06751 === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.529Z [INFO] TestTokenListCommand_Pretty.server: New leader elected: payload=Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090 === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.538Z [INFO] TestTokenListCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-3fc23864-d826-a0f9-8eec-ff08c2e06751.dc1 === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.530Z [INFO] TestTokenListCommand_Pretty.server: initializing acls === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.538Z [INFO] TestTokenListCommand_JSON.server: Updating LAN server: server="Node-3fc23864-d826-a0f9-8eec-ff08c2e06751 (Addr: tcp/127.0.0.1:22258) (DC: dc1)" writer.go:29: 2021-01-29T19:32:33.540Z [INFO] TestTokenListCommand_JSON.server: Handled event for server in area: event=member-update server=Node-3fc23864-d826-a0f9-8eec-ff08c2e06751.dc1 area=wan === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.543Z [INFO] TestTokenListCommand_Pretty.server: Created ACL 'global-management' policy === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.542Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.544Z [WARN] TestTokenListCommand_Pretty.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:33.549Z [INFO] TestTokenListCommand_Pretty.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:33.551Z [INFO] TestTokenListCommand_Pretty.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:33.553Z [INFO] TestTokenListCommand_Pretty.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:33.556Z [INFO] TestTokenListCommand_Pretty.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:33.559Z [INFO] TestTokenListCommand_Pretty.server.serf.lan: serf: EventMemberUpdate: Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090 writer.go:29: 2021-01-29T19:32:33.563Z [INFO] TestTokenListCommand_Pretty.server: Updating LAN server: server="Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090 (Addr: tcp/127.0.0.1:22264) (DC: dc1)" writer.go:29: 2021-01-29T19:32:33.564Z [INFO] TestTokenListCommand_Pretty.server.serf.wan: serf: EventMemberUpdate: Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090.dc1 writer.go:29: 2021-01-29T19:32:33.568Z [INFO] TestTokenListCommand_Pretty.server: Handled event for server in area: event=member-update server=Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090.dc1 area=wan writer.go:29: 2021-01-29T19:32:33.576Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.586Z [INFO] TestTokenListCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:33.588Z [INFO] TestTokenListCommand_JSON.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:33.590Z [DEBUG] TestTokenListCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-3fc23864-d826-a0f9-8eec-ff08c2e06751 === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.584Z [INFO] TestTokenListCommand_Pretty.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:33.593Z [INFO] TestTokenListCommand_Pretty.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:33.594Z [DEBUG] TestTokenListCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090 === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.592Z [INFO] TestTokenListCommand_JSON.server: member joined, marking health alive: member=Node-3fc23864-d826-a0f9-8eec-ff08c2e06751 === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.596Z [INFO] TestTokenListCommand_Pretty.server: member joined, marking health alive: member=Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090 === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.598Z [DEBUG] TestTokenListCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-3fc23864-d826-a0f9-8eec-ff08c2e06751 === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.599Z [DEBUG] TestTokenListCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-e2ecca89-c793-ba61-f2e1-1955b9ed5090 === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.620Z [DEBUG] TestTokenListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:48172 latency=9.020886ms writer.go:29: 2021-01-29T19:32:33.636Z [DEBUG] TestTokenListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:48172 latency=1.8411ms writer.go:29: 2021-01-29T19:32:33.650Z [DEBUG] TestTokenListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:48172 latency=7.357315ms writer.go:29: 2021-01-29T19:32:33.658Z [DEBUG] TestTokenListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:48172 latency=2.465372ms writer.go:29: 2021-01-29T19:32:33.662Z [DEBUG] TestTokenListCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:48172 latency=2.030166ms writer.go:29: 2021-01-29T19:32:33.672Z [DEBUG] TestTokenListCommand_JSON.http: Request finished: method=GET url=/v1/acl/tokens from=127.0.0.1:48176 latency=823.382µs writer.go:29: 2021-01-29T19:32:33.678Z [INFO] TestTokenListCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:32:33.681Z [INFO] TestTokenListCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:33.683Z [DEBUG] TestTokenListCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:33.685Z [DEBUG] TestTokenListCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:33.688Z [DEBUG] TestTokenListCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:33.688Z [ERROR] TestTokenListCommand_JSON.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:33.688Z [DEBUG] TestTokenListCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:33.689Z [DEBUG] TestTokenListCommand_JSON.leader: stopped routine: routine="acl token reaping" === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.696Z [DEBUG] TestTokenListCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:53390 latency=27.833306ms === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.691Z [WARN] TestTokenListCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:33.695Z [DEBUG] TestTokenListCommand_JSON.leader: stopped routine: routine="CA root pruning" === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.706Z [DEBUG] TestTokenListCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:53390 latency=1.073793ms === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.707Z [WARN] TestTokenListCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:33.709Z [INFO] TestTokenListCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:33.709Z [INFO] TestTokenListCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:33.711Z [INFO] TestTokenListCommand_JSON: shutdown complete === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.710Z [DEBUG] TestTokenListCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:53390 latency=2.105033ms === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.712Z [INFO] TestTokenListCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:22253 network=tcp === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.711Z [DEBUG] TestTokenListCommand_Pretty: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:33.713Z [INFO] TestTokenListCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:22253 network=udp writer.go:29: 2021-01-29T19:32:33.713Z [INFO] TestTokenListCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:22254 network=tcp === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:33.713Z [INFO] TestTokenListCommand_Pretty: Synced node info writer.go:29: 2021-01-29T19:32:33.716Z [DEBUG] TestTokenListCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:53390 latency=575.229µs writer.go:29: 2021-01-29T19:32:33.718Z [DEBUG] TestTokenListCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:53390 latency=583.211µs writer.go:29: 2021-01-29T19:32:33.722Z [DEBUG] TestTokenListCommand_Pretty.http: Request finished: method=GET url=/v1/acl/tokens from=127.0.0.1:53394 latency=95.539µs writer.go:29: 2021-01-29T19:32:33.727Z [INFO] TestTokenListCommand_Pretty: Requesting shutdown writer.go:29: 2021-01-29T19:32:33.728Z [INFO] TestTokenListCommand_Pretty.server: shutting down server writer.go:29: 2021-01-29T19:32:33.729Z [DEBUG] TestTokenListCommand_Pretty.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:33.730Z [DEBUG] TestTokenListCommand_Pretty.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:33.731Z [DEBUG] TestTokenListCommand_Pretty.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:33.732Z [WARN] TestTokenListCommand_Pretty.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:33.730Z [DEBUG] TestTokenListCommand_Pretty.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:33.731Z [DEBUG] TestTokenListCommand_Pretty.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:33.732Z [DEBUG] TestTokenListCommand_Pretty.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:33.734Z [WARN] TestTokenListCommand_Pretty.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:33.739Z [INFO] TestTokenListCommand_Pretty.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:33.740Z [INFO] TestTokenListCommand_Pretty: consul server down writer.go:29: 2021-01-29T19:32:33.741Z [INFO] TestTokenListCommand_Pretty: shutdown complete writer.go:29: 2021-01-29T19:32:33.742Z [INFO] TestTokenListCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:22259 network=tcp writer.go:29: 2021-01-29T19:32:33.743Z [INFO] TestTokenListCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:22259 network=udp writer.go:29: 2021-01-29T19:32:33.743Z [INFO] TestTokenListCommand_Pretty: Stopping server: protocol=HTTP address=127.0.0.1:22260 network=tcp === CONT TestTokenListCommand_JSON writer.go:29: 2021-01-29T19:32:34.214Z [INFO] TestTokenListCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:34.215Z [INFO] TestTokenListCommand_JSON: Endpoints down --- PASS: TestTokenListCommand_JSON (0.82s) === CONT TestTokenListCommand_Pretty writer.go:29: 2021-01-29T19:32:34.245Z [INFO] TestTokenListCommand_Pretty: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:34.246Z [INFO] TestTokenListCommand_Pretty: Endpoints down --- PASS: TestTokenListCommand_Pretty (0.85s) PASS ok github.com/hashicorp/consul/command/acl/token/list 1.001s === RUN TestTokenReadCommand_noTabs === PAUSE TestTokenReadCommand_noTabs === RUN TestTokenReadCommand_Pretty === PAUSE TestTokenReadCommand_Pretty === RUN TestTokenReadCommand_JSON === PAUSE TestTokenReadCommand_JSON === CONT TestTokenReadCommand_noTabs === CONT TestTokenReadCommand_JSON [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestTokenReadCommand_Pretty [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestTokenReadCommand_noTabs (0.01s) === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.453Z [WARN] TestTokenReadCommand_Pretty: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:36.455Z [DEBUG] TestTokenReadCommand_Pretty.tlsutil: Update: version=1 === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.454Z [WARN] TestTokenReadCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:36.459Z [DEBUG] TestTokenReadCommand_JSON.tlsutil: Update: version=1 === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.459Z [DEBUG] TestTokenReadCommand_Pretty.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:36.464Z [INFO] TestTokenReadCommand_Pretty.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:dd30eaa3-6501-4744-2e8c-f9e516217b1a Address:127.0.0.1:16138}]" writer.go:29: 2021-01-29T19:32:36.466Z [INFO] TestTokenReadCommand_Pretty.server.raft: entering follower state: follower="Node at 127.0.0.1:16138 [Follower]" leader= === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.464Z [DEBUG] TestTokenReadCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.467Z [INFO] TestTokenReadCommand_Pretty.server.serf.wan: serf: EventMemberJoin: Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:36.476Z [INFO] TestTokenReadCommand_Pretty.server.serf.lan: serf: EventMemberJoin: Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a 127.0.0.1 === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.477Z [INFO] TestTokenReadCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a6d17a97-181b-5a10-ec96-48dd4f86c28c Address:127.0.0.1:16132}]" === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.480Z [INFO] TestTokenReadCommand_Pretty: Started DNS server: address=127.0.0.1:16133 network=udp writer.go:29: 2021-01-29T19:32:36.482Z [INFO] TestTokenReadCommand_Pretty: Started DNS server: address=127.0.0.1:16133 network=tcp === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.480Z [INFO] TestTokenReadCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:16132 [Follower]" leader= writer.go:29: 2021-01-29T19:32:36.486Z [INFO] TestTokenReadCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c.dc1 127.0.0.1 === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.484Z [INFO] TestTokenReadCommand_Pretty: Started HTTP server: address=127.0.0.1:16134 network=tcp writer.go:29: 2021-01-29T19:32:36.488Z [INFO] TestTokenReadCommand_Pretty: started state syncer writer.go:29: 2021-01-29T19:32:36.485Z [INFO] TestTokenReadCommand_Pretty.server: Adding LAN server: server="Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a (Addr: tcp/127.0.0.1:16138) (DC: dc1)" writer.go:29: 2021-01-29T19:32:36.485Z [INFO] TestTokenReadCommand_Pretty.server: Handled event for server in area: event=member-join server=Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a.dc1 area=wan === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.512Z [INFO] TestTokenReadCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c 127.0.0.1 writer.go:29: 2021-01-29T19:32:36.514Z [INFO] TestTokenReadCommand_JSON.server: Handled event for server in area: event=member-join server=Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c.dc1 area=wan writer.go:29: 2021-01-29T19:32:36.515Z [INFO] TestTokenReadCommand_JSON: Started DNS server: address=127.0.0.1:16127 network=udp writer.go:29: 2021-01-29T19:32:36.516Z [INFO] TestTokenReadCommand_JSON: Started DNS server: address=127.0.0.1:16127 network=tcp writer.go:29: 2021-01-29T19:32:36.515Z [INFO] TestTokenReadCommand_JSON.server: Adding LAN server: server="Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c (Addr: tcp/127.0.0.1:16132) (DC: dc1)" writer.go:29: 2021-01-29T19:32:36.518Z [INFO] TestTokenReadCommand_JSON: Started HTTP server: address=127.0.0.1:16128 network=tcp writer.go:29: 2021-01-29T19:32:36.521Z [INFO] TestTokenReadCommand_JSON: started state syncer === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.525Z [WARN] TestTokenReadCommand_Pretty.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:36.527Z [INFO] TestTokenReadCommand_Pretty.server.raft: entering candidate state: node="Node at 127.0.0.1:16138 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:36.530Z [DEBUG] TestTokenReadCommand_Pretty.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:36.535Z [DEBUG] TestTokenReadCommand_Pretty.server.raft: vote granted: from=dd30eaa3-6501-4744-2e8c-f9e516217b1a term=2 tally=1 writer.go:29: 2021-01-29T19:32:36.540Z [INFO] TestTokenReadCommand_Pretty.server.raft: election won: tally=1 === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.544Z [WARN] TestTokenReadCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:36.545Z [INFO] TestTokenReadCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:16132 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:36.547Z [DEBUG] TestTokenReadCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:36.549Z [DEBUG] TestTokenReadCommand_JSON.server.raft: vote granted: from=a6d17a97-181b-5a10-ec96-48dd4f86c28c term=2 tally=1 writer.go:29: 2021-01-29T19:32:36.549Z [INFO] TestTokenReadCommand_JSON.server.raft: election won: tally=1 === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.535Z [DEBUG] TestTokenReadCommand_Pretty.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.550Z [INFO] TestTokenReadCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:16132 [Leader]" === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.544Z [INFO] TestTokenReadCommand_Pretty.server.raft: entering leader state: leader="Node at 127.0.0.1:16138 [Leader]" === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.552Z [INFO] TestTokenReadCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:36.553Z [INFO] TestTokenReadCommand_JSON.server: New leader elected: payload=Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.554Z [INFO] TestTokenReadCommand_Pretty.server: cluster leadership acquired === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.554Z [INFO] TestTokenReadCommand_JSON.server: initializing acls === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.554Z [INFO] TestTokenReadCommand_Pretty.server: New leader elected: payload=Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a writer.go:29: 2021-01-29T19:32:36.554Z [INFO] TestTokenReadCommand_Pretty.server: initializing acls === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.557Z [INFO] TestTokenReadCommand_JSON.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:36.558Z [WARN] TestTokenReadCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.559Z [INFO] TestTokenReadCommand_Pretty.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:36.559Z [WARN] TestTokenReadCommand_Pretty.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:36.561Z [INFO] TestTokenReadCommand_Pretty.server: Bootstrapped ACL master token from configuration === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.562Z [INFO] TestTokenReadCommand_JSON.server: Bootstrapped ACL master token from configuration === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.563Z [INFO] TestTokenReadCommand_Pretty.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:36.564Z [INFO] TestTokenReadCommand_Pretty.leader: started routine: routine="legacy ACL token upgrade" === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.564Z [INFO] TestTokenReadCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:36.565Z [INFO] TestTokenReadCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.565Z [INFO] TestTokenReadCommand_Pretty.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:36.566Z [INFO] TestTokenReadCommand_Pretty.server.serf.lan: serf: EventMemberUpdate: Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.565Z [INFO] TestTokenReadCommand_JSON.leader: started routine: routine="acl token reaping" === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.567Z [INFO] TestTokenReadCommand_Pretty.server.serf.wan: serf: EventMemberUpdate: Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a.dc1 writer.go:29: 2021-01-29T19:32:36.569Z [INFO] TestTokenReadCommand_Pretty.server: Updating LAN server: server="Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a (Addr: tcp/127.0.0.1:16138) (DC: dc1)" === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.568Z [INFO] TestTokenReadCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c writer.go:29: 2021-01-29T19:32:36.570Z [INFO] TestTokenReadCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c.dc1 writer.go:29: 2021-01-29T19:32:36.570Z [INFO] TestTokenReadCommand_JSON.server: Updating LAN server: server="Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c (Addr: tcp/127.0.0.1:16132) (DC: dc1)" === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.569Z [INFO] TestTokenReadCommand_Pretty.server: Handled event for server in area: event=member-update server=Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a.dc1 area=wan writer.go:29: 2021-01-29T19:32:36.574Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.573Z [INFO] TestTokenReadCommand_JSON.server: Handled event for server in area: event=member-update server=Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c.dc1 area=wan writer.go:29: 2021-01-29T19:32:36.575Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:36.611Z [INFO] TestTokenReadCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.611Z [INFO] TestTokenReadCommand_Pretty.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.613Z [INFO] TestTokenReadCommand_JSON.leader: started routine: routine="CA root pruning" === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.613Z [INFO] TestTokenReadCommand_Pretty.leader: started routine: routine="CA root pruning" === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.615Z [DEBUG] TestTokenReadCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.615Z [DEBUG] TestTokenReadCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.616Z [INFO] TestTokenReadCommand_JSON.server: member joined, marking health alive: member=Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.617Z [INFO] TestTokenReadCommand_Pretty.server: member joined, marking health alive: member=Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a writer.go:29: 2021-01-29T19:32:36.621Z [DEBUG] TestTokenReadCommand_Pretty.server: Skipping self join check for node since the cluster is too small: node=Node-dd30eaa3-6501-4744-2e8c-f9e516217b1a === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.625Z [DEBUG] TestTokenReadCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-a6d17a97-181b-5a10-ec96-48dd4f86c28c writer.go:29: 2021-01-29T19:32:36.646Z [DEBUG] TestTokenReadCommand_JSON: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:36.650Z [INFO] TestTokenReadCommand_JSON: Synced node info === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:36.751Z [DEBUG] TestTokenReadCommand_Pretty.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:36606 latency=7.297654ms writer.go:29: 2021-01-29T19:32:36.763Z [DEBUG] TestTokenReadCommand_Pretty.http: Request finished: method=GET url=/v1/acl/token/7be7a8b5-cbfb-a5d7-9950-95cb48819fd3 from=127.0.0.1:36608 latency=5.21209ms writer.go:29: 2021-01-29T19:32:36.768Z [INFO] TestTokenReadCommand_Pretty: Requesting shutdown writer.go:29: 2021-01-29T19:32:36.769Z [INFO] TestTokenReadCommand_Pretty.server: shutting down server writer.go:29: 2021-01-29T19:32:36.769Z [DEBUG] TestTokenReadCommand_Pretty.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:36.770Z [DEBUG] TestTokenReadCommand_Pretty.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:36.771Z [DEBUG] TestTokenReadCommand_Pretty.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:36.782Z [WARN] TestTokenReadCommand_Pretty.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:36.776Z [ERROR] TestTokenReadCommand_Pretty.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:36.777Z [DEBUG] TestTokenReadCommand_Pretty.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:36.777Z [DEBUG] TestTokenReadCommand_Pretty.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:36.786Z [DEBUG] TestTokenReadCommand_Pretty.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:36.787Z [WARN] TestTokenReadCommand_Pretty.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:36.790Z [INFO] TestTokenReadCommand_Pretty.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:36.792Z [INFO] TestTokenReadCommand_Pretty: consul server down writer.go:29: 2021-01-29T19:32:36.793Z [INFO] TestTokenReadCommand_Pretty: shutdown complete writer.go:29: 2021-01-29T19:32:36.794Z [INFO] TestTokenReadCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:16133 network=tcp writer.go:29: 2021-01-29T19:32:36.796Z [INFO] TestTokenReadCommand_Pretty: Stopping server: protocol=DNS address=127.0.0.1:16133 network=udp writer.go:29: 2021-01-29T19:32:36.797Z [INFO] TestTokenReadCommand_Pretty: Stopping server: protocol=HTTP address=127.0.0.1:16134 network=tcp === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:36.956Z [DEBUG] TestTokenReadCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:47750 latency=3.762204ms writer.go:29: 2021-01-29T19:32:36.963Z [DEBUG] TestTokenReadCommand_JSON.http: Request finished: method=GET url=/v1/acl/token/e7fb62fc-6b4d-0ae0-54b0-b362a92357ff from=127.0.0.1:47752 latency=2.51841ms writer.go:29: 2021-01-29T19:32:36.966Z [INFO] TestTokenReadCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:32:36.967Z [INFO] TestTokenReadCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:36.968Z [DEBUG] TestTokenReadCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:36.968Z [DEBUG] TestTokenReadCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:36.969Z [DEBUG] TestTokenReadCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:36.970Z [WARN] TestTokenReadCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:36.968Z [DEBUG] TestTokenReadCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:36.969Z [DEBUG] TestTokenReadCommand_JSON.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:36.970Z [DEBUG] TestTokenReadCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:36.973Z [WARN] TestTokenReadCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:36.992Z [INFO] TestTokenReadCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:36.993Z [INFO] TestTokenReadCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:36.997Z [INFO] TestTokenReadCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:32:36.999Z [INFO] TestTokenReadCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:16127 network=tcp writer.go:29: 2021-01-29T19:32:37.002Z [INFO] TestTokenReadCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:16127 network=udp writer.go:29: 2021-01-29T19:32:37.004Z [INFO] TestTokenReadCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:16128 network=tcp === CONT TestTokenReadCommand_Pretty writer.go:29: 2021-01-29T19:32:37.298Z [INFO] TestTokenReadCommand_Pretty: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:37.299Z [INFO] TestTokenReadCommand_Pretty: Endpoints down --- PASS: TestTokenReadCommand_Pretty (0.89s) === CONT TestTokenReadCommand_JSON writer.go:29: 2021-01-29T19:32:37.506Z [INFO] TestTokenReadCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:37.508Z [INFO] TestTokenReadCommand_JSON: Endpoints down --- PASS: TestTokenReadCommand_JSON (1.10s) PASS ok github.com/hashicorp/consul/command/acl/token/read 1.255s === RUN TestTokenUpdateCommand_noTabs === PAUSE TestTokenUpdateCommand_noTabs === RUN TestTokenUpdateCommand === PAUSE TestTokenUpdateCommand === RUN TestTokenUpdateCommand_JSON === PAUSE TestTokenUpdateCommand_JSON === CONT TestTokenUpdateCommand_noTabs === CONT TestTokenUpdateCommand_JSON === CONT TestTokenUpdateCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestTokenUpdateCommand_noTabs (0.01s) === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.568Z [WARN] TestTokenUpdateCommand_JSON: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:37.570Z [DEBUG] TestTokenUpdateCommand_JSON.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:37.576Z [DEBUG] TestTokenUpdateCommand_JSON.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.578Z [WARN] TestTokenUpdateCommand: bootstrap = true: do not enable unless necessary === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.582Z [INFO] TestTokenUpdateCommand_JSON.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f02d2bd6-ac47-beab-2f5c-a4ff85afd588 Address:127.0.0.1:25327}]" === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.582Z [DEBUG] TestTokenUpdateCommand.tlsutil: Update: version=1 === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.584Z [INFO] TestTokenUpdateCommand_JSON.server.raft: entering follower state: follower="Node at 127.0.0.1:25327 [Follower]" leader= === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.585Z [DEBUG] TestTokenUpdateCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.589Z [INFO] TestTokenUpdateCommand_JSON.server.serf.wan: serf: EventMemberJoin: Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:37.598Z [INFO] TestTokenUpdateCommand_JSON.server.serf.lan: serf: EventMemberJoin: Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 127.0.0.1 === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.603Z [INFO] TestTokenUpdateCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e47d9683-d471-f96d-5d20-f7dc36be2480 Address:127.0.0.1:25321}]" === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.605Z [INFO] TestTokenUpdateCommand_JSON.server: Adding LAN server: server="Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 (Addr: tcp/127.0.0.1:25327) (DC: dc1)" writer.go:29: 2021-01-29T19:32:37.606Z [INFO] TestTokenUpdateCommand_JSON.server: Handled event for server in area: event=member-join server=Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588.dc1 area=wan === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.606Z [INFO] TestTokenUpdateCommand.server.serf.wan: serf: EventMemberJoin: Node-e47d9683-d471-f96d-5d20-f7dc36be2480.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:37.607Z [INFO] TestTokenUpdateCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:25321 [Follower]" leader= writer.go:29: 2021-01-29T19:32:37.611Z [INFO] TestTokenUpdateCommand.server.serf.lan: serf: EventMemberJoin: Node-e47d9683-d471-f96d-5d20-f7dc36be2480 127.0.0.1 === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.612Z [INFO] TestTokenUpdateCommand_JSON: Started DNS server: address=127.0.0.1:25322 network=udp === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.614Z [INFO] TestTokenUpdateCommand: Started DNS server: address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:32:37.615Z [INFO] TestTokenUpdateCommand.server: Adding LAN server: server="Node-e47d9683-d471-f96d-5d20-f7dc36be2480 (Addr: tcp/127.0.0.1:25321) (DC: dc1)" === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.614Z [INFO] TestTokenUpdateCommand_JSON: Started DNS server: address=127.0.0.1:25322 network=tcp === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.615Z [INFO] TestTokenUpdateCommand.server: Handled event for server in area: event=member-join server=Node-e47d9683-d471-f96d-5d20-f7dc36be2480.dc1 area=wan writer.go:29: 2021-01-29T19:32:37.618Z [INFO] TestTokenUpdateCommand: Started DNS server: address=127.0.0.1:25316 network=tcp writer.go:29: 2021-01-29T19:32:37.620Z [INFO] TestTokenUpdateCommand: Started HTTP server: address=127.0.0.1:25317 network=tcp === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.620Z [INFO] TestTokenUpdateCommand_JSON: Started HTTP server: address=127.0.0.1:25323 network=tcp writer.go:29: 2021-01-29T19:32:37.622Z [INFO] TestTokenUpdateCommand_JSON: started state syncer === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.621Z [INFO] TestTokenUpdateCommand: started state syncer === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.642Z [WARN] TestTokenUpdateCommand_JSON.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:37.643Z [INFO] TestTokenUpdateCommand_JSON.server.raft: entering candidate state: node="Node at 127.0.0.1:25327 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:37.645Z [DEBUG] TestTokenUpdateCommand_JSON.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:37.647Z [DEBUG] TestTokenUpdateCommand_JSON.server.raft: vote granted: from=f02d2bd6-ac47-beab-2f5c-a4ff85afd588 term=2 tally=1 writer.go:29: 2021-01-29T19:32:37.649Z [INFO] TestTokenUpdateCommand_JSON.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:37.651Z [INFO] TestTokenUpdateCommand_JSON.server.raft: entering leader state: leader="Node at 127.0.0.1:25327 [Leader]" writer.go:29: 2021-01-29T19:32:37.653Z [INFO] TestTokenUpdateCommand_JSON.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:37.654Z [INFO] TestTokenUpdateCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:32:37.654Z [INFO] TestTokenUpdateCommand_JSON.server: New leader elected: payload=Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 writer.go:29: 2021-01-29T19:32:37.656Z [INFO] TestTokenUpdateCommand_JSON.server: initializing acls writer.go:29: 2021-01-29T19:32:37.664Z [INFO] TestTokenUpdateCommand_JSON.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:32:37.666Z [WARN] TestTokenUpdateCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.668Z [DEBUG] TestTokenUpdateCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.666Z [INFO] TestTokenUpdateCommand_JSON.server: Created ACL 'global-management' policy === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.674Z [WARN] TestTokenUpdateCommand.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.674Z [WARN] TestTokenUpdateCommand_JSON.server: Configuring a non-UUID master token is deprecated === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.677Z [INFO] TestTokenUpdateCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:25321 [Candidate]" term=2 === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.671Z [INFO] TestTokenUpdateCommand_JSON.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:37.683Z [INFO] TestTokenUpdateCommand_JSON.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:37.684Z [INFO] TestTokenUpdateCommand_JSON.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:37.685Z [INFO] TestTokenUpdateCommand_JSON.leader: started routine: routine="acl token reaping" === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.684Z [DEBUG] TestTokenUpdateCommand.server.raft: votes: needed=1 === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.686Z [DEBUG] TestTokenUpdateCommand_JSON.server: transitioning out of legacy ACL mode === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.687Z [DEBUG] TestTokenUpdateCommand.server.raft: vote granted: from=e47d9683-d471-f96d-5d20-f7dc36be2480 term=2 tally=1 writer.go:29: 2021-01-29T19:32:37.690Z [INFO] TestTokenUpdateCommand.server.raft: election won: tally=1 === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.689Z [INFO] TestTokenUpdateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.692Z [INFO] TestTokenUpdateCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:25321 [Leader]" === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.694Z [INFO] TestTokenUpdateCommand_JSON.server: Updating LAN server: server="Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 (Addr: tcp/127.0.0.1:25327) (DC: dc1)" === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.695Z [INFO] TestTokenUpdateCommand.server: cluster leadership acquired === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.696Z [INFO] TestTokenUpdateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588.dc1 === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.698Z [INFO] TestTokenUpdateCommand.server: New leader elected: payload=Node-e47d9683-d471-f96d-5d20-f7dc36be2480 writer.go:29: 2021-01-29T19:32:37.719Z [INFO] TestTokenUpdateCommand.server: initializing acls === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.699Z [INFO] TestTokenUpdateCommand_JSON.server.serf.lan: serf: EventMemberUpdate: Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 writer.go:29: 2021-01-29T19:32:37.732Z [INFO] TestTokenUpdateCommand_JSON.server.serf.wan: serf: EventMemberUpdate: Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588.dc1 writer.go:29: 2021-01-29T19:32:37.700Z [INFO] TestTokenUpdateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588.dc1 area=wan writer.go:29: 2021-01-29T19:32:37.733Z [INFO] TestTokenUpdateCommand_JSON.server: Updating LAN server: server="Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 (Addr: tcp/127.0.0.1:25327) (DC: dc1)" writer.go:29: 2021-01-29T19:32:37.744Z [INFO] TestTokenUpdateCommand_JSON.server: Handled event for server in area: event=member-update server=Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588.dc1 area=wan === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.746Z [INFO] TestTokenUpdateCommand.server: Created ACL 'global-management' policy === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.752Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.751Z [WARN] TestTokenUpdateCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:37.760Z [INFO] TestTokenUpdateCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:32:37.771Z [INFO] TestTokenUpdateCommand.server: initializing acls writer.go:29: 2021-01-29T19:32:37.774Z [WARN] TestTokenUpdateCommand.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:32:37.772Z [INFO] TestTokenUpdateCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:32:37.776Z [INFO] TestTokenUpdateCommand.leader: started routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:37.779Z [INFO] TestTokenUpdateCommand.leader: started routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:37.784Z [DEBUG] TestTokenUpdateCommand.server: transitioning out of legacy ACL mode writer.go:29: 2021-01-29T19:32:37.786Z [INFO] TestTokenUpdateCommand.server.serf.lan: serf: EventMemberUpdate: Node-e47d9683-d471-f96d-5d20-f7dc36be2480 writer.go:29: 2021-01-29T19:32:37.792Z [INFO] TestTokenUpdateCommand.server.serf.wan: serf: EventMemberUpdate: Node-e47d9683-d471-f96d-5d20-f7dc36be2480.dc1 writer.go:29: 2021-01-29T19:32:37.792Z [INFO] TestTokenUpdateCommand.server: Updating LAN server: server="Node-e47d9683-d471-f96d-5d20-f7dc36be2480 (Addr: tcp/127.0.0.1:25321) (DC: dc1)" writer.go:29: 2021-01-29T19:32:37.792Z [INFO] TestTokenUpdateCommand.server.serf.lan: serf: EventMemberUpdate: Node-e47d9683-d471-f96d-5d20-f7dc36be2480 writer.go:29: 2021-01-29T19:32:37.794Z [INFO] TestTokenUpdateCommand.server: Handled event for server in area: event=member-update server=Node-e47d9683-d471-f96d-5d20-f7dc36be2480.dc1 area=wan writer.go:29: 2021-01-29T19:32:37.799Z [INFO] TestTokenUpdateCommand.server.serf.wan: serf: EventMemberUpdate: Node-e47d9683-d471-f96d-5d20-f7dc36be2480.dc1 writer.go:29: 2021-01-29T19:32:37.801Z [INFO] TestTokenUpdateCommand.server: Handled event for server in area: event=member-update server=Node-e47d9683-d471-f96d-5d20-f7dc36be2480.dc1 area=wan writer.go:29: 2021-01-29T19:32:37.799Z [INFO] TestTokenUpdateCommand.server: Updating LAN server: server="Node-e47d9683-d471-f96d-5d20-f7dc36be2480 (Addr: tcp/127.0.0.1:25321) (DC: dc1)" writer.go:29: 2021-01-29T19:32:37.801Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.814Z [INFO] TestTokenUpdateCommand_JSON.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:37.815Z [INFO] TestTokenUpdateCommand_JSON.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:37.817Z [DEBUG] TestTokenUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 writer.go:29: 2021-01-29T19:32:37.817Z [INFO] TestTokenUpdateCommand_JSON.server: member joined, marking health alive: member=Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 writer.go:29: 2021-01-29T19:32:37.819Z [DEBUG] TestTokenUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.819Z [INFO] TestTokenUpdateCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.820Z [DEBUG] TestTokenUpdateCommand_JSON.server: Skipping self join check for node since the cluster is too small: node=Node-f02d2bd6-ac47-beab-2f5c-a4ff85afd588 === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.821Z [INFO] TestTokenUpdateCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:37.822Z [DEBUG] TestTokenUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-e47d9683-d471-f96d-5d20-f7dc36be2480 writer.go:29: 2021-01-29T19:32:37.822Z [INFO] TestTokenUpdateCommand.server: member joined, marking health alive: member=Node-e47d9683-d471-f96d-5d20-f7dc36be2480 writer.go:29: 2021-01-29T19:32:37.824Z [DEBUG] TestTokenUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-e47d9683-d471-f96d-5d20-f7dc36be2480 writer.go:29: 2021-01-29T19:32:37.825Z [DEBUG] TestTokenUpdateCommand.server: Skipping self join check for node since the cluster is too small: node=Node-e47d9683-d471-f96d-5d20-f7dc36be2480 === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:37.847Z [DEBUG] TestTokenUpdateCommand_JSON: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:37.849Z [INFO] TestTokenUpdateCommand_JSON: Synced node info writer.go:29: 2021-01-29T19:32:37.851Z [DEBUG] TestTokenUpdateCommand_JSON: Node info in sync === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:37.860Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:54964 latency=4.992319ms writer.go:29: 2021-01-29T19:32:37.868Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:54964 latency=1.511964ms writer.go:29: 2021-01-29T19:32:37.872Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=PUT url=/v1/acl/create from=127.0.0.1:54964 latency=938.838µs writer.go:29: 2021-01-29T19:32:37.882Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/51e227e5-b2c8-b202-6064-cebfd994b5bd from=127.0.0.1:54966 latency=3.309005ms writer.go:29: 2021-01-29T19:32:37.897Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=PUT url=/v1/acl/token/51e227e5-b2c8-b202-6064-cebfd994b5bd from=127.0.0.1:54966 latency=12.950419ms writer.go:29: 2021-01-29T19:32:37.903Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/51e227e5-b2c8-b202-6064-cebfd994b5bd from=127.0.0.1:54964 latency=1.174523ms writer.go:29: 2021-01-29T19:32:37.911Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/51e227e5-b2c8-b202-6064-cebfd994b5bd from=127.0.0.1:54968 latency=1.31667ms writer.go:29: 2021-01-29T19:32:37.916Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=PUT url=/v1/acl/token/51e227e5-b2c8-b202-6064-cebfd994b5bd from=127.0.0.1:54968 latency=3.301187ms writer.go:29: 2021-01-29T19:32:37.932Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/51e227e5-b2c8-b202-6064-cebfd994b5bd from=127.0.0.1:54964 latency=155.44µs writer.go:29: 2021-01-29T19:32:37.940Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/51e227e5-b2c8-b202-6064-cebfd994b5bd from=127.0.0.1:54970 latency=1.085167ms writer.go:29: 2021-01-29T19:32:37.943Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=PUT url=/v1/acl/token/51e227e5-b2c8-b202-6064-cebfd994b5bd from=127.0.0.1:54970 latency=1.432791ms writer.go:29: 2021-01-29T19:32:37.947Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/51e227e5-b2c8-b202-6064-cebfd994b5bd from=127.0.0.1:54964 latency=242.222µs writer.go:29: 2021-01-29T19:32:37.953Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=96.969µs writer.go:29: 2021-01-29T19:32:37.970Z [DEBUG] TestTokenUpdateCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:37.973Z [INFO] TestTokenUpdateCommand: Synced node info writer.go:29: 2021-01-29T19:32:37.985Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=140.946µs writer.go:29: 2021-01-29T19:32:38.014Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=138.603µs === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:38.036Z [DEBUG] TestTokenUpdateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/policy from=127.0.0.1:34194 latency=3.531846ms writer.go:29: 2021-01-29T19:32:38.039Z [DEBUG] TestTokenUpdateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:34194 latency=852.411µs === RUN TestTokenUpdateCommand_JSON/update_with_policy_by_name === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:38.045Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=91.825µs === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:38.050Z [DEBUG] TestTokenUpdateCommand_JSON.http: Request finished: method=GET url=/v1/acl/token/6d5e04aa-2818-6e16-7a42-aeb0070263f9 from=127.0.0.1:34196 latency=1.179312ms writer.go:29: 2021-01-29T19:32:38.052Z [DEBUG] TestTokenUpdateCommand_JSON.http: Request finished: method=PUT url=/v1/acl/token/6d5e04aa-2818-6e16-7a42-aeb0070263f9 from=127.0.0.1:34196 latency=1.112299ms writer.go:29: 2021-01-29T19:32:38.056Z [INFO] TestTokenUpdateCommand_JSON: Requesting shutdown writer.go:29: 2021-01-29T19:32:38.058Z [INFO] TestTokenUpdateCommand_JSON.server: shutting down server writer.go:29: 2021-01-29T19:32:38.060Z [DEBUG] TestTokenUpdateCommand_JSON.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:38.063Z [DEBUG] TestTokenUpdateCommand_JSON.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:38.064Z [DEBUG] TestTokenUpdateCommand_JSON.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:38.063Z [DEBUG] TestTokenUpdateCommand_JSON.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:38.064Z [DEBUG] TestTokenUpdateCommand_JSON.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:38.072Z [WARN] TestTokenUpdateCommand_JSON.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:38.072Z [DEBUG] TestTokenUpdateCommand_JSON.leader: stopped routine: routine="acl token reaping" === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:38.075Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=83.032µs === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:38.077Z [WARN] TestTokenUpdateCommand_JSON.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:38.079Z [INFO] TestTokenUpdateCommand_JSON.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:38.080Z [INFO] TestTokenUpdateCommand_JSON: consul server down writer.go:29: 2021-01-29T19:32:38.081Z [INFO] TestTokenUpdateCommand_JSON: shutdown complete writer.go:29: 2021-01-29T19:32:38.082Z [INFO] TestTokenUpdateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:25322 network=tcp writer.go:29: 2021-01-29T19:32:38.083Z [INFO] TestTokenUpdateCommand_JSON: Stopping server: protocol=DNS address=127.0.0.1:25322 network=udp writer.go:29: 2021-01-29T19:32:38.084Z [INFO] TestTokenUpdateCommand_JSON: Stopping server: protocol=HTTP address=127.0.0.1:25323 network=tcp === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:38.104Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=159.482µs writer.go:29: 2021-01-29T19:32:38.134Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=114.423µs writer.go:29: 2021-01-29T19:32:38.165Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=163.994µs writer.go:29: 2021-01-29T19:32:38.209Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=130.145µs writer.go:29: 2021-01-29T19:32:38.240Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=164.592µs writer.go:29: 2021-01-29T19:32:38.270Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=167.661µs writer.go:29: 2021-01-29T19:32:38.302Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=95.46µs writer.go:29: 2021-01-29T19:32:38.340Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=170.511µs writer.go:29: 2021-01-29T19:32:38.371Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=183.692µs writer.go:29: 2021-01-29T19:32:38.406Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=92.103µs writer.go:29: 2021-01-29T19:32:38.438Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=149.623µs writer.go:29: 2021-01-29T19:32:38.468Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=166.34µs writer.go:29: 2021-01-29T19:32:38.500Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=184.597µs writer.go:29: 2021-01-29T19:32:38.542Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=136.088µs writer.go:29: 2021-01-29T19:32:38.572Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=169.312µs === CONT TestTokenUpdateCommand_JSON writer.go:29: 2021-01-29T19:32:38.585Z [INFO] TestTokenUpdateCommand_JSON: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:38.586Z [INFO] TestTokenUpdateCommand_JSON: Endpoints down --- PASS: TestTokenUpdateCommand_JSON (1.07s) --- PASS: TestTokenUpdateCommand_JSON/update_with_policy_by_name (0.01s) === CONT TestTokenUpdateCommand writer.go:29: 2021-01-29T19:32:38.605Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=172.335µs writer.go:29: 2021-01-29T19:32:38.638Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=169.742µs writer.go:29: 2021-01-29T19:32:38.668Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=81.915µs writer.go:29: 2021-01-29T19:32:38.699Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=168.962µs writer.go:29: 2021-01-29T19:32:38.731Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=338.024µs writer.go:29: 2021-01-29T19:32:38.765Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=155.554µs writer.go:29: 2021-01-29T19:32:38.796Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/self from=127.0.0.1:54964 latency=100.274µs writer.go:29: 2021-01-29T19:32:38.805Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/435874f3-b658-e463-67bc-fd40d8c5033e from=127.0.0.1:54976 latency=1.537828ms writer.go:29: 2021-01-29T19:32:38.808Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=PUT url=/v1/acl/token/435874f3-b658-e463-67bc-fd40d8c5033e from=127.0.0.1:54976 latency=1.190078ms writer.go:29: 2021-01-29T19:32:38.813Z [DEBUG] TestTokenUpdateCommand.http: Request finished: method=GET url=/v1/acl/token/435874f3-b658-e463-67bc-fd40d8c5033e from=127.0.0.1:54964 latency=107.726µs writer.go:29: 2021-01-29T19:32:38.818Z [INFO] TestTokenUpdateCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:38.820Z [INFO] TestTokenUpdateCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:38.822Z [DEBUG] TestTokenUpdateCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:38.825Z [DEBUG] TestTokenUpdateCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:38.828Z [DEBUG] TestTokenUpdateCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:38.825Z [DEBUG] TestTokenUpdateCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:32:38.828Z [DEBUG] TestTokenUpdateCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:32:38.831Z [WARN] TestTokenUpdateCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:38.831Z [DEBUG] TestTokenUpdateCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:38.840Z [WARN] TestTokenUpdateCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:38.845Z [INFO] TestTokenUpdateCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:38.845Z [INFO] TestTokenUpdateCommand: consul server down writer.go:29: 2021-01-29T19:32:38.847Z [INFO] TestTokenUpdateCommand: shutdown complete writer.go:29: 2021-01-29T19:32:38.848Z [INFO] TestTokenUpdateCommand: Stopping server: protocol=DNS address=127.0.0.1:25316 network=tcp writer.go:29: 2021-01-29T19:32:38.849Z [INFO] TestTokenUpdateCommand: Stopping server: protocol=DNS address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:32:38.849Z [INFO] TestTokenUpdateCommand: Stopping server: protocol=HTTP address=127.0.0.1:25317 network=tcp writer.go:29: 2021-01-29T19:32:39.350Z [INFO] TestTokenUpdateCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:39.352Z [INFO] TestTokenUpdateCommand: Endpoints down --- PASS: TestTokenUpdateCommand (1.84s) PASS ok github.com/hashicorp/consul/command/acl/token/update 1.985s === RUN TestConfigFail === PAUSE TestConfigFail === RUN TestRetryJoin [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:32:40.789Z [WARN] TestRetryJoin: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:40.790Z [DEBUG] TestRetryJoin.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:40.800Z [DEBUG] TestRetryJoin.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:40.804Z [INFO] TestRetryJoin.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:12ad262c-8aed-856f-501d-c56a8ab3ff64 Address:127.0.0.1:30426}]" writer.go:29: 2021-01-29T19:32:40.807Z [INFO] TestRetryJoin.server.serf.wan: serf: EventMemberJoin: Node-12ad262c-8aed-856f-501d-c56a8ab3ff64.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:40.808Z [INFO] TestRetryJoin.server.raft: entering follower state: follower="Node at 127.0.0.1:30426 [Follower]" leader= writer.go:29: 2021-01-29T19:32:40.811Z [INFO] TestRetryJoin.server.serf.lan: serf: EventMemberJoin: Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 127.0.0.1 writer.go:29: 2021-01-29T19:32:40.816Z [INFO] TestRetryJoin.server: Adding LAN server: server="Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 (Addr: tcp/127.0.0.1:30426) (DC: dc1)" writer.go:29: 2021-01-29T19:32:40.816Z [INFO] TestRetryJoin: Started DNS server: address=127.0.0.1:30421 network=udp writer.go:29: 2021-01-29T19:32:40.818Z [INFO] TestRetryJoin: Started DNS server: address=127.0.0.1:30421 network=tcp writer.go:29: 2021-01-29T19:32:40.816Z [INFO] TestRetryJoin.server: Handled event for server in area: event=member-join server=Node-12ad262c-8aed-856f-501d-c56a8ab3ff64.dc1 area=wan writer.go:29: 2021-01-29T19:32:40.820Z [INFO] TestRetryJoin: Started HTTP server: address=127.0.0.1:30422 network=tcp writer.go:29: 2021-01-29T19:32:40.822Z [INFO] TestRetryJoin: started state syncer writer.go:29: 2021-01-29T19:32:40.878Z [WARN] TestRetryJoin.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:40.879Z [INFO] TestRetryJoin.server.raft: entering candidate state: node="Node at 127.0.0.1:30426 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:40.881Z [DEBUG] TestRetryJoin.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:40.885Z [DEBUG] TestRetryJoin.server.raft: vote granted: from=12ad262c-8aed-856f-501d-c56a8ab3ff64 term=2 tally=1 writer.go:29: 2021-01-29T19:32:40.886Z [INFO] TestRetryJoin.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:40.888Z [INFO] TestRetryJoin.server.raft: entering leader state: leader="Node at 127.0.0.1:30426 [Leader]" writer.go:29: 2021-01-29T19:32:40.890Z [INFO] TestRetryJoin.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:40.891Z [INFO] TestRetryJoin.server: New leader elected: payload=Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 writer.go:29: 2021-01-29T19:32:40.892Z [DEBUG] TestRetryJoin.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30426 writer.go:29: 2021-01-29T19:32:40.899Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:40.976Z [INFO] TestRetryJoin.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:40.976Z [INFO] TestRetryJoin.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:40.977Z [DEBUG] TestRetryJoin.server: Skipping self join check for node since the cluster is too small: node=Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 writer.go:29: 2021-01-29T19:32:40.978Z [INFO] TestRetryJoin.server: member joined, marking health alive: member=Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 writer.go:29: 2021-01-29T19:32:41.139Z [DEBUG] TestRetryJoin: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:41.142Z [INFO] TestRetryJoin: Synced node info writer.go:29: 2021-01-29T19:32:41.182Z [WARN] TestRetryJoin: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:41.183Z [DEBUG] TestRetryJoin.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:41.185Z [DEBUG] TestRetryJoin.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:41.188Z [INFO] TestRetryJoin.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ccd4137f-4e7b-2c49-5a21-583f09d20e13 Address:127.0.0.1:30432}]" writer.go:29: 2021-01-29T19:32:41.189Z [INFO] TestRetryJoin.server.raft: entering follower state: follower="Node at 127.0.0.1:30432 [Follower]" leader= writer.go:29: 2021-01-29T19:32:41.189Z [INFO] TestRetryJoin.server.serf.wan: serf: EventMemberJoin: Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:41.194Z [INFO] TestRetryJoin.server.serf.lan: serf: EventMemberJoin: Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 127.0.0.1 writer.go:29: 2021-01-29T19:32:41.196Z [INFO] TestRetryJoin.server: Adding LAN server: server="Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 (Addr: tcp/127.0.0.1:30432) (DC: dc1)" writer.go:29: 2021-01-29T19:32:41.196Z [INFO] TestRetryJoin: Started DNS server: address=127.0.0.1:30427 network=udp writer.go:29: 2021-01-29T19:32:41.197Z [INFO] TestRetryJoin.server: Handled event for server in area: event=member-join server=Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13.dc1 area=wan writer.go:29: 2021-01-29T19:32:41.198Z [INFO] TestRetryJoin: Started DNS server: address=127.0.0.1:30427 network=tcp writer.go:29: 2021-01-29T19:32:41.199Z [INFO] TestRetryJoin: Started HTTP server: address=127.0.0.1:30428 network=tcp writer.go:29: 2021-01-29T19:32:41.200Z [INFO] TestRetryJoin: started state syncer writer.go:29: 2021-01-29T19:32:41.200Z [INFO] TestRetryJoin: Retry join is supported for the following discovery methods: cluster=LAN discovery_methods="aliyun aws gce mdns os packet" writer.go:29: 2021-01-29T19:32:41.202Z [INFO] TestRetryJoin: Joining cluster...: cluster=LAN writer.go:29: 2021-01-29T19:32:41.202Z [INFO] TestRetryJoin: (LAN) joining: lan_addresses=[127.0.0.1:30424] writer.go:29: 2021-01-29T19:32:41.201Z [INFO] TestRetryJoin: Retry join is supported for the following discovery methods: cluster=WAN discovery_methods="aliyun aws gce mdns os packet" writer.go:29: 2021-01-29T19:32:41.204Z [INFO] TestRetryJoin: Joining cluster...: cluster=WAN writer.go:29: 2021-01-29T19:32:41.205Z [INFO] TestRetryJoin: (WAN) joining: wan_addresses=[127.0.0.1:30425] writer.go:29: 2021-01-29T19:32:41.206Z [DEBUG] TestRetryJoin.server.memberlist.wan: memberlist: Initiating push/pull sync with: 127.0.0.1:30425 writer.go:29: 2021-01-29T19:32:41.206Z [DEBUG] TestRetryJoin.server.memberlist.wan: memberlist: Stream connection from=127.0.0.1:34224 writer.go:29: 2021-01-29T19:32:41.206Z [DEBUG] TestRetryJoin.server.memberlist.lan: memberlist: Initiating push/pull sync with: 127.0.0.1:30424 writer.go:29: 2021-01-29T19:32:41.209Z [INFO] TestRetryJoin.server.serf.wan: serf: EventMemberJoin: Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:41.209Z [INFO] TestRetryJoin.server.serf.wan: serf: EventMemberJoin: Node-12ad262c-8aed-856f-501d-c56a8ab3ff64.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:41.207Z [DEBUG] TestRetryJoin.server.memberlist.lan: memberlist: Stream connection from=127.0.0.1:49208 writer.go:29: 2021-01-29T19:32:41.214Z [INFO] TestRetryJoin: (WAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:32:41.212Z [INFO] TestRetryJoin.server: Handled event for server in area: event=member-join server=Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13.dc1 area=wan writer.go:29: 2021-01-29T19:32:41.217Z [INFO] TestRetryJoin: Join cluster completed. Synced with initial agents: cluster=WAN num_agents=1 writer.go:29: 2021-01-29T19:32:41.214Z [INFO] TestRetryJoin.server: Handled event for server in area: event=member-join server=Node-12ad262c-8aed-856f-501d-c56a8ab3ff64.dc1 area=wan writer.go:29: 2021-01-29T19:32:41.217Z [INFO] TestRetryJoin.server.serf.lan: serf: EventMemberJoin: Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 127.0.0.1 writer.go:29: 2021-01-29T19:32:41.217Z [INFO] TestRetryJoin.server.serf.lan: serf: EventMemberJoin: Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 127.0.0.1 writer.go:29: 2021-01-29T19:32:41.222Z [INFO] TestRetryJoin.server: Adding LAN server: server="Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 (Addr: tcp/127.0.0.1:30432) (DC: dc1)" writer.go:29: 2021-01-29T19:32:41.224Z [INFO] TestRetryJoin: (LAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:32:41.226Z [DEBUG] TestRetryJoin: systemd notify failed: error="No socket" writer.go:29: 2021-01-29T19:32:41.224Z [INFO] TestRetryJoin.server: Adding LAN server: server="Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 (Addr: tcp/127.0.0.1:30426) (DC: dc1)" writer.go:29: 2021-01-29T19:32:41.227Z [INFO] TestRetryJoin: Join cluster completed. Synced with initial agents: cluster=LAN num_agents=1 writer.go:29: 2021-01-29T19:32:41.225Z [ERROR] TestRetryJoin.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 other=Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 writer.go:29: 2021-01-29T19:32:41.231Z [INFO] TestRetryJoin.server: member joined, marking health alive: member=Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 writer.go:29: 2021-01-29T19:32:41.248Z [WARN] TestRetryJoin.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:41.249Z [INFO] TestRetryJoin.server.raft: entering candidate state: node="Node at 127.0.0.1:30432 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:41.250Z [DEBUG] TestRetryJoin.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:41.251Z [DEBUG] TestRetryJoin.server.raft: vote granted: from=ccd4137f-4e7b-2c49-5a21-583f09d20e13 term=2 tally=1 writer.go:29: 2021-01-29T19:32:41.252Z [INFO] TestRetryJoin.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:41.252Z [INFO] TestRetryJoin.server.raft: entering leader state: leader="Node at 127.0.0.1:30432 [Leader]" writer.go:29: 2021-01-29T19:32:41.253Z [INFO] TestRetryJoin.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:41.254Z [INFO] TestRetryJoin.server: New leader elected: payload=Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 writer.go:29: 2021-01-29T19:32:41.255Z [DEBUG] TestRetryJoin.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30432 writer.go:29: 2021-01-29T19:32:41.259Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:41.265Z [INFO] TestRetryJoin.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:41.266Z [INFO] TestRetryJoin.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:41.267Z [ERROR] TestRetryJoin.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 other=Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 writer.go:29: 2021-01-29T19:32:41.268Z [INFO] TestRetryJoin.server: member joined, marking health alive: member=Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 writer.go:29: 2021-01-29T19:32:41.270Z [ERROR] TestRetryJoin.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 other=Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 writer.go:29: 2021-01-29T19:32:41.270Z [INFO] TestRetryJoin.server: member joined, marking health alive: member=Node-12ad262c-8aed-856f-501d-c56a8ab3ff64 writer.go:29: 2021-01-29T19:32:41.366Z [INFO] TestRetryJoin: Requesting shutdown writer.go:29: 2021-01-29T19:32:41.367Z [INFO] TestRetryJoin.server: shutting down server writer.go:29: 2021-01-29T19:32:41.368Z [DEBUG] TestRetryJoin.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:41.368Z [WARN] TestRetryJoin.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:41.370Z [ERROR] TestRetryJoin.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:41.370Z [DEBUG] TestRetryJoin.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:41.372Z [WARN] TestRetryJoin.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:41.376Z [INFO] TestRetryJoin.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:41.376Z [INFO] TestRetryJoin: consul server down writer.go:29: 2021-01-29T19:32:41.379Z [INFO] TestRetryJoin: shutdown complete writer.go:29: 2021-01-29T19:32:41.380Z [INFO] TestRetryJoin: Stopping server: protocol=DNS address=127.0.0.1:30427 network=tcp writer.go:29: 2021-01-29T19:32:41.381Z [INFO] TestRetryJoin: Stopping server: protocol=DNS address=127.0.0.1:30427 network=udp writer.go:29: 2021-01-29T19:32:41.397Z [INFO] TestRetryJoin: Stopping server: protocol=HTTP address=127.0.0.1:30428 network=tcp writer.go:29: 2021-01-29T19:32:41.808Z [DEBUG] TestRetryJoin: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:41.810Z [DEBUG] TestRetryJoin: Node info in sync writer.go:29: 2021-01-29T19:32:41.811Z [DEBUG] TestRetryJoin: Node info in sync writer.go:29: 2021-01-29T19:32:41.899Z [INFO] TestRetryJoin: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:41.900Z [INFO] TestRetryJoin: Endpoints down writer.go:29: 2021-01-29T19:32:41.901Z [INFO] TestRetryJoin: Requesting shutdown writer.go:29: 2021-01-29T19:32:41.903Z [INFO] TestRetryJoin.server: shutting down server writer.go:29: 2021-01-29T19:32:41.904Z [DEBUG] TestRetryJoin.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:41.905Z [WARN] TestRetryJoin.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:41.907Z [DEBUG] TestRetryJoin.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:41.909Z [WARN] TestRetryJoin.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:41.912Z [INFO] TestRetryJoin.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:41.912Z [INFO] TestRetryJoin: consul server down writer.go:29: 2021-01-29T19:32:41.915Z [INFO] TestRetryJoin: shutdown complete writer.go:29: 2021-01-29T19:32:41.916Z [INFO] TestRetryJoin: Stopping server: protocol=DNS address=127.0.0.1:30421 network=tcp writer.go:29: 2021-01-29T19:32:41.917Z [INFO] TestRetryJoin: Stopping server: protocol=DNS address=127.0.0.1:30421 network=udp writer.go:29: 2021-01-29T19:32:41.918Z [INFO] TestRetryJoin: Stopping server: protocol=HTTP address=127.0.0.1:30422 network=tcp writer.go:29: 2021-01-29T19:32:42.316Z [DEBUG] TestRetryJoin.server.memberlist.lan: memberlist: Failed ping: Node-ccd4137f-4e7b-2c49-5a21-583f09d20e13 (timeout reached) writer.go:29: 2021-01-29T19:32:42.420Z [INFO] TestRetryJoin: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:42.421Z [INFO] TestRetryJoin: Endpoints down --- PASS: TestRetryJoin (1.66s) === RUN TestRetryJoinFail === PAUSE TestRetryJoinFail === RUN TestRetryJoinWanFail === PAUSE TestRetryJoinWanFail === RUN TestProtectDataDir === PAUSE TestProtectDataDir === RUN TestBadDataDirPermissions === PAUSE TestBadDataDirPermissions === RUN TestReloadLoggerFail writer.go:29: 2021-01-29T19:32:42.458Z [WARN] TestReloadLoggerFail: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:42.458Z [DEBUG] TestReloadLoggerFail.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:42.460Z [DEBUG] TestReloadLoggerFail.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:42.468Z [INFO] TestReloadLoggerFail.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:16def4a8-b7d9-017f-c192-cf706b45e9c1 Address:127.0.0.1:30438}]" writer.go:29: 2021-01-29T19:32:42.471Z [INFO] TestReloadLoggerFail.server.raft: entering follower state: follower="Node at 127.0.0.1:30438 [Follower]" leader= writer.go:29: 2021-01-29T19:32:42.473Z [INFO] TestReloadLoggerFail.server.serf.wan: serf: EventMemberJoin: Node-16def4a8-b7d9-017f-c192-cf706b45e9c1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:42.487Z [INFO] TestReloadLoggerFail.server.serf.lan: serf: EventMemberJoin: Node-16def4a8-b7d9-017f-c192-cf706b45e9c1 127.0.0.1 writer.go:29: 2021-01-29T19:32:42.490Z [INFO] TestReloadLoggerFail.server: Adding LAN server: server="Node-16def4a8-b7d9-017f-c192-cf706b45e9c1 (Addr: tcp/127.0.0.1:30438) (DC: dc1)" writer.go:29: 2021-01-29T19:32:42.491Z [INFO] TestReloadLoggerFail.server: Handled event for server in area: event=member-join server=Node-16def4a8-b7d9-017f-c192-cf706b45e9c1.dc1 area=wan writer.go:29: 2021-01-29T19:32:42.493Z [INFO] TestReloadLoggerFail: Started DNS server: address=127.0.0.1:30433 network=tcp writer.go:29: 2021-01-29T19:32:42.494Z [INFO] TestReloadLoggerFail: Started DNS server: address=127.0.0.1:30433 network=udp writer.go:29: 2021-01-29T19:32:42.496Z [INFO] TestReloadLoggerFail: Started HTTP server: address=127.0.0.1:30434 network=tcp writer.go:29: 2021-01-29T19:32:42.497Z [INFO] TestReloadLoggerFail: started state syncer writer.go:29: 2021-01-29T19:32:42.531Z [WARN] TestReloadLoggerFail.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:42.532Z [INFO] TestReloadLoggerFail.server.raft: entering candidate state: node="Node at 127.0.0.1:30438 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:42.536Z [DEBUG] TestReloadLoggerFail.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:42.537Z [DEBUG] TestReloadLoggerFail.server.raft: vote granted: from=16def4a8-b7d9-017f-c192-cf706b45e9c1 term=2 tally=1 writer.go:29: 2021-01-29T19:32:42.538Z [INFO] TestReloadLoggerFail.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:42.539Z [INFO] TestReloadLoggerFail.server.raft: entering leader state: leader="Node at 127.0.0.1:30438 [Leader]" writer.go:29: 2021-01-29T19:32:42.540Z [INFO] TestReloadLoggerFail.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:42.541Z [INFO] TestReloadLoggerFail.server: New leader elected: payload=Node-16def4a8-b7d9-017f-c192-cf706b45e9c1 writer.go:29: 2021-01-29T19:32:42.552Z [DEBUG] TestReloadLoggerFail.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30438 writer.go:29: 2021-01-29T19:32:42.557Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:42.569Z [INFO] TestReloadLoggerFail.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:42.570Z [INFO] TestReloadLoggerFail.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:42.572Z [DEBUG] TestReloadLoggerFail.server: Skipping self join check for node since the cluster is too small: node=Node-16def4a8-b7d9-017f-c192-cf706b45e9c1 writer.go:29: 2021-01-29T19:32:42.574Z [INFO] TestReloadLoggerFail.server: member joined, marking health alive: member=Node-16def4a8-b7d9-017f-c192-cf706b45e9c1 writer.go:29: 2021-01-29T19:32:42.655Z [DEBUG] TestReloadLoggerFail: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:42.657Z [INFO] TestReloadLoggerFail: Synced node info writer.go:29: 2021-01-29T19:32:42.734Z [DEBUG] TestReloadLoggerFail: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:42.735Z [DEBUG] TestReloadLoggerFail: Node info in sync writer.go:29: 2021-01-29T19:32:42.736Z [DEBUG] TestReloadLoggerFail: Node info in sync writer.go:29: 2021-01-29T19:32:42.872Z [INFO] TestReloadLoggerFail: Reloading configuration... writer.go:29: 2021-01-29T19:32:42.878Z [DEBUG] TestReloadLoggerFail.tlsutil: Update: version=2 writer.go:29: 2021-01-29T19:32:42.879Z [INFO] TestReloadLoggerFail: Requesting shutdown writer.go:29: 2021-01-29T19:32:42.879Z [DEBUG] TestReloadLoggerFail: Node info in sync writer.go:29: 2021-01-29T19:32:42.881Z [DEBUG] TestReloadLoggerFail: Node info in sync writer.go:29: 2021-01-29T19:32:42.880Z [INFO] TestReloadLoggerFail.server: shutting down server writer.go:29: 2021-01-29T19:32:42.883Z [DEBUG] TestReloadLoggerFail.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:42.884Z [WARN] TestReloadLoggerFail.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:42.886Z [DEBUG] TestReloadLoggerFail.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:42.889Z [WARN] TestReloadLoggerFail.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:42.890Z [INFO] TestReloadLoggerFail.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:42.890Z [INFO] TestReloadLoggerFail: consul server down writer.go:29: 2021-01-29T19:32:42.893Z [INFO] TestReloadLoggerFail: shutdown complete writer.go:29: 2021-01-29T19:32:42.894Z [INFO] TestReloadLoggerFail: Stopping server: protocol=DNS address=127.0.0.1:30433 network=tcp writer.go:29: 2021-01-29T19:32:42.896Z [INFO] TestReloadLoggerFail: Stopping server: protocol=DNS address=127.0.0.1:30433 network=udp writer.go:29: 2021-01-29T19:32:42.897Z [INFO] TestReloadLoggerFail: Stopping server: protocol=HTTP address=127.0.0.1:30434 network=tcp writer.go:29: 2021-01-29T19:32:43.398Z [INFO] TestReloadLoggerFail: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:43.400Z [INFO] TestReloadLoggerFail: Endpoints down --- PASS: TestReloadLoggerFail (0.98s) === RUN TestReloadLoggerSuccess writer.go:29: 2021-01-29T19:32:43.431Z [WARN] TestReloadLoggerSuccess: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:43.435Z [DEBUG] TestReloadLoggerSuccess.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:43.442Z [DEBUG] TestReloadLoggerSuccess.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:43.464Z [INFO] TestReloadLoggerSuccess.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6c467759-3a3e-48af-46c1-eeeea5715693 Address:127.0.0.1:30444}]" writer.go:29: 2021-01-29T19:32:43.467Z [INFO] TestReloadLoggerSuccess.server.raft: entering follower state: follower="Node at 127.0.0.1:30444 [Follower]" leader= writer.go:29: 2021-01-29T19:32:43.469Z [INFO] TestReloadLoggerSuccess.server.serf.wan: serf: EventMemberJoin: Node-6c467759-3a3e-48af-46c1-eeeea5715693.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:43.473Z [INFO] TestReloadLoggerSuccess.server.serf.lan: serf: EventMemberJoin: Node-6c467759-3a3e-48af-46c1-eeeea5715693 127.0.0.1 writer.go:29: 2021-01-29T19:32:43.476Z [INFO] TestReloadLoggerSuccess.server: Adding LAN server: server="Node-6c467759-3a3e-48af-46c1-eeeea5715693 (Addr: tcp/127.0.0.1:30444) (DC: dc1)" writer.go:29: 2021-01-29T19:32:43.477Z [INFO] TestReloadLoggerSuccess.server: Handled event for server in area: event=member-join server=Node-6c467759-3a3e-48af-46c1-eeeea5715693.dc1 area=wan writer.go:29: 2021-01-29T19:32:43.478Z [INFO] TestReloadLoggerSuccess: Started DNS server: address=127.0.0.1:30439 network=tcp writer.go:29: 2021-01-29T19:32:43.479Z [INFO] TestReloadLoggerSuccess: Started DNS server: address=127.0.0.1:30439 network=udp writer.go:29: 2021-01-29T19:32:43.487Z [INFO] TestReloadLoggerSuccess: Started HTTP server: address=127.0.0.1:30440 network=tcp writer.go:29: 2021-01-29T19:32:43.493Z [INFO] TestReloadLoggerSuccess: started state syncer writer.go:29: 2021-01-29T19:32:43.537Z [WARN] TestReloadLoggerSuccess.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:43.538Z [INFO] TestReloadLoggerSuccess.server.raft: entering candidate state: node="Node at 127.0.0.1:30444 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:43.540Z [DEBUG] TestReloadLoggerSuccess.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:43.540Z [DEBUG] TestReloadLoggerSuccess.server.raft: vote granted: from=6c467759-3a3e-48af-46c1-eeeea5715693 term=2 tally=1 writer.go:29: 2021-01-29T19:32:43.541Z [INFO] TestReloadLoggerSuccess.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:43.542Z [INFO] TestReloadLoggerSuccess.server.raft: entering leader state: leader="Node at 127.0.0.1:30444 [Leader]" writer.go:29: 2021-01-29T19:32:43.543Z [INFO] TestReloadLoggerSuccess.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:43.543Z [INFO] TestReloadLoggerSuccess.server: New leader elected: payload=Node-6c467759-3a3e-48af-46c1-eeeea5715693 writer.go:29: 2021-01-29T19:32:43.545Z [DEBUG] TestReloadLoggerSuccess.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30444 writer.go:29: 2021-01-29T19:32:43.549Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:43.554Z [INFO] TestReloadLoggerSuccess.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:43.555Z [INFO] TestReloadLoggerSuccess.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:43.557Z [DEBUG] TestReloadLoggerSuccess.server: Skipping self join check for node since the cluster is too small: node=Node-6c467759-3a3e-48af-46c1-eeeea5715693 writer.go:29: 2021-01-29T19:32:43.559Z [INFO] TestReloadLoggerSuccess.server: member joined, marking health alive: member=Node-6c467759-3a3e-48af-46c1-eeeea5715693 writer.go:29: 2021-01-29T19:32:43.618Z [INFO] TestReloadLoggerSuccess: Reloading configuration... writer.go:29: 2021-01-29T19:32:43.632Z [DEBUG] TestReloadLoggerSuccess: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:32:43.638Z [INFO] TestReloadLoggerSuccess: Synced node info writer.go:29: 2021-01-29T19:32:43.641Z [DEBUG] TestReloadLoggerSuccess: Node info in sync writer.go:29: 2021-01-29T19:32:43.662Z [DEBUG] TestReloadLoggerSuccess.tlsutil: Update: version=2 writer.go:29: 2021-01-29T19:32:43.663Z [INFO] TestReloadLoggerSuccess: Requesting shutdown writer.go:29: 2021-01-29T19:32:43.664Z [INFO] TestReloadLoggerSuccess.server: shutting down server writer.go:29: 2021-01-29T19:32:43.665Z [DEBUG] TestReloadLoggerSuccess.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:43.666Z [WARN] TestReloadLoggerSuccess.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:43.667Z [DEBUG] TestReloadLoggerSuccess.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:43.668Z [WARN] TestReloadLoggerSuccess.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:43.671Z [INFO] TestReloadLoggerSuccess.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:43.671Z [INFO] TestReloadLoggerSuccess: consul server down writer.go:29: 2021-01-29T19:32:43.675Z [INFO] TestReloadLoggerSuccess: shutdown complete writer.go:29: 2021-01-29T19:32:43.677Z [INFO] TestReloadLoggerSuccess: Stopping server: protocol=DNS address=127.0.0.1:30439 network=tcp writer.go:29: 2021-01-29T19:32:43.678Z [INFO] TestReloadLoggerSuccess: Stopping server: protocol=DNS address=127.0.0.1:30439 network=udp writer.go:29: 2021-01-29T19:32:43.680Z [INFO] TestReloadLoggerSuccess: Stopping server: protocol=HTTP address=127.0.0.1:30440 network=tcp writer.go:29: 2021-01-29T19:32:44.182Z [INFO] TestReloadLoggerSuccess: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:44.183Z [INFO] TestReloadLoggerSuccess: Endpoints down --- PASS: TestReloadLoggerSuccess (0.78s) === CONT TestConfigFail === RUN TestConfigFail/agent_-server_-bind=10.0.0.1_-datacenter= === CONT TestProtectDataDir === CONT TestRetryJoinWanFail === CONT TestRetryJoinFail --- PASS: TestProtectDataDir (0.02s) === CONT TestBadDataDirPermissions --- PASS: TestRetryJoinFail (0.06s) --- PASS: TestBadDataDirPermissions (0.06s) === RUN TestConfigFail/agent_-server_-bind=10.0.0.1_-datacenter=foo_some-other-arg === RUN TestConfigFail/agent_-server_-bind=10.0.0.1 --- PASS: TestRetryJoinWanFail (0.59s) === RUN TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise_0.0.0.0_-bind_10.0.0.1 === RUN TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise_::_-bind_10.0.0.1 === RUN TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise_[::]_-bind_10.0.0.1 === RUN TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise-wan_0.0.0.0_-bind_10.0.0.1 === RUN TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise-wan_::_-bind_10.0.0.1 === RUN TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise-wan_[::]_-bind_10.0.0.1 --- PASS: TestConfigFail (2.59s) --- PASS: TestConfigFail/agent_-server_-bind=10.0.0.1_-datacenter= (0.24s) --- PASS: TestConfigFail/agent_-server_-bind=10.0.0.1_-datacenter=foo_some-other-arg (0.15s) --- PASS: TestConfigFail/agent_-server_-bind=10.0.0.1 (0.29s) --- PASS: TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise_0.0.0.0_-bind_10.0.0.1 (0.33s) --- PASS: TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise_::_-bind_10.0.0.1 (0.39s) --- PASS: TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise_[::]_-bind_10.0.0.1 (0.34s) --- PASS: TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise-wan_0.0.0.0_-bind_10.0.0.1 (0.30s) --- PASS: TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise-wan_::_-bind_10.0.0.1 (0.34s) --- PASS: TestConfigFail/agent_-server_-data-dir_/tmp/consul-test/TestConfigFail-consul223265800_-advertise-wan_[::]_-bind_10.0.0.1 (0.20s) PASS ok github.com/hashicorp/consul/command/agent 6.092s === RUN TestCatalogCommand_noTabs === PAUSE TestCatalogCommand_noTabs === CONT TestCatalogCommand_noTabs --- PASS: TestCatalogCommand_noTabs (0.00s) PASS ok github.com/hashicorp/consul/command/catalog 0.014s === RUN TestCatalogListDatacentersCommand_noTabs === PAUSE TestCatalogListDatacentersCommand_noTabs === RUN TestCatalogListDatacentersCommand_Validation === PAUSE TestCatalogListDatacentersCommand_Validation === RUN TestCatalogListDatacentersCommand === PAUSE TestCatalogListDatacentersCommand === CONT TestCatalogListDatacentersCommand_noTabs === CONT TestCatalogListDatacentersCommand --- PASS: TestCatalogListDatacentersCommand_noTabs (0.00s) === CONT TestCatalogListDatacentersCommand_Validation --- PASS: TestCatalogListDatacentersCommand_Validation (0.00s) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestCatalogListDatacentersCommand writer.go:29: 2021-01-29T19:32:45.959Z [WARN] TestCatalogListDatacentersCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:45.962Z [DEBUG] TestCatalogListDatacentersCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:45.967Z [DEBUG] TestCatalogListDatacentersCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:45.973Z [INFO] TestCatalogListDatacentersCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6bd30dcb-9583-a676-f06b-033311f068fa Address:127.0.0.1:17153}]" writer.go:29: 2021-01-29T19:32:45.976Z [INFO] TestCatalogListDatacentersCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:17153 [Follower]" leader= writer.go:29: 2021-01-29T19:32:45.982Z [INFO] TestCatalogListDatacentersCommand.server.serf.wan: serf: EventMemberJoin: Node-6bd30dcb-9583-a676-f06b-033311f068fa.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:45.986Z [INFO] TestCatalogListDatacentersCommand.server.serf.lan: serf: EventMemberJoin: Node-6bd30dcb-9583-a676-f06b-033311f068fa 127.0.0.1 writer.go:29: 2021-01-29T19:32:45.990Z [INFO] TestCatalogListDatacentersCommand: Started DNS server: address=127.0.0.1:17148 network=udp writer.go:29: 2021-01-29T19:32:45.990Z [INFO] TestCatalogListDatacentersCommand.server: Handled event for server in area: event=member-join server=Node-6bd30dcb-9583-a676-f06b-033311f068fa.dc1 area=wan writer.go:29: 2021-01-29T19:32:45.991Z [INFO] TestCatalogListDatacentersCommand.server: Adding LAN server: server="Node-6bd30dcb-9583-a676-f06b-033311f068fa (Addr: tcp/127.0.0.1:17153) (DC: dc1)" writer.go:29: 2021-01-29T19:32:45.993Z [INFO] TestCatalogListDatacentersCommand: Started DNS server: address=127.0.0.1:17148 network=tcp writer.go:29: 2021-01-29T19:32:46.013Z [INFO] TestCatalogListDatacentersCommand: Started HTTP server: address=127.0.0.1:17149 network=tcp writer.go:29: 2021-01-29T19:32:46.014Z [INFO] TestCatalogListDatacentersCommand: started state syncer writer.go:29: 2021-01-29T19:32:46.043Z [WARN] TestCatalogListDatacentersCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:46.045Z [INFO] TestCatalogListDatacentersCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:17153 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:46.049Z [DEBUG] TestCatalogListDatacentersCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:46.052Z [DEBUG] TestCatalogListDatacentersCommand.server.raft: vote granted: from=6bd30dcb-9583-a676-f06b-033311f068fa term=2 tally=1 writer.go:29: 2021-01-29T19:32:46.054Z [INFO] TestCatalogListDatacentersCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:46.058Z [INFO] TestCatalogListDatacentersCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:17153 [Leader]" writer.go:29: 2021-01-29T19:32:46.062Z [INFO] TestCatalogListDatacentersCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:46.063Z [INFO] TestCatalogListDatacentersCommand.server: New leader elected: payload=Node-6bd30dcb-9583-a676-f06b-033311f068fa writer.go:29: 2021-01-29T19:32:46.067Z [DEBUG] TestCatalogListDatacentersCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:17153 writer.go:29: 2021-01-29T19:32:46.081Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:46.146Z [INFO] TestCatalogListDatacentersCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:46.148Z [INFO] TestCatalogListDatacentersCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:46.152Z [DEBUG] TestCatalogListDatacentersCommand.server: Skipping self join check for node since the cluster is too small: node=Node-6bd30dcb-9583-a676-f06b-033311f068fa writer.go:29: 2021-01-29T19:32:46.157Z [INFO] TestCatalogListDatacentersCommand.server: member joined, marking health alive: member=Node-6bd30dcb-9583-a676-f06b-033311f068fa writer.go:29: 2021-01-29T19:32:46.330Z [DEBUG] TestCatalogListDatacentersCommand.http: Request finished: method=GET url=/v1/catalog/datacenters from=127.0.0.1:59906 latency=3.365778ms writer.go:29: 2021-01-29T19:32:46.333Z [INFO] TestCatalogListDatacentersCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:46.334Z [INFO] TestCatalogListDatacentersCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:46.335Z [DEBUG] TestCatalogListDatacentersCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:46.335Z [ERROR] TestCatalogListDatacentersCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:46.336Z [WARN] TestCatalogListDatacentersCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:46.336Z [DEBUG] TestCatalogListDatacentersCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:46.342Z [WARN] TestCatalogListDatacentersCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:46.347Z [INFO] TestCatalogListDatacentersCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:46.347Z [INFO] TestCatalogListDatacentersCommand: consul server down writer.go:29: 2021-01-29T19:32:46.349Z [INFO] TestCatalogListDatacentersCommand: shutdown complete writer.go:29: 2021-01-29T19:32:46.350Z [INFO] TestCatalogListDatacentersCommand: Stopping server: protocol=DNS address=127.0.0.1:17148 network=tcp writer.go:29: 2021-01-29T19:32:46.350Z [INFO] TestCatalogListDatacentersCommand: Stopping server: protocol=DNS address=127.0.0.1:17148 network=udp writer.go:29: 2021-01-29T19:32:46.352Z [INFO] TestCatalogListDatacentersCommand: Stopping server: protocol=HTTP address=127.0.0.1:17149 network=tcp writer.go:29: 2021-01-29T19:32:46.853Z [INFO] TestCatalogListDatacentersCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:46.855Z [INFO] TestCatalogListDatacentersCommand: Endpoints down --- PASS: TestCatalogListDatacentersCommand (1.00s) PASS ok github.com/hashicorp/consul/command/catalog/list/dc 1.261s === RUN TestCatalogListNodesCommand_noTabs === PAUSE TestCatalogListNodesCommand_noTabs === RUN TestCatalogListNodesCommand_Validation === PAUSE TestCatalogListNodesCommand_Validation === RUN TestCatalogListNodesCommand === PAUSE TestCatalogListNodesCommand === RUN TestCatalogListNodesCommand_verticalBar catalog_list_nodes_test.go:168: DM-skipped --- SKIP: TestCatalogListNodesCommand_verticalBar (0.00s) === CONT TestCatalogListNodesCommand_noTabs === CONT TestCatalogListNodesCommand === CONT TestCatalogListNodesCommand_Validation --- PASS: TestCatalogListNodesCommand_noTabs (0.00s) --- PASS: TestCatalogListNodesCommand_Validation (0.00s) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestCatalogListNodesCommand writer.go:29: 2021-01-29T19:32:48.043Z [WARN] TestCatalogListNodesCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:48.045Z [DEBUG] TestCatalogListNodesCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:48.072Z [DEBUG] TestCatalogListNodesCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:48.082Z [INFO] TestCatalogListNodesCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:932e30e5-6bf5-c98d-bdda-53e1896f431b Address:127.0.0.1:18174}]" writer.go:29: 2021-01-29T19:32:48.084Z [INFO] TestCatalogListNodesCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:18174 [Follower]" leader= writer.go:29: 2021-01-29T19:32:48.085Z [INFO] TestCatalogListNodesCommand.server.serf.wan: serf: EventMemberJoin: Node-932e30e5-6bf5-c98d-bdda-53e1896f431b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:48.088Z [INFO] TestCatalogListNodesCommand.server.serf.lan: serf: EventMemberJoin: Node-932e30e5-6bf5-c98d-bdda-53e1896f431b 127.0.0.1 writer.go:29: 2021-01-29T19:32:48.090Z [INFO] TestCatalogListNodesCommand.server: Adding LAN server: server="Node-932e30e5-6bf5-c98d-bdda-53e1896f431b (Addr: tcp/127.0.0.1:18174) (DC: dc1)" writer.go:29: 2021-01-29T19:32:48.091Z [INFO] TestCatalogListNodesCommand.server: Handled event for server in area: event=member-join server=Node-932e30e5-6bf5-c98d-bdda-53e1896f431b.dc1 area=wan writer.go:29: 2021-01-29T19:32:48.091Z [INFO] TestCatalogListNodesCommand: Started DNS server: address=127.0.0.1:18169 network=udp writer.go:29: 2021-01-29T19:32:48.093Z [INFO] TestCatalogListNodesCommand: Started DNS server: address=127.0.0.1:18169 network=tcp writer.go:29: 2021-01-29T19:32:48.108Z [INFO] TestCatalogListNodesCommand: Started HTTP server: address=127.0.0.1:18170 network=tcp writer.go:29: 2021-01-29T19:32:48.109Z [INFO] TestCatalogListNodesCommand: started state syncer writer.go:29: 2021-01-29T19:32:48.147Z [WARN] TestCatalogListNodesCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:48.148Z [INFO] TestCatalogListNodesCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:18174 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:48.151Z [DEBUG] TestCatalogListNodesCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:48.153Z [DEBUG] TestCatalogListNodesCommand.server.raft: vote granted: from=932e30e5-6bf5-c98d-bdda-53e1896f431b term=2 tally=1 writer.go:29: 2021-01-29T19:32:48.155Z [INFO] TestCatalogListNodesCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:48.157Z [INFO] TestCatalogListNodesCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:18174 [Leader]" writer.go:29: 2021-01-29T19:32:48.159Z [INFO] TestCatalogListNodesCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:48.160Z [INFO] TestCatalogListNodesCommand.server: New leader elected: payload=Node-932e30e5-6bf5-c98d-bdda-53e1896f431b writer.go:29: 2021-01-29T19:32:48.172Z [DEBUG] TestCatalogListNodesCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:18174 writer.go:29: 2021-01-29T19:32:48.180Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:48.283Z [INFO] TestCatalogListNodesCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:48.284Z [INFO] TestCatalogListNodesCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:48.285Z [DEBUG] TestCatalogListNodesCommand.server: Skipping self join check for node since the cluster is too small: node=Node-932e30e5-6bf5-c98d-bdda-53e1896f431b writer.go:29: 2021-01-29T19:32:48.286Z [INFO] TestCatalogListNodesCommand.server: member joined, marking health alive: member=Node-932e30e5-6bf5-c98d-bdda-53e1896f431b writer.go:29: 2021-01-29T19:32:48.382Z [DEBUG] TestCatalogListNodesCommand: Skipping remote check since it is managed automatically: check=serfHealth === RUN TestCatalogListNodesCommand/simple === CONT TestCatalogListNodesCommand writer.go:29: 2021-01-29T19:32:48.389Z [INFO] TestCatalogListNodesCommand: Synced node info writer.go:29: 2021-01-29T19:32:48.389Z [DEBUG] TestCatalogListNodesCommand: Node info in sync writer.go:29: 2021-01-29T19:32:48.392Z [DEBUG] TestCatalogListNodesCommand.http: Request finished: method=GET url=/v1/catalog/nodes from=127.0.0.1:42464 latency=2.993677ms === RUN TestCatalogListNodesCommand/detailed === CONT TestCatalogListNodesCommand writer.go:29: 2021-01-29T19:32:48.406Z [DEBUG] TestCatalogListNodesCommand.http: Request finished: method=GET url=/v1/catalog/nodes from=127.0.0.1:42466 latency=269.633µs === RUN TestCatalogListNodesCommand/node-meta === CONT TestCatalogListNodesCommand writer.go:29: 2021-01-29T19:32:48.422Z [DEBUG] TestCatalogListNodesCommand.http: Request finished: method=GET url=/v1/catalog/nodes?node-meta=foo%3Abar from=127.0.0.1:42468 latency=196.179µs === RUN TestCatalogListNodesCommand/filter === CONT TestCatalogListNodesCommand writer.go:29: 2021-01-29T19:32:48.435Z [DEBUG] TestCatalogListNodesCommand.http: Request finished: method=GET url=/v1/catalog/nodes?filter=Meta.foo+%3D%3D+bar from=127.0.0.1:42470 latency=5.568102ms === RUN TestCatalogListNodesCommand/near === CONT TestCatalogListNodesCommand writer.go:29: 2021-01-29T19:32:48.440Z [DEBUG] TestCatalogListNodesCommand.http: Request finished: method=GET url=/v1/catalog/nodes?near=_agent from=127.0.0.1:42472 latency=173.798µs === RUN TestCatalogListNodesCommand/service_present === CONT TestCatalogListNodesCommand writer.go:29: 2021-01-29T19:32:48.453Z [DEBUG] TestCatalogListNodesCommand.http: Request finished: method=GET url=/v1/catalog/service/consul from=127.0.0.1:42474 latency=7.767571ms === RUN TestCatalogListNodesCommand/service_missing === CONT TestCatalogListNodesCommand writer.go:29: 2021-01-29T19:32:48.463Z [DEBUG] TestCatalogListNodesCommand.http: Request finished: method=GET url=/v1/catalog/service/this-service-will-literally-never-exist from=127.0.0.1:42476 latency=1.379541ms writer.go:29: 2021-01-29T19:32:48.464Z [INFO] TestCatalogListNodesCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:48.467Z [INFO] TestCatalogListNodesCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:48.470Z [DEBUG] TestCatalogListNodesCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:48.473Z [WARN] TestCatalogListNodesCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:48.473Z [DEBUG] TestCatalogListNodesCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:48.479Z [WARN] TestCatalogListNodesCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:48.502Z [INFO] TestCatalogListNodesCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:48.505Z [INFO] TestCatalogListNodesCommand: consul server down writer.go:29: 2021-01-29T19:32:48.506Z [INFO] TestCatalogListNodesCommand: shutdown complete writer.go:29: 2021-01-29T19:32:48.507Z [INFO] TestCatalogListNodesCommand: Stopping server: protocol=DNS address=127.0.0.1:18169 network=tcp writer.go:29: 2021-01-29T19:32:48.509Z [INFO] TestCatalogListNodesCommand: Stopping server: protocol=DNS address=127.0.0.1:18169 network=udp writer.go:29: 2021-01-29T19:32:48.510Z [INFO] TestCatalogListNodesCommand: Stopping server: protocol=HTTP address=127.0.0.1:18170 network=tcp writer.go:29: 2021-01-29T19:32:49.011Z [INFO] TestCatalogListNodesCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:49.013Z [INFO] TestCatalogListNodesCommand: Endpoints down --- PASS: TestCatalogListNodesCommand (1.03s) --- PASS: TestCatalogListNodesCommand/simple (0.01s) --- PASS: TestCatalogListNodesCommand/detailed (0.02s) --- PASS: TestCatalogListNodesCommand/node-meta (0.01s) --- PASS: TestCatalogListNodesCommand/filter (0.01s) --- PASS: TestCatalogListNodesCommand/near (0.01s) --- PASS: TestCatalogListNodesCommand/service_present (0.02s) --- PASS: TestCatalogListNodesCommand/service_missing (0.01s) PASS ok github.com/hashicorp/consul/command/catalog/list/nodes 1.239s === RUN TestCatalogListServicesCommand_noTabs === PAUSE TestCatalogListServicesCommand_noTabs === RUN TestCatalogListServicesCommand_Validation === PAUSE TestCatalogListServicesCommand_Validation === RUN TestCatalogListServicesCommand === PAUSE TestCatalogListServicesCommand === CONT TestCatalogListServicesCommand_noTabs === CONT TestCatalogListServicesCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestCatalogListServicesCommand_noTabs (0.00s) === CONT TestCatalogListServicesCommand_Validation --- PASS: TestCatalogListServicesCommand_Validation (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestCatalogListServicesCommand writer.go:29: 2021-01-29T19:32:56.491Z [WARN] TestCatalogListServicesCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:56.492Z [DEBUG] TestCatalogListServicesCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:56.495Z [DEBUG] TestCatalogListServicesCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:56.501Z [INFO] TestCatalogListServicesCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:688eab80-70bf-1b43-f644-6d9e4841e801 Address:127.0.0.1:21237}]" writer.go:29: 2021-01-29T19:32:56.504Z [INFO] TestCatalogListServicesCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:21237 [Follower]" leader= writer.go:29: 2021-01-29T19:32:56.507Z [INFO] TestCatalogListServicesCommand.server.serf.wan: serf: EventMemberJoin: Node-688eab80-70bf-1b43-f644-6d9e4841e801.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:56.511Z [INFO] TestCatalogListServicesCommand.server.serf.lan: serf: EventMemberJoin: Node-688eab80-70bf-1b43-f644-6d9e4841e801 127.0.0.1 writer.go:29: 2021-01-29T19:32:56.513Z [INFO] TestCatalogListServicesCommand.server: Adding LAN server: server="Node-688eab80-70bf-1b43-f644-6d9e4841e801 (Addr: tcp/127.0.0.1:21237) (DC: dc1)" writer.go:29: 2021-01-29T19:32:56.513Z [INFO] TestCatalogListServicesCommand: Started DNS server: address=127.0.0.1:21232 network=udp writer.go:29: 2021-01-29T19:32:56.518Z [INFO] TestCatalogListServicesCommand: Started DNS server: address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:32:56.513Z [INFO] TestCatalogListServicesCommand.server: Handled event for server in area: event=member-join server=Node-688eab80-70bf-1b43-f644-6d9e4841e801.dc1 area=wan writer.go:29: 2021-01-29T19:32:56.519Z [INFO] TestCatalogListServicesCommand: Started HTTP server: address=127.0.0.1:21233 network=tcp writer.go:29: 2021-01-29T19:32:56.522Z [INFO] TestCatalogListServicesCommand: started state syncer writer.go:29: 2021-01-29T19:32:56.562Z [WARN] TestCatalogListServicesCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:56.564Z [INFO] TestCatalogListServicesCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:21237 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:56.567Z [DEBUG] TestCatalogListServicesCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:56.569Z [DEBUG] TestCatalogListServicesCommand.server.raft: vote granted: from=688eab80-70bf-1b43-f644-6d9e4841e801 term=2 tally=1 writer.go:29: 2021-01-29T19:32:56.570Z [INFO] TestCatalogListServicesCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:56.572Z [INFO] TestCatalogListServicesCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:21237 [Leader]" writer.go:29: 2021-01-29T19:32:56.574Z [INFO] TestCatalogListServicesCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:56.574Z [INFO] TestCatalogListServicesCommand.server: New leader elected: payload=Node-688eab80-70bf-1b43-f644-6d9e4841e801 writer.go:29: 2021-01-29T19:32:56.575Z [DEBUG] TestCatalogListServicesCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:21237 writer.go:29: 2021-01-29T19:32:56.580Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:56.620Z [INFO] TestCatalogListServicesCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:56.621Z [INFO] TestCatalogListServicesCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.622Z [DEBUG] TestCatalogListServicesCommand.server: Skipping self join check for node since the cluster is too small: node=Node-688eab80-70bf-1b43-f644-6d9e4841e801 writer.go:29: 2021-01-29T19:32:56.623Z [INFO] TestCatalogListServicesCommand.server: member joined, marking health alive: member=Node-688eab80-70bf-1b43-f644-6d9e4841e801 writer.go:29: 2021-01-29T19:32:56.765Z [INFO] TestCatalogListServicesCommand: Synced node info writer.go:29: 2021-01-29T19:32:56.769Z [INFO] TestCatalogListServicesCommand: Synced service: service=testing writer.go:29: 2021-01-29T19:32:56.771Z [DEBUG] TestCatalogListServicesCommand.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:43008 latency=21.754938ms === RUN TestCatalogListServicesCommand/simple === CONT TestCatalogListServicesCommand writer.go:29: 2021-01-29T19:32:56.784Z [DEBUG] TestCatalogListServicesCommand.http: Request finished: method=GET url=/v1/catalog/services from=127.0.0.1:43010 latency=253.467µs === RUN TestCatalogListServicesCommand/tags === CONT TestCatalogListServicesCommand writer.go:29: 2021-01-29T19:32:56.794Z [DEBUG] TestCatalogListServicesCommand.http: Request finished: method=GET url=/v1/catalog/services from=127.0.0.1:43012 latency=1.306707ms === RUN TestCatalogListServicesCommand/node_missing === CONT TestCatalogListServicesCommand writer.go:29: 2021-01-29T19:32:56.812Z [DEBUG] TestCatalogListServicesCommand.http: Request finished: method=GET url=/v1/catalog/node/not-a-real-node from=127.0.0.1:43014 latency=12.135897ms === RUN TestCatalogListServicesCommand/node_present === CONT TestCatalogListServicesCommand writer.go:29: 2021-01-29T19:32:56.823Z [DEBUG] TestCatalogListServicesCommand.http: Request finished: method=GET url=/v1/catalog/node/Node-688eab80-70bf-1b43-f644-6d9e4841e801 from=127.0.0.1:43016 latency=374.974µs === RUN TestCatalogListServicesCommand/node-meta === CONT TestCatalogListServicesCommand writer.go:29: 2021-01-29T19:32:56.833Z [DEBUG] TestCatalogListServicesCommand.http: Request finished: method=GET url=/v1/catalog/services?node-meta=foo%3Abar from=127.0.0.1:43018 latency=311.851µs writer.go:29: 2021-01-29T19:32:56.834Z [INFO] TestCatalogListServicesCommand: Requesting shutdown writer.go:29: 2021-01-29T19:32:56.837Z [INFO] TestCatalogListServicesCommand.server: shutting down server writer.go:29: 2021-01-29T19:32:56.839Z [DEBUG] TestCatalogListServicesCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.839Z [ERROR] TestCatalogListServicesCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:56.842Z [WARN] TestCatalogListServicesCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:56.842Z [DEBUG] TestCatalogListServicesCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:56.845Z [WARN] TestCatalogListServicesCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:56.848Z [INFO] TestCatalogListServicesCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:56.848Z [INFO] TestCatalogListServicesCommand: consul server down writer.go:29: 2021-01-29T19:32:56.850Z [INFO] TestCatalogListServicesCommand: shutdown complete writer.go:29: 2021-01-29T19:32:56.851Z [INFO] TestCatalogListServicesCommand: Stopping server: protocol=DNS address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:32:56.852Z [INFO] TestCatalogListServicesCommand: Stopping server: protocol=DNS address=127.0.0.1:21232 network=udp writer.go:29: 2021-01-29T19:32:56.854Z [INFO] TestCatalogListServicesCommand: Stopping server: protocol=HTTP address=127.0.0.1:21233 network=tcp writer.go:29: 2021-01-29T19:32:57.356Z [INFO] TestCatalogListServicesCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:57.357Z [INFO] TestCatalogListServicesCommand: Endpoints down --- PASS: TestCatalogListServicesCommand (0.91s) --- PASS: TestCatalogListServicesCommand/simple (0.01s) --- PASS: TestCatalogListServicesCommand/tags (0.01s) --- PASS: TestCatalogListServicesCommand/node_missing (0.02s) --- PASS: TestCatalogListServicesCommand/node_present (0.01s) --- PASS: TestCatalogListServicesCommand/node-meta (0.01s) PASS ok github.com/hashicorp/consul/command/catalog/list/services 1.083s ? github.com/hashicorp/consul/command/config [no test files] === RUN TestConfigDelete_noTabs === PAUSE TestConfigDelete_noTabs === RUN TestConfigDelete === PAUSE TestConfigDelete === RUN TestConfigDelete_InvalidArgs === PAUSE TestConfigDelete_InvalidArgs === CONT TestConfigDelete [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestConfigDelete_InvalidArgs === CONT TestConfigDelete_noTabs === RUN TestConfigDelete_InvalidArgs/no_kind --- PASS: TestConfigDelete_noTabs (0.01s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === RUN TestConfigDelete_InvalidArgs/no_name --- PASS: TestConfigDelete_InvalidArgs (0.06s) --- PASS: TestConfigDelete_InvalidArgs/no_kind (0.05s) --- PASS: TestConfigDelete_InvalidArgs/no_name (0.01s) === CONT TestConfigDelete writer.go:29: 2021-01-29T19:32:55.495Z [WARN] TestConfigDelete: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:55.502Z [DEBUG] TestConfigDelete.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:55.505Z [DEBUG] TestConfigDelete.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:55.517Z [INFO] TestConfigDelete.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9e7d78d8-63fb-da43-5d40-eae54dafb0a8 Address:127.0.0.1:15111}]" writer.go:29: 2021-01-29T19:32:55.520Z [INFO] TestConfigDelete.server.raft: entering follower state: follower="Node at 127.0.0.1:15111 [Follower]" leader= writer.go:29: 2021-01-29T19:32:55.522Z [INFO] TestConfigDelete.server.serf.wan: serf: EventMemberJoin: Node-9e7d78d8-63fb-da43-5d40-eae54dafb0a8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:55.531Z [INFO] TestConfigDelete.server.serf.lan: serf: EventMemberJoin: Node-9e7d78d8-63fb-da43-5d40-eae54dafb0a8 127.0.0.1 writer.go:29: 2021-01-29T19:32:55.540Z [INFO] TestConfigDelete.server: Adding LAN server: server="Node-9e7d78d8-63fb-da43-5d40-eae54dafb0a8 (Addr: tcp/127.0.0.1:15111) (DC: dc1)" writer.go:29: 2021-01-29T19:32:55.543Z [INFO] TestConfigDelete.server: Handled event for server in area: event=member-join server=Node-9e7d78d8-63fb-da43-5d40-eae54dafb0a8.dc1 area=wan writer.go:29: 2021-01-29T19:32:55.547Z [INFO] TestConfigDelete: Started DNS server: address=127.0.0.1:15106 network=tcp writer.go:29: 2021-01-29T19:32:55.549Z [INFO] TestConfigDelete: Started DNS server: address=127.0.0.1:15106 network=udp writer.go:29: 2021-01-29T19:32:55.566Z [INFO] TestConfigDelete: Started HTTP server: address=127.0.0.1:15107 network=tcp writer.go:29: 2021-01-29T19:32:55.569Z [INFO] TestConfigDelete: started state syncer writer.go:29: 2021-01-29T19:32:55.578Z [WARN] TestConfigDelete.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:55.579Z [INFO] TestConfigDelete.server.raft: entering candidate state: node="Node at 127.0.0.1:15111 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:55.582Z [DEBUG] TestConfigDelete.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:55.583Z [DEBUG] TestConfigDelete.server.raft: vote granted: from=9e7d78d8-63fb-da43-5d40-eae54dafb0a8 term=2 tally=1 writer.go:29: 2021-01-29T19:32:55.584Z [INFO] TestConfigDelete.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:55.585Z [INFO] TestConfigDelete.server.raft: entering leader state: leader="Node at 127.0.0.1:15111 [Leader]" writer.go:29: 2021-01-29T19:32:55.586Z [INFO] TestConfigDelete.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:55.587Z [INFO] TestConfigDelete.server: New leader elected: payload=Node-9e7d78d8-63fb-da43-5d40-eae54dafb0a8 writer.go:29: 2021-01-29T19:32:55.589Z [DEBUG] TestConfigDelete.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15111 writer.go:29: 2021-01-29T19:32:55.603Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:55.652Z [INFO] TestConfigDelete.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:55.654Z [INFO] TestConfigDelete.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.656Z [DEBUG] TestConfigDelete.server: Skipping self join check for node since the cluster is too small: node=Node-9e7d78d8-63fb-da43-5d40-eae54dafb0a8 writer.go:29: 2021-01-29T19:32:55.658Z [INFO] TestConfigDelete.server: member joined, marking health alive: member=Node-9e7d78d8-63fb-da43-5d40-eae54dafb0a8 writer.go:29: 2021-01-29T19:32:55.834Z [DEBUG] TestConfigDelete.http: Request finished: method=PUT url=/v1/config from=127.0.0.1:37252 latency=4.028665ms writer.go:29: 2021-01-29T19:32:55.844Z [DEBUG] TestConfigDelete.http: Request finished: method=DELETE url=/v1/config/service-defaults/web from=127.0.0.1:37254 latency=5.557067ms writer.go:29: 2021-01-29T19:32:55.847Z [ERROR] TestConfigDelete.http: Request error: method=GET url=/v1/config/service-defaults/web from=127.0.0.1:37252 error="Config entry not found for "service-defaults" / "web"" writer.go:29: 2021-01-29T19:32:55.848Z [DEBUG] TestConfigDelete.http: Request finished: method=GET url=/v1/config/service-defaults/web from=127.0.0.1:37252 latency=1.261732ms writer.go:29: 2021-01-29T19:32:55.851Z [INFO] TestConfigDelete: Requesting shutdown writer.go:29: 2021-01-29T19:32:55.851Z [INFO] TestConfigDelete.server: shutting down server writer.go:29: 2021-01-29T19:32:55.852Z [DEBUG] TestConfigDelete.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.853Z [WARN] TestConfigDelete.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:55.855Z [ERROR] TestConfigDelete.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:32:55.855Z [DEBUG] TestConfigDelete.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:55.855Z [WARN] TestConfigDelete.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:55.859Z [INFO] TestConfigDelete.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:55.860Z [INFO] TestConfigDelete: consul server down writer.go:29: 2021-01-29T19:32:55.862Z [INFO] TestConfigDelete: shutdown complete writer.go:29: 2021-01-29T19:32:55.862Z [INFO] TestConfigDelete: Stopping server: protocol=DNS address=127.0.0.1:15106 network=tcp writer.go:29: 2021-01-29T19:32:55.863Z [INFO] TestConfigDelete: Stopping server: protocol=DNS address=127.0.0.1:15106 network=udp writer.go:29: 2021-01-29T19:32:55.864Z [INFO] TestConfigDelete: Stopping server: protocol=HTTP address=127.0.0.1:15107 network=tcp writer.go:29: 2021-01-29T19:32:56.365Z [INFO] TestConfigDelete: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:56.367Z [INFO] TestConfigDelete: Endpoints down --- PASS: TestConfigDelete (0.93s) PASS ok github.com/hashicorp/consul/command/config/delete 1.112s === RUN TestConfigList_noTabs === PAUSE TestConfigList_noTabs === RUN TestConfigList [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:32:57.226Z [WARN] TestConfigList: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:32:57.227Z [DEBUG] TestConfigList.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:32:57.229Z [DEBUG] TestConfigList.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:32:57.234Z [INFO] TestConfigList.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f806030f-a33d-e195-1614-a2938d10a98f Address:127.0.0.1:10006}]" writer.go:29: 2021-01-29T19:32:57.236Z [INFO] TestConfigList.server.raft: entering follower state: follower="Node at 127.0.0.1:10006 [Follower]" leader= writer.go:29: 2021-01-29T19:32:57.238Z [INFO] TestConfigList.server.serf.wan: serf: EventMemberJoin: Node-f806030f-a33d-e195-1614-a2938d10a98f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:32:57.243Z [INFO] TestConfigList.server.serf.lan: serf: EventMemberJoin: Node-f806030f-a33d-e195-1614-a2938d10a98f 127.0.0.1 writer.go:29: 2021-01-29T19:32:57.249Z [INFO] TestConfigList.server: Adding LAN server: server="Node-f806030f-a33d-e195-1614-a2938d10a98f (Addr: tcp/127.0.0.1:10006) (DC: dc1)" writer.go:29: 2021-01-29T19:32:57.250Z [INFO] TestConfigList: Started DNS server: address=127.0.0.1:10001 network=udp writer.go:29: 2021-01-29T19:32:57.250Z [INFO] TestConfigList.server: Handled event for server in area: event=member-join server=Node-f806030f-a33d-e195-1614-a2938d10a98f.dc1 area=wan writer.go:29: 2021-01-29T19:32:57.253Z [INFO] TestConfigList: Started DNS server: address=127.0.0.1:10001 network=tcp writer.go:29: 2021-01-29T19:32:57.255Z [INFO] TestConfigList: Started HTTP server: address=127.0.0.1:10002 network=tcp writer.go:29: 2021-01-29T19:32:57.257Z [INFO] TestConfigList: started state syncer writer.go:29: 2021-01-29T19:32:57.279Z [WARN] TestConfigList.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:32:57.284Z [INFO] TestConfigList.server.raft: entering candidate state: node="Node at 127.0.0.1:10006 [Candidate]" term=2 writer.go:29: 2021-01-29T19:32:57.287Z [DEBUG] TestConfigList.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:32:57.288Z [DEBUG] TestConfigList.server.raft: vote granted: from=f806030f-a33d-e195-1614-a2938d10a98f term=2 tally=1 writer.go:29: 2021-01-29T19:32:57.290Z [INFO] TestConfigList.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:32:57.291Z [INFO] TestConfigList.server.raft: entering leader state: leader="Node at 127.0.0.1:10006 [Leader]" writer.go:29: 2021-01-29T19:32:57.293Z [INFO] TestConfigList.server: cluster leadership acquired writer.go:29: 2021-01-29T19:32:57.294Z [INFO] TestConfigList.server: New leader elected: payload=Node-f806030f-a33d-e195-1614-a2938d10a98f writer.go:29: 2021-01-29T19:32:57.300Z [DEBUG] TestConfigList.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:10006 writer.go:29: 2021-01-29T19:32:57.308Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:32:57.344Z [INFO] TestConfigList.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:32:57.346Z [INFO] TestConfigList.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.347Z [DEBUG] TestConfigList.server: Skipping self join check for node since the cluster is too small: node=Node-f806030f-a33d-e195-1614-a2938d10a98f writer.go:29: 2021-01-29T19:32:57.345Z [INFO] TestConfigList: Synced node info writer.go:29: 2021-01-29T19:32:57.350Z [DEBUG] TestConfigList: Node info in sync writer.go:29: 2021-01-29T19:32:57.349Z [INFO] TestConfigList.server: member joined, marking health alive: member=Node-f806030f-a33d-e195-1614-a2938d10a98f writer.go:29: 2021-01-29T19:32:57.409Z [DEBUG] TestConfigList.http: Request finished: method=PUT url=/v1/config from=127.0.0.1:42928 latency=4.190585ms writer.go:29: 2021-01-29T19:32:57.428Z [DEBUG] TestConfigList.http: Request finished: method=PUT url=/v1/config from=127.0.0.1:42928 latency=5.780615ms writer.go:29: 2021-01-29T19:32:57.445Z [DEBUG] TestConfigList.http: Request finished: method=PUT url=/v1/config from=127.0.0.1:42928 latency=7.639554ms writer.go:29: 2021-01-29T19:32:57.458Z [DEBUG] TestConfigList.http: Request finished: method=GET url=/v1/config/service-defaults from=127.0.0.1:42930 latency=2.980714ms writer.go:29: 2021-01-29T19:32:57.464Z [INFO] TestConfigList: Requesting shutdown writer.go:29: 2021-01-29T19:32:57.468Z [INFO] TestConfigList.server: shutting down server writer.go:29: 2021-01-29T19:32:57.469Z [DEBUG] TestConfigList.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.472Z [WARN] TestConfigList.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:57.472Z [DEBUG] TestConfigList.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:32:57.475Z [WARN] TestConfigList.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:32:57.480Z [INFO] TestConfigList.server.router.manager: shutting down writer.go:29: 2021-01-29T19:32:57.481Z [INFO] TestConfigList: consul server down writer.go:29: 2021-01-29T19:32:57.482Z [INFO] TestConfigList: shutdown complete writer.go:29: 2021-01-29T19:32:57.483Z [INFO] TestConfigList: Stopping server: protocol=DNS address=127.0.0.1:10001 network=tcp writer.go:29: 2021-01-29T19:32:57.484Z [INFO] TestConfigList: Stopping server: protocol=DNS address=127.0.0.1:10001 network=udp writer.go:29: 2021-01-29T19:32:57.484Z [INFO] TestConfigList: Stopping server: protocol=HTTP address=127.0.0.1:10002 network=tcp writer.go:29: 2021-01-29T19:32:57.985Z [INFO] TestConfigList: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:32:57.987Z [INFO] TestConfigList: Endpoints down --- PASS: TestConfigList (0.82s) === RUN TestConfigList_InvalidArgs === PAUSE TestConfigList_InvalidArgs === CONT TestConfigList_noTabs --- PASS: TestConfigList_noTabs (0.00s) === CONT TestConfigList_InvalidArgs === RUN TestConfigList_InvalidArgs/no_kind --- PASS: TestConfigList_InvalidArgs (0.00s) --- PASS: TestConfigList_InvalidArgs/no_kind (0.00s) PASS ok github.com/hashicorp/consul/command/config/list 0.976s === RUN TestConfigRead_noTabs === PAUSE TestConfigRead_noTabs === RUN TestConfigRead === PAUSE TestConfigRead === RUN TestConfigRead_InvalidArgs === PAUSE TestConfigRead_InvalidArgs === CONT TestConfigRead_noTabs === CONT TestConfigRead [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestConfigRead_InvalidArgs [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestConfigRead_noTabs (0.01s) === RUN TestConfigRead_InvalidArgs/no_kind === RUN TestConfigRead_InvalidArgs/no_name --- PASS: TestConfigRead_InvalidArgs (0.01s) --- PASS: TestConfigRead_InvalidArgs/no_kind (0.00s) --- PASS: TestConfigRead_InvalidArgs/no_name (0.00s) === CONT TestConfigRead writer.go:29: 2021-01-29T19:33:03.955Z [WARN] TestConfigRead: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:03.970Z [DEBUG] TestConfigRead.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:03.971Z [DEBUG] TestConfigRead.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:03.982Z [INFO] TestConfigRead.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2a1767a3-6a72-8c57-0ac6-ebbddb630e15 Address:127.0.0.1:26342}]" writer.go:29: 2021-01-29T19:33:03.984Z [INFO] TestConfigRead.server.serf.wan: serf: EventMemberJoin: Node-2a1767a3-6a72-8c57-0ac6-ebbddb630e15.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:03.986Z [INFO] TestConfigRead.server.serf.lan: serf: EventMemberJoin: Node-2a1767a3-6a72-8c57-0ac6-ebbddb630e15 127.0.0.1 writer.go:29: 2021-01-29T19:33:03.988Z [INFO] TestConfigRead: Started DNS server: address=127.0.0.1:26337 network=udp writer.go:29: 2021-01-29T19:33:03.989Z [INFO] TestConfigRead.server.raft: entering follower state: follower="Node at 127.0.0.1:26342 [Follower]" leader= writer.go:29: 2021-01-29T19:33:03.995Z [INFO] TestConfigRead.server: Handled event for server in area: event=member-join server=Node-2a1767a3-6a72-8c57-0ac6-ebbddb630e15.dc1 area=wan writer.go:29: 2021-01-29T19:33:03.995Z [INFO] TestConfigRead.server: Adding LAN server: server="Node-2a1767a3-6a72-8c57-0ac6-ebbddb630e15 (Addr: tcp/127.0.0.1:26342) (DC: dc1)" writer.go:29: 2021-01-29T19:33:03.997Z [INFO] TestConfigRead: Started DNS server: address=127.0.0.1:26337 network=tcp writer.go:29: 2021-01-29T19:33:04.000Z [INFO] TestConfigRead: Started HTTP server: address=127.0.0.1:26338 network=tcp writer.go:29: 2021-01-29T19:33:04.003Z [INFO] TestConfigRead: started state syncer writer.go:29: 2021-01-29T19:33:04.041Z [WARN] TestConfigRead.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:04.042Z [INFO] TestConfigRead.server.raft: entering candidate state: node="Node at 127.0.0.1:26342 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:04.052Z [DEBUG] TestConfigRead.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:04.053Z [DEBUG] TestConfigRead.server.raft: vote granted: from=2a1767a3-6a72-8c57-0ac6-ebbddb630e15 term=2 tally=1 writer.go:29: 2021-01-29T19:33:04.054Z [INFO] TestConfigRead.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:04.055Z [INFO] TestConfigRead.server.raft: entering leader state: leader="Node at 127.0.0.1:26342 [Leader]" writer.go:29: 2021-01-29T19:33:04.057Z [INFO] TestConfigRead.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:04.058Z [INFO] TestConfigRead.server: New leader elected: payload=Node-2a1767a3-6a72-8c57-0ac6-ebbddb630e15 writer.go:29: 2021-01-29T19:33:04.058Z [DEBUG] TestConfigRead.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:26342 writer.go:29: 2021-01-29T19:33:04.096Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:04.134Z [INFO] TestConfigRead.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:04.135Z [INFO] TestConfigRead.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.136Z [DEBUG] TestConfigRead.server: Skipping self join check for node since the cluster is too small: node=Node-2a1767a3-6a72-8c57-0ac6-ebbddb630e15 writer.go:29: 2021-01-29T19:33:04.136Z [INFO] TestConfigRead.server: member joined, marking health alive: member=Node-2a1767a3-6a72-8c57-0ac6-ebbddb630e15 writer.go:29: 2021-01-29T19:33:04.277Z [DEBUG] TestConfigRead: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:04.280Z [INFO] TestConfigRead: Synced node info writer.go:29: 2021-01-29T19:33:04.281Z [DEBUG] TestConfigRead: Node info in sync writer.go:29: 2021-01-29T19:33:04.374Z [DEBUG] TestConfigRead.http: Request finished: method=PUT url=/v1/config from=127.0.0.1:57766 latency=3.325107ms writer.go:29: 2021-01-29T19:33:04.381Z [DEBUG] TestConfigRead.http: Request finished: method=GET url=/v1/config/service-defaults/web from=127.0.0.1:57768 latency=1.296209ms writer.go:29: 2021-01-29T19:33:04.386Z [INFO] TestConfigRead: Requesting shutdown writer.go:29: 2021-01-29T19:33:04.387Z [INFO] TestConfigRead.server: shutting down server writer.go:29: 2021-01-29T19:33:04.389Z [DEBUG] TestConfigRead.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.391Z [WARN] TestConfigRead.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:04.391Z [DEBUG] TestConfigRead.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:04.395Z [WARN] TestConfigRead.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:04.397Z [INFO] TestConfigRead.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:04.397Z [INFO] TestConfigRead: consul server down writer.go:29: 2021-01-29T19:33:04.399Z [INFO] TestConfigRead: shutdown complete writer.go:29: 2021-01-29T19:33:04.399Z [INFO] TestConfigRead: Stopping server: protocol=DNS address=127.0.0.1:26337 network=tcp writer.go:29: 2021-01-29T19:33:04.400Z [INFO] TestConfigRead: Stopping server: protocol=DNS address=127.0.0.1:26337 network=udp writer.go:29: 2021-01-29T19:33:04.401Z [INFO] TestConfigRead: Stopping server: protocol=HTTP address=127.0.0.1:26338 network=tcp writer.go:29: 2021-01-29T19:33:04.902Z [INFO] TestConfigRead: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:04.903Z [INFO] TestConfigRead: Endpoints down --- PASS: TestConfigRead (1.01s) PASS ok github.com/hashicorp/consul/command/config/read 1.146s === RUN TestConfigWrite_noTabs === PAUSE TestConfigWrite_noTabs === RUN TestConfigWrite === PAUSE TestConfigWrite === RUN TestParseConfigEntry === PAUSE TestParseConfigEntry === CONT TestParseConfigEntry === CONT TestConfigWrite === CONT TestConfigWrite_noTabs === RUN TestParseConfigEntry/proxy-defaults:_extra_fields_or_typo_(hcl_snake_case) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestConfigWrite_noTabs (0.00s) === RUN TestParseConfigEntry/proxy-defaults:_extra_fields_or_typo_(hcl_camel_case) === RUN TestParseConfigEntry/proxy-defaults:_extra_fields_or_typo_(json_snake_case) === RUN TestParseConfigEntry/proxy-defaults:_extra_fields_or_typo_(json_camel_case) === RUN TestParseConfigEntry/proxy-defaults_(hcl_snake_case) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === RUN TestParseConfigEntry/proxy-defaults_(hcl_camel_case) === RUN TestParseConfigEntry/proxy-defaults_(json_snake_case) === RUN TestParseConfigEntry/proxy-defaults_(json_camel_case) === RUN TestParseConfigEntry/service-defaults_(hcl_snake_case) === RUN TestParseConfigEntry/service-defaults_(hcl_camel_case) === RUN TestParseConfigEntry/service-defaults_(json_snake_case) === RUN TestParseConfigEntry/service-defaults_(json_camel_case) === RUN TestParseConfigEntry/service-router:_kitchen_sink_(hcl_snake_case) === RUN TestParseConfigEntry/service-router:_kitchen_sink_(hcl_camel_case) === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.113Z [WARN] TestConfigWrite: bootstrap = true: do not enable unless necessary === RUN TestParseConfigEntry/service-router:_kitchen_sink_(json_snake_case) === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.114Z [DEBUG] TestConfigWrite.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:05.118Z [DEBUG] TestConfigWrite.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:05.124Z [INFO] TestConfigWrite.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6b5529ef-b663-c503-c136-483a7d557311 Address:127.0.0.1:22258}]" writer.go:29: 2021-01-29T19:33:05.125Z [INFO] TestConfigWrite.server.raft: entering follower state: follower="Node at 127.0.0.1:22258 [Follower]" leader= writer.go:29: 2021-01-29T19:33:05.130Z [INFO] TestConfigWrite.server.serf.wan: serf: EventMemberJoin: Node-6b5529ef-b663-c503-c136-483a7d557311.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:05.133Z [INFO] TestConfigWrite.server.serf.lan: serf: EventMemberJoin: Node-6b5529ef-b663-c503-c136-483a7d557311 127.0.0.1 writer.go:29: 2021-01-29T19:33:05.136Z [INFO] TestConfigWrite: Started DNS server: address=127.0.0.1:22253 network=udp writer.go:29: 2021-01-29T19:33:05.136Z [INFO] TestConfigWrite.server: Adding LAN server: server="Node-6b5529ef-b663-c503-c136-483a7d557311 (Addr: tcp/127.0.0.1:22258) (DC: dc1)" writer.go:29: 2021-01-29T19:33:05.138Z [INFO] TestConfigWrite: Started DNS server: address=127.0.0.1:22253 network=tcp === RUN TestParseConfigEntry/service-router:_kitchen_sink_(json_camel_case) === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.141Z [INFO] TestConfigWrite: Started HTTP server: address=127.0.0.1:22254 network=tcp writer.go:29: 2021-01-29T19:33:05.136Z [INFO] TestConfigWrite.server: Handled event for server in area: event=member-join server=Node-6b5529ef-b663-c503-c136-483a7d557311.dc1 area=wan writer.go:29: 2021-01-29T19:33:05.142Z [INFO] TestConfigWrite: started state syncer === RUN TestParseConfigEntry/service-splitter:_kitchen_sink_(hcl_snake_case) === RUN TestParseConfigEntry/service-splitter:_kitchen_sink_(hcl_camel_case) === RUN TestParseConfigEntry/service-splitter:_kitchen_sink_(json_snake_case) === RUN TestParseConfigEntry/service-splitter:_kitchen_sink_(json_camel_case) === RUN TestParseConfigEntry/service-resolver:_subsets_with_failover_(hcl_snake_case) === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.181Z [WARN] TestConfigWrite.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:05.182Z [INFO] TestConfigWrite.server.raft: entering candidate state: node="Node at 127.0.0.1:22258 [Candidate]" term=2 === RUN TestParseConfigEntry/service-resolver:_subsets_with_failover_(hcl_camel_case) === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.183Z [DEBUG] TestConfigWrite.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:05.184Z [DEBUG] TestConfigWrite.server.raft: vote granted: from=6b5529ef-b663-c503-c136-483a7d557311 term=2 tally=1 writer.go:29: 2021-01-29T19:33:05.185Z [INFO] TestConfigWrite.server.raft: election won: tally=1 === RUN TestParseConfigEntry/service-resolver:_subsets_with_failover_(json_snake_case) === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.187Z [INFO] TestConfigWrite.server.raft: entering leader state: leader="Node at 127.0.0.1:22258 [Leader]" === RUN TestParseConfigEntry/service-resolver:_subsets_with_failover_(json_camel_case) === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.188Z [INFO] TestConfigWrite.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:05.189Z [INFO] TestConfigWrite.server: New leader elected: payload=Node-6b5529ef-b663-c503-c136-483a7d557311 === RUN TestParseConfigEntry/service-resolver:_redirect_(hcl_snake_case) === RUN TestParseConfigEntry/service-resolver:_redirect_(hcl_camel_case) === RUN TestParseConfigEntry/service-resolver:_redirect_(json_snake_case) === RUN TestParseConfigEntry/service-resolver:_redirect_(json_camel_case) === RUN TestConfigWrite/File === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.190Z [DEBUG] TestConfigWrite.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22258 writer.go:29: 2021-01-29T19:33:05.225Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === RUN TestParseConfigEntry/service-resolver:_default_(hcl_snake_case) === RUN TestParseConfigEntry/service-resolver:_default_(hcl_camel_case) === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.284Z [DEBUG] TestConfigWrite.http: Request finished: method=PUT url=/v1/config from=127.0.0.1:48262 latency=19.107665ms === RUN TestParseConfigEntry/service-resolver:_default_(json_snake_case) === RUN TestParseConfigEntry/service-resolver:_default_(json_camel_case) === RUN TestParseConfigEntry/expose_paths:_kitchen_sink_proxy_defaults_(hcl_snake_case) === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.323Z [DEBUG] TestConfigWrite.http: Request finished: method=GET url=/v1/config/service-defaults/web from=127.0.0.1:48264 latency=27.956599ms === RUN TestParseConfigEntry/expose_paths:_kitchen_sink_proxy_defaults_(hcl_camel_case) === RUN TestParseConfigEntry/expose_paths:_kitchen_sink_proxy_defaults_(json_snake_case) === RUN TestConfigWrite/Stdin === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.345Z [INFO] TestConfigWrite.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:05.346Z [INFO] TestConfigWrite.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.347Z [DEBUG] TestConfigWrite.server: Skipping self join check for node since the cluster is too small: node=Node-6b5529ef-b663-c503-c136-483a7d557311 writer.go:29: 2021-01-29T19:33:05.347Z [INFO] TestConfigWrite.server: member joined, marking health alive: member=Node-6b5529ef-b663-c503-c136-483a7d557311 writer.go:29: 2021-01-29T19:33:05.346Z [DEBUG] TestConfigWrite.http: Request finished: method=PUT url=/v1/config from=127.0.0.1:48266 latency=2.669277ms writer.go:29: 2021-01-29T19:33:05.351Z [DEBUG] TestConfigWrite.http: Request finished: method=GET url=/v1/config/proxy-defaults/global from=127.0.0.1:48264 latency=104.508µs === RUN TestParseConfigEntry/expose_paths:_kitchen_sink_proxy_defaults_(json_camel_case) === RUN TestConfigWrite/No_config --- PASS: TestParseConfigEntry (0.33s) --- PASS: TestParseConfigEntry/proxy-defaults:_extra_fields_or_typo_(hcl_snake_case) (0.00s) --- PASS: TestParseConfigEntry/proxy-defaults:_extra_fields_or_typo_(hcl_camel_case) (0.00s) --- PASS: TestParseConfigEntry/proxy-defaults:_extra_fields_or_typo_(json_snake_case) (0.00s) --- PASS: TestParseConfigEntry/proxy-defaults:_extra_fields_or_typo_(json_camel_case) (0.00s) --- PASS: TestParseConfigEntry/proxy-defaults_(hcl_snake_case) (0.00s) --- PASS: TestParseConfigEntry/proxy-defaults_(hcl_camel_case) (0.00s) --- PASS: TestParseConfigEntry/proxy-defaults_(json_snake_case) (0.00s) --- PASS: TestParseConfigEntry/proxy-defaults_(json_camel_case) (0.00s) --- PASS: TestParseConfigEntry/service-defaults_(hcl_snake_case) (0.00s) --- PASS: TestParseConfigEntry/service-defaults_(hcl_camel_case) (0.01s) --- PASS: TestParseConfigEntry/service-defaults_(json_snake_case) (0.01s) --- PASS: TestParseConfigEntry/service-defaults_(json_camel_case) (0.01s) --- PASS: TestParseConfigEntry/service-router:_kitchen_sink_(hcl_snake_case) (0.02s) --- PASS: TestParseConfigEntry/service-router:_kitchen_sink_(hcl_camel_case) (0.00s) --- PASS: TestParseConfigEntry/service-router:_kitchen_sink_(json_snake_case) (0.02s) --- PASS: TestParseConfigEntry/service-router:_kitchen_sink_(json_camel_case) (0.01s) --- PASS: TestParseConfigEntry/service-splitter:_kitchen_sink_(hcl_snake_case) (0.00s) --- PASS: TestParseConfigEntry/service-splitter:_kitchen_sink_(hcl_camel_case) (0.00s) --- PASS: TestParseConfigEntry/service-splitter:_kitchen_sink_(json_snake_case) (0.01s) --- PASS: TestParseConfigEntry/service-splitter:_kitchen_sink_(json_camel_case) (0.00s) --- PASS: TestParseConfigEntry/service-resolver:_subsets_with_failover_(hcl_snake_case) (0.01s) --- PASS: TestParseConfigEntry/service-resolver:_subsets_with_failover_(hcl_camel_case) (0.00s) --- PASS: TestParseConfigEntry/service-resolver:_subsets_with_failover_(json_snake_case) (0.00s) --- PASS: TestParseConfigEntry/service-resolver:_subsets_with_failover_(json_camel_case) (0.01s) --- PASS: TestParseConfigEntry/service-resolver:_redirect_(hcl_snake_case) (0.00s) --- PASS: TestParseConfigEntry/service-resolver:_redirect_(hcl_camel_case) (0.00s) --- PASS: TestParseConfigEntry/service-resolver:_redirect_(json_snake_case) (0.00s) --- PASS: TestParseConfigEntry/service-resolver:_redirect_(json_camel_case) (0.03s) --- PASS: TestParseConfigEntry/service-resolver:_default_(hcl_snake_case) (0.01s) --- PASS: TestParseConfigEntry/service-resolver:_default_(hcl_camel_case) (0.00s) --- PASS: TestParseConfigEntry/service-resolver:_default_(json_snake_case) (0.00s) --- PASS: TestParseConfigEntry/service-resolver:_default_(json_camel_case) (0.00s) --- PASS: TestParseConfigEntry/expose_paths:_kitchen_sink_proxy_defaults_(hcl_snake_case) (0.00s) --- PASS: TestParseConfigEntry/expose_paths:_kitchen_sink_proxy_defaults_(hcl_camel_case) (0.00s) --- PASS: TestParseConfigEntry/expose_paths:_kitchen_sink_proxy_defaults_(json_snake_case) (0.02s) --- PASS: TestParseConfigEntry/expose_paths:_kitchen_sink_proxy_defaults_(json_camel_case) (0.00s) === CONT TestConfigWrite writer.go:29: 2021-01-29T19:33:05.357Z [INFO] TestConfigWrite: Requesting shutdown writer.go:29: 2021-01-29T19:33:05.359Z [INFO] TestConfigWrite.server: shutting down server writer.go:29: 2021-01-29T19:33:05.360Z [DEBUG] TestConfigWrite.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.360Z [WARN] TestConfigWrite.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:05.362Z [ERROR] TestConfigWrite.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:05.362Z [DEBUG] TestConfigWrite.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:05.362Z [WARN] TestConfigWrite.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:05.374Z [INFO] TestConfigWrite.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:05.375Z [INFO] TestConfigWrite: consul server down writer.go:29: 2021-01-29T19:33:05.376Z [INFO] TestConfigWrite: shutdown complete writer.go:29: 2021-01-29T19:33:05.377Z [INFO] TestConfigWrite: Stopping server: protocol=DNS address=127.0.0.1:22253 network=tcp writer.go:29: 2021-01-29T19:33:05.377Z [INFO] TestConfigWrite: Stopping server: protocol=DNS address=127.0.0.1:22253 network=udp writer.go:29: 2021-01-29T19:33:05.378Z [INFO] TestConfigWrite: Stopping server: protocol=HTTP address=127.0.0.1:22254 network=tcp writer.go:29: 2021-01-29T19:33:05.879Z [INFO] TestConfigWrite: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:05.882Z [INFO] TestConfigWrite: Endpoints down --- PASS: TestConfigWrite (0.86s) --- PASS: TestConfigWrite/File (0.12s) --- PASS: TestConfigWrite/Stdin (0.02s) --- PASS: TestConfigWrite/No_config (0.00s) PASS ok github.com/hashicorp/consul/command/config/write 1.035s === RUN TestConnectCommand_noTabs === PAUSE TestConnectCommand_noTabs === CONT TestConnectCommand_noTabs --- PASS: TestConnectCommand_noTabs (0.00s) PASS ok github.com/hashicorp/consul/command/connect 0.015s === RUN TestCatalogCommand_noTabs === PAUSE TestCatalogCommand_noTabs === CONT TestCatalogCommand_noTabs --- PASS: TestCatalogCommand_noTabs (0.00s) PASS ok github.com/hashicorp/consul/command/connect/ca 0.011s === RUN TestConnectCAGetConfigCommand_noTabs === PAUSE TestConnectCAGetConfigCommand_noTabs === RUN TestConnectCAGetConfigCommand === PAUSE TestConnectCAGetConfigCommand === CONT TestConnectCAGetConfigCommand_noTabs === CONT TestConnectCAGetConfigCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestConnectCAGetConfigCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestConnectCAGetConfigCommand writer.go:29: 2021-01-29T19:33:07.430Z [WARN] TestConnectCAGetConfigCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:07.432Z [DEBUG] TestConnectCAGetConfigCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:07.436Z [DEBUG] TestConnectCAGetConfigCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:07.445Z [INFO] TestConnectCAGetConfigCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5bd01576-3dae-a322-c39b-9ed65d1e6cb9 Address:127.0.0.1:22258}]" writer.go:29: 2021-01-29T19:33:07.448Z [INFO] TestConnectCAGetConfigCommand.server.serf.wan: serf: EventMemberJoin: Node-5bd01576-3dae-a322-c39b-9ed65d1e6cb9.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:07.450Z [INFO] TestConnectCAGetConfigCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:22258 [Follower]" leader= writer.go:29: 2021-01-29T19:33:07.454Z [INFO] TestConnectCAGetConfigCommand.server.serf.lan: serf: EventMemberJoin: Node-5bd01576-3dae-a322-c39b-9ed65d1e6cb9 127.0.0.1 writer.go:29: 2021-01-29T19:33:07.458Z [INFO] TestConnectCAGetConfigCommand: Started DNS server: address=127.0.0.1:22253 network=udp writer.go:29: 2021-01-29T19:33:07.459Z [INFO] TestConnectCAGetConfigCommand.server: Handled event for server in area: event=member-join server=Node-5bd01576-3dae-a322-c39b-9ed65d1e6cb9.dc1 area=wan writer.go:29: 2021-01-29T19:33:07.459Z [INFO] TestConnectCAGetConfigCommand.server: Adding LAN server: server="Node-5bd01576-3dae-a322-c39b-9ed65d1e6cb9 (Addr: tcp/127.0.0.1:22258) (DC: dc1)" writer.go:29: 2021-01-29T19:33:07.460Z [INFO] TestConnectCAGetConfigCommand: Started DNS server: address=127.0.0.1:22253 network=tcp writer.go:29: 2021-01-29T19:33:07.462Z [INFO] TestConnectCAGetConfigCommand: Started HTTP server: address=127.0.0.1:22254 network=tcp writer.go:29: 2021-01-29T19:33:07.463Z [INFO] TestConnectCAGetConfigCommand: started state syncer writer.go:29: 2021-01-29T19:33:07.525Z [WARN] TestConnectCAGetConfigCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:07.526Z [INFO] TestConnectCAGetConfigCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:22258 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:07.531Z [DEBUG] TestConnectCAGetConfigCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:07.533Z [DEBUG] TestConnectCAGetConfigCommand.server.raft: vote granted: from=5bd01576-3dae-a322-c39b-9ed65d1e6cb9 term=2 tally=1 writer.go:29: 2021-01-29T19:33:07.535Z [INFO] TestConnectCAGetConfigCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:07.537Z [INFO] TestConnectCAGetConfigCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:22258 [Leader]" writer.go:29: 2021-01-29T19:33:07.543Z [INFO] TestConnectCAGetConfigCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:07.543Z [INFO] TestConnectCAGetConfigCommand.server: New leader elected: payload=Node-5bd01576-3dae-a322-c39b-9ed65d1e6cb9 writer.go:29: 2021-01-29T19:33:07.549Z [DEBUG] TestConnectCAGetConfigCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22258 writer.go:29: 2021-01-29T19:33:07.549Z [INFO] TestConnectCAGetConfigCommand: Synced node info writer.go:29: 2021-01-29T19:33:07.554Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:07.626Z [INFO] TestConnectCAGetConfigCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:07.627Z [INFO] TestConnectCAGetConfigCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.627Z [DEBUG] TestConnectCAGetConfigCommand.server: Skipping self join check for node since the cluster is too small: node=Node-5bd01576-3dae-a322-c39b-9ed65d1e6cb9 writer.go:29: 2021-01-29T19:33:07.628Z [INFO] TestConnectCAGetConfigCommand.server: member joined, marking health alive: member=Node-5bd01576-3dae-a322-c39b-9ed65d1e6cb9 writer.go:29: 2021-01-29T19:33:07.642Z [DEBUG] TestConnectCAGetConfigCommand.http: Request finished: method=GET url=/v1/connect/ca/configuration from=127.0.0.1:48268 latency=3.135252ms writer.go:29: 2021-01-29T19:33:07.646Z [INFO] TestConnectCAGetConfigCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:07.647Z [INFO] TestConnectCAGetConfigCommand.server: shutting down server writer.go:29: 2021-01-29T19:33:07.654Z [DEBUG] TestConnectCAGetConfigCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.656Z [WARN] TestConnectCAGetConfigCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.661Z [DEBUG] TestConnectCAGetConfigCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:07.673Z [WARN] TestConnectCAGetConfigCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:07.676Z [INFO] TestConnectCAGetConfigCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:07.677Z [INFO] TestConnectCAGetConfigCommand: consul server down writer.go:29: 2021-01-29T19:33:07.679Z [INFO] TestConnectCAGetConfigCommand: shutdown complete writer.go:29: 2021-01-29T19:33:07.680Z [INFO] TestConnectCAGetConfigCommand: Stopping server: protocol=DNS address=127.0.0.1:22253 network=tcp writer.go:29: 2021-01-29T19:33:07.681Z [INFO] TestConnectCAGetConfigCommand: Stopping server: protocol=DNS address=127.0.0.1:22253 network=udp writer.go:29: 2021-01-29T19:33:07.682Z [INFO] TestConnectCAGetConfigCommand: Stopping server: protocol=HTTP address=127.0.0.1:22254 network=tcp writer.go:29: 2021-01-29T19:33:08.183Z [INFO] TestConnectCAGetConfigCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:08.185Z [INFO] TestConnectCAGetConfigCommand: Endpoints down --- PASS: TestConnectCAGetConfigCommand (0.78s) PASS ok github.com/hashicorp/consul/command/connect/ca/get 0.937s === RUN TestConnectCASetConfigCommand_noTabs === PAUSE TestConnectCASetConfigCommand_noTabs === RUN TestConnectCASetConfigCommand === PAUSE TestConnectCASetConfigCommand === CONT TestConnectCASetConfigCommand_noTabs === CONT TestConnectCASetConfigCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestConnectCASetConfigCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestConnectCASetConfigCommand writer.go:29: 2021-01-29T19:33:12.324Z [WARN] TestConnectCASetConfigCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:12.326Z [DEBUG] TestConnectCASetConfigCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:12.330Z [DEBUG] TestConnectCASetConfigCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:12.336Z [INFO] TestConnectCASetConfigCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2a622dcb-1419-ee4c-b518-4846b2dd6fef Address:127.0.0.1:13069}]" writer.go:29: 2021-01-29T19:33:12.340Z [INFO] TestConnectCASetConfigCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:13069 [Follower]" leader= writer.go:29: 2021-01-29T19:33:12.341Z [INFO] TestConnectCASetConfigCommand.server.serf.wan: serf: EventMemberJoin: Node-2a622dcb-1419-ee4c-b518-4846b2dd6fef.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:12.344Z [INFO] TestConnectCASetConfigCommand.server.serf.lan: serf: EventMemberJoin: Node-2a622dcb-1419-ee4c-b518-4846b2dd6fef 127.0.0.1 writer.go:29: 2021-01-29T19:33:12.348Z [INFO] TestConnectCASetConfigCommand.server: Adding LAN server: server="Node-2a622dcb-1419-ee4c-b518-4846b2dd6fef (Addr: tcp/127.0.0.1:13069) (DC: dc1)" writer.go:29: 2021-01-29T19:33:12.349Z [INFO] TestConnectCASetConfigCommand.server: Handled event for server in area: event=member-join server=Node-2a622dcb-1419-ee4c-b518-4846b2dd6fef.dc1 area=wan writer.go:29: 2021-01-29T19:33:12.351Z [INFO] TestConnectCASetConfigCommand: Started DNS server: address=127.0.0.1:13064 network=udp writer.go:29: 2021-01-29T19:33:12.354Z [INFO] TestConnectCASetConfigCommand: Started DNS server: address=127.0.0.1:13064 network=tcp writer.go:29: 2021-01-29T19:33:12.357Z [INFO] TestConnectCASetConfigCommand: Started HTTP server: address=127.0.0.1:13065 network=tcp writer.go:29: 2021-01-29T19:33:12.358Z [INFO] TestConnectCASetConfigCommand: started state syncer writer.go:29: 2021-01-29T19:33:12.407Z [WARN] TestConnectCASetConfigCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:12.409Z [INFO] TestConnectCASetConfigCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:13069 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:12.411Z [DEBUG] TestConnectCASetConfigCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:12.412Z [DEBUG] TestConnectCASetConfigCommand.server.raft: vote granted: from=2a622dcb-1419-ee4c-b518-4846b2dd6fef term=2 tally=1 writer.go:29: 2021-01-29T19:33:12.413Z [INFO] TestConnectCASetConfigCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:12.414Z [INFO] TestConnectCASetConfigCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:13069 [Leader]" writer.go:29: 2021-01-29T19:33:12.415Z [INFO] TestConnectCASetConfigCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:12.416Z [INFO] TestConnectCASetConfigCommand.server: New leader elected: payload=Node-2a622dcb-1419-ee4c-b518-4846b2dd6fef writer.go:29: 2021-01-29T19:33:12.419Z [DEBUG] TestConnectCASetConfigCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:13069 writer.go:29: 2021-01-29T19:33:12.427Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:12.481Z [INFO] TestConnectCASetConfigCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:12.483Z [INFO] TestConnectCASetConfigCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.485Z [DEBUG] TestConnectCASetConfigCommand.server: Skipping self join check for node since the cluster is too small: node=Node-2a622dcb-1419-ee4c-b518-4846b2dd6fef writer.go:29: 2021-01-29T19:33:12.487Z [INFO] TestConnectCASetConfigCommand.server: member joined, marking health alive: member=Node-2a622dcb-1419-ee4c-b518-4846b2dd6fef writer.go:29: 2021-01-29T19:33:12.540Z [INFO] TestConnectCASetConfigCommand.server.connect: CA provider config updated writer.go:29: 2021-01-29T19:33:12.543Z [DEBUG] TestConnectCASetConfigCommand.http: Request finished: method=PUT url=/v1/connect/ca/configuration from=127.0.0.1:59814 latency=24.800813ms writer.go:29: 2021-01-29T19:33:12.553Z [INFO] TestConnectCASetConfigCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:12.554Z [INFO] TestConnectCASetConfigCommand.server: shutting down server writer.go:29: 2021-01-29T19:33:12.556Z [DEBUG] TestConnectCASetConfigCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.557Z [WARN] TestConnectCASetConfigCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:12.556Z [ERROR] TestConnectCASetConfigCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:12.557Z [DEBUG] TestConnectCASetConfigCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:12.561Z [WARN] TestConnectCASetConfigCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:12.563Z [INFO] TestConnectCASetConfigCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:12.563Z [INFO] TestConnectCASetConfigCommand: consul server down writer.go:29: 2021-01-29T19:33:12.567Z [INFO] TestConnectCASetConfigCommand: shutdown complete writer.go:29: 2021-01-29T19:33:12.569Z [INFO] TestConnectCASetConfigCommand: Stopping server: protocol=DNS address=127.0.0.1:13064 network=tcp writer.go:29: 2021-01-29T19:33:12.571Z [INFO] TestConnectCASetConfigCommand: Stopping server: protocol=DNS address=127.0.0.1:13064 network=udp writer.go:29: 2021-01-29T19:33:12.573Z [INFO] TestConnectCASetConfigCommand: Stopping server: protocol=HTTP address=127.0.0.1:13065 network=tcp writer.go:29: 2021-01-29T19:33:13.078Z [INFO] TestConnectCASetConfigCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:13.080Z [INFO] TestConnectCASetConfigCommand: Endpoints down --- PASS: TestConnectCASetConfigCommand (0.90s) PASS ok github.com/hashicorp/consul/command/connect/ca/set 1.150s === RUN TestBootstrapConfig_ConfigureArgs === RUN TestBootstrapConfig_ConfigureArgs/defaults === RUN TestBootstrapConfig_ConfigureArgs/extra-stats-sinks === RUN TestBootstrapConfig_ConfigureArgs/simple-statsd-sink === RUN TestBootstrapConfig_ConfigureArgs/simple-statsd-sink-plus-extra === RUN TestBootstrapConfig_ConfigureArgs/simple-statsd-sink-env === RUN TestBootstrapConfig_ConfigureArgs/simple-dogstatsd-sink === RUN TestBootstrapConfig_ConfigureArgs/simple-dogstatsd-unix-sink === RUN TestBootstrapConfig_ConfigureArgs/simple-dogstatsd-sink-env === RUN TestBootstrapConfig_ConfigureArgs/stats-config-override === RUN TestBootstrapConfig_ConfigureArgs/simple-tags === RUN TestBootstrapConfig_ConfigureArgs/prometheus-bind-addr === RUN TestBootstrapConfig_ConfigureArgs/prometheus-bind-addr-with-overrides === RUN TestBootstrapConfig_ConfigureArgs/stats-bind-addr === RUN TestBootstrapConfig_ConfigureArgs/stats-bind-addr-with-overrides === RUN TestBootstrapConfig_ConfigureArgs/stats-flush-interval === RUN TestBootstrapConfig_ConfigureArgs/override-tracing === RUN TestBootstrapConfig_ConfigureArgs/err-bad-prometheus-addr === RUN TestBootstrapConfig_ConfigureArgs/err-bad-stats-addr === RUN TestBootstrapConfig_ConfigureArgs/err-bad-statsd-addr === RUN TestBootstrapConfig_ConfigureArgs/err-bad-dogstatsd-addr --- PASS: TestBootstrapConfig_ConfigureArgs (0.53s) --- PASS: TestBootstrapConfig_ConfigureArgs/defaults (0.04s) --- PASS: TestBootstrapConfig_ConfigureArgs/extra-stats-sinks (0.07s) --- PASS: TestBootstrapConfig_ConfigureArgs/simple-statsd-sink (0.02s) --- PASS: TestBootstrapConfig_ConfigureArgs/simple-statsd-sink-plus-extra (0.04s) --- PASS: TestBootstrapConfig_ConfigureArgs/simple-statsd-sink-env (0.02s) --- PASS: TestBootstrapConfig_ConfigureArgs/simple-dogstatsd-sink (0.03s) --- PASS: TestBootstrapConfig_ConfigureArgs/simple-dogstatsd-unix-sink (0.04s) --- PASS: TestBootstrapConfig_ConfigureArgs/simple-dogstatsd-sink-env (0.02s) --- PASS: TestBootstrapConfig_ConfigureArgs/stats-config-override (0.03s) --- PASS: TestBootstrapConfig_ConfigureArgs/simple-tags (0.02s) --- PASS: TestBootstrapConfig_ConfigureArgs/prometheus-bind-addr (0.03s) --- PASS: TestBootstrapConfig_ConfigureArgs/prometheus-bind-addr-with-overrides (0.03s) --- PASS: TestBootstrapConfig_ConfigureArgs/stats-bind-addr (0.03s) --- PASS: TestBootstrapConfig_ConfigureArgs/stats-bind-addr-with-overrides (0.03s) --- PASS: TestBootstrapConfig_ConfigureArgs/stats-flush-interval (0.03s) --- PASS: TestBootstrapConfig_ConfigureArgs/override-tracing (0.05s) --- PASS: TestBootstrapConfig_ConfigureArgs/err-bad-prometheus-addr (0.00s) --- PASS: TestBootstrapConfig_ConfigureArgs/err-bad-stats-addr (0.00s) --- PASS: TestBootstrapConfig_ConfigureArgs/err-bad-statsd-addr (0.00s) --- PASS: TestBootstrapConfig_ConfigureArgs/err-bad-dogstatsd-addr (0.00s) === RUN TestEnvoyCommand_noTabs === PAUSE TestEnvoyCommand_noTabs === RUN TestGenerateConfig === RUN TestGenerateConfig/no-args === RUN TestGenerateConfig/defaults { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/token-arg { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "c9a52720-bf6c-4aa6-b8bc-66881a5ade95" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/token-env { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "c9a52720-bf6c-4aa6-b8bc-66881a5ade95" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/token-file-arg { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "c9a52720-bf6c-4aa6-b8bc-66881a5ade95" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/token-file-env { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "c9a52720-bf6c-4aa6-b8bc-66881a5ade95" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/grpc-addr-flag { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 9999 } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/grpc-addr-env { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 9999 } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/grpc-addr-unix { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "pipe": { "path": "/var/run/consul.sock" } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/grpc-addr-config { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 9999 } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/access-log-path { "admin": { "access_log_path": "/some/path/access.log", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/missing-ca-file === RUN TestGenerateConfig/existing-ca-file { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "tls_context": { "common_tls_context": { "validation_context": { "trusted_ca": { "inline_string": "-----BEGIN CERTIFICATE-----\nMIIEtzCCA5+gAwIBAgIJAIewRMI8OnvTMA0GCSqGSIb3DQEBBQUAMIGYMQswCQYD\nVQQGEwJVUzELMAkGA1UECBMCQ0ExFjAUBgNVBAcTDVNhbiBGcmFuY2lzY28xHDAa\nBgNVBAoTE0hhc2hpQ29ycCBUZXN0IENlcnQxDDAKBgNVBAsTA0RldjEWMBQGA1UE\nAxMNdGVzdC5pbnRlcm5hbDEgMB4GCSqGSIb3DQEJARYRdGVzdEBpbnRlcm5hbC5j\nb20wHhcNMTQwNDA3MTkwMTA4WhcNMjQwNDA0MTkwMTA4WjCBmDELMAkGA1UEBhMC\nVVMxCzAJBgNVBAgTAkNBMRYwFAYDVQQHEw1TYW4gRnJhbmNpc2NvMRwwGgYDVQQK\nExNIYXNoaUNvcnAgVGVzdCBDZXJ0MQwwCgYDVQQLEwNEZXYxFjAUBgNVBAMTDXRl\nc3QuaW50ZXJuYWwxIDAeBgkqhkiG9w0BCQEWEXRlc3RAaW50ZXJuYWwuY29tMIIB\nIjANBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAxrs6JK4NpiOItxrpNR/1ppUU\nmH7p2BgLCBZ6eHdclle9J56i68adt8J85zaqphCfz6VDP58DsFx+N50PZyjQaDsU\nd0HejRqfHRMtg2O+UQkv4Z66+Vo+gc6uGuANi2xMtSYDVTAqqzF48OOPQDgYkzcG\nxcFZzTRFFZt2vPnyHj8cHcaFo/NMNVh7C3yTXevRGNm9u2mrbxCEeiHzFC2WUnvg\nU2jQuC7Fhnl33Zd3B6d3mQH6O23ncmwxTcPUJe6xZaIRrDuzwUcyhLj5Z3faag/f\npFIIcHSiHRfoqHLGsGg+3swId/zVJSSDHr7pJUu7Cre+vZa63FqDaooqvnisrQID\nAQABo4IBADCB/TAdBgNVHQ4EFgQUo/nrOfqvbee2VklVKIFlyQEbuJUwgc0GA1Ud\nIwSBxTCBwoAUo/nrOfqvbee2VklVKIFlyQEbuJWhgZ6kgZswgZgxCzAJBgNVBAYT\nAlVTMQswCQYDVQQIEwJDQTEWMBQGA1UEBxMNU2FuIEZyYW5jaXNjbzEcMBoGA1UE\nChMTSGFzaGlDb3JwIFRlc3QgQ2VydDEMMAoGA1UECxMDRGV2MRYwFAYDVQQDEw10\nZXN0LmludGVybmFsMSAwHgYJKoZIhvcNAQkBFhF0ZXN0QGludGVybmFsLmNvbYIJ\nAIewRMI8OnvTMAwGA1UdEwQFMAMBAf8wDQYJKoZIhvcNAQEFBQADggEBADa9fV9h\ngjapBlkNmu64WX0Ufub5dsJrdHS8672P30S7ILB7Mk0W8sL65IezRsZnG898yHf9\n2uzmz5OvNTM9K380g7xFlyobSVq+6yqmmSAlA/ptAcIIZT727P5jig/DB7fzJM3g\njctDlEGOmEe50GQXc25VKpcpjAsNQi5ER5gowQ0v3IXNZs+yU+LvxLHc0rUJ/XSp\nlFCAMOqd5uRoMOejnT51G6krvLNzPaQ3N9jQfNVY4Q0zfs0M+6dRWvqfqB9Vyq8/\nPOLMld+HyAZEBk9zK3ZVIXx6XS4dkDnSNR91njLq7eouf6M7+7s/oMQZZRtAfQ6r\nwlW975rYa1ZqEdA=\n-----END CERTIFICATE-----\n" } } } }, "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/custom-bootstrap { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy" }, "custom_field": "foo" }=== RUN TestGenerateConfig/extra_-single { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] }, { "name": "fake_cluster_1" } ], "listeners": [ { "name": "fake_listener_1" } ] }, "stats_sinks": [ { "name": "fake_sink_1" } ], "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/extra_-multiple { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] }, { "name": "fake_cluster_1" }, { "name": "fake_cluster_2" } ], "listeners": [ { "name": "fake_listener_1" }, { "name": "fake_listener_2" } ] }, "stats_sinks": [ { "name": "fake_sink_1" }, { "name": "fake_sink_2" } ], "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/stats-config-override { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] } ] }, "stats_config": { "name": "fake_config" }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } === RUN TestGenerateConfig/zipkin-tracing-config { "admin": { "access_log_path": "/dev/null", "address": { "socket_address": { "address": "127.0.0.1", "port_value": 19000 } } }, "node": { "cluster": "test-proxy", "id": "test-proxy", "metadata": { "namespace": "default", "envoy_version": "1.13.0" } }, "static_resources": { "clusters": [ { "name": "local_agent", "connect_timeout": "1s", "type": "STATIC", "http2_protocol_options": {}, "hosts": [ { "socket_address": { "address": "127.0.0.1", "port_value": 8502 } } ] }, { "name": "zipkin", "type": "STRICT_DNS", "connect_timeout": "5s", "load_assignment": { "cluster_name": "zipkin", "endpoints": [ { "lb_endpoints": [ { "endpoint": { "address": { "socket_address": { "address": "zipkin.service.consul", "port_value": 9411 } } } } ] } ] } } ] }, "stats_config": { "stats_tags": [ { "regex": "^cluster\\.((?:([^.]+)~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.custom_hash" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:([^.]+)\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service_subset" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.service" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.namespace" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.datacenter" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.[^.]+\\.consul\\.)", "tag_name": "consul.routing_type" }, { "regex": "^cluster\\.((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.([^.]+)\\.consul\\.)", "tag_name": "consul.trust_domain" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+)\\.[^.]+\\.[^.]+\\.consul\\.)", "tag_name": "consul.target" }, { "regex": "^cluster\\.(((?:[^.]+~)?(?:[^.]+\\.)?[^.]+\\.[^.]+\\.[^.]+\\.[^.]+\\.[^.]+)\\.consul\\.)", "tag_name": "consul.full_target" }, { "tag_name": "local_cluster", "fixed_value": "test-proxy" } ], "use_all_default_tags": true }, "tracing": { "http": { "name": "envoy.zipkin", "config": { "collector_cluster": "zipkin", "collector_endpoint": "/api/v1/spans" } } }, "dynamic_resources": { "lds_config": { "ads": {} }, "cds_config": { "ads": {} }, "ads_config": { "api_type": "GRPC", "grpc_services": { "initial_metadata": [ { "key": "x-consul-token", "value": "" } ], "envoy_grpc": { "cluster_name": "local_agent" } } } }, "layered_runtime": { "layers": [ { "name": "static_layer", "static_layer": { "envoy.deprecated_features:envoy.api.v2.Cluster.tls_context": true, "envoy.deprecated_features:envoy.config.trace.v2.ZipkinConfig.HTTP_JSON_V1": true, "envoy.deprecated_features:envoy.config.filter.network.http_connection_manager.v2.HttpConnectionManager.Tracing.operation_name": true } } ] } } --- PASS: TestGenerateConfig (0.45s) --- PASS: TestGenerateConfig/no-args (0.00s) --- PASS: TestGenerateConfig/defaults (0.05s) --- PASS: TestGenerateConfig/token-arg (0.03s) --- PASS: TestGenerateConfig/token-env (0.02s) --- PASS: TestGenerateConfig/token-file-arg (0.02s) --- PASS: TestGenerateConfig/token-file-env (0.03s) --- PASS: TestGenerateConfig/grpc-addr-flag (0.02s) --- PASS: TestGenerateConfig/grpc-addr-env (0.02s) --- PASS: TestGenerateConfig/grpc-addr-unix (0.02s) --- PASS: TestGenerateConfig/grpc-addr-config (0.03s) --- PASS: TestGenerateConfig/access-log-path (0.04s) --- PASS: TestGenerateConfig/missing-ca-file (0.01s) --- PASS: TestGenerateConfig/existing-ca-file (0.05s) --- PASS: TestGenerateConfig/custom-bootstrap (0.03s) --- PASS: TestGenerateConfig/extra_-single (0.02s) --- PASS: TestGenerateConfig/extra_-multiple (0.02s) --- PASS: TestGenerateConfig/stats-config-override (0.02s) --- PASS: TestGenerateConfig/zipkin-tracing-config (0.02s) === RUN TestEnvoyCommand_canBindInternal === PAUSE TestEnvoyCommand_canBindInternal === RUN TestExecEnvoy === RUN TestExecEnvoy/default === RUN TestExecEnvoy/hot-restart-epoch === RUN TestExecEnvoy/hot-restart-version === RUN TestExecEnvoy/hot-restart-version#01 === RUN TestExecEnvoy/hot-restart-version#02 --- PASS: TestExecEnvoy (2.62s) --- PASS: TestExecEnvoy/default (0.50s) --- PASS: TestExecEnvoy/hot-restart-epoch (0.60s) --- PASS: TestExecEnvoy/hot-restart-version (0.41s) --- PASS: TestExecEnvoy/hot-restart-version#01 (0.40s) --- PASS: TestExecEnvoy/hot-restart-version#02 (0.71s) === RUN TestHelperProcess --- PASS: TestHelperProcess (0.00s) === CONT TestEnvoyCommand_noTabs === CONT TestEnvoyCommand_canBindInternal --- PASS: TestEnvoyCommand_noTabs (0.00s) === RUN TestEnvoyCommand_canBindInternal/IPNet === RUN TestEnvoyCommand_canBindInternal/IPNet/ipv4 === RUN TestEnvoyCommand_canBindInternal/IPNet/secondary_ipv4 === RUN TestEnvoyCommand_canBindInternal/IPNet/ipv6 === RUN TestEnvoyCommand_canBindInternal/IPNet/ipv4_not_found === RUN TestEnvoyCommand_canBindInternal/IPNet/ipv6_not_found === RUN TestEnvoyCommand_canBindInternal/IPAddr === RUN TestEnvoyCommand_canBindInternal/IPAddr/ipv4_not_found === RUN TestEnvoyCommand_canBindInternal/IPAddr/ipv6_not_found === RUN TestEnvoyCommand_canBindInternal/IPAddr/ipv4 === RUN TestEnvoyCommand_canBindInternal/IPAddr/secondary_ipv4 === RUN TestEnvoyCommand_canBindInternal/IPAddr/ipv6 --- PASS: TestEnvoyCommand_canBindInternal (0.02s) --- PASS: TestEnvoyCommand_canBindInternal/IPNet (0.01s) --- PASS: TestEnvoyCommand_canBindInternal/IPNet/ipv4 (0.00s) --- PASS: TestEnvoyCommand_canBindInternal/IPNet/secondary_ipv4 (0.00s) --- PASS: TestEnvoyCommand_canBindInternal/IPNet/ipv6 (0.00s) --- PASS: TestEnvoyCommand_canBindInternal/IPNet/ipv4_not_found (0.00s) --- PASS: TestEnvoyCommand_canBindInternal/IPNet/ipv6_not_found (0.00s) --- PASS: TestEnvoyCommand_canBindInternal/IPAddr (0.01s) --- PASS: TestEnvoyCommand_canBindInternal/IPAddr/ipv4_not_found (0.00s) --- PASS: TestEnvoyCommand_canBindInternal/IPAddr/ipv6_not_found (0.00s) --- PASS: TestEnvoyCommand_canBindInternal/IPAddr/ipv4 (0.00s) --- PASS: TestEnvoyCommand_canBindInternal/IPAddr/secondary_ipv4 (0.00s) --- PASS: TestEnvoyCommand_canBindInternal/IPAddr/ipv6 (0.00s) PASS ok github.com/hashicorp/consul/command/connect/envoy 3.959s === RUN TestConnectEnvoyPipeBootstrapCommand_noTabs === PAUSE TestConnectEnvoyPipeBootstrapCommand_noTabs === CONT TestConnectEnvoyPipeBootstrapCommand_noTabs --- PASS: TestConnectEnvoyPipeBootstrapCommand_noTabs (0.00s) PASS ok github.com/hashicorp/consul/command/connect/envoy/pipe-bootstrap 0.008s === RUN TestFlagUpstreams_impl --- PASS: TestFlagUpstreams_impl (0.00s) === RUN TestFlagUpstreams === RUN TestFlagUpstreams/bad_format === RUN TestFlagUpstreams/port_not_int === RUN TestFlagUpstreams/4_parts === RUN TestFlagUpstreams/single_value === RUN TestFlagUpstreams/single_value_prepared_query === RUN TestFlagUpstreams/invalid_type === RUN TestFlagUpstreams/address_specified === RUN TestFlagUpstreams/repeat_value,_overwrite --- PASS: TestFlagUpstreams (0.02s) --- PASS: TestFlagUpstreams/bad_format (0.00s) --- PASS: TestFlagUpstreams/port_not_int (0.00s) --- PASS: TestFlagUpstreams/4_parts (0.00s) --- PASS: TestFlagUpstreams/single_value (0.00s) --- PASS: TestFlagUpstreams/single_value_prepared_query (0.00s) --- PASS: TestFlagUpstreams/invalid_type (0.01s) --- PASS: TestFlagUpstreams/address_specified (0.00s) --- PASS: TestFlagUpstreams/repeat_value,_overwrite (0.00s) === RUN TestCommandConfigWatcher === PAUSE TestCommandConfigWatcher === RUN TestCatalogCommand_noTabs === PAUSE TestCatalogCommand_noTabs === RUN TestRegisterMonitor_good === PAUSE TestRegisterMonitor_good === RUN TestRegisterMonitor_heartbeat === PAUSE TestRegisterMonitor_heartbeat === CONT TestCommandConfigWatcher === CONT TestRegisterMonitor_heartbeat === CONT TestCatalogCommand_noTabs === CONT TestRegisterMonitor_good === RUN TestCommandConfigWatcher/-service_flag_only [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestCatalogCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:16.891Z [WARN] TestCommandConfigWatcher/-service_flag_only: bootstrap = true: do not enable unless necessary === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:16.899Z [WARN] TestRegisterMonitor_good: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:16.909Z [DEBUG] TestRegisterMonitor_good.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:16.911Z [DEBUG] TestRegisterMonitor_good.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:16.911Z [WARN] TestRegisterMonitor_heartbeat: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:16.913Z [DEBUG] TestRegisterMonitor_heartbeat.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:16.914Z [DEBUG] TestRegisterMonitor_heartbeat.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:16.915Z [INFO] TestRegisterMonitor_good.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:149cf329-ed02-d454-0ef6-7a8aba32a68f Address:127.0.0.1:25327}]" === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:16.917Z [DEBUG] TestCommandConfigWatcher/-service_flag_only.tlsutil: Update: version=1 === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:16.917Z [INFO] TestRegisterMonitor_good.server.raft: entering follower state: follower="Node at 127.0.0.1:25327 [Follower]" leader= === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:16.918Z [DEBUG] TestCommandConfigWatcher/-service_flag_only.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:16.920Z [INFO] TestRegisterMonitor_good.server.serf.wan: serf: EventMemberJoin: Node-149cf329-ed02-d454-0ef6-7a8aba32a68f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:16.961Z [INFO] TestRegisterMonitor_good.server.serf.lan: serf: EventMemberJoin: Node-149cf329-ed02-d454-0ef6-7a8aba32a68f 127.0.0.1 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:16.961Z [INFO] TestRegisterMonitor_heartbeat.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1d57126e-afac-6c89-b6ce-bf94bd8d93e4 Address:127.0.0.1:25321}]" === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:16.965Z [WARN] TestRegisterMonitor_good.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:16.965Z [INFO] TestRegisterMonitor_heartbeat.server.serf.wan: serf: EventMemberJoin: Node-1d57126e-afac-6c89-b6ce-bf94bd8d93e4.dc1 127.0.0.1 === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:16.967Z [INFO] TestRegisterMonitor_good.server.raft: entering candidate state: node="Node at 127.0.0.1:25327 [Candidate]" term=2 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:16.970Z [INFO] TestRegisterMonitor_heartbeat.server.raft: entering follower state: follower="Node at 127.0.0.1:25321 [Follower]" leader= === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:16.976Z [INFO] TestRegisterMonitor_good.server: Adding LAN server: server="Node-149cf329-ed02-d454-0ef6-7a8aba32a68f (Addr: tcp/127.0.0.1:25327) (DC: dc1)" writer.go:29: 2021-01-29T19:33:16.978Z [INFO] TestRegisterMonitor_good.server: Handled event for server in area: event=member-join server=Node-149cf329-ed02-d454-0ef6-7a8aba32a68f.dc1 area=wan writer.go:29: 2021-01-29T19:33:16.979Z [INFO] TestRegisterMonitor_good: Started DNS server: address=127.0.0.1:25322 network=udp === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:16.970Z [INFO] TestRegisterMonitor_heartbeat.server.serf.lan: serf: EventMemberJoin: Node-1d57126e-afac-6c89-b6ce-bf94bd8d93e4 127.0.0.1 === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:16.990Z [INFO] TestRegisterMonitor_good: Started DNS server: address=127.0.0.1:25322 network=tcp writer.go:29: 2021-01-29T19:33:16.984Z [WARN] TestRegisterMonitor_good.server.raft: unable to get address for sever, using fallback address: id=149cf329-ed02-d454-0ef6-7a8aba32a68f fallback=127.0.0.1:25327 error="Could not find address for server id 149cf329-ed02-d454-0ef6-7a8aba32a68f" writer.go:29: 2021-01-29T19:33:16.996Z [INFO] TestRegisterMonitor_good: Started HTTP server: address=127.0.0.1:25323 network=tcp writer.go:29: 2021-01-29T19:33:16.997Z [INFO] TestRegisterMonitor_good: started state syncer writer.go:29: 2021-01-29T19:33:17.003Z [DEBUG] TestRegisterMonitor_good.server.raft: votes: needed=1 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.001Z [INFO] TestRegisterMonitor_heartbeat.server: Handled event for server in area: event=member-join server=Node-1d57126e-afac-6c89-b6ce-bf94bd8d93e4.dc1 area=wan === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.005Z [INFO] TestCommandConfigWatcher/-service_flag_only.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:766fc9f0-39f2-02d4-9a8c-a0fee86cfe48 Address:127.0.0.1:25333}]" === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.005Z [DEBUG] TestRegisterMonitor_good.server.raft: vote granted: from=149cf329-ed02-d454-0ef6-7a8aba32a68f term=2 tally=1 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.005Z [INFO] TestRegisterMonitor_heartbeat.server: Adding LAN server: server="Node-1d57126e-afac-6c89-b6ce-bf94bd8d93e4 (Addr: tcp/127.0.0.1:25321) (DC: dc1)" === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.007Z [INFO] TestRegisterMonitor_good.server.raft: election won: tally=1 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.008Z [INFO] TestRegisterMonitor_heartbeat: Started DNS server: address=127.0.0.1:25316 network=udp === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.008Z [INFO] TestRegisterMonitor_good.server.raft: entering leader state: leader="Node at 127.0.0.1:25327 [Leader]" writer.go:29: 2021-01-29T19:33:17.010Z [INFO] TestRegisterMonitor_good.server: cluster leadership acquired === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.008Z [INFO] TestCommandConfigWatcher/-service_flag_only.server.serf.wan: serf: EventMemberJoin: Node-766fc9f0-39f2-02d4-9a8c-a0fee86cfe48.dc1 127.0.0.1 === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.022Z [INFO] TestRegisterMonitor_good.server: New leader elected: payload=Node-149cf329-ed02-d454-0ef6-7a8aba32a68f writer.go:29: 2021-01-29T19:33:17.022Z [DEBUG] TestRegisterMonitor_good.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25327 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.010Z [INFO] TestRegisterMonitor_heartbeat: Started DNS server: address=127.0.0.1:25316 network=tcp === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.026Z [INFO] TestRegisterMonitor_good: Synced node info === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.010Z [INFO] TestCommandConfigWatcher/-service_flag_only.server.raft: entering follower state: follower="Node at 127.0.0.1:25333 [Follower]" leader= === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.027Z [WARN] TestRegisterMonitor_heartbeat.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:17.030Z [INFO] TestRegisterMonitor_heartbeat.server.raft: entering candidate state: node="Node at 127.0.0.1:25321 [Candidate]" term=2 === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.029Z [DEBUG] TestRegisterMonitor_good: Node info in sync === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.019Z [INFO] TestCommandConfigWatcher/-service_flag_only.server.serf.lan: serf: EventMemberJoin: Node-766fc9f0-39f2-02d4-9a8c-a0fee86cfe48 127.0.0.1 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.032Z [INFO] TestRegisterMonitor_heartbeat: Started HTTP server: address=127.0.0.1:25317 network=tcp === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.033Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.033Z [DEBUG] TestRegisterMonitor_heartbeat.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:17.037Z [DEBUG] TestRegisterMonitor_heartbeat.server.raft: vote granted: from=1d57126e-afac-6c89-b6ce-bf94bd8d93e4 term=2 tally=1 === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.041Z [INFO] TestCommandConfigWatcher/-service_flag_only.server: Adding LAN server: server="Node-766fc9f0-39f2-02d4-9a8c-a0fee86cfe48 (Addr: tcp/127.0.0.1:25333) (DC: dc1)" === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.041Z [INFO] TestRegisterMonitor_heartbeat.server.raft: election won: tally=1 === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.045Z [INFO] TestCommandConfigWatcher/-service_flag_only.server: Handled event for server in area: event=member-join server=Node-766fc9f0-39f2-02d4-9a8c-a0fee86cfe48.dc1 area=wan === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.035Z [INFO] TestRegisterMonitor_heartbeat: started state syncer === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.052Z [INFO] TestCommandConfigWatcher/-service_flag_only: Started DNS server: address=127.0.0.1:25328 network=tcp === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.051Z [INFO] TestRegisterMonitor_heartbeat.server.raft: entering leader state: leader="Node at 127.0.0.1:25321 [Leader]" writer.go:29: 2021-01-29T19:33:17.071Z [INFO] TestRegisterMonitor_heartbeat.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:17.073Z [INFO] TestRegisterMonitor_heartbeat.server: New leader elected: payload=Node-1d57126e-afac-6c89-b6ce-bf94bd8d93e4 writer.go:29: 2021-01-29T19:33:17.077Z [DEBUG] TestRegisterMonitor_heartbeat.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25321 writer.go:29: 2021-01-29T19:33:17.081Z [INFO] TestRegisterMonitor_heartbeat: Synced node info writer.go:29: 2021-01-29T19:33:17.083Z [DEBUG] TestRegisterMonitor_heartbeat: Node info in sync === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.084Z [WARN] TestCommandConfigWatcher/-service_flag_only: Check socket connection failed: check=service:two-sidecars-sidecar-proxy:1 error="dial tcp 127.0.0.1:21000: connect: connection refused" writer.go:29: 2021-01-29T19:33:17.086Z [WARN] TestCommandConfigWatcher/-service_flag_only: Check is now critical: check=service:two-sidecars-sidecar-proxy:1 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.093Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.092Z [WARN] TestCommandConfigWatcher/-service_flag_only.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:17.097Z [INFO] TestCommandConfigWatcher/-service_flag_only.server.raft: entering candidate state: node="Node at 127.0.0.1:25333 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:17.096Z [INFO] TestCommandConfigWatcher/-service_flag_only: Started DNS server: address=127.0.0.1:25328 network=udp writer.go:29: 2021-01-29T19:33:17.101Z [INFO] TestCommandConfigWatcher/-service_flag_only: Started HTTP server: address=127.0.0.1:25329 network=tcp === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.104Z [INFO] TestRegisterMonitor_heartbeat.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:17.105Z [INFO] TestRegisterMonitor_heartbeat.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.107Z [DEBUG] TestRegisterMonitor_heartbeat.server: Skipping self join check for node since the cluster is too small: node=Node-1d57126e-afac-6c89-b6ce-bf94bd8d93e4 writer.go:29: 2021-01-29T19:33:17.108Z [INFO] TestRegisterMonitor_heartbeat.server: member joined, marking health alive: member=Node-1d57126e-afac-6c89-b6ce-bf94bd8d93e4 === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.102Z [INFO] TestCommandConfigWatcher/-service_flag_only: started state syncer === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.124Z [INFO] TestRegisterMonitor_good.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.103Z [DEBUG] TestCommandConfigWatcher/-service_flag_only.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:17.128Z [DEBUG] TestCommandConfigWatcher/-service_flag_only.server.raft: vote granted: from=766fc9f0-39f2-02d4-9a8c-a0fee86cfe48 term=2 tally=1 === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.128Z [INFO] TestRegisterMonitor_good.leader: started routine: routine="CA root pruning" === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.135Z [INFO] TestCommandConfigWatcher/-service_flag_only.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:17.142Z [INFO] TestCommandConfigWatcher/-service_flag_only.server.raft: entering leader state: leader="Node at 127.0.0.1:25333 [Leader]" writer.go:29: 2021-01-29T19:33:17.150Z [INFO] TestCommandConfigWatcher/-service_flag_only.server: cluster leadership acquired === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.136Z [DEBUG] TestRegisterMonitor_good.server: Skipping self join check for node since the cluster is too small: node=Node-149cf329-ed02-d454-0ef6-7a8aba32a68f writer.go:29: 2021-01-29T19:33:17.155Z [INFO] TestRegisterMonitor_good.server: member joined, marking health alive: member=Node-149cf329-ed02-d454-0ef6-7a8aba32a68f === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.157Z [INFO] TestCommandConfigWatcher/-service_flag_only.server: New leader elected: payload=Node-766fc9f0-39f2-02d4-9a8c-a0fee86cfe48 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.159Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:55092 latency=22.690053ms === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.163Z [DEBUG] TestRegisterMonitor_good.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:34316 latency=18.818459ms === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.167Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=GET url=/v1/catalog/service/foo-proxy?stale= from=127.0.0.1:55090 latency=32.440216ms === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.167Z [DEBUG] TestCommandConfigWatcher/-service_flag_only.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25333 === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.169Z [DEBUG] TestRegisterMonitor_good.http: Request finished: method=GET url=/v1/catalog/service/foo-proxy?stale= from=127.0.0.1:34318 latency=7.965652ms writer.go:29: 2021-01-29T19:33:17.215Z [DEBUG] TestRegisterMonitor_good.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:34316 latency=142.894µs writer.go:29: 2021-01-29T19:33:17.222Z [ERROR] TestRegisterMonitor_good.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.224Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:55092 latency=1.209941ms === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.227Z [DEBUG] TestRegisterMonitor_good: Node info in sync writer.go:29: 2021-01-29T19:33:17.231Z [INFO] TestRegisterMonitor_good: Synced service: service=foo-proxy writer.go:29: 2021-01-29T19:33:17.233Z [DEBUG] TestRegisterMonitor_good: Check in sync: check=foo-proxy-ttl writer.go:29: 2021-01-29T19:33:17.235Z [DEBUG] TestRegisterMonitor_good.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:34318 latency=62.725364ms === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.237Z [ERROR] TestRegisterMonitor_heartbeat.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.237Z [INFO] TestCommandConfigWatcher/-service_flag_only: Requesting shutdown === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.245Z [INFO] TestRegisterMonitor_good: registered Consul service: service=foo-proxy === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.248Z [DEBUG] TestRegisterMonitor_heartbeat: Node info in sync === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.243Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.247Z [INFO] TestRegisterMonitor_good: stop request received, deregistering === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.253Z [INFO] TestRegisterMonitor_heartbeat: Synced service: service=foo-proxy writer.go:29: 2021-01-29T19:33:17.254Z [DEBUG] TestRegisterMonitor_heartbeat: Check in sync: check=foo-proxy-ttl === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.245Z [INFO] TestCommandConfigWatcher/-service_flag_only.server: shutting down server writer.go:29: 2021-01-29T19:33:17.260Z [WARN] TestCommandConfigWatcher/-service_flag_only.server.serf.lan: serf: Shutdown without a Leave === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.255Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:55090 latency=81.959278ms === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.262Z [ERROR] TestCommandConfigWatcher/-service_flag_only.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.260Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=GET url=/v1/agent/checks from=127.0.0.1:55092 latency=7.828691ms writer.go:29: 2021-01-29T19:33:17.266Z [INFO] TestRegisterMonitor_heartbeat: registered Consul service: service=foo-proxy === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.267Z [DEBUG] TestRegisterMonitor_good: removed check: check=foo-proxy-ttl writer.go:29: 2021-01-29T19:33:17.268Z [DEBUG] TestRegisterMonitor_good: removed service: service=foo-proxy === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.267Z [WARN] TestCommandConfigWatcher/-service_flag_only.server.serf.wan: serf: Shutdown without a Leave === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.271Z [DEBUG] TestRegisterMonitor_heartbeat: Check status updated: check=foo-proxy-ttl status=critical writer.go:29: 2021-01-29T19:33:17.272Z [DEBUG] TestRegisterMonitor_heartbeat: Node info in sync writer.go:29: 2021-01-29T19:33:17.273Z [DEBUG] TestRegisterMonitor_heartbeat: Service in sync: service=foo-proxy === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.270Z [INFO] TestCommandConfigWatcher/-service_flag_only.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.275Z [INFO] TestRegisterMonitor_heartbeat: Synced check: check=foo-proxy-ttl writer.go:29: 2021-01-29T19:33:17.276Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=PUT url=/v1/agent/check/fail/foo-proxy-ttl?note= from=127.0.0.1:55092 latency=5.57643ms === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.275Z [INFO] TestCommandConfigWatcher/-service_flag_only.leader: started routine: routine="CA root pruning" === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.270Z [DEBUG] TestRegisterMonitor_good: Node info in sync writer.go:29: 2021-01-29T19:33:17.280Z [INFO] TestRegisterMonitor_good: Deregistered service: service=foo-proxy === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.278Z [DEBUG] TestCommandConfigWatcher/-service_flag_only.server: Skipping self join check for node since the cluster is too small: node=Node-766fc9f0-39f2-02d4-9a8c-a0fee86cfe48 === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.281Z [DEBUG] TestRegisterMonitor_good.http: Request finished: method=PUT url=/v1/agent/service/deregister/foo-proxy from=127.0.0.1:34318 latency=14.089182ms === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.282Z [INFO] TestCommandConfigWatcher/-service_flag_only.server: member joined, marking health alive: member=Node-766fc9f0-39f2-02d4-9a8c-a0fee86cfe48 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.284Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=GET url=/v1/agent/checks from=127.0.0.1:55092 latency=156.058µs === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.285Z [DEBUG] TestRegisterMonitor_good.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:34318 latency=1.60715ms === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.285Z [ERROR] TestCommandConfigWatcher/-service_flag_only.server: failed to reconcile member: member="{Node-766fc9f0-39f2-02d4-9a8c-a0fee86cfe48 127.0.0.1 25331 map[acls:0 bootstrap:1 build:1.7.4: dc:dc1 id:766fc9f0-39f2-02d4-9a8c-a0fee86cfe48 port:25333 raft_vsn:3 role:consul segment: vsn:2 vsn_max:3 vsn_min:2 wan_join_port:25332] alive 1 5 2 2 5 4}" error="raft is already shutdown" === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.288Z [INFO] TestRegisterMonitor_good: Requesting shutdown writer.go:29: 2021-01-29T19:33:17.288Z [INFO] TestRegisterMonitor_good.server: shutting down server writer.go:29: 2021-01-29T19:33:17.289Z [DEBUG] TestRegisterMonitor_good.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.290Z [WARN] TestRegisterMonitor_good.server.serf.lan: serf: Shutdown without a Leave === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.275Z [INFO] TestCommandConfigWatcher/-service_flag_only.server.router.manager: shutting down === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.291Z [DEBUG] TestRegisterMonitor_good.leader: stopped routine: routine="CA root pruning" === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.276Z [INFO] TestCommandConfigWatcher/-service_flag_only: consul server down writer.go:29: 2021-01-29T19:33:17.290Z [DEBUG] TestCommandConfigWatcher/-service_flag_only.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.299Z [DEBUG] TestCommandConfigWatcher/-service_flag_only.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.295Z [INFO] TestCommandConfigWatcher/-service_flag_only: shutdown complete writer.go:29: 2021-01-29T19:33:17.302Z [INFO] TestCommandConfigWatcher/-service_flag_only: Stopping server: protocol=DNS address=127.0.0.1:25328 network=tcp writer.go:29: 2021-01-29T19:33:17.303Z [INFO] TestCommandConfigWatcher/-service_flag_only: Stopping server: protocol=DNS address=127.0.0.1:25328 network=udp writer.go:29: 2021-01-29T19:33:17.304Z [INFO] TestCommandConfigWatcher/-service_flag_only: Stopping server: protocol=HTTP address=127.0.0.1:25329 network=tcp === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.313Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=GET url=/v1/agent/checks from=127.0.0.1:55092 latency=194.015µs === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.317Z [WARN] TestRegisterMonitor_good.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:17.320Z [INFO] TestRegisterMonitor_good.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:17.322Z [INFO] TestRegisterMonitor_good: consul server down writer.go:29: 2021-01-29T19:33:17.323Z [INFO] TestRegisterMonitor_good: shutdown complete writer.go:29: 2021-01-29T19:33:17.325Z [INFO] TestRegisterMonitor_good: Stopping server: protocol=DNS address=127.0.0.1:25322 network=tcp writer.go:29: 2021-01-29T19:33:17.326Z [INFO] TestRegisterMonitor_good: Stopping server: protocol=DNS address=127.0.0.1:25322 network=udp writer.go:29: 2021-01-29T19:33:17.327Z [INFO] TestRegisterMonitor_good: Stopping server: protocol=HTTP address=127.0.0.1:25323 network=tcp === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.334Z [DEBUG] TestRegisterMonitor_heartbeat: Check status updated: check=foo-proxy-ttl status=passing writer.go:29: 2021-01-29T19:33:17.336Z [DEBUG] TestRegisterMonitor_heartbeat: Node info in sync writer.go:29: 2021-01-29T19:33:17.338Z [DEBUG] TestRegisterMonitor_heartbeat: Service in sync: service=foo-proxy writer.go:29: 2021-01-29T19:33:17.341Z [INFO] TestRegisterMonitor_heartbeat: Synced check: check=foo-proxy-ttl writer.go:29: 2021-01-29T19:33:17.343Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=PUT url=/v1/agent/check/pass/foo-proxy-ttl?note= from=127.0.0.1:55092 latency=9.141941ms writer.go:29: 2021-01-29T19:33:17.343Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=GET url=/v1/agent/checks from=127.0.0.1:55090 latency=612.045µs writer.go:29: 2021-01-29T19:33:17.352Z [INFO] TestRegisterMonitor_heartbeat: stop request received, deregistering writer.go:29: 2021-01-29T19:33:17.354Z [DEBUG] TestRegisterMonitor_heartbeat: removed check: check=foo-proxy-ttl writer.go:29: 2021-01-29T19:33:17.356Z [DEBUG] TestRegisterMonitor_heartbeat: removed service: service=foo-proxy writer.go:29: 2021-01-29T19:33:17.358Z [DEBUG] TestRegisterMonitor_heartbeat: Node info in sync writer.go:29: 2021-01-29T19:33:17.362Z [INFO] TestRegisterMonitor_heartbeat: Deregistered service: service=foo-proxy writer.go:29: 2021-01-29T19:33:17.399Z [DEBUG] TestRegisterMonitor_heartbeat.http: Request finished: method=PUT url=/v1/agent/service/deregister/foo-proxy from=127.0.0.1:55090 latency=46.245498ms writer.go:29: 2021-01-29T19:33:17.407Z [INFO] TestRegisterMonitor_heartbeat: Requesting shutdown writer.go:29: 2021-01-29T19:33:17.409Z [INFO] TestRegisterMonitor_heartbeat.server: shutting down server writer.go:29: 2021-01-29T19:33:17.410Z [DEBUG] TestRegisterMonitor_heartbeat.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.411Z [WARN] TestRegisterMonitor_heartbeat.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:17.411Z [DEBUG] TestRegisterMonitor_heartbeat.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.420Z [WARN] TestRegisterMonitor_heartbeat.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:17.422Z [INFO] TestRegisterMonitor_heartbeat.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:17.423Z [INFO] TestRegisterMonitor_heartbeat: consul server down writer.go:29: 2021-01-29T19:33:17.424Z [INFO] TestRegisterMonitor_heartbeat: shutdown complete writer.go:29: 2021-01-29T19:33:17.425Z [INFO] TestRegisterMonitor_heartbeat: Stopping server: protocol=DNS address=127.0.0.1:25316 network=tcp writer.go:29: 2021-01-29T19:33:17.425Z [INFO] TestRegisterMonitor_heartbeat: Stopping server: protocol=DNS address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:33:17.426Z [INFO] TestRegisterMonitor_heartbeat: Stopping server: protocol=HTTP address=127.0.0.1:25317 network=tcp === CONT TestCommandConfigWatcher/-service_flag_only writer.go:29: 2021-01-29T19:33:17.806Z [INFO] TestCommandConfigWatcher/-service_flag_only: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:17.808Z [INFO] TestCommandConfigWatcher/-service_flag_only: Endpoints down === RUN TestCommandConfigWatcher/-service_flag_with_upstreams === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.832Z [INFO] TestRegisterMonitor_good: Waiting for endpoints to shut down === CONT TestCommandConfigWatcher/-service_flag_with_upstreams writer.go:29: 2021-01-29T19:33:17.848Z [WARN] TestCommandConfigWatcher/-service_flag_with_upstreams: bootstrap = true: do not enable unless necessary === CONT TestRegisterMonitor_good writer.go:29: 2021-01-29T19:33:17.843Z [INFO] TestRegisterMonitor_good: Endpoints down --- PASS: TestRegisterMonitor_good (1.01s) === CONT TestCommandConfigWatcher/-service_flag_with_upstreams writer.go:29: 2021-01-29T19:33:17.850Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:17.852Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:17.858Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:cf73899b-ee7f-45fe-e8be-1c80ad89efc0 Address:127.0.0.1:25339}]" writer.go:29: 2021-01-29T19:33:17.860Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server.raft: entering follower state: follower="Node at 127.0.0.1:25339 [Follower]" leader= writer.go:29: 2021-01-29T19:33:17.876Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server.serf.wan: serf: EventMemberJoin: Node-cf73899b-ee7f-45fe-e8be-1c80ad89efc0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:17.880Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server.serf.lan: serf: EventMemberJoin: Node-cf73899b-ee7f-45fe-e8be-1c80ad89efc0 127.0.0.1 writer.go:29: 2021-01-29T19:33:17.884Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Started DNS server: address=127.0.0.1:25334 network=udp writer.go:29: 2021-01-29T19:33:17.886Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server: Handled event for server in area: event=member-join server=Node-cf73899b-ee7f-45fe-e8be-1c80ad89efc0.dc1 area=wan writer.go:29: 2021-01-29T19:33:17.886Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server: Adding LAN server: server="Node-cf73899b-ee7f-45fe-e8be-1c80ad89efc0 (Addr: tcp/127.0.0.1:25339) (DC: dc1)" writer.go:29: 2021-01-29T19:33:17.890Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Started DNS server: address=127.0.0.1:25334 network=tcp writer.go:29: 2021-01-29T19:33:17.903Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Started HTTP server: address=127.0.0.1:25335 network=tcp writer.go:29: 2021-01-29T19:33:17.907Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: started state syncer writer.go:29: 2021-01-29T19:33:17.921Z [WARN] TestCommandConfigWatcher/-service_flag_with_upstreams.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:17.925Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server.raft: entering candidate state: node="Node at 127.0.0.1:25339 [Candidate]" term=2 === CONT TestRegisterMonitor_heartbeat writer.go:29: 2021-01-29T19:33:17.927Z [INFO] TestRegisterMonitor_heartbeat: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:17.929Z [INFO] TestRegisterMonitor_heartbeat: Endpoints down --- PASS: TestRegisterMonitor_heartbeat (1.09s) === CONT TestCommandConfigWatcher/-service_flag_with_upstreams writer.go:29: 2021-01-29T19:33:17.929Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:17.932Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams.server.raft: vote granted: from=cf73899b-ee7f-45fe-e8be-1c80ad89efc0 term=2 tally=1 writer.go:29: 2021-01-29T19:33:17.934Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:17.936Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server.raft: entering leader state: leader="Node at 127.0.0.1:25339 [Leader]" writer.go:29: 2021-01-29T19:33:17.939Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:17.939Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server: New leader elected: payload=Node-cf73899b-ee7f-45fe-e8be-1c80ad89efc0 writer.go:29: 2021-01-29T19:33:17.953Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25339 writer.go:29: 2021-01-29T19:33:17.974Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:17.981Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:17.983Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:17.984Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams.server: Skipping self join check for node since the cluster is too small: node=Node-cf73899b-ee7f-45fe-e8be-1c80ad89efc0 writer.go:29: 2021-01-29T19:33:17.985Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server: member joined, marking health alive: member=Node-cf73899b-ee7f-45fe-e8be-1c80ad89efc0 writer.go:29: 2021-01-29T19:33:18.076Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:18.079Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Synced node info writer.go:29: 2021-01-29T19:33:18.086Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Synced service: service=two-sidecars-sidecar-proxy writer.go:29: 2021-01-29T19:33:18.095Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Synced service: service=other-sidecar-for-two-sidecars writer.go:29: 2021-01-29T19:33:18.099Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Synced service: service=no-sidecar writer.go:29: 2021-01-29T19:33:18.104Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Synced service: service=one-sidecar writer.go:29: 2021-01-29T19:33:18.118Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Synced service: service=one-sidecar-sidecar-proxy writer.go:29: 2021-01-29T19:33:18.130Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Synced service: service=two-sidecars writer.go:29: 2021-01-29T19:33:18.132Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Check in sync: check=service:one-sidecar-sidecar-proxy:1 writer.go:29: 2021-01-29T19:33:18.133Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Check in sync: check=service:one-sidecar-sidecar-proxy:2 writer.go:29: 2021-01-29T19:33:18.136Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Check in sync: check=service:two-sidecars-sidecar-proxy:1 writer.go:29: 2021-01-29T19:33:18.144Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Check in sync: check=service:two-sidecars-sidecar-proxy:2 writer.go:29: 2021-01-29T19:33:18.153Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Node info in sync writer.go:29: 2021-01-29T19:33:18.157Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Service in sync: service=one-sidecar-sidecar-proxy writer.go:29: 2021-01-29T19:33:18.159Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Service in sync: service=two-sidecars writer.go:29: 2021-01-29T19:33:18.161Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Service in sync: service=two-sidecars-sidecar-proxy writer.go:29: 2021-01-29T19:33:18.168Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Service in sync: service=other-sidecar-for-two-sidecars writer.go:29: 2021-01-29T19:33:18.171Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Service in sync: service=no-sidecar writer.go:29: 2021-01-29T19:33:18.173Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Service in sync: service=one-sidecar writer.go:29: 2021-01-29T19:33:18.172Z [WARN] TestCommandConfigWatcher/-service_flag_with_upstreams: Check socket connection failed: check=service:two-sidecars-sidecar-proxy:1 error="dial tcp 127.0.0.1:21000: connect: connection refused" writer.go:29: 2021-01-29T19:33:18.177Z [WARN] TestCommandConfigWatcher/-service_flag_with_upstreams: Check is now critical: check=service:two-sidecars-sidecar-proxy:1 writer.go:29: 2021-01-29T19:33:18.174Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Check in sync: check=service:one-sidecar-sidecar-proxy:2 writer.go:29: 2021-01-29T19:33:18.175Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Requesting shutdown writer.go:29: 2021-01-29T19:33:18.180Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Check in sync: check=service:two-sidecars-sidecar-proxy:1 writer.go:29: 2021-01-29T19:33:18.191Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Check in sync: check=service:two-sidecars-sidecar-proxy:2 writer.go:29: 2021-01-29T19:33:18.192Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams: Check in sync: check=service:one-sidecar-sidecar-proxy:1 writer.go:29: 2021-01-29T19:33:18.192Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server: shutting down server writer.go:29: 2021-01-29T19:33:18.193Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.194Z [WARN] TestCommandConfigWatcher/-service_flag_with_upstreams.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:18.196Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_upstreams.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.212Z [WARN] TestCommandConfigWatcher/-service_flag_with_upstreams.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:18.215Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:18.216Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: consul server down writer.go:29: 2021-01-29T19:33:18.217Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: shutdown complete writer.go:29: 2021-01-29T19:33:18.218Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Stopping server: protocol=DNS address=127.0.0.1:25334 network=tcp writer.go:29: 2021-01-29T19:33:18.219Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Stopping server: protocol=DNS address=127.0.0.1:25334 network=udp writer.go:29: 2021-01-29T19:33:18.220Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Stopping server: protocol=HTTP address=127.0.0.1:25335 network=tcp writer.go:29: 2021-01-29T19:33:18.721Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:18.722Z [INFO] TestCommandConfigWatcher/-service_flag_with_upstreams: Endpoints down === RUN TestCommandConfigWatcher/-service_flag_with_-service-addr writer.go:29: 2021-01-29T19:33:18.736Z [WARN] TestCommandConfigWatcher/-service_flag_with_-service-addr: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:18.738Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_-service-addr.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:18.740Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_-service-addr.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:18.754Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b3681d20-d863-81ed-417d-ed07d9f3d3ed Address:127.0.0.1:25345}]" writer.go:29: 2021-01-29T19:33:18.757Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.raft: entering follower state: follower="Node at 127.0.0.1:25345 [Follower]" leader= writer.go:29: 2021-01-29T19:33:18.758Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.serf.wan: serf: EventMemberJoin: Node-b3681d20-d863-81ed-417d-ed07d9f3d3ed.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.762Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.serf.lan: serf: EventMemberJoin: Node-b3681d20-d863-81ed-417d-ed07d9f3d3ed 127.0.0.1 writer.go:29: 2021-01-29T19:33:18.765Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server: Adding LAN server: server="Node-b3681d20-d863-81ed-417d-ed07d9f3d3ed (Addr: tcp/127.0.0.1:25345) (DC: dc1)" writer.go:29: 2021-01-29T19:33:18.766Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server: Handled event for server in area: event=member-join server=Node-b3681d20-d863-81ed-417d-ed07d9f3d3ed.dc1 area=wan writer.go:29: 2021-01-29T19:33:18.768Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: Started DNS server: address=127.0.0.1:25340 network=tcp writer.go:29: 2021-01-29T19:33:18.773Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: Started DNS server: address=127.0.0.1:25340 network=udp writer.go:29: 2021-01-29T19:33:18.775Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: Started HTTP server: address=127.0.0.1:25341 network=tcp writer.go:29: 2021-01-29T19:33:18.777Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: started state syncer writer.go:29: 2021-01-29T19:33:18.806Z [WARN] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:18.807Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.raft: entering candidate state: node="Node at 127.0.0.1:25345 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:18.809Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:18.809Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.raft: vote granted: from=b3681d20-d863-81ed-417d-ed07d9f3d3ed term=2 tally=1 writer.go:29: 2021-01-29T19:33:18.810Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:18.811Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.raft: entering leader state: leader="Node at 127.0.0.1:25345 [Leader]" writer.go:29: 2021-01-29T19:33:18.813Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:18.814Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server: New leader elected: payload=Node-b3681d20-d863-81ed-417d-ed07d9f3d3ed writer.go:29: 2021-01-29T19:33:18.816Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_-service-addr.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25345 writer.go:29: 2021-01-29T19:33:18.821Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:18.844Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:18.846Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:18.848Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_-service-addr.server: Skipping self join check for node since the cluster is too small: node=Node-b3681d20-d863-81ed-417d-ed07d9f3d3ed writer.go:29: 2021-01-29T19:33:18.850Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server: member joined, marking health alive: member=Node-b3681d20-d863-81ed-417d-ed07d9f3d3ed writer.go:29: 2021-01-29T19:33:19.034Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: Requesting shutdown writer.go:29: 2021-01-29T19:33:19.037Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server: shutting down server writer.go:29: 2021-01-29T19:33:19.038Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_-service-addr.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.040Z [WARN] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:19.041Z [ERROR] TestCommandConfigWatcher/-service_flag_with_-service-addr.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:19.041Z [DEBUG] TestCommandConfigWatcher/-service_flag_with_-service-addr.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.047Z [WARN] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:19.049Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:19.052Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: consul server down writer.go:29: 2021-01-29T19:33:19.053Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: shutdown complete writer.go:29: 2021-01-29T19:33:19.054Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: Stopping server: protocol=DNS address=127.0.0.1:25340 network=tcp writer.go:29: 2021-01-29T19:33:19.054Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: Stopping server: protocol=DNS address=127.0.0.1:25340 network=udp writer.go:29: 2021-01-29T19:33:19.055Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: Stopping server: protocol=HTTP address=127.0.0.1:25341 network=tcp writer.go:29: 2021-01-29T19:33:19.556Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:19.557Z [INFO] TestCommandConfigWatcher/-service_flag_with_-service-addr: Endpoints down === RUN TestCommandConfigWatcher/-service,_-service-addr,_-listen writer.go:29: 2021-01-29T19:33:19.588Z [WARN] TestCommandConfigWatcher/-service,_-service-addr,_-listen: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:19.590Z [DEBUG] TestCommandConfigWatcher/-service,_-service-addr,_-listen.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:19.594Z [DEBUG] TestCommandConfigWatcher/-service,_-service-addr,_-listen.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:19.603Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:57279a09-074c-8f09-524d-7b8d3f53ff02 Address:127.0.0.1:25351}]" writer.go:29: 2021-01-29T19:33:19.606Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.raft: entering follower state: follower="Node at 127.0.0.1:25351 [Follower]" leader= writer.go:29: 2021-01-29T19:33:19.608Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.serf.wan: serf: EventMemberJoin: Node-57279a09-074c-8f09-524d-7b8d3f53ff02.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:19.617Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.serf.lan: serf: EventMemberJoin: Node-57279a09-074c-8f09-524d-7b8d3f53ff02 127.0.0.1 writer.go:29: 2021-01-29T19:33:19.622Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server: Adding LAN server: server="Node-57279a09-074c-8f09-524d-7b8d3f53ff02 (Addr: tcp/127.0.0.1:25351) (DC: dc1)" writer.go:29: 2021-01-29T19:33:19.623Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: Started DNS server: address=127.0.0.1:25346 network=udp writer.go:29: 2021-01-29T19:33:19.623Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server: Handled event for server in area: event=member-join server=Node-57279a09-074c-8f09-524d-7b8d3f53ff02.dc1 area=wan writer.go:29: 2021-01-29T19:33:19.627Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: Started DNS server: address=127.0.0.1:25346 network=tcp writer.go:29: 2021-01-29T19:33:19.634Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: Started HTTP server: address=127.0.0.1:25347 network=tcp writer.go:29: 2021-01-29T19:33:19.635Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: started state syncer writer.go:29: 2021-01-29T19:33:19.650Z [WARN] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:19.652Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.raft: entering candidate state: node="Node at 127.0.0.1:25351 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:19.655Z [DEBUG] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:19.657Z [DEBUG] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.raft: vote granted: from=57279a09-074c-8f09-524d-7b8d3f53ff02 term=2 tally=1 writer.go:29: 2021-01-29T19:33:19.659Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:19.660Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.raft: entering leader state: leader="Node at 127.0.0.1:25351 [Leader]" writer.go:29: 2021-01-29T19:33:19.662Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:19.663Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server: New leader elected: payload=Node-57279a09-074c-8f09-524d-7b8d3f53ff02 writer.go:29: 2021-01-29T19:33:19.667Z [DEBUG] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25351 writer.go:29: 2021-01-29T19:33:19.682Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:19.688Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:19.690Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.691Z [DEBUG] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server: Skipping self join check for node since the cluster is too small: node=Node-57279a09-074c-8f09-524d-7b8d3f53ff02 writer.go:29: 2021-01-29T19:33:19.692Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server: member joined, marking health alive: member=Node-57279a09-074c-8f09-524d-7b8d3f53ff02 writer.go:29: 2021-01-29T19:33:19.792Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: Requesting shutdown writer.go:29: 2021-01-29T19:33:19.795Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server: shutting down server writer.go:29: 2021-01-29T19:33:19.797Z [DEBUG] TestCommandConfigWatcher/-service,_-service-addr,_-listen.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.798Z [ERROR] TestCommandConfigWatcher/-service,_-service-addr,_-listen.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:19.802Z [DEBUG] TestCommandConfigWatcher/-service,_-service-addr,_-listen.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:19.804Z [WARN] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:19.809Z [WARN] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:19.812Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:19.813Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: consul server down writer.go:29: 2021-01-29T19:33:19.815Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: shutdown complete writer.go:29: 2021-01-29T19:33:19.817Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: Stopping server: protocol=DNS address=127.0.0.1:25346 network=tcp writer.go:29: 2021-01-29T19:33:19.818Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: Stopping server: protocol=DNS address=127.0.0.1:25346 network=udp writer.go:29: 2021-01-29T19:33:19.819Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: Stopping server: protocol=HTTP address=127.0.0.1:25347 network=tcp writer.go:29: 2021-01-29T19:33:20.321Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:20.322Z [INFO] TestCommandConfigWatcher/-service,_-service-addr,_-listen: Endpoints down === RUN TestCommandConfigWatcher/-sidecar-for,_no_sidecar writer.go:29: 2021-01-29T19:33:20.362Z [WARN] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:20.363Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:20.364Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:20.369Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b60573c8-99a3-b2a6-00fe-ded590e17688 Address:127.0.0.1:25357}]" writer.go:29: 2021-01-29T19:33:20.374Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:25357 [Follower]" leader= writer.go:29: 2021-01-29T19:33:20.380Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.serf.wan: serf: EventMemberJoin: Node-b60573c8-99a3-b2a6-00fe-ded590e17688.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:20.387Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.serf.lan: serf: EventMemberJoin: Node-b60573c8-99a3-b2a6-00fe-ded590e17688 127.0.0.1 writer.go:29: 2021-01-29T19:33:20.392Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server: Adding LAN server: server="Node-b60573c8-99a3-b2a6-00fe-ded590e17688 (Addr: tcp/127.0.0.1:25357) (DC: dc1)" writer.go:29: 2021-01-29T19:33:20.393Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server: Handled event for server in area: event=member-join server=Node-b60573c8-99a3-b2a6-00fe-ded590e17688.dc1 area=wan writer.go:29: 2021-01-29T19:33:20.414Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: Started DNS server: address=127.0.0.1:25352 network=udp writer.go:29: 2021-01-29T19:33:20.415Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: Started DNS server: address=127.0.0.1:25352 network=tcp writer.go:29: 2021-01-29T19:33:20.416Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: Started HTTP server: address=127.0.0.1:25353 network=tcp writer.go:29: 2021-01-29T19:33:20.419Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: started state syncer writer.go:29: 2021-01-29T19:33:20.422Z [WARN] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:20.424Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:25357 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:20.428Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:20.430Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.raft: vote granted: from=b60573c8-99a3-b2a6-00fe-ded590e17688 term=2 tally=1 writer.go:29: 2021-01-29T19:33:20.433Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:20.435Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:25357 [Leader]" writer.go:29: 2021-01-29T19:33:20.437Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:20.438Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server: New leader elected: payload=Node-b60573c8-99a3-b2a6-00fe-ded590e17688 writer.go:29: 2021-01-29T19:33:20.442Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:47318 latency=1.751179ms writer.go:29: 2021-01-29T19:33:20.446Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25357 writer.go:29: 2021-01-29T19:33:20.450Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:33:20.456Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:33:20.457Z [WARN] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:20.462Z [ERROR] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:20.496Z [WARN] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:20.498Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:20.503Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:20.509Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: consul server down writer.go:29: 2021-01-29T19:33:20.511Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: shutdown complete writer.go:29: 2021-01-29T19:33:20.509Z [ERROR] TestCommandConfigWatcher/-sidecar-for,_no_sidecar.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: leadership lost while committing log" writer.go:29: 2021-01-29T19:33:20.513Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: Stopping server: protocol=DNS address=127.0.0.1:25352 network=tcp writer.go:29: 2021-01-29T19:33:20.515Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: Stopping server: protocol=DNS address=127.0.0.1:25352 network=udp writer.go:29: 2021-01-29T19:33:20.516Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:25353 network=tcp writer.go:29: 2021-01-29T19:33:21.018Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:21.019Z [INFO] TestCommandConfigWatcher/-sidecar-for,_no_sidecar: Endpoints down === RUN TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars writer.go:29: 2021-01-29T19:33:21.035Z [WARN] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:21.037Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:21.040Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:21.045Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:52ceb504-ce07-623b-f3d4-0e54ed3c6d3e Address:127.0.0.1:25363}]" writer.go:29: 2021-01-29T19:33:21.048Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.raft: entering follower state: follower="Node at 127.0.0.1:25363 [Follower]" leader= writer.go:29: 2021-01-29T19:33:21.053Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.serf.wan: serf: EventMemberJoin: Node-52ceb504-ce07-623b-f3d4-0e54ed3c6d3e.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.057Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.serf.lan: serf: EventMemberJoin: Node-52ceb504-ce07-623b-f3d4-0e54ed3c6d3e 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.061Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server: Adding LAN server: server="Node-52ceb504-ce07-623b-f3d4-0e54ed3c6d3e (Addr: tcp/127.0.0.1:25363) (DC: dc1)" writer.go:29: 2021-01-29T19:33:21.064Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server: Handled event for server in area: event=member-join server=Node-52ceb504-ce07-623b-f3d4-0e54ed3c6d3e.dc1 area=wan writer.go:29: 2021-01-29T19:33:21.069Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Started DNS server: address=127.0.0.1:25358 network=udp writer.go:29: 2021-01-29T19:33:21.070Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Started DNS server: address=127.0.0.1:25358 network=tcp writer.go:29: 2021-01-29T19:33:21.073Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Started HTTP server: address=127.0.0.1:25359 network=tcp writer.go:29: 2021-01-29T19:33:21.074Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: started state syncer writer.go:29: 2021-01-29T19:33:21.118Z [WARN] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:21.120Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.raft: entering candidate state: node="Node at 127.0.0.1:25363 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:21.123Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:21.124Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.raft: vote granted: from=52ceb504-ce07-623b-f3d4-0e54ed3c6d3e term=2 tally=1 writer.go:29: 2021-01-29T19:33:21.125Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:21.127Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.raft: entering leader state: leader="Node at 127.0.0.1:25363 [Leader]" writer.go:29: 2021-01-29T19:33:21.128Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:21.129Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server: New leader elected: payload=Node-52ceb504-ce07-623b-f3d4-0e54ed3c6d3e writer.go:29: 2021-01-29T19:33:21.135Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25363 writer.go:29: 2021-01-29T19:33:21.141Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:21.147Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:21.149Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.151Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server: Skipping self join check for node since the cluster is too small: node=Node-52ceb504-ce07-623b-f3d4-0e54ed3c6d3e writer.go:29: 2021-01-29T19:33:21.152Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server: member joined, marking health alive: member=Node-52ceb504-ce07-623b-f3d4-0e54ed3c6d3e writer.go:29: 2021-01-29T19:33:21.241Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:21.242Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Synced node info writer.go:29: 2021-01-29T19:33:21.244Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Synced service: service=no-sidecar writer.go:29: 2021-01-29T19:33:21.245Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Synced service: service=one-sidecar writer.go:29: 2021-01-29T19:33:21.247Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Synced service: service=one-sidecar-sidecar-proxy writer.go:29: 2021-01-29T19:33:21.249Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Synced service: service=two-sidecars writer.go:29: 2021-01-29T19:33:21.251Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Synced service: service=two-sidecars-sidecar-proxy writer.go:29: 2021-01-29T19:33:21.253Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Synced service: service=other-sidecar-for-two-sidecars writer.go:29: 2021-01-29T19:33:21.254Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Check in sync: check=service:one-sidecar-sidecar-proxy:1 writer.go:29: 2021-01-29T19:33:21.255Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Check in sync: check=service:one-sidecar-sidecar-proxy:2 writer.go:29: 2021-01-29T19:33:21.256Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Check in sync: check=service:two-sidecars-sidecar-proxy:1 writer.go:29: 2021-01-29T19:33:21.256Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Check in sync: check=service:two-sidecars-sidecar-proxy:2 writer.go:29: 2021-01-29T19:33:21.258Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Node info in sync writer.go:29: 2021-01-29T19:33:21.260Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Service in sync: service=two-sidecars writer.go:29: 2021-01-29T19:33:21.262Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Service in sync: service=two-sidecars-sidecar-proxy writer.go:29: 2021-01-29T19:33:21.264Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Service in sync: service=other-sidecar-for-two-sidecars writer.go:29: 2021-01-29T19:33:21.265Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Service in sync: service=no-sidecar writer.go:29: 2021-01-29T19:33:21.267Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Service in sync: service=one-sidecar writer.go:29: 2021-01-29T19:33:21.269Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Service in sync: service=one-sidecar-sidecar-proxy writer.go:29: 2021-01-29T19:33:21.271Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Check in sync: check=service:one-sidecar-sidecar-proxy:1 writer.go:29: 2021-01-29T19:33:21.273Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Check in sync: check=service:one-sidecar-sidecar-proxy:2 writer.go:29: 2021-01-29T19:33:21.274Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Check in sync: check=service:two-sidecars-sidecar-proxy:1 writer.go:29: 2021-01-29T19:33:21.276Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Check in sync: check=service:two-sidecars-sidecar-proxy:2 writer.go:29: 2021-01-29T19:33:21.303Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:42380 latency=1.111089ms writer.go:29: 2021-01-29T19:33:21.313Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Requesting shutdown writer.go:29: 2021-01-29T19:33:21.315Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server: shutting down server writer.go:29: 2021-01-29T19:33:21.319Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.324Z [WARN] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:21.324Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:21.329Z [WARN] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:21.331Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:21.333Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: consul server down writer.go:29: 2021-01-29T19:33:21.334Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: shutdown complete writer.go:29: 2021-01-29T19:33:21.335Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Stopping server: protocol=DNS address=127.0.0.1:25358 network=tcp writer.go:29: 2021-01-29T19:33:21.337Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Stopping server: protocol=DNS address=127.0.0.1:25358 network=udp writer.go:29: 2021-01-29T19:33:21.338Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Stopping server: protocol=HTTP address=127.0.0.1:25359 network=tcp writer.go:29: 2021-01-29T19:33:21.840Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:21.841Z [INFO] TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars: Endpoints down === RUN TestCommandConfigWatcher/-sidecar-for,_non-existent writer.go:29: 2021-01-29T19:33:21.908Z [WARN] TestCommandConfigWatcher/-sidecar-for,_non-existent: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:21.911Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_non-existent.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:21.922Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_non-existent.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:21.928Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:70c6ff6b-e0f9-e17c-b941-0ddcd251d77f Address:127.0.0.1:25369}]" writer.go:29: 2021-01-29T19:33:21.930Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.raft: entering follower state: follower="Node at 127.0.0.1:25369 [Follower]" leader= writer.go:29: 2021-01-29T19:33:21.932Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.serf.wan: serf: EventMemberJoin: Node-70c6ff6b-e0f9-e17c-b941-0ddcd251d77f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.944Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.serf.lan: serf: EventMemberJoin: Node-70c6ff6b-e0f9-e17c-b941-0ddcd251d77f 127.0.0.1 writer.go:29: 2021-01-29T19:33:21.948Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: Started DNS server: address=127.0.0.1:25364 network=udp writer.go:29: 2021-01-29T19:33:21.948Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server: Adding LAN server: server="Node-70c6ff6b-e0f9-e17c-b941-0ddcd251d77f (Addr: tcp/127.0.0.1:25369) (DC: dc1)" writer.go:29: 2021-01-29T19:33:21.952Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: Started DNS server: address=127.0.0.1:25364 network=tcp writer.go:29: 2021-01-29T19:33:21.954Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: Started HTTP server: address=127.0.0.1:25365 network=tcp writer.go:29: 2021-01-29T19:33:21.955Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: started state syncer writer.go:29: 2021-01-29T19:33:21.949Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server: Handled event for server in area: event=member-join server=Node-70c6ff6b-e0f9-e17c-b941-0ddcd251d77f.dc1 area=wan writer.go:29: 2021-01-29T19:33:21.996Z [WARN] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:21.997Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.raft: entering candidate state: node="Node at 127.0.0.1:25369 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:21.998Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:21.999Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.raft: vote granted: from=70c6ff6b-e0f9-e17c-b941-0ddcd251d77f term=2 tally=1 writer.go:29: 2021-01-29T19:33:22.000Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:22.001Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.raft: entering leader state: leader="Node at 127.0.0.1:25369 [Leader]" writer.go:29: 2021-01-29T19:33:22.002Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:22.002Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server: New leader elected: payload=Node-70c6ff6b-e0f9-e17c-b941-0ddcd251d77f writer.go:29: 2021-01-29T19:33:22.004Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_non-existent.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25369 writer.go:29: 2021-01-29T19:33:22.014Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:22.029Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:22.030Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.031Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_non-existent.server: Skipping self join check for node since the cluster is too small: node=Node-70c6ff6b-e0f9-e17c-b941-0ddcd251d77f writer.go:29: 2021-01-29T19:33:22.032Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server: member joined, marking health alive: member=Node-70c6ff6b-e0f9-e17c-b941-0ddcd251d77f writer.go:29: 2021-01-29T19:33:22.346Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_non-existent.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:59028 latency=201.903µs writer.go:29: 2021-01-29T19:33:22.349Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: Requesting shutdown writer.go:29: 2021-01-29T19:33:22.350Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server: shutting down server writer.go:29: 2021-01-29T19:33:22.351Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_non-existent.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.352Z [WARN] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:22.351Z [ERROR] TestCommandConfigWatcher/-sidecar-for,_non-existent.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:22.352Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_non-existent.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:22.357Z [WARN] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:22.375Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:22.375Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: consul server down writer.go:29: 2021-01-29T19:33:22.378Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: shutdown complete writer.go:29: 2021-01-29T19:33:22.379Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: Stopping server: protocol=DNS address=127.0.0.1:25364 network=tcp writer.go:29: 2021-01-29T19:33:22.381Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: Stopping server: protocol=DNS address=127.0.0.1:25364 network=udp writer.go:29: 2021-01-29T19:33:22.382Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: Stopping server: protocol=HTTP address=127.0.0.1:25365 network=tcp writer.go:29: 2021-01-29T19:33:22.884Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:22.886Z [INFO] TestCommandConfigWatcher/-sidecar-for,_non-existent: Endpoints down === RUN TestCommandConfigWatcher/-sidecar-for,_one_sidecar writer.go:29: 2021-01-29T19:33:22.904Z [WARN] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:22.906Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:22.908Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:22.922Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5e4b3531-b326-9ea1-3573-723cda6869e8 Address:127.0.0.1:25375}]" writer.go:29: 2021-01-29T19:33:22.925Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.raft: entering follower state: follower="Node at 127.0.0.1:25375 [Follower]" leader= writer.go:29: 2021-01-29T19:33:22.940Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.serf.wan: serf: EventMemberJoin: Node-5e4b3531-b326-9ea1-3573-723cda6869e8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:22.945Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.serf.lan: serf: EventMemberJoin: Node-5e4b3531-b326-9ea1-3573-723cda6869e8 127.0.0.1 writer.go:29: 2021-01-29T19:33:22.951Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: Started DNS server: address=127.0.0.1:25370 network=udp writer.go:29: 2021-01-29T19:33:22.959Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: Started DNS server: address=127.0.0.1:25370 network=tcp writer.go:29: 2021-01-29T19:33:22.951Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server: Handled event for server in area: event=member-join server=Node-5e4b3531-b326-9ea1-3573-723cda6869e8.dc1 area=wan writer.go:29: 2021-01-29T19:33:22.962Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server: Adding LAN server: server="Node-5e4b3531-b326-9ea1-3573-723cda6869e8 (Addr: tcp/127.0.0.1:25375) (DC: dc1)" writer.go:29: 2021-01-29T19:33:22.967Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: Started HTTP server: address=127.0.0.1:25371 network=tcp writer.go:29: 2021-01-29T19:33:22.982Z [WARN] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:22.991Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.raft: entering candidate state: node="Node at 127.0.0.1:25375 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:22.993Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: started state syncer writer.go:29: 2021-01-29T19:33:22.995Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:23.008Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.raft: vote granted: from=5e4b3531-b326-9ea1-3573-723cda6869e8 term=2 tally=1 writer.go:29: 2021-01-29T19:33:23.010Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:23.011Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.raft: entering leader state: leader="Node at 127.0.0.1:25375 [Leader]" writer.go:29: 2021-01-29T19:33:23.013Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:23.014Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25375 writer.go:29: 2021-01-29T19:33:23.014Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server: New leader elected: payload=Node-5e4b3531-b326-9ea1-3573-723cda6869e8 writer.go:29: 2021-01-29T19:33:23.026Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:23.043Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:23.044Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:23.045Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server: Skipping self join check for node since the cluster is too small: node=Node-5e4b3531-b326-9ea1-3573-723cda6869e8 writer.go:29: 2021-01-29T19:33:23.050Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server: member joined, marking health alive: member=Node-5e4b3531-b326-9ea1-3573-723cda6869e8 writer.go:29: 2021-01-29T19:33:23.161Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:35390 latency=467.126µs writer.go:29: 2021-01-29T19:33:23.169Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.http: Request finished: method=GET url=/v1/agent/service/one-sidecar-sidecar-proxy from=127.0.0.1:35390 latency=1.662056ms writer.go:29: 2021-01-29T19:33:23.171Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.http: Request finished: method=GET url=/v1/agent/service/one-sidecar-sidecar-proxy from=127.0.0.1:35392 latency=1.80979ms writer.go:29: 2021-01-29T19:33:23.174Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: Requesting shutdown writer.go:29: 2021-01-29T19:33:23.176Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server: shutting down server writer.go:29: 2021-01-29T19:33:23.178Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:23.182Z [WARN] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:23.179Z [ERROR] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:23.182Z [DEBUG] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:23.197Z [WARN] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:23.212Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:23.212Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: consul server down writer.go:29: 2021-01-29T19:33:23.215Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: shutdown complete writer.go:29: 2021-01-29T19:33:23.217Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: Stopping server: protocol=DNS address=127.0.0.1:25370 network=tcp writer.go:29: 2021-01-29T19:33:23.218Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: Stopping server: protocol=DNS address=127.0.0.1:25370 network=udp writer.go:29: 2021-01-29T19:33:23.220Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: Stopping server: protocol=HTTP address=127.0.0.1:25371 network=tcp writer.go:29: 2021-01-29T19:33:24.222Z [WARN] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: Timeout stopping server: protocol=HTTP address=127.0.0.1:25371 network=tcp writer.go:29: 2021-01-29T19:33:24.223Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:24.224Z [INFO] TestCommandConfigWatcher/-sidecar-for,_one_sidecar: Endpoints down --- PASS: TestCommandConfigWatcher (7.39s) --- PASS: TestCommandConfigWatcher/-service_flag_only (0.97s) --- PASS: TestCommandConfigWatcher/-service_flag_with_upstreams (0.91s) --- PASS: TestCommandConfigWatcher/-service_flag_with_-service-addr (0.84s) --- PASS: TestCommandConfigWatcher/-service,_-service-addr,_-listen (0.76s) --- PASS: TestCommandConfigWatcher/-sidecar-for,_no_sidecar (0.70s) --- PASS: TestCommandConfigWatcher/-sidecar-for,_multiple_sidecars (0.82s) --- PASS: TestCommandConfigWatcher/-sidecar-for,_non-existent (1.05s) --- PASS: TestCommandConfigWatcher/-sidecar-for,_one_sidecar (1.34s) PASS ok github.com/hashicorp/consul/command/connect/proxy 7.531s === RUN TestEventCommand_noTabs === PAUSE TestEventCommand_noTabs === RUN TestEventCommand === PAUSE TestEventCommand === CONT TestEventCommand_noTabs === CONT TestEventCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestEventCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestEventCommand writer.go:29: 2021-01-29T19:33:20.506Z [WARN] TestEventCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:20.508Z [DEBUG] TestEventCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:20.510Z [DEBUG] TestEventCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:20.514Z [INFO] TestEventCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bb112f20-a4be-48cc-1bff-699cc8aac2b4 Address:127.0.0.1:26342}]" writer.go:29: 2021-01-29T19:33:20.515Z [INFO] TestEventCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:26342 [Follower]" leader= writer.go:29: 2021-01-29T19:33:20.517Z [INFO] TestEventCommand.server.serf.wan: serf: EventMemberJoin: Node-bb112f20-a4be-48cc-1bff-699cc8aac2b4.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:20.520Z [INFO] TestEventCommand.server.serf.lan: serf: EventMemberJoin: Node-bb112f20-a4be-48cc-1bff-699cc8aac2b4 127.0.0.1 writer.go:29: 2021-01-29T19:33:20.522Z [INFO] TestEventCommand.server: Handled event for server in area: event=member-join server=Node-bb112f20-a4be-48cc-1bff-699cc8aac2b4.dc1 area=wan writer.go:29: 2021-01-29T19:33:20.523Z [INFO] TestEventCommand.server: Adding LAN server: server="Node-bb112f20-a4be-48cc-1bff-699cc8aac2b4 (Addr: tcp/127.0.0.1:26342) (DC: dc1)" writer.go:29: 2021-01-29T19:33:20.525Z [INFO] TestEventCommand: Started DNS server: address=127.0.0.1:26337 network=udp writer.go:29: 2021-01-29T19:33:20.525Z [INFO] TestEventCommand: Started DNS server: address=127.0.0.1:26337 network=tcp writer.go:29: 2021-01-29T19:33:20.528Z [INFO] TestEventCommand: Started HTTP server: address=127.0.0.1:26338 network=tcp writer.go:29: 2021-01-29T19:33:20.529Z [INFO] TestEventCommand: started state syncer writer.go:29: 2021-01-29T19:33:20.561Z [WARN] TestEventCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:20.562Z [INFO] TestEventCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:26342 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:20.564Z [DEBUG] TestEventCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:20.566Z [DEBUG] TestEventCommand.server.raft: vote granted: from=bb112f20-a4be-48cc-1bff-699cc8aac2b4 term=2 tally=1 writer.go:29: 2021-01-29T19:33:20.567Z [INFO] TestEventCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:20.569Z [INFO] TestEventCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:26342 [Leader]" writer.go:29: 2021-01-29T19:33:20.570Z [INFO] TestEventCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:20.570Z [INFO] TestEventCommand.server: New leader elected: payload=Node-bb112f20-a4be-48cc-1bff-699cc8aac2b4 writer.go:29: 2021-01-29T19:33:20.570Z [DEBUG] TestEventCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:26342 writer.go:29: 2021-01-29T19:33:20.576Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:20.601Z [DEBUG] TestEventCommand.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:57844 latency=7.078091ms writer.go:29: 2021-01-29T19:33:20.619Z [DEBUG] TestEventCommand.server: User event: event=cmd writer.go:29: 2021-01-29T19:33:20.620Z [DEBUG] TestEventCommand: new event: event_name=cmd event_id=e4e65844-e34c-2719-1097-c4ad9049f840 writer.go:29: 2021-01-29T19:33:20.620Z [DEBUG] TestEventCommand.http: Request finished: method=PUT url=/v1/event/fire/cmd from=127.0.0.1:57844 latency=1.231683ms writer.go:29: 2021-01-29T19:33:20.627Z [INFO] TestEventCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:20.630Z [INFO] TestEventCommand.server: shutting down server writer.go:29: 2021-01-29T19:33:20.635Z [WARN] TestEventCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:20.635Z [ERROR] TestEventCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:20.650Z [WARN] TestEventCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:20.670Z [INFO] TestEventCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:20.672Z [ERROR] TestEventCommand.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: leadership lost while committing log" writer.go:29: 2021-01-29T19:33:20.672Z [INFO] TestEventCommand: consul server down writer.go:29: 2021-01-29T19:33:20.674Z [INFO] TestEventCommand: shutdown complete writer.go:29: 2021-01-29T19:33:20.676Z [INFO] TestEventCommand: Stopping server: protocol=DNS address=127.0.0.1:26337 network=tcp writer.go:29: 2021-01-29T19:33:20.677Z [INFO] TestEventCommand: Stopping server: protocol=DNS address=127.0.0.1:26337 network=udp writer.go:29: 2021-01-29T19:33:20.679Z [INFO] TestEventCommand: Stopping server: protocol=HTTP address=127.0.0.1:26338 network=tcp writer.go:29: 2021-01-29T19:33:21.180Z [INFO] TestEventCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:21.181Z [INFO] TestEventCommand: Endpoints down --- PASS: TestEventCommand (0.72s) PASS ok github.com/hashicorp/consul/command/event 0.848s === RUN TestExecCommand_noTabs === PAUSE TestExecCommand_noTabs === RUN TestExecCommand === PAUSE TestExecCommand === RUN TestExecCommand_NoShell === PAUSE TestExecCommand_NoShell === RUN TestExecCommand_CrossDC exec_test.go:70: DM-skipped --- SKIP: TestExecCommand_CrossDC (0.00s) === RUN TestExecCommand_Validate === PAUSE TestExecCommand_Validate === RUN TestExecCommand_Sessions === PAUSE TestExecCommand_Sessions === RUN TestExecCommand_Sessions_Foreign === PAUSE TestExecCommand_Sessions_Foreign === RUN TestExecCommand_UploadDestroy === PAUSE TestExecCommand_UploadDestroy === RUN TestExecCommand_StreamResults === PAUSE TestExecCommand_StreamResults === CONT TestExecCommand_noTabs --- PASS: TestExecCommand_noTabs (0.00s) === CONT TestExecCommand_StreamResults [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestExecCommand_Sessions === CONT TestExecCommand_UploadDestroy === CONT TestExecCommand_Sessions_Foreign [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:23.966Z [WARN] TestExecCommand_StreamResults: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:23.968Z [DEBUG] TestExecCommand_StreamResults.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:23.970Z [DEBUG] TestExecCommand_StreamResults.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:23.985Z [WARN] TestExecCommand_Sessions_Foreign: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:23.986Z [DEBUG] TestExecCommand_Sessions_Foreign.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:23.992Z [DEBUG] TestExecCommand_Sessions_Foreign.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:23.995Z [INFO] TestExecCommand_StreamResults.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:72b247e6-c5d4-3c5e-a8ec-107a78556a9a Address:127.0.0.1:11027}]" === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:23.997Z [WARN] TestExecCommand_Sessions: bootstrap = true: do not enable unless necessary === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:23.997Z [WARN] TestExecCommand_UploadDestroy: bootstrap = true: do not enable unless necessary === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:23.997Z [INFO] TestExecCommand_StreamResults.server.serf.wan: serf: EventMemberJoin: Node-72b247e6-c5d4-3c5e-a8ec-107a78556a9a.dc1 127.0.0.1 === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:23.998Z [DEBUG] TestExecCommand_Sessions.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:24.000Z [DEBUG] TestExecCommand_Sessions.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.014Z [INFO] TestExecCommand_StreamResults.server.raft: entering follower state: follower="Node at 127.0.0.1:11027 [Follower]" leader= writer.go:29: 2021-01-29T19:33:24.016Z [INFO] TestExecCommand_StreamResults.server.serf.lan: serf: EventMemberJoin: Node-72b247e6-c5d4-3c5e-a8ec-107a78556a9a 127.0.0.1 === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:23.998Z [DEBUG] TestExecCommand_UploadDestroy.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:24.019Z [DEBUG] TestExecCommand_UploadDestroy.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.021Z [INFO] TestExecCommand_StreamResults.server: Adding LAN server: server="Node-72b247e6-c5d4-3c5e-a8ec-107a78556a9a (Addr: tcp/127.0.0.1:11027) (DC: dc1)" === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.024Z [INFO] TestExecCommand_Sessions_Foreign.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2d471161-ba43-9134-9242-6ffb9e3ca112 Address:127.0.0.1:11045}]" === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.034Z [INFO] TestExecCommand_StreamResults.server: Handled event for server in area: event=member-join server=Node-72b247e6-c5d4-3c5e-a8ec-107a78556a9a.dc1 area=wan writer.go:29: 2021-01-29T19:33:24.036Z [INFO] TestExecCommand_StreamResults: Started DNS server: address=127.0.0.1:11022 network=tcp === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.039Z [INFO] TestExecCommand_Sessions_Foreign.server.raft: entering follower state: follower="Node at 127.0.0.1:11045 [Follower]" leader= === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.039Z [INFO] TestExecCommand_Sessions.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d67f8281-c84b-9173-03d9-4513bc1999de Address:127.0.0.1:11033}]" writer.go:29: 2021-01-29T19:33:24.044Z [INFO] TestExecCommand_Sessions.server.serf.wan: serf: EventMemberJoin: Node-d67f8281-c84b-9173-03d9-4513bc1999de.dc1 127.0.0.1 === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.049Z [INFO] TestExecCommand_StreamResults: Started DNS server: address=127.0.0.1:11022 network=udp === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.061Z [INFO] TestExecCommand_Sessions_Foreign.server.serf.wan: serf: EventMemberJoin: Node-2d471161-ba43-9134-9242-6ffb9e3ca112.dc1 127.0.0.1 === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.053Z [INFO] TestExecCommand_Sessions.server.raft: entering follower state: follower="Node at 127.0.0.1:11033 [Follower]" leader= === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.066Z [INFO] TestExecCommand_Sessions_Foreign.server.serf.lan: serf: EventMemberJoin: Node-2d471161-ba43-9134-9242-6ffb9e3ca112 127.0.0.1 === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.068Z [INFO] TestExecCommand_StreamResults: Started HTTP server: address=127.0.0.1:11023 network=tcp writer.go:29: 2021-01-29T19:33:24.069Z [INFO] TestExecCommand_StreamResults: started state syncer === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.070Z [INFO] TestExecCommand_UploadDestroy.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:46155e46-c888-6896-4196-75cc5d574315 Address:127.0.0.1:11039}]" === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.061Z [INFO] TestExecCommand_Sessions.server.serf.lan: serf: EventMemberJoin: Node-d67f8281-c84b-9173-03d9-4513bc1999de 127.0.0.1 === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.072Z [INFO] TestExecCommand_UploadDestroy.server.raft: entering follower state: follower="Node at 127.0.0.1:11039 [Follower]" leader= === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.075Z [WARN] TestExecCommand_StreamResults.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:24.077Z [INFO] TestExecCommand_StreamResults.server.raft: entering candidate state: node="Node at 127.0.0.1:11027 [Candidate]" term=2 === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.083Z [INFO] TestExecCommand_Sessions_Foreign.server: Adding LAN server: server="Node-2d471161-ba43-9134-9242-6ffb9e3ca112 (Addr: tcp/127.0.0.1:11045) (DC: dc1)" writer.go:29: 2021-01-29T19:33:24.088Z [INFO] TestExecCommand_Sessions_Foreign: Started DNS server: address=127.0.0.1:11040 network=udp writer.go:29: 2021-01-29T19:33:24.088Z [INFO] TestExecCommand_Sessions_Foreign.server: Handled event for server in area: event=member-join server=Node-2d471161-ba43-9134-9242-6ffb9e3ca112.dc1 area=wan writer.go:29: 2021-01-29T19:33:24.092Z [INFO] TestExecCommand_Sessions_Foreign: Started DNS server: address=127.0.0.1:11040 network=tcp writer.go:29: 2021-01-29T19:33:24.095Z [INFO] TestExecCommand_Sessions_Foreign: Started HTTP server: address=127.0.0.1:11041 network=tcp writer.go:29: 2021-01-29T19:33:24.096Z [INFO] TestExecCommand_Sessions_Foreign: started state syncer === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.099Z [INFO] TestExecCommand_Sessions.server: Adding LAN server: server="Node-d67f8281-c84b-9173-03d9-4513bc1999de (Addr: tcp/127.0.0.1:11033) (DC: dc1)" === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.099Z [WARN] TestExecCommand_Sessions_Foreign.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.100Z [INFO] TestExecCommand_Sessions.server: Handled event for server in area: event=member-join server=Node-d67f8281-c84b-9173-03d9-4513bc1999de.dc1 area=wan === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.100Z [INFO] TestExecCommand_Sessions_Foreign.server.raft: entering candidate state: node="Node at 127.0.0.1:11045 [Candidate]" term=2 === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.102Z [INFO] TestExecCommand_Sessions: Started DNS server: address=127.0.0.1:11028 network=udp writer.go:29: 2021-01-29T19:33:24.104Z [INFO] TestExecCommand_Sessions: Started DNS server: address=127.0.0.1:11028 network=tcp === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.103Z [INFO] TestExecCommand_UploadDestroy.server.serf.wan: serf: EventMemberJoin: Node-46155e46-c888-6896-4196-75cc5d574315.dc1 127.0.0.1 === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.109Z [INFO] TestExecCommand_Sessions: Started HTTP server: address=127.0.0.1:11029 network=tcp writer.go:29: 2021-01-29T19:33:24.111Z [INFO] TestExecCommand_Sessions: started state syncer === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.112Z [INFO] TestExecCommand_UploadDestroy.server.serf.lan: serf: EventMemberJoin: Node-46155e46-c888-6896-4196-75cc5d574315 127.0.0.1 === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.115Z [DEBUG] TestExecCommand_Sessions_Foreign.server.raft: votes: needed=1 === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.115Z [DEBUG] TestExecCommand_StreamResults.server.raft: votes: needed=1 === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.117Z [DEBUG] TestExecCommand_Sessions_Foreign.server.raft: vote granted: from=2d471161-ba43-9134-9242-6ffb9e3ca112 term=2 tally=1 === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.117Z [DEBUG] TestExecCommand_StreamResults.server.raft: vote granted: from=72b247e6-c5d4-3c5e-a8ec-107a78556a9a term=2 tally=1 === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.118Z [INFO] TestExecCommand_Sessions_Foreign.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:24.120Z [INFO] TestExecCommand_Sessions_Foreign.server.raft: entering leader state: leader="Node at 127.0.0.1:11045 [Leader]" === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.120Z [INFO] TestExecCommand_StreamResults.server.raft: election won: tally=1 === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.121Z [INFO] TestExecCommand_UploadDestroy.server: Adding LAN server: server="Node-46155e46-c888-6896-4196-75cc5d574315 (Addr: tcp/127.0.0.1:11039) (DC: dc1)" writer.go:29: 2021-01-29T19:33:24.121Z [INFO] TestExecCommand_UploadDestroy: Started DNS server: address=127.0.0.1:11034 network=udp === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.124Z [INFO] TestExecCommand_StreamResults.server.raft: entering leader state: leader="Node at 127.0.0.1:11027 [Leader]" === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.127Z [INFO] TestExecCommand_UploadDestroy: Started DNS server: address=127.0.0.1:11034 network=tcp === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.128Z [INFO] TestExecCommand_StreamResults.server: cluster leadership acquired === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.129Z [WARN] TestExecCommand_Sessions.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.121Z [INFO] TestExecCommand_UploadDestroy.server: Handled event for server in area: event=member-join server=Node-46155e46-c888-6896-4196-75cc5d574315.dc1 area=wan === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.130Z [INFO] TestExecCommand_StreamResults.server: New leader elected: payload=Node-72b247e6-c5d4-3c5e-a8ec-107a78556a9a === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.130Z [INFO] TestExecCommand_Sessions.server.raft: entering candidate state: node="Node at 127.0.0.1:11033 [Candidate]" term=2 === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.123Z [INFO] TestExecCommand_Sessions_Foreign.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:24.124Z [INFO] TestExecCommand_Sessions_Foreign.server: New leader elected: payload=Node-2d471161-ba43-9134-9242-6ffb9e3ca112 === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.132Z [INFO] TestExecCommand_UploadDestroy: Started HTTP server: address=127.0.0.1:11035 network=tcp === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.137Z [DEBUG] TestExecCommand_StreamResults.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:11027 === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.133Z [DEBUG] TestExecCommand_Sessions_Foreign.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:11045 writer.go:29: 2021-01-29T19:33:24.144Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.145Z [DEBUG] TestExecCommand_Sessions.server.raft: votes: needed=1 === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.139Z [INFO] TestExecCommand_UploadDestroy: started state syncer === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.147Z [DEBUG] TestExecCommand_Sessions.server.raft: vote granted: from=d67f8281-c84b-9173-03d9-4513bc1999de term=2 tally=1 === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.137Z [WARN] TestExecCommand_UploadDestroy.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.150Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.149Z [INFO] TestExecCommand_Sessions.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:24.152Z [INFO] TestExecCommand_Sessions.server.raft: entering leader state: leader="Node at 127.0.0.1:11033 [Leader]" writer.go:29: 2021-01-29T19:33:24.154Z [INFO] TestExecCommand_Sessions.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:24.154Z [INFO] TestExecCommand_Sessions.server: New leader elected: payload=Node-d67f8281-c84b-9173-03d9-4513bc1999de === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.151Z [INFO] TestExecCommand_UploadDestroy.server.raft: entering candidate state: node="Node at 127.0.0.1:11039 [Candidate]" term=2 === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.158Z [DEBUG] TestExecCommand_Sessions.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:11033 === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.164Z [DEBUG] TestExecCommand_UploadDestroy.server.raft: votes: needed=1 === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.166Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.166Z [DEBUG] TestExecCommand_UploadDestroy.server.raft: vote granted: from=46155e46-c888-6896-4196-75cc5d574315 term=2 tally=1 writer.go:29: 2021-01-29T19:33:24.169Z [INFO] TestExecCommand_UploadDestroy.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:24.172Z [INFO] TestExecCommand_UploadDestroy.server.raft: entering leader state: leader="Node at 127.0.0.1:11039 [Leader]" writer.go:29: 2021-01-29T19:33:24.176Z [INFO] TestExecCommand_UploadDestroy.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:24.176Z [INFO] TestExecCommand_UploadDestroy.server: New leader elected: payload=Node-46155e46-c888-6896-4196-75cc5d574315 writer.go:29: 2021-01-29T19:33:24.177Z [DEBUG] TestExecCommand_UploadDestroy.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:11039 === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.202Z [INFO] TestExecCommand_Sessions_Foreign: Synced node info === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.204Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.204Z [DEBUG] TestExecCommand_Sessions_Foreign: Node info in sync writer.go:29: 2021-01-29T19:33:24.216Z [INFO] TestExecCommand_Sessions_Foreign.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.219Z [INFO] TestExecCommand_Sessions.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:24.220Z [INFO] TestExecCommand_Sessions.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.220Z [DEBUG] TestExecCommand_Sessions.server: Skipping self join check for node since the cluster is too small: node=Node-d67f8281-c84b-9173-03d9-4513bc1999de === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.219Z [INFO] TestExecCommand_Sessions_Foreign.leader: started routine: routine="CA root pruning" === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.215Z [INFO] TestExecCommand_StreamResults.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.222Z [DEBUG] TestExecCommand_Sessions_Foreign.server: Skipping self join check for node since the cluster is too small: node=Node-2d471161-ba43-9134-9242-6ffb9e3ca112 === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.221Z [INFO] TestExecCommand_Sessions.server: member joined, marking health alive: member=Node-d67f8281-c84b-9173-03d9-4513bc1999de === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.225Z [INFO] TestExecCommand_Sessions_Foreign.server: member joined, marking health alive: member=Node-2d471161-ba43-9134-9242-6ffb9e3ca112 === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.230Z [INFO] TestExecCommand_UploadDestroy.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:24.232Z [INFO] TestExecCommand_UploadDestroy.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.234Z [DEBUG] TestExecCommand_UploadDestroy.server: Skipping self join check for node since the cluster is too small: node=Node-46155e46-c888-6896-4196-75cc5d574315 writer.go:29: 2021-01-29T19:33:24.236Z [INFO] TestExecCommand_UploadDestroy.server: member joined, marking health alive: member=Node-46155e46-c888-6896-4196-75cc5d574315 === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.224Z [INFO] TestExecCommand_StreamResults.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.242Z [DEBUG] TestExecCommand_StreamResults.server: Skipping self join check for node since the cluster is too small: node=Node-72b247e6-c5d4-3c5e-a8ec-107a78556a9a writer.go:29: 2021-01-29T19:33:24.247Z [INFO] TestExecCommand_StreamResults.server: member joined, marking health alive: member=Node-72b247e6-c5d4-3c5e-a8ec-107a78556a9a === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.252Z [DEBUG] TestExecCommand_Sessions: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:24.256Z [INFO] TestExecCommand_Sessions: Synced node info === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.325Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:54430 latency=4.15808ms writer.go:29: 2021-01-29T19:33:24.330Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=GET url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/?keys=&wait=2000ms from=127.0.0.1:54430 latency=65.95µs writer.go:29: 2021-01-29T19:33:24.340Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=PUT url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/foo/ack?acquire=2a4c9efd-02c0-80fb-fbab-641479ae3615 from=127.0.0.1:54434 latency=7.651321ms === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.342Z [DEBUG] TestExecCommand_UploadDestroy.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:48880 latency=12.524795ms === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.350Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=GET url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/?index=1&keys=&wait=2000ms from=127.0.0.1:54430 latency=506.736µs === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.354Z [DEBUG] TestExecCommand_UploadDestroy.http: Request finished: method=PUT url=/v1/kv/_rexec/1675c257-c31a-427a-ce02-f28efc404748/job?acquire=1675c257-c31a-427a-ce02-f28efc404748 from=127.0.0.1:48880 latency=4.53838ms === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.361Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=GET url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/?index=12&keys=&wait=2000ms from=127.0.0.1:54430 latency=9.149305ms writer.go:29: 2021-01-29T19:33:24.361Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=PUT url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/foo/exit?acquire=2a4c9efd-02c0-80fb-fbab-641479ae3615 from=127.0.0.1:54436 latency=995.919µs writer.go:29: 2021-01-29T19:33:24.367Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=GET url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/foo/exit from=127.0.0.1:54430 latency=176.843µs === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.368Z [DEBUG] TestExecCommand_UploadDestroy.http: Request finished: method=GET url=/v1/kv/_rexec/1675c257-c31a-427a-ce02-f28efc404748/job from=127.0.0.1:48886 latency=150.532µs writer.go:29: 2021-01-29T19:33:24.372Z [DEBUG] TestExecCommand_UploadDestroy.http: Request finished: method=DELETE url=/v1/kv/_rexec/1675c257-c31a-427a-ce02-f28efc404748?recurse= from=127.0.0.1:48880 latency=672.686µs === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.376Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=PUT url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/foo/random?acquire=2a4c9efd-02c0-80fb-fbab-641479ae3615 from=127.0.0.1:54440 latency=1.018615ms writer.go:29: 2021-01-29T19:33:24.376Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=GET url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/?index=13&keys=&wait=2000ms from=127.0.0.1:54430 latency=6.565978ms === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.413Z [DEBUG] TestExecCommand_UploadDestroy.http: Request finished: method=GET url=/v1/kv/_rexec/1675c257-c31a-427a-ce02-f28efc404748/job from=127.0.0.1:48890 latency=35.724µs writer.go:29: 2021-01-29T19:33:24.415Z [INFO] TestExecCommand_UploadDestroy: Requesting shutdown writer.go:29: 2021-01-29T19:33:24.417Z [INFO] TestExecCommand_UploadDestroy.server: shutting down server === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.417Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=PUT url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/foo/out/00000?acquire=2a4c9efd-02c0-80fb-fbab-641479ae3615 from=127.0.0.1:54446 latency=812.769µs === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.419Z [DEBUG] TestExecCommand_UploadDestroy.leader: stopping routine: routine="CA root pruning" === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.417Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=GET url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/?index=14&keys=&wait=2000ms from=127.0.0.1:54430 latency=29.154315ms === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.421Z [WARN] TestExecCommand_UploadDestroy.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.423Z [ERROR] TestExecCommand_UploadDestroy.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:24.424Z [DEBUG] TestExecCommand_UploadDestroy.leader: stopped routine: routine="CA root pruning" === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.423Z [DEBUG] TestExecCommand_Sessions_Foreign.http: Request finished: method=GET url=/v1/health/service/consul?passing=1 from=127.0.0.1:43280 latency=7.46406ms === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.424Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=GET url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/foo/out/00000 from=127.0.0.1:54430 latency=161.643µs === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.427Z [WARN] TestExecCommand_UploadDestroy.server.serf.wan: serf: Shutdown without a Leave === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.448Z [DEBUG] TestExecCommand_StreamResults: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.449Z [INFO] TestExecCommand_UploadDestroy.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:24.451Z [INFO] TestExecCommand_UploadDestroy: consul server down writer.go:29: 2021-01-29T19:33:24.451Z [INFO] TestExecCommand_UploadDestroy: shutdown complete writer.go:29: 2021-01-29T19:33:24.452Z [INFO] TestExecCommand_UploadDestroy: Stopping server: protocol=DNS address=127.0.0.1:11034 network=tcp writer.go:29: 2021-01-29T19:33:24.453Z [INFO] TestExecCommand_UploadDestroy: Stopping server: protocol=DNS address=127.0.0.1:11034 network=udp writer.go:29: 2021-01-29T19:33:24.454Z [INFO] TestExecCommand_UploadDestroy: Stopping server: protocol=HTTP address=127.0.0.1:11035 network=tcp === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.457Z [DEBUG] TestExecCommand_Sessions_Foreign.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:43280 latency=6.417117ms === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.459Z [DEBUG] TestExecCommand_Sessions.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:39160 latency=6.083568ms === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.466Z [INFO] TestExecCommand_StreamResults: Synced node info writer.go:29: 2021-01-29T19:33:24.468Z [DEBUG] TestExecCommand_StreamResults: Node info in sync === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.470Z [DEBUG] TestExecCommand_Sessions_Foreign.http: Request finished: method=GET url=/v1/session/info/b927a3c0-3a64-0920-b1fe-81defea406e2 from=127.0.0.1:43290 latency=67.822µs === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.470Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=PUT url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/foo/out/00001?acquire=2a4c9efd-02c0-80fb-fbab-641479ae3615 from=127.0.0.1:54450 latency=6.276855ms writer.go:29: 2021-01-29T19:33:24.472Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=GET url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/?index=15&keys=&wait=2000ms from=127.0.0.1:54430 latency=20.948593ms writer.go:29: 2021-01-29T19:33:24.475Z [DEBUG] TestExecCommand_StreamResults.http: Request finished: method=GET url=/v1/kv/_rexec/2a4c9efd-02c0-80fb-fbab-641479ae3615/foo/out/00001 from=127.0.0.1:54430 latency=68.086µs writer.go:29: 2021-01-29T19:33:24.476Z [INFO] TestExecCommand_StreamResults: Requesting shutdown writer.go:29: 2021-01-29T19:33:24.479Z [INFO] TestExecCommand_StreamResults.server: shutting down server === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.476Z [DEBUG] TestExecCommand_Sessions.http: Request finished: method=GET url=/v1/session/info/1cbab6f8-ed03-a16a-cb1c-bf78e51f4788 from=127.0.0.1:39164 latency=8.121466ms === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.480Z [DEBUG] TestExecCommand_StreamResults.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.482Z [WARN] TestExecCommand_StreamResults.server.serf.lan: serf: Shutdown without a Leave === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.484Z [DEBUG] TestExecCommand_Sessions_Foreign.http: Request finished: method=PUT url=/v1/session/destroy/b927a3c0-3a64-0920-b1fe-81defea406e2 from=127.0.0.1:43280 latency=9.540814ms === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.486Z [DEBUG] TestExecCommand_Sessions.http: Request finished: method=PUT url=/v1/session/destroy/1cbab6f8-ed03-a16a-cb1c-bf78e51f4788 from=127.0.0.1:39160 latency=2.237998ms === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.489Z [DEBUG] TestExecCommand_StreamResults.leader: stopped routine: routine="CA root pruning" === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.497Z [DEBUG] TestExecCommand_Sessions_Foreign.http: Request finished: method=GET url=/v1/session/info/b927a3c0-3a64-0920-b1fe-81defea406e2 from=127.0.0.1:43292 latency=93.186µs === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.496Z [WARN] TestExecCommand_StreamResults.server.serf.wan: serf: Shutdown without a Leave === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.498Z [INFO] TestExecCommand_Sessions_Foreign: Requesting shutdown writer.go:29: 2021-01-29T19:33:24.499Z [INFO] TestExecCommand_Sessions_Foreign.server: shutting down server writer.go:29: 2021-01-29T19:33:24.500Z [DEBUG] TestExecCommand_Sessions_Foreign.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:24.502Z [WARN] TestExecCommand_Sessions_Foreign.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.504Z [DEBUG] TestExecCommand_Sessions_Foreign.leader: stopped routine: routine="CA root pruning" === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.510Z [INFO] TestExecCommand_StreamResults.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:24.513Z [INFO] TestExecCommand_StreamResults: consul server down === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.511Z [DEBUG] TestExecCommand_Sessions.http: Request finished: method=GET url=/v1/session/info/1cbab6f8-ed03-a16a-cb1c-bf78e51f4788 from=127.0.0.1:39170 latency=390.736µs === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.519Z [INFO] TestExecCommand_StreamResults: shutdown complete === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.523Z [INFO] TestExecCommand_Sessions: Requesting shutdown writer.go:29: 2021-01-29T19:33:24.524Z [INFO] TestExecCommand_Sessions.server: shutting down server === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.520Z [WARN] TestExecCommand_Sessions_Foreign.server.serf.wan: serf: Shutdown without a Leave === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.527Z [DEBUG] TestExecCommand_Sessions.leader: stopping routine: routine="CA root pruning" === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.524Z [INFO] TestExecCommand_StreamResults: Stopping server: protocol=DNS address=127.0.0.1:11022 network=tcp writer.go:29: 2021-01-29T19:33:24.529Z [INFO] TestExecCommand_StreamResults: Stopping server: protocol=DNS address=127.0.0.1:11022 network=udp === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.529Z [WARN] TestExecCommand_Sessions.server.serf.lan: serf: Shutdown without a Leave === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:24.530Z [INFO] TestExecCommand_StreamResults: Stopping server: protocol=HTTP address=127.0.0.1:11023 network=tcp === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.527Z [INFO] TestExecCommand_Sessions_Foreign.server.router.manager: shutting down === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.531Z [DEBUG] TestExecCommand_Sessions.leader: stopped routine: routine="CA root pruning" === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:24.533Z [INFO] TestExecCommand_Sessions_Foreign: consul server down writer.go:29: 2021-01-29T19:33:24.537Z [INFO] TestExecCommand_Sessions_Foreign: shutdown complete writer.go:29: 2021-01-29T19:33:24.538Z [INFO] TestExecCommand_Sessions_Foreign: Stopping server: protocol=DNS address=127.0.0.1:11040 network=tcp writer.go:29: 2021-01-29T19:33:24.540Z [INFO] TestExecCommand_Sessions_Foreign: Stopping server: protocol=DNS address=127.0.0.1:11040 network=udp writer.go:29: 2021-01-29T19:33:24.541Z [INFO] TestExecCommand_Sessions_Foreign: Stopping server: protocol=HTTP address=127.0.0.1:11041 network=tcp === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:24.557Z [DEBUG] TestExecCommand_Sessions: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:24.561Z [DEBUG] TestExecCommand_Sessions: Node info in sync writer.go:29: 2021-01-29T19:33:24.560Z [WARN] TestExecCommand_Sessions.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:24.565Z [DEBUG] TestExecCommand_Sessions: Node info in sync writer.go:29: 2021-01-29T19:33:24.570Z [INFO] TestExecCommand_Sessions.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:24.570Z [INFO] TestExecCommand_Sessions: consul server down writer.go:29: 2021-01-29T19:33:24.573Z [INFO] TestExecCommand_Sessions: shutdown complete writer.go:29: 2021-01-29T19:33:24.576Z [INFO] TestExecCommand_Sessions: Stopping server: protocol=DNS address=127.0.0.1:11028 network=tcp writer.go:29: 2021-01-29T19:33:24.577Z [INFO] TestExecCommand_Sessions: Stopping server: protocol=DNS address=127.0.0.1:11028 network=udp writer.go:29: 2021-01-29T19:33:24.578Z [INFO] TestExecCommand_Sessions: Stopping server: protocol=HTTP address=127.0.0.1:11029 network=tcp === CONT TestExecCommand_UploadDestroy writer.go:29: 2021-01-29T19:33:24.961Z [INFO] TestExecCommand_UploadDestroy: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:24.962Z [INFO] TestExecCommand_UploadDestroy: Endpoints down --- PASS: TestExecCommand_UploadDestroy (1.07s) === CONT TestExecCommand_NoShell writer.go:29: 2021-01-29T19:33:24.975Z [WARN] TestExecCommand_NoShell: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:24.976Z [DEBUG] TestExecCommand_NoShell.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:24.978Z [DEBUG] TestExecCommand_NoShell.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:24.989Z [INFO] TestExecCommand_NoShell.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c12ff575-c730-1963-9393-0ebd50950ae6 Address:127.0.0.1:11051}]" writer.go:29: 2021-01-29T19:33:24.991Z [INFO] TestExecCommand_NoShell.server.raft: entering follower state: follower="Node at 127.0.0.1:11051 [Follower]" leader= writer.go:29: 2021-01-29T19:33:24.993Z [INFO] TestExecCommand_NoShell.server.serf.wan: serf: EventMemberJoin: Node-c12ff575-c730-1963-9393-0ebd50950ae6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.011Z [INFO] TestExecCommand_NoShell.server.serf.lan: serf: EventMemberJoin: Node-c12ff575-c730-1963-9393-0ebd50950ae6 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.013Z [INFO] TestExecCommand_NoShell.server: Adding LAN server: server="Node-c12ff575-c730-1963-9393-0ebd50950ae6 (Addr: tcp/127.0.0.1:11051) (DC: dc1)" writer.go:29: 2021-01-29T19:33:25.015Z [INFO] TestExecCommand_NoShell.server: Handled event for server in area: event=member-join server=Node-c12ff575-c730-1963-9393-0ebd50950ae6.dc1 area=wan writer.go:29: 2021-01-29T19:33:25.017Z [INFO] TestExecCommand_NoShell: Started DNS server: address=127.0.0.1:11046 network=tcp writer.go:29: 2021-01-29T19:33:25.018Z [INFO] TestExecCommand_NoShell: Started DNS server: address=127.0.0.1:11046 network=udp writer.go:29: 2021-01-29T19:33:25.019Z [INFO] TestExecCommand_NoShell: Started HTTP server: address=127.0.0.1:11047 network=tcp writer.go:29: 2021-01-29T19:33:25.020Z [INFO] TestExecCommand_NoShell: started state syncer writer.go:29: 2021-01-29T19:33:25.038Z [WARN] TestExecCommand_NoShell.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:25.039Z [INFO] TestExecCommand_NoShell.server.raft: entering candidate state: node="Node at 127.0.0.1:11051 [Candidate]" term=2 === CONT TestExecCommand_Sessions_Foreign writer.go:29: 2021-01-29T19:33:25.043Z [INFO] TestExecCommand_Sessions_Foreign: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:25.045Z [INFO] TestExecCommand_Sessions_Foreign: Endpoints down --- PASS: TestExecCommand_Sessions_Foreign (1.15s) === CONT TestExecCommand_Validate --- PASS: TestExecCommand_Validate (0.00s) === CONT TestExecCommand === CONT TestExecCommand_NoShell writer.go:29: 2021-01-29T19:33:25.042Z [DEBUG] TestExecCommand_NoShell.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:25.052Z [DEBUG] TestExecCommand_NoShell.server.raft: vote granted: from=c12ff575-c730-1963-9393-0ebd50950ae6 term=2 tally=1 writer.go:29: 2021-01-29T19:33:25.053Z [INFO] TestExecCommand_NoShell.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:25.054Z [INFO] TestExecCommand_NoShell.server.raft: entering leader state: leader="Node at 127.0.0.1:11051 [Leader]" writer.go:29: 2021-01-29T19:33:25.055Z [INFO] TestExecCommand_NoShell.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:25.056Z [INFO] TestExecCommand_NoShell.server: New leader elected: payload=Node-c12ff575-c730-1963-9393-0ebd50950ae6 writer.go:29: 2021-01-29T19:33:25.070Z [DEBUG] TestExecCommand_NoShell.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:11051 writer.go:29: 2021-01-29T19:33:25.075Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:25.078Z [INFO] TestExecCommand_NoShell.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:25.079Z [INFO] TestExecCommand_NoShell.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.080Z [DEBUG] TestExecCommand_NoShell.server: Skipping self join check for node since the cluster is too small: node=Node-c12ff575-c730-1963-9393-0ebd50950ae6 writer.go:29: 2021-01-29T19:33:25.081Z [INFO] TestExecCommand_NoShell.server: member joined, marking health alive: member=Node-c12ff575-c730-1963-9393-0ebd50950ae6 === CONT TestExecCommand_Sessions writer.go:29: 2021-01-29T19:33:25.082Z [INFO] TestExecCommand_Sessions: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:25.083Z [INFO] TestExecCommand_Sessions: Endpoints down --- PASS: TestExecCommand_Sessions (1.19s) === CONT TestExecCommand writer.go:29: 2021-01-29T19:33:25.097Z [WARN] TestExecCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:25.099Z [DEBUG] TestExecCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:25.100Z [DEBUG] TestExecCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:25.115Z [INFO] TestExecCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:07bc65a7-8623-ae31-42be-bac597329b50 Address:127.0.0.1:11057}]" writer.go:29: 2021-01-29T19:33:25.127Z [INFO] TestExecCommand.server.serf.wan: serf: EventMemberJoin: Node-07bc65a7-8623-ae31-42be-bac597329b50.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.127Z [INFO] TestExecCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:11057 [Follower]" leader= writer.go:29: 2021-01-29T19:33:25.128Z [INFO] TestExecCommand.server.serf.lan: serf: EventMemberJoin: Node-07bc65a7-8623-ae31-42be-bac597329b50 127.0.0.1 writer.go:29: 2021-01-29T19:33:25.135Z [INFO] TestExecCommand: Started DNS server: address=127.0.0.1:11052 network=udp writer.go:29: 2021-01-29T19:33:25.137Z [INFO] TestExecCommand: Started DNS server: address=127.0.0.1:11052 network=tcp writer.go:29: 2021-01-29T19:33:25.135Z [INFO] TestExecCommand.server: Handled event for server in area: event=member-join server=Node-07bc65a7-8623-ae31-42be-bac597329b50.dc1 area=wan writer.go:29: 2021-01-29T19:33:25.148Z [INFO] TestExecCommand: Started HTTP server: address=127.0.0.1:11053 network=tcp writer.go:29: 2021-01-29T19:33:25.150Z [INFO] TestExecCommand: started state syncer writer.go:29: 2021-01-29T19:33:25.155Z [INFO] TestExecCommand.server: Adding LAN server: server="Node-07bc65a7-8623-ae31-42be-bac597329b50 (Addr: tcp/127.0.0.1:11057) (DC: dc1)" writer.go:29: 2021-01-29T19:33:25.177Z [WARN] TestExecCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:25.178Z [INFO] TestExecCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:11057 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:25.180Z [DEBUG] TestExecCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:25.181Z [DEBUG] TestExecCommand.server.raft: vote granted: from=07bc65a7-8623-ae31-42be-bac597329b50 term=2 tally=1 writer.go:29: 2021-01-29T19:33:25.182Z [INFO] TestExecCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:25.182Z [INFO] TestExecCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:11057 [Leader]" writer.go:29: 2021-01-29T19:33:25.183Z [INFO] TestExecCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:25.185Z [DEBUG] TestExecCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:11057 writer.go:29: 2021-01-29T19:33:25.185Z [INFO] TestExecCommand.server: New leader elected: payload=Node-07bc65a7-8623-ae31-42be-bac597329b50 === CONT TestExecCommand_NoShell writer.go:29: 2021-01-29T19:33:25.190Z [DEBUG] TestExecCommand_NoShell: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:25.193Z [INFO] TestExecCommand_NoShell: Synced node info writer.go:29: 2021-01-29T19:33:25.195Z [DEBUG] TestExecCommand_NoShell: Node info in sync === CONT TestExecCommand writer.go:29: 2021-01-29T19:33:25.199Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:25.205Z [INFO] TestExecCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:25.206Z [INFO] TestExecCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:25.222Z [DEBUG] TestExecCommand.server: Skipping self join check for node since the cluster is too small: node=Node-07bc65a7-8623-ae31-42be-bac597329b50 writer.go:29: 2021-01-29T19:33:25.227Z [INFO] TestExecCommand.server: member joined, marking health alive: member=Node-07bc65a7-8623-ae31-42be-bac597329b50 writer.go:29: 2021-01-29T19:33:25.247Z [DEBUG] TestExecCommand.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:43704 latency=4.553687ms writer.go:29: 2021-01-29T19:33:25.256Z [DEBUG] TestExecCommand.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:43704 latency=1.210716ms writer.go:29: 2021-01-29T19:33:25.259Z [DEBUG] TestExecCommand.http: Request finished: method=PUT url=/v1/kv/_rexec/2784fa0b-5be5-99c1-8b55-c14845e1a2f3/job?acquire=2784fa0b-5be5-99c1-8b55-c14845e1a2f3 from=127.0.0.1:43704 latency=594.128µs === CONT TestExecCommand_NoShell writer.go:29: 2021-01-29T19:33:25.346Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:37560 latency=878.938µs writer.go:29: 2021-01-29T19:33:25.361Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:37560 latency=1.464399ms writer.go:29: 2021-01-29T19:33:25.364Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=PUT url=/v1/kv/_rexec/7ed83dd2-8cf6-874b-16c5-ba67d351a249/job?acquire=7ed83dd2-8cf6-874b-16c5-ba67d351a249 from=127.0.0.1:37560 latency=935.591µs === CONT TestExecCommand writer.go:29: 2021-01-29T19:33:25.435Z [DEBUG] TestExecCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:25.437Z [INFO] TestExecCommand: Synced node info writer.go:29: 2021-01-29T19:33:25.438Z [DEBUG] TestExecCommand: Node info in sync writer.go:29: 2021-01-29T19:33:25.461Z [DEBUG] TestExecCommand.server: User event: event=_rexec writer.go:29: 2021-01-29T19:33:25.461Z [DEBUG] TestExecCommand.http: Request finished: method=PUT url=/v1/event/fire/_rexec from=127.0.0.1:43704 latency=403.888µs writer.go:29: 2021-01-29T19:33:25.463Z [DEBUG] TestExecCommand: received remote exec event: id=8ad59e3b-cfdf-3907-ca6a-6f1ef93a3bc9 writer.go:29: 2021-01-29T19:33:25.465Z [DEBUG] TestExecCommand.http: Request finished: method=GET url=/v1/kv/_rexec/2784fa0b-5be5-99c1-8b55-c14845e1a2f3/?keys=&wait=1000ms from=127.0.0.1:43704 latency=150.58µs writer.go:29: 2021-01-29T19:33:25.468Z [INFO] TestExecCommand: remote exec script: script=uptime writer.go:29: 2021-01-29T19:33:25.469Z [DEBUG] TestExecCommand.http: Request finished: method=GET url=/v1/kv/_rexec/2784fa0b-5be5-99c1-8b55-c14845e1a2f3/?index=12&keys=&wait=1000ms from=127.0.0.1:43704 latency=158.245µs writer.go:29: 2021-01-29T19:33:25.511Z [DEBUG] TestExecCommand.http: Request finished: method=GET url=/v1/kv/_rexec/2784fa0b-5be5-99c1-8b55-c14845e1a2f3/?index=14&keys=&wait=1000ms from=127.0.0.1:43704 latency=37.700165ms writer.go:29: 2021-01-29T19:33:25.514Z [DEBUG] TestExecCommand.http: Request finished: method=GET url=/v1/kv/_rexec/2784fa0b-5be5-99c1-8b55-c14845e1a2f3/Node-07bc65a7-8623-ae31-42be-bac597329b50/out/00000 from=127.0.0.1:43704 latency=68.556µs writer.go:29: 2021-01-29T19:33:25.517Z [DEBUG] TestExecCommand.http: Request finished: method=GET url=/v1/kv/_rexec/2784fa0b-5be5-99c1-8b55-c14845e1a2f3/?index=15&keys=&wait=1000ms from=127.0.0.1:43704 latency=203.396µs writer.go:29: 2021-01-29T19:33:25.520Z [DEBUG] TestExecCommand.http: Request finished: method=GET url=/v1/kv/_rexec/2784fa0b-5be5-99c1-8b55-c14845e1a2f3/Node-07bc65a7-8623-ae31-42be-bac597329b50/exit from=127.0.0.1:43704 latency=162.716µs === CONT TestExecCommand_StreamResults writer.go:29: 2021-01-29T19:33:25.531Z [WARN] TestExecCommand_StreamResults: Timeout stopping server: protocol=HTTP address=127.0.0.1:11023 network=tcp writer.go:29: 2021-01-29T19:33:25.533Z [INFO] TestExecCommand_StreamResults: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:25.534Z [INFO] TestExecCommand_StreamResults: Endpoints down --- PASS: TestExecCommand_StreamResults (1.65s) === CONT TestExecCommand_NoShell writer.go:29: 2021-01-29T19:33:25.567Z [DEBUG] TestExecCommand_NoShell.server: User event: event=_rexec writer.go:29: 2021-01-29T19:33:25.567Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=PUT url=/v1/event/fire/_rexec from=127.0.0.1:37560 latency=594.039µs writer.go:29: 2021-01-29T19:33:25.569Z [DEBUG] TestExecCommand_NoShell: received remote exec event: id=745a5f18-b558-6d41-6b42-cd433b90c560 writer.go:29: 2021-01-29T19:33:25.574Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=GET url=/v1/kv/_rexec/7ed83dd2-8cf6-874b-16c5-ba67d351a249/?keys=&wait=1000ms from=127.0.0.1:37560 latency=1.694626ms writer.go:29: 2021-01-29T19:33:25.575Z [INFO] TestExecCommand_NoShell: remote exec script: script= writer.go:29: 2021-01-29T19:33:25.576Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=GET url=/v1/kv/_rexec/7ed83dd2-8cf6-874b-16c5-ba67d351a249/?index=13&keys=&wait=1000ms from=127.0.0.1:37560 latency=154.524µs writer.go:29: 2021-01-29T19:33:25.584Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=GET url=/v1/kv/_rexec/7ed83dd2-8cf6-874b-16c5-ba67d351a249/?index=14&keys=&wait=1000ms from=127.0.0.1:37560 latency=4.84329ms writer.go:29: 2021-01-29T19:33:25.587Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=GET url=/v1/kv/_rexec/7ed83dd2-8cf6-874b-16c5-ba67d351a249/Node-c12ff575-c730-1963-9393-0ebd50950ae6/out/00000 from=127.0.0.1:37560 latency=86.475µs writer.go:29: 2021-01-29T19:33:25.590Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=GET url=/v1/kv/_rexec/7ed83dd2-8cf6-874b-16c5-ba67d351a249/?index=15&keys=&wait=1000ms from=127.0.0.1:37560 latency=95.831µs writer.go:29: 2021-01-29T19:33:25.592Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=GET url=/v1/kv/_rexec/7ed83dd2-8cf6-874b-16c5-ba67d351a249/Node-c12ff575-c730-1963-9393-0ebd50950ae6/exit from=127.0.0.1:37560 latency=54.438µs writer.go:29: 2021-01-29T19:33:25.612Z [DEBUG] TestExecCommand_NoShell: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:25.613Z [DEBUG] TestExecCommand_NoShell: Node info in sync === CONT TestExecCommand writer.go:29: 2021-01-29T19:33:26.528Z [DEBUG] TestExecCommand.http: Request finished: method=PUT url=/v1/session/destroy/2784fa0b-5be5-99c1-8b55-c14845e1a2f3 from=127.0.0.1:43708 latency=1.075587ms writer.go:29: 2021-01-29T19:33:26.530Z [DEBUG] TestExecCommand.http: Request finished: method=GET url=/v1/kv/_rexec/2784fa0b-5be5-99c1-8b55-c14845e1a2f3/?index=16&keys=&wait=1000ms from=127.0.0.1:43704 latency=1.006866435s writer.go:29: 2021-01-29T19:33:26.530Z [DEBUG] TestExecCommand.http: Request finished: method=DELETE url=/v1/kv/_rexec/2784fa0b-5be5-99c1-8b55-c14845e1a2f3?recurse= from=127.0.0.1:43708 latency=651.488µs writer.go:29: 2021-01-29T19:33:26.534Z [DEBUG] TestExecCommand.http: Request finished: method=PUT url=/v1/session/destroy/2784fa0b-5be5-99c1-8b55-c14845e1a2f3 from=127.0.0.1:43708 latency=594.296µs writer.go:29: 2021-01-29T19:33:26.538Z [INFO] TestExecCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:26.540Z [INFO] TestExecCommand.server: shutting down server writer.go:29: 2021-01-29T19:33:26.541Z [DEBUG] TestExecCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.543Z [WARN] TestExecCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.543Z [DEBUG] TestExecCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.547Z [WARN] TestExecCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.550Z [INFO] TestExecCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:26.551Z [INFO] TestExecCommand: consul server down writer.go:29: 2021-01-29T19:33:26.555Z [INFO] TestExecCommand: shutdown complete writer.go:29: 2021-01-29T19:33:26.556Z [INFO] TestExecCommand: Stopping server: protocol=DNS address=127.0.0.1:11052 network=tcp writer.go:29: 2021-01-29T19:33:26.558Z [INFO] TestExecCommand: Stopping server: protocol=DNS address=127.0.0.1:11052 network=udp writer.go:29: 2021-01-29T19:33:26.559Z [INFO] TestExecCommand: Stopping server: protocol=HTTP address=127.0.0.1:11053 network=tcp === CONT TestExecCommand_NoShell writer.go:29: 2021-01-29T19:33:26.598Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=PUT url=/v1/session/destroy/7ed83dd2-8cf6-874b-16c5-ba67d351a249 from=127.0.0.1:37564 latency=777.833µs writer.go:29: 2021-01-29T19:33:26.600Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=GET url=/v1/kv/_rexec/7ed83dd2-8cf6-874b-16c5-ba67d351a249/?index=16&keys=&wait=1000ms from=127.0.0.1:37560 latency=1.005279641s writer.go:29: 2021-01-29T19:33:26.601Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=DELETE url=/v1/kv/_rexec/7ed83dd2-8cf6-874b-16c5-ba67d351a249?recurse= from=127.0.0.1:37564 latency=458.483µs writer.go:29: 2021-01-29T19:33:26.605Z [DEBUG] TestExecCommand_NoShell.http: Request finished: method=PUT url=/v1/session/destroy/7ed83dd2-8cf6-874b-16c5-ba67d351a249 from=127.0.0.1:37564 latency=495.753µs writer.go:29: 2021-01-29T19:33:26.606Z [INFO] TestExecCommand_NoShell: Requesting shutdown writer.go:29: 2021-01-29T19:33:26.608Z [INFO] TestExecCommand_NoShell.server: shutting down server writer.go:29: 2021-01-29T19:33:26.609Z [DEBUG] TestExecCommand_NoShell.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.610Z [WARN] TestExecCommand_NoShell.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.612Z [DEBUG] TestExecCommand_NoShell.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:26.613Z [WARN] TestExecCommand_NoShell.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:26.617Z [INFO] TestExecCommand_NoShell.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:26.618Z [INFO] TestExecCommand_NoShell: consul server down writer.go:29: 2021-01-29T19:33:26.621Z [INFO] TestExecCommand_NoShell: shutdown complete writer.go:29: 2021-01-29T19:33:26.622Z [INFO] TestExecCommand_NoShell: Stopping server: protocol=DNS address=127.0.0.1:11046 network=tcp writer.go:29: 2021-01-29T19:33:26.623Z [INFO] TestExecCommand_NoShell: Stopping server: protocol=DNS address=127.0.0.1:11046 network=udp writer.go:29: 2021-01-29T19:33:26.625Z [INFO] TestExecCommand_NoShell: Stopping server: protocol=HTTP address=127.0.0.1:11047 network=tcp === CONT TestExecCommand writer.go:29: 2021-01-29T19:33:27.061Z [INFO] TestExecCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:27.062Z [INFO] TestExecCommand: Endpoints down --- PASS: TestExecCommand (2.02s) === CONT TestExecCommand_NoShell writer.go:29: 2021-01-29T19:33:27.127Z [INFO] TestExecCommand_NoShell: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:27.128Z [INFO] TestExecCommand_NoShell: Endpoints down --- PASS: TestExecCommand_NoShell (2.17s) PASS ok github.com/hashicorp/consul/command/exec 3.508s === RUN TestConfigUtil_Values === PAUSE TestConfigUtil_Values === RUN TestConfigUtil_Visit === PAUSE TestConfigUtil_Visit === RUN TestFlagMapValueSet === PAUSE TestFlagMapValueSet === RUN TestAppendSliceValue_implements === PAUSE TestAppendSliceValue_implements === RUN TestAppendSliceValueSet === PAUSE TestAppendSliceValueSet === RUN TestHTTPFlagsSetToken --- PASS: TestHTTPFlagsSetToken (0.00s) === CONT TestConfigUtil_Values === CONT TestAppendSliceValueSet --- PASS: TestAppendSliceValueSet (0.00s) === CONT TestConfigUtil_Visit --- PASS: TestConfigUtil_Visit (0.00s) === CONT TestConfigUtil_Values config_test.go:85: (case 3) err: 1 error(s) decoding: * error decoding 'duration': time: invalid duration "nope" --- FAIL: TestConfigUtil_Values (0.00s) === CONT TestAppendSliceValue_implements --- PASS: TestAppendSliceValue_implements (0.00s) === CONT TestFlagMapValueSet --- PASS: TestFlagMapValueSet (0.00s) FAIL FAIL github.com/hashicorp/consul/command/flags 0.004s === RUN TestForceLeaveCommand_noTabs === PAUSE TestForceLeaveCommand_noTabs === RUN TestForceLeaveCommand === PAUSE TestForceLeaveCommand === RUN TestForceLeaveCommand_NoNodeWithName === PAUSE TestForceLeaveCommand_NoNodeWithName === RUN TestForceLeaveCommand_prune === PAUSE TestForceLeaveCommand_prune === RUN TestForceLeaveCommand_noAddrs === PAUSE TestForceLeaveCommand_noAddrs === CONT TestForceLeaveCommand_noTabs === CONT TestForceLeaveCommand_prune === CONT TestForceLeaveCommand_noAddrs [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestForceLeaveCommand_noTabs (0.00s) === CONT TestForceLeaveCommand --- PASS: TestForceLeaveCommand_noAddrs (0.00s) === CONT TestForceLeaveCommand_NoNodeWithName [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:33:29.580Z [WARN] TestForceLeaveCommand_NoNodeWithName: bootstrap = true: do not enable unless necessary === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.583Z [WARN] TestForceLeaveCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:29.585Z [DEBUG] TestForceLeaveCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:29.587Z [DEBUG] TestForceLeaveCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.583Z [DEBUG] TestForceLeaveCommand_NoNodeWithName.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:29.598Z [DEBUG] TestForceLeaveCommand_NoNodeWithName.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.606Z [INFO] TestForceLeaveCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a217b1d9-f8a7-9d51-ee54-88860362c0d6 Address:127.0.0.1:20228}]" writer.go:29: 2021-01-29T19:33:29.610Z [INFO] TestForceLeaveCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:20228 [Follower]" leader= === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.632Z [WARN] TestForceLeaveCommand_prune-a1: bootstrap = true: do not enable unless necessary === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.638Z [INFO] TestForceLeaveCommand_NoNodeWithName.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:796816b1-4176-5383-9ea8-d1a767ad7b1d Address:127.0.0.1:20222}]" === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.634Z [DEBUG] TestForceLeaveCommand_prune-a1.tlsutil: Update: version=1 === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.641Z [INFO] TestForceLeaveCommand_NoNodeWithName.server.raft: entering follower state: follower="Node at 127.0.0.1:20222 [Follower]" leader= === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.645Z [DEBUG] TestForceLeaveCommand_prune-a1.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.643Z [INFO] TestForceLeaveCommand_NoNodeWithName.server.serf.wan: serf: EventMemberJoin: Node-796816b1-4176-5383-9ea8-d1a767ad7b1d.dc1 127.0.0.1 === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.649Z [INFO] TestForceLeaveCommand_prune-a1.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:257ef726-5f48-6a4d-2824-18a501d62843 Address:127.0.0.1:20216}]" === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.649Z [INFO] TestForceLeaveCommand_NoNodeWithName.server.serf.lan: serf: EventMemberJoin: Node-796816b1-4176-5383-9ea8-d1a767ad7b1d 127.0.0.1 writer.go:29: 2021-01-29T19:33:29.653Z [INFO] TestForceLeaveCommand_NoNodeWithName: Started DNS server: address=127.0.0.1:20217 network=udp === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.652Z [INFO] TestForceLeaveCommand_prune-a1.server.raft: entering follower state: follower="Node at 127.0.0.1:20216 [Follower]" leader= === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.655Z [INFO] TestForceLeaveCommand_NoNodeWithName.server: Handled event for server in area: event=member-join server=Node-796816b1-4176-5383-9ea8-d1a767ad7b1d.dc1 area=wan === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.652Z [INFO] TestForceLeaveCommand_prune-a1.server.serf.wan: serf: EventMemberJoin: Node-257ef726-5f48-6a4d-2824-18a501d62843.dc1 127.0.0.1 === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.655Z [INFO] TestForceLeaveCommand_NoNodeWithName.server: Adding LAN server: server="Node-796816b1-4176-5383-9ea8-d1a767ad7b1d (Addr: tcp/127.0.0.1:20222) (DC: dc1)" writer.go:29: 2021-01-29T19:33:29.657Z [INFO] TestForceLeaveCommand_NoNodeWithName: Started DNS server: address=127.0.0.1:20217 network=tcp writer.go:29: 2021-01-29T19:33:29.659Z [INFO] TestForceLeaveCommand_NoNodeWithName: Started HTTP server: address=127.0.0.1:20218 network=tcp === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.659Z [INFO] TestForceLeaveCommand_prune-a1.server.serf.lan: serf: EventMemberJoin: Node-257ef726-5f48-6a4d-2824-18a501d62843 127.0.0.1 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.647Z [INFO] TestForceLeaveCommand.server.serf.wan: serf: EventMemberJoin: Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6.dc1 127.0.0.1 === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.666Z [INFO] TestForceLeaveCommand_prune-a1.server: Adding LAN server: server="Node-257ef726-5f48-6a4d-2824-18a501d62843 (Addr: tcp/127.0.0.1:20216) (DC: dc1)" === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.663Z [INFO] TestForceLeaveCommand_NoNodeWithName: started state syncer === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.668Z [INFO] TestForceLeaveCommand_prune-a1.server: Handled event for server in area: event=member-join server=Node-257ef726-5f48-6a4d-2824-18a501d62843.dc1 area=wan writer.go:29: 2021-01-29T19:33:29.669Z [INFO] TestForceLeaveCommand_prune-a1: Started DNS server: address=127.0.0.1:20211 network=tcp === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.671Z [INFO] TestForceLeaveCommand.server.serf.lan: serf: EventMemberJoin: Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6 127.0.0.1 === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.673Z [INFO] TestForceLeaveCommand_prune-a1: Started DNS server: address=127.0.0.1:20211 network=udp === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.682Z [INFO] TestForceLeaveCommand.server: Adding LAN server: server="Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6 (Addr: tcp/127.0.0.1:20228) (DC: dc1)" === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.682Z [INFO] TestForceLeaveCommand_prune-a1: Started HTTP server: address=127.0.0.1:20212 network=tcp === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.683Z [WARN] TestForceLeaveCommand.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.684Z [INFO] TestForceLeaveCommand_prune-a1: started state syncer === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.685Z [INFO] TestForceLeaveCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:20228 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:29.684Z [INFO] TestForceLeaveCommand.server: Handled event for server in area: event=member-join server=Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6.dc1 area=wan writer.go:29: 2021-01-29T19:33:29.689Z [INFO] TestForceLeaveCommand: Started DNS server: address=127.0.0.1:20223 network=tcp writer.go:29: 2021-01-29T19:33:29.691Z [DEBUG] TestForceLeaveCommand.server.raft: votes: needed=1 === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.693Z [WARN] TestForceLeaveCommand_NoNodeWithName.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.693Z [DEBUG] TestForceLeaveCommand.server.raft: vote granted: from=a217b1d9-f8a7-9d51-ee54-88860362c0d6 term=2 tally=1 === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.694Z [INFO] TestForceLeaveCommand_NoNodeWithName.server.raft: entering candidate state: node="Node at 127.0.0.1:20222 [Candidate]" term=2 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.695Z [INFO] TestForceLeaveCommand.server.raft: election won: tally=1 === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.697Z [DEBUG] TestForceLeaveCommand_NoNodeWithName.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:29.698Z [DEBUG] TestForceLeaveCommand_NoNodeWithName.server.raft: vote granted: from=796816b1-4176-5383-9ea8-d1a767ad7b1d term=2 tally=1 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.697Z [INFO] TestForceLeaveCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:20228 [Leader]" === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.700Z [INFO] TestForceLeaveCommand_NoNodeWithName.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:29.702Z [INFO] TestForceLeaveCommand_NoNodeWithName.server.raft: entering leader state: leader="Node at 127.0.0.1:20222 [Leader]" === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.699Z [INFO] TestForceLeaveCommand: Started DNS server: address=127.0.0.1:20223 network=udp === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.703Z [INFO] TestForceLeaveCommand_NoNodeWithName.server: cluster leadership acquired === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.703Z [INFO] TestForceLeaveCommand.server: cluster leadership acquired === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.705Z [INFO] TestForceLeaveCommand_NoNodeWithName.server: New leader elected: payload=Node-796816b1-4176-5383-9ea8-d1a767ad7b1d === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.704Z [INFO] TestForceLeaveCommand.server: New leader elected: payload=Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6 writer.go:29: 2021-01-29T19:33:29.705Z [DEBUG] TestForceLeaveCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:20228 === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.714Z [WARN] TestForceLeaveCommand_prune-a1.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.715Z [DEBUG] TestForceLeaveCommand_NoNodeWithName.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:20222 === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.716Z [INFO] TestForceLeaveCommand_prune-a1.server.raft: entering candidate state: node="Node at 127.0.0.1:20216 [Candidate]" term=2 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.707Z [INFO] TestForceLeaveCommand: Started HTTP server: address=127.0.0.1:20224 network=tcp === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.719Z [DEBUG] TestForceLeaveCommand_prune-a1.server.raft: votes: needed=1 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.719Z [INFO] TestForceLeaveCommand: started state syncer === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.721Z [DEBUG] TestForceLeaveCommand_prune-a1.server.raft: vote granted: from=257ef726-5f48-6a4d-2824-18a501d62843 term=2 tally=1 writer.go:29: 2021-01-29T19:33:29.722Z [INFO] TestForceLeaveCommand_prune-a1.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:29.724Z [INFO] TestForceLeaveCommand_prune-a1.server.raft: entering leader state: leader="Node at 127.0.0.1:20216 [Leader]" writer.go:29: 2021-01-29T19:33:29.738Z [INFO] TestForceLeaveCommand_prune-a1.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:29.741Z [INFO] TestForceLeaveCommand_prune-a1.server: New leader elected: payload=Node-257ef726-5f48-6a4d-2824-18a501d62843 === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.765Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.772Z [DEBUG] TestForceLeaveCommand_prune-a1.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:20216 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.811Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:29.813Z [INFO] TestForceLeaveCommand: Synced node info writer.go:29: 2021-01-29T19:33:29.814Z [DEBUG] TestForceLeaveCommand: Node info in sync === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.816Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.850Z [INFO] TestForceLeaveCommand_NoNodeWithName.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.838Z [WARN] TestForceLeaveCommand: bootstrap = true: do not enable unless necessary === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.852Z [INFO] TestForceLeaveCommand_NoNodeWithName.leader: started routine: routine="CA root pruning" === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.852Z [INFO] TestForceLeaveCommand_prune-a1.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.853Z [DEBUG] TestForceLeaveCommand_NoNodeWithName.server: Skipping self join check for node since the cluster is too small: node=Node-796816b1-4176-5383-9ea8-d1a767ad7b1d === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.853Z [INFO] TestForceLeaveCommand_prune-a1.leader: started routine: routine="CA root pruning" === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.852Z [DEBUG] TestForceLeaveCommand.tlsutil: Update: version=1 === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:29.855Z [INFO] TestForceLeaveCommand_NoNodeWithName.server: member joined, marking health alive: member=Node-796816b1-4176-5383-9ea8-d1a767ad7b1d === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:29.855Z [DEBUG] TestForceLeaveCommand_prune-a1.server: Skipping self join check for node since the cluster is too small: node=Node-257ef726-5f48-6a4d-2824-18a501d62843 writer.go:29: 2021-01-29T19:33:29.857Z [INFO] TestForceLeaveCommand_prune-a1.server: member joined, marking health alive: member=Node-257ef726-5f48-6a4d-2824-18a501d62843 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:29.856Z [DEBUG] TestForceLeaveCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:29.851Z [INFO] TestForceLeaveCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:29.861Z [INFO] TestForceLeaveCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b0a7184d-8c23-2869-3f4b-8809128c159b Address:127.0.0.1:20234}]" writer.go:29: 2021-01-29T19:33:29.861Z [INFO] TestForceLeaveCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:29.865Z [DEBUG] TestForceLeaveCommand.server: Skipping self join check for node since the cluster is too small: node=Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6 writer.go:29: 2021-01-29T19:33:29.864Z [INFO] TestForceLeaveCommand.server.serf.wan: serf: EventMemberJoin: Node-b0a7184d-8c23-2869-3f4b-8809128c159b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:29.866Z [INFO] TestForceLeaveCommand.server: member joined, marking health alive: member=Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6 writer.go:29: 2021-01-29T19:33:29.865Z [INFO] TestForceLeaveCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:20234 [Follower]" leader= writer.go:29: 2021-01-29T19:33:29.872Z [INFO] TestForceLeaveCommand.server.serf.lan: serf: EventMemberJoin: Node-b0a7184d-8c23-2869-3f4b-8809128c159b 127.0.0.1 writer.go:29: 2021-01-29T19:33:29.892Z [INFO] TestForceLeaveCommand.server: Adding LAN server: server="Node-b0a7184d-8c23-2869-3f4b-8809128c159b (Addr: tcp/127.0.0.1:20234) (DC: dc1)" writer.go:29: 2021-01-29T19:33:29.893Z [INFO] TestForceLeaveCommand.server: Handled event for server in area: event=member-join server=Node-b0a7184d-8c23-2869-3f4b-8809128c159b.dc1 area=wan writer.go:29: 2021-01-29T19:33:29.894Z [INFO] TestForceLeaveCommand: Started DNS server: address=127.0.0.1:20229 network=udp writer.go:29: 2021-01-29T19:33:29.895Z [INFO] TestForceLeaveCommand: Started DNS server: address=127.0.0.1:20229 network=tcp writer.go:29: 2021-01-29T19:33:29.897Z [INFO] TestForceLeaveCommand: Started HTTP server: address=127.0.0.1:20230 network=tcp writer.go:29: 2021-01-29T19:33:29.898Z [INFO] TestForceLeaveCommand: started state syncer writer.go:29: 2021-01-29T19:33:29.911Z [WARN] TestForceLeaveCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:29.912Z [INFO] TestForceLeaveCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:20234 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:29.913Z [DEBUG] TestForceLeaveCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:29.914Z [DEBUG] TestForceLeaveCommand.server.raft: vote granted: from=b0a7184d-8c23-2869-3f4b-8809128c159b term=2 tally=1 writer.go:29: 2021-01-29T19:33:29.915Z [INFO] TestForceLeaveCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:29.916Z [INFO] TestForceLeaveCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:20234 [Leader]" writer.go:29: 2021-01-29T19:33:29.917Z [INFO] TestForceLeaveCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:29.918Z [INFO] TestForceLeaveCommand.server: New leader elected: payload=Node-b0a7184d-8c23-2869-3f4b-8809128c159b writer.go:29: 2021-01-29T19:33:29.920Z [DEBUG] TestForceLeaveCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:20234 writer.go:29: 2021-01-29T19:33:29.929Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:29.977Z [INFO] TestForceLeaveCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:29.979Z [INFO] TestForceLeaveCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:29.980Z [DEBUG] TestForceLeaveCommand.server: Skipping self join check for node since the cluster is too small: node=Node-b0a7184d-8c23-2869-3f4b-8809128c159b writer.go:29: 2021-01-29T19:33:29.982Z [INFO] TestForceLeaveCommand.server: member joined, marking health alive: member=Node-b0a7184d-8c23-2869-3f4b-8809128c159b === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.001Z [DEBUG] TestForceLeaveCommand_prune-a1: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:30.004Z [INFO] TestForceLeaveCommand_prune-a1: Synced node info === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:30.039Z [DEBUG] TestForceLeaveCommand_NoNodeWithName: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:30.044Z [INFO] TestForceLeaveCommand_NoNodeWithName: Synced node info writer.go:29: 2021-01-29T19:33:30.064Z [INFO] TestForceLeaveCommand_NoNodeWithName: Force leaving node: node=garbage-name writer.go:29: 2021-01-29T19:33:30.066Z [ERROR] TestForceLeaveCommand_NoNodeWithName.http: Request error: method=PUT url=/v1/agent/force-leave/garbage-name from=127.0.0.1:51866 error="agent: No node found with name 'garbage-name'" writer.go:29: 2021-01-29T19:33:30.072Z [DEBUG] TestForceLeaveCommand_NoNodeWithName.http: Request finished: method=PUT url=/v1/agent/force-leave/garbage-name from=127.0.0.1:51866 latency=7.925947ms writer.go:29: 2021-01-29T19:33:30.083Z [INFO] TestForceLeaveCommand_NoNodeWithName: Requesting shutdown writer.go:29: 2021-01-29T19:33:30.092Z [INFO] TestForceLeaveCommand_NoNodeWithName.server: shutting down server writer.go:29: 2021-01-29T19:33:30.094Z [DEBUG] TestForceLeaveCommand_NoNodeWithName.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.100Z [WARN] TestForceLeaveCommand_NoNodeWithName.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:30.112Z [DEBUG] TestForceLeaveCommand_NoNodeWithName.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.143Z [WARN] TestForceLeaveCommand_NoNodeWithName.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:30.162Z [INFO] TestForceLeaveCommand_NoNodeWithName.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:30.164Z [INFO] TestForceLeaveCommand_NoNodeWithName: consul server down writer.go:29: 2021-01-29T19:33:30.166Z [INFO] TestForceLeaveCommand_NoNodeWithName: shutdown complete writer.go:29: 2021-01-29T19:33:30.168Z [INFO] TestForceLeaveCommand_NoNodeWithName: Stopping server: protocol=DNS address=127.0.0.1:20217 network=tcp writer.go:29: 2021-01-29T19:33:30.169Z [INFO] TestForceLeaveCommand_NoNodeWithName: Stopping server: protocol=DNS address=127.0.0.1:20217 network=udp === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.176Z [WARN] TestForceLeaveCommand_prune-a2: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:30.178Z [DEBUG] TestForceLeaveCommand_prune-a2.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:30.182Z [DEBUG] TestForceLeaveCommand_prune-a2.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:30.171Z [INFO] TestForceLeaveCommand_NoNodeWithName: Stopping server: protocol=HTTP address=127.0.0.1:20218 network=tcp === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.202Z [INFO] TestForceLeaveCommand_prune-a2.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 Address:127.0.0.1:20240}]" writer.go:29: 2021-01-29T19:33:30.205Z [INFO] TestForceLeaveCommand_prune-a2.server.serf.wan: serf: EventMemberJoin: Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.210Z [INFO] TestForceLeaveCommand_prune-a2.server.serf.lan: serf: EventMemberJoin: Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.218Z [INFO] TestForceLeaveCommand_prune-a2: Started DNS server: address=127.0.0.1:20235 network=udp writer.go:29: 2021-01-29T19:33:30.214Z [INFO] TestForceLeaveCommand_prune-a2.server.raft: entering follower state: follower="Node at 127.0.0.1:20240 [Follower]" leader= writer.go:29: 2021-01-29T19:33:30.263Z [INFO] TestForceLeaveCommand_prune-a2.server: Handled event for server in area: event=member-join server=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6.dc1 area=wan writer.go:29: 2021-01-29T19:33:30.263Z [INFO] TestForceLeaveCommand_prune-a2.server: Adding LAN server: server="Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 (Addr: tcp/127.0.0.1:20240) (DC: dc1)" writer.go:29: 2021-01-29T19:33:30.272Z [INFO] TestForceLeaveCommand_prune-a2: Started DNS server: address=127.0.0.1:20235 network=tcp writer.go:29: 2021-01-29T19:33:30.277Z [INFO] TestForceLeaveCommand_prune-a2: Started HTTP server: address=127.0.0.1:20236 network=tcp === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.276Z [INFO] TestForceLeaveCommand: (LAN) joining: lan_addresses=[127.0.0.1:20226] === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.278Z [INFO] TestForceLeaveCommand_prune-a2: started state syncer === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.279Z [DEBUG] TestForceLeaveCommand.server.memberlist.lan: memberlist: Initiating push/pull sync with: 127.0.0.1:20226 writer.go:29: 2021-01-29T19:33:30.280Z [DEBUG] TestForceLeaveCommand.server.memberlist.lan: memberlist: Stream connection from=127.0.0.1:35538 writer.go:29: 2021-01-29T19:33:30.284Z [INFO] TestForceLeaveCommand.server.serf.lan: serf: EventMemberJoin: Node-b0a7184d-8c23-2869-3f4b-8809128c159b 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.285Z [INFO] TestForceLeaveCommand.server.serf.lan: serf: EventMemberJoin: Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.288Z [INFO] TestForceLeaveCommand: (LAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:33:30.286Z [INFO] TestForceLeaveCommand.server: Adding LAN server: server="Node-b0a7184d-8c23-2869-3f4b-8809128c159b (Addr: tcp/127.0.0.1:20234) (DC: dc1)" writer.go:29: 2021-01-29T19:33:30.291Z [INFO] TestForceLeaveCommand.server: New leader elected: payload=Node-b0a7184d-8c23-2869-3f4b-8809128c159b writer.go:29: 2021-01-29T19:33:30.295Z [DEBUG] TestForceLeaveCommand: systemd notify failed: error="No socket" writer.go:29: 2021-01-29T19:33:30.295Z [DEBUG] TestForceLeaveCommand.server.serf.lan: serf: messageUserEventType: consul:new-leader writer.go:29: 2021-01-29T19:33:30.301Z [DEBUG] TestForceLeaveCommand.server.serf.lan: serf: messageJoinType: Node-b0a7184d-8c23-2869-3f4b-8809128c159b writer.go:29: 2021-01-29T19:33:30.296Z [ERROR] TestForceLeaveCommand.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-b0a7184d-8c23-2869-3f4b-8809128c159b other=Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6 writer.go:29: 2021-01-29T19:33:30.303Z [INFO] TestForceLeaveCommand.server: member joined, marking health alive: member=Node-b0a7184d-8c23-2869-3f4b-8809128c159b writer.go:29: 2021-01-29T19:33:30.289Z [INFO] TestForceLeaveCommand.server: Adding LAN server: server="Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6 (Addr: tcp/127.0.0.1:20228) (DC: dc1)" writer.go:29: 2021-01-29T19:33:30.301Z [WARN] TestForceLeaveCommand.server.memberlist.wan: memberlist: Failed to resolve Node-b0a7184d-8c23-2869-3f4b-8809128c159b.dc1/127.0.0.1:20233: lookup Node-b0a7184d-8c23-2869-3f4b-8809128c159b.dc1/127.0.0.1: no such host === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.317Z [WARN] TestForceLeaveCommand_prune-a2.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:30.319Z [INFO] TestForceLeaveCommand_prune-a2.server.raft: entering candidate state: node="Node at 127.0.0.1:20240 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:30.322Z [DEBUG] TestForceLeaveCommand_prune-a2.server.raft: votes: needed=1 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.307Z [ERROR] TestForceLeaveCommand.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6 other=Node-b0a7184d-8c23-2869-3f4b-8809128c159b writer.go:29: 2021-01-29T19:33:30.310Z [DEBUG] TestForceLeaveCommand.server: Failed to flood-join server at address: server=Node-b0a7184d-8c23-2869-3f4b-8809128c159b.dc1 address=127.0.0.1:20233 error="1 error occurred: * Failed to resolve Node-b0a7184d-8c23-2869-3f4b-8809128c159b.dc1/127.0.0.1:20233: lookup Node-b0a7184d-8c23-2869-3f4b-8809128c159b.dc1/127.0.0.1: no such host " === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.327Z [DEBUG] TestForceLeaveCommand_prune-a2.server.raft: vote granted: from=df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 term=2 tally=1 writer.go:29: 2021-01-29T19:33:30.331Z [INFO] TestForceLeaveCommand_prune-a2.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:30.337Z [INFO] TestForceLeaveCommand_prune-a2.server.raft: entering leader state: leader="Node at 127.0.0.1:20240 [Leader]" writer.go:29: 2021-01-29T19:33:30.343Z [INFO] TestForceLeaveCommand_prune-a2.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:30.343Z [INFO] TestForceLeaveCommand_prune-a2.server: New leader elected: payload=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 writer.go:29: 2021-01-29T19:33:30.344Z [DEBUG] TestForceLeaveCommand_prune-a2.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:20240 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.300Z [INFO] TestForceLeaveCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:30.308Z [DEBUG] TestForceLeaveCommand: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.349Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.311Z [WARN] TestForceLeaveCommand.server.memberlist.wan: memberlist: Failed to resolve Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6.dc1/127.0.0.1:20227: lookup Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6.dc1/127.0.0.1: no such host writer.go:29: 2021-01-29T19:33:30.342Z [INFO] TestForceLeaveCommand.server: member joined, marking health alive: member=Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6 writer.go:29: 2021-01-29T19:33:30.347Z [INFO] TestForceLeaveCommand.server: shutting down server writer.go:29: 2021-01-29T19:33:30.352Z [DEBUG] TestForceLeaveCommand.server: Failed to flood-join server at address: server=Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6.dc1 address=127.0.0.1:20227 error="1 error occurred: * Failed to resolve Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6.dc1/127.0.0.1:20227: lookup Node-a217b1d9-f8a7-9d51-ee54-88860362c0d6.dc1/127.0.0.1: no such host " === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.357Z [INFO] TestForceLeaveCommand_prune-a2.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.356Z [DEBUG] TestForceLeaveCommand.leader: stopping routine: routine="CA root pruning" === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.358Z [INFO] TestForceLeaveCommand_prune-a2.leader: started routine: routine="CA root pruning" === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.359Z [WARN] TestForceLeaveCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:30.359Z [DEBUG] TestForceLeaveCommand.leader: stopped routine: routine="CA root pruning" === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.360Z [DEBUG] TestForceLeaveCommand_prune-a2.server: Skipping self join check for node since the cluster is too small: node=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.365Z [WARN] TestForceLeaveCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:30.367Z [INFO] TestForceLeaveCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:30.368Z [INFO] TestForceLeaveCommand: Synced node info writer.go:29: 2021-01-29T19:33:30.368Z [INFO] TestForceLeaveCommand: consul server down === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.365Z [INFO] TestForceLeaveCommand_prune-a2.server: member joined, marking health alive: member=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.370Z [INFO] TestForceLeaveCommand: shutdown complete writer.go:29: 2021-01-29T19:33:30.371Z [INFO] TestForceLeaveCommand: Stopping server: protocol=DNS address=127.0.0.1:20229 network=tcp writer.go:29: 2021-01-29T19:33:30.370Z [DEBUG] TestForceLeaveCommand: Node info in sync writer.go:29: 2021-01-29T19:33:30.382Z [INFO] TestForceLeaveCommand: Stopping server: protocol=DNS address=127.0.0.1:20229 network=udp writer.go:29: 2021-01-29T19:33:30.384Z [INFO] TestForceLeaveCommand: Stopping server: protocol=HTTP address=127.0.0.1:20230 network=tcp === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.664Z [INFO] TestForceLeaveCommand_prune-a2: (LAN) joining: lan_addresses=[127.0.0.1:20214] writer.go:29: 2021-01-29T19:33:30.668Z [DEBUG] TestForceLeaveCommand_prune-a2.server.memberlist.lan: memberlist: Initiating push/pull sync with: 127.0.0.1:20214 writer.go:29: 2021-01-29T19:33:30.668Z [DEBUG] TestForceLeaveCommand_prune-a1.server.memberlist.lan: memberlist: Stream connection from=127.0.0.1:43868 writer.go:29: 2021-01-29T19:33:30.672Z [INFO] TestForceLeaveCommand_prune-a1.server.serf.lan: serf: EventMemberJoin: Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.674Z [INFO] TestForceLeaveCommand_prune-a1.server: Adding LAN server: server="Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 (Addr: tcp/127.0.0.1:20240) (DC: dc1)" writer.go:29: 2021-01-29T19:33:30.675Z [INFO] TestForceLeaveCommand_prune-a1.server: New leader elected: payload=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 writer.go:29: 2021-01-29T19:33:30.672Z [INFO] TestForceLeaveCommand_prune-a2.server.serf.lan: serf: EventMemberJoin: Node-257ef726-5f48-6a4d-2824-18a501d62843 127.0.0.1 writer.go:29: 2021-01-29T19:33:30.675Z [ERROR] TestForceLeaveCommand_prune-a1.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 other=Node-257ef726-5f48-6a4d-2824-18a501d62843 writer.go:29: 2021-01-29T19:33:30.684Z [INFO] TestForceLeaveCommand_prune-a1.server: member joined, marking health alive: member=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 writer.go:29: 2021-01-29T19:33:30.682Z [INFO] TestForceLeaveCommand_prune-a2: (LAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:33:30.686Z [DEBUG] TestForceLeaveCommand_prune-a2: systemd notify failed: error="No socket" writer.go:29: 2021-01-29T19:33:30.687Z [INFO] TestForceLeaveCommand_prune-a2: Requesting shutdown writer.go:29: 2021-01-29T19:33:30.683Z [INFO] TestForceLeaveCommand_prune-a2.server: Adding LAN server: server="Node-257ef726-5f48-6a4d-2824-18a501d62843 (Addr: tcp/127.0.0.1:20216) (DC: dc1)" writer.go:29: 2021-01-29T19:33:30.689Z [INFO] TestForceLeaveCommand_prune-a2.server: shutting down server writer.go:29: 2021-01-29T19:33:30.690Z [WARN] TestForceLeaveCommand_prune-a1.server.memberlist.wan: memberlist: Failed to resolve Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6.dc1/127.0.0.1:20239: lookup Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6.dc1/127.0.0.1: no such host writer.go:29: 2021-01-29T19:33:30.694Z [DEBUG] TestForceLeaveCommand_prune-a1.server: Failed to flood-join server at address: server=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6.dc1 address=127.0.0.1:20239 error="1 error occurred: * Failed to resolve Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6.dc1/127.0.0.1:20239: lookup Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6.dc1/127.0.0.1: no such host " writer.go:29: 2021-01-29T19:33:30.692Z [DEBUG] TestForceLeaveCommand_prune-a2.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.690Z [ERROR] TestForceLeaveCommand_prune-a2.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-257ef726-5f48-6a4d-2824-18a501d62843 other=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 writer.go:29: 2021-01-29T19:33:30.692Z [ERROR] TestForceLeaveCommand_prune-a2.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestForceLeaveCommand_NoNodeWithName writer.go:29: 2021-01-29T19:33:30.703Z [INFO] TestForceLeaveCommand_NoNodeWithName: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:30.704Z [INFO] TestForceLeaveCommand_NoNodeWithName: Endpoints down --- PASS: TestForceLeaveCommand_NoNodeWithName (1.18s) === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.694Z [WARN] TestForceLeaveCommand_prune-a2.server.memberlist.wan: memberlist: Failed to resolve Node-257ef726-5f48-6a4d-2824-18a501d62843.dc1/127.0.0.1:20215: lookup Node-257ef726-5f48-6a4d-2824-18a501d62843.dc1/127.0.0.1: no such host writer.go:29: 2021-01-29T19:33:30.706Z [DEBUG] TestForceLeaveCommand_prune-a2.server: Failed to flood-join server at address: server=Node-257ef726-5f48-6a4d-2824-18a501d62843.dc1 address=127.0.0.1:20215 error="1 error occurred: * Failed to resolve Node-257ef726-5f48-6a4d-2824-18a501d62843.dc1/127.0.0.1:20215: lookup Node-257ef726-5f48-6a4d-2824-18a501d62843.dc1/127.0.0.1: no such host " writer.go:29: 2021-01-29T19:33:30.702Z [INFO] TestForceLeaveCommand_prune-a2.server: member joined, marking health alive: member=Node-257ef726-5f48-6a4d-2824-18a501d62843 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.707Z [DEBUG] TestForceLeaveCommand: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.702Z [WARN] TestForceLeaveCommand_prune-a2.server.serf.lan: serf: Shutdown without a Leave === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.710Z [DEBUG] TestForceLeaveCommand: Node info in sync === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:30.702Z [DEBUG] TestForceLeaveCommand_prune-a2.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:30.712Z [WARN] TestForceLeaveCommand_prune-a2.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:30.716Z [INFO] TestForceLeaveCommand_prune-a2.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:30.716Z [INFO] TestForceLeaveCommand_prune-a2: consul server down writer.go:29: 2021-01-29T19:33:30.719Z [INFO] TestForceLeaveCommand_prune-a2: shutdown complete writer.go:29: 2021-01-29T19:33:30.721Z [INFO] TestForceLeaveCommand_prune-a2: Stopping server: protocol=DNS address=127.0.0.1:20235 network=tcp writer.go:29: 2021-01-29T19:33:30.722Z [INFO] TestForceLeaveCommand_prune-a2: Stopping server: protocol=DNS address=127.0.0.1:20235 network=udp writer.go:29: 2021-01-29T19:33:30.723Z [INFO] TestForceLeaveCommand_prune-a2: Stopping server: protocol=HTTP address=127.0.0.1:20236 network=tcp === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:30.886Z [INFO] TestForceLeaveCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:30.888Z [INFO] TestForceLeaveCommand: Endpoints down writer.go:29: 2021-01-29T19:33:30.894Z [INFO] TestForceLeaveCommand: Force leaving node: node=Node-b0a7184d-8c23-2869-3f4b-8809128c159b writer.go:29: 2021-01-29T19:33:30.896Z [DEBUG] TestForceLeaveCommand.http: Request finished: method=PUT url=/v1/agent/force-leave/Node-b0a7184d-8c23-2869-3f4b-8809128c159b from=127.0.0.1:50264 latency=2.431254ms writer.go:29: 2021-01-29T19:33:31.181Z [DEBUG] TestForceLeaveCommand.server.memberlist.lan: memberlist: Failed ping: Node-b0a7184d-8c23-2869-3f4b-8809128c159b (timeout reached) === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:31.224Z [INFO] TestForceLeaveCommand_prune-a2: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:31.226Z [INFO] TestForceLeaveCommand_prune-a2: Endpoints down writer.go:29: 2021-01-29T19:33:31.232Z [INFO] TestForceLeaveCommand_prune-a1: Force leaving node: node=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 === CONT TestForceLeaveCommand writer.go:29: 2021-01-29T19:33:31.680Z [INFO] TestForceLeaveCommand.server.memberlist.lan: memberlist: Suspect Node-b0a7184d-8c23-2869-3f4b-8809128c159b has failed, no acks received writer.go:29: 2021-01-29T19:33:31.722Z [DEBUG] TestForceLeaveCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:31.723Z [DEBUG] TestForceLeaveCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:31.724Z [WARN] TestForceLeaveCommand: error getting server health from server: server=Node-b0a7184d-8c23-2869-3f4b-8809128c159b error="rpc error getting client: failed to get conn: dial tcp 127.0.0.1:0->127.0.0.1:20234: connect: connection refused" writer.go:29: 2021-01-29T19:33:32.722Z [WARN] TestForceLeaveCommand: error getting server health from server: server=Node-b0a7184d-8c23-2869-3f4b-8809128c159b error="context deadline exceeded" writer.go:29: 2021-01-29T19:33:33.179Z [DEBUG] TestForceLeaveCommand.server.memberlist.lan: memberlist: Failed ping: Node-b0a7184d-8c23-2869-3f4b-8809128c159b (timeout reached) writer.go:29: 2021-01-29T19:33:33.335Z [DEBUG] TestForceLeaveCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:33.337Z [DEBUG] TestForceLeaveCommand: Node info in sync writer.go:29: 2021-01-29T19:33:33.722Z [DEBUG] TestForceLeaveCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:33.725Z [WARN] TestForceLeaveCommand: error getting server health from server: server=Node-b0a7184d-8c23-2869-3f4b-8809128c159b error="rpc error getting client: failed to get conn: dial tcp 127.0.0.1:0->127.0.0.1:20234: connect: connection refused" writer.go:29: 2021-01-29T19:33:34.679Z [INFO] TestForceLeaveCommand.server.memberlist.lan: memberlist: Suspect Node-b0a7184d-8c23-2869-3f4b-8809128c159b has failed, no acks received writer.go:29: 2021-01-29T19:33:34.721Z [WARN] TestForceLeaveCommand: error getting server health from server: server=Node-b0a7184d-8c23-2869-3f4b-8809128c159b error="context deadline exceeded" writer.go:29: 2021-01-29T19:33:35.683Z [INFO] TestForceLeaveCommand.server.memberlist.lan: memberlist: Marking Node-b0a7184d-8c23-2869-3f4b-8809128c159b as failed, suspect timeout reached (0 peer confirmations) writer.go:29: 2021-01-29T19:33:35.686Z [INFO] TestForceLeaveCommand.server.serf.lan: serf: EventMemberLeave: Node-b0a7184d-8c23-2869-3f4b-8809128c159b 127.0.0.1 writer.go:29: 2021-01-29T19:33:35.692Z [INFO] TestForceLeaveCommand.server: Removing LAN server: server="Node-b0a7184d-8c23-2869-3f4b-8809128c159b (Addr: tcp/127.0.0.1:20234) (DC: dc1)" writer.go:29: 2021-01-29T19:33:35.694Z [INFO] TestForceLeaveCommand.server: deregistering member: member=Node-b0a7184d-8c23-2869-3f4b-8809128c159b reason=left writer.go:29: 2021-01-29T19:33:35.711Z [INFO] TestForceLeaveCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:35.714Z [INFO] TestForceLeaveCommand.server: shutting down server writer.go:29: 2021-01-29T19:33:35.716Z [DEBUG] TestForceLeaveCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.718Z [WARN] TestForceLeaveCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:35.718Z [DEBUG] TestForceLeaveCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:35.722Z [WARN] TestForceLeaveCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:35.724Z [INFO] TestForceLeaveCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:35.725Z [INFO] TestForceLeaveCommand: consul server down writer.go:29: 2021-01-29T19:33:35.727Z [INFO] TestForceLeaveCommand: shutdown complete writer.go:29: 2021-01-29T19:33:35.728Z [INFO] TestForceLeaveCommand: Stopping server: protocol=DNS address=127.0.0.1:20223 network=tcp writer.go:29: 2021-01-29T19:33:35.729Z [INFO] TestForceLeaveCommand: Stopping server: protocol=DNS address=127.0.0.1:20223 network=udp writer.go:29: 2021-01-29T19:33:35.730Z [INFO] TestForceLeaveCommand: Stopping server: protocol=HTTP address=127.0.0.1:20224 network=tcp writer.go:29: 2021-01-29T19:33:36.180Z [DEBUG] TestForceLeaveCommand.server.memberlist.lan: memberlist: Failed ping: Node-b0a7184d-8c23-2869-3f4b-8809128c159b (timeout reached) writer.go:29: 2021-01-29T19:33:36.232Z [INFO] TestForceLeaveCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:36.233Z [INFO] TestForceLeaveCommand: Endpoints down --- PASS: TestForceLeaveCommand (6.71s) === CONT TestForceLeaveCommand_prune writer.go:29: 2021-01-29T19:33:39.236Z [INFO] TestForceLeaveCommand_prune-a1.server.serf.lan: serf: EventMemberReap (forced): Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.238Z [DEBUG] TestForceLeaveCommand_prune-a1.http: Request finished: method=PUT url=/v1/agent/force-leave/Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6?prune=1 from=127.0.0.1:44748 latency=8.006252448s writer.go:29: 2021-01-29T19:33:39.239Z [INFO] TestForceLeaveCommand_prune-a1.server: Removing LAN server: server="Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 (Addr: tcp/127.0.0.1:20240) (DC: dc1)" writer.go:29: 2021-01-29T19:33:39.240Z [INFO] TestForceLeaveCommand_prune-a1.server: deregistering member: member=Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 reason=reaped writer.go:29: 2021-01-29T19:33:39.241Z [DEBUG] TestForceLeaveCommand_prune-a1.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:39.240Z [INFO] TestForceLeaveCommand_prune-a1: Requesting shutdown writer.go:29: 2021-01-29T19:33:39.246Z [INFO] TestForceLeaveCommand_prune-a1.server: shutting down server writer.go:29: 2021-01-29T19:33:39.242Z [DEBUG] TestForceLeaveCommand_prune-a1: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:39.243Z [INFO] TestForceLeaveCommand_prune-a1.server.memberlist.lan: memberlist: Suspect Node-df0aaf25-9d5c-f1b2-7e28-df52c6a5fdf6 has failed, no acks received writer.go:29: 2021-01-29T19:33:39.248Z [DEBUG] TestForceLeaveCommand_prune-a1.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.250Z [DEBUG] TestForceLeaveCommand_prune-a1: Node info in sync writer.go:29: 2021-01-29T19:33:39.255Z [WARN] TestForceLeaveCommand_prune-a1.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:39.255Z [DEBUG] TestForceLeaveCommand_prune-a1.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.256Z [WARN] TestForceLeaveCommand_prune-a1: error getting server health from server: server=Node-257ef726-5f48-6a4d-2824-18a501d62843 error="rpc error making call: EOF" writer.go:29: 2021-01-29T19:33:39.272Z [DEBUG] TestForceLeaveCommand_prune-a1: Node info in sync writer.go:29: 2021-01-29T19:33:39.276Z [WARN] TestForceLeaveCommand_prune-a1.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:39.281Z [INFO] TestForceLeaveCommand_prune-a1.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:39.281Z [INFO] TestForceLeaveCommand_prune-a1: consul server down writer.go:29: 2021-01-29T19:33:39.283Z [INFO] TestForceLeaveCommand_prune-a1: shutdown complete writer.go:29: 2021-01-29T19:33:39.284Z [INFO] TestForceLeaveCommand_prune-a1: Stopping server: protocol=DNS address=127.0.0.1:20211 network=tcp writer.go:29: 2021-01-29T19:33:39.285Z [INFO] TestForceLeaveCommand_prune-a1: Stopping server: protocol=DNS address=127.0.0.1:20211 network=udp writer.go:29: 2021-01-29T19:33:39.286Z [INFO] TestForceLeaveCommand_prune-a1: Stopping server: protocol=HTTP address=127.0.0.1:20212 network=tcp writer.go:29: 2021-01-29T19:33:39.787Z [INFO] TestForceLeaveCommand_prune-a1: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:39.788Z [INFO] TestForceLeaveCommand_prune-a1: Endpoints down --- PASS: TestForceLeaveCommand_prune (10.27s) PASS ok github.com/hashicorp/consul/command/forceleave 10.559s ? github.com/hashicorp/consul/command/helpers [no test files] === RUN TestInfoCommand_noTabs === PAUSE TestInfoCommand_noTabs === RUN TestInfoCommand === PAUSE TestInfoCommand === CONT TestInfoCommand_noTabs === CONT TestInfoCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestInfoCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestInfoCommand writer.go:29: 2021-01-29T19:33:32.346Z [WARN] TestInfoCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:32.348Z [DEBUG] TestInfoCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:32.350Z [DEBUG] TestInfoCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:32.371Z [INFO] TestInfoCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:47d839b5-3f2d-2283-24df-ca0e2b529483 Address:127.0.0.1:19195}]" writer.go:29: 2021-01-29T19:33:32.374Z [INFO] TestInfoCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:19195 [Follower]" leader= writer.go:29: 2021-01-29T19:33:32.375Z [INFO] TestInfoCommand.server.serf.wan: serf: EventMemberJoin: Node-47d839b5-3f2d-2283-24df-ca0e2b529483.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:32.383Z [INFO] TestInfoCommand.server.serf.lan: serf: EventMemberJoin: Node-47d839b5-3f2d-2283-24df-ca0e2b529483 127.0.0.1 writer.go:29: 2021-01-29T19:33:32.391Z [INFO] TestInfoCommand.server: Handled event for server in area: event=member-join server=Node-47d839b5-3f2d-2283-24df-ca0e2b529483.dc1 area=wan writer.go:29: 2021-01-29T19:33:32.391Z [INFO] TestInfoCommand.server: Adding LAN server: server="Node-47d839b5-3f2d-2283-24df-ca0e2b529483 (Addr: tcp/127.0.0.1:19195) (DC: dc1)" writer.go:29: 2021-01-29T19:33:32.395Z [INFO] TestInfoCommand: Started DNS server: address=127.0.0.1:19190 network=tcp writer.go:29: 2021-01-29T19:33:32.398Z [INFO] TestInfoCommand: Started DNS server: address=127.0.0.1:19190 network=udp writer.go:29: 2021-01-29T19:33:32.400Z [INFO] TestInfoCommand: Started HTTP server: address=127.0.0.1:19191 network=tcp writer.go:29: 2021-01-29T19:33:32.402Z [INFO] TestInfoCommand: started state syncer writer.go:29: 2021-01-29T19:33:32.427Z [WARN] TestInfoCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:32.430Z [INFO] TestInfoCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:19195 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:32.440Z [DEBUG] TestInfoCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:32.443Z [DEBUG] TestInfoCommand.server.raft: vote granted: from=47d839b5-3f2d-2283-24df-ca0e2b529483 term=2 tally=1 writer.go:29: 2021-01-29T19:33:32.445Z [INFO] TestInfoCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:32.446Z [INFO] TestInfoCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:19195 [Leader]" writer.go:29: 2021-01-29T19:33:32.449Z [INFO] TestInfoCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:32.450Z [INFO] TestInfoCommand.server: New leader elected: payload=Node-47d839b5-3f2d-2283-24df-ca0e2b529483 writer.go:29: 2021-01-29T19:33:32.456Z [DEBUG] TestInfoCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19195 writer.go:29: 2021-01-29T19:33:32.472Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:32.532Z [INFO] TestInfoCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:32.533Z [INFO] TestInfoCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.535Z [DEBUG] TestInfoCommand.server: Skipping self join check for node since the cluster is too small: node=Node-47d839b5-3f2d-2283-24df-ca0e2b529483 writer.go:29: 2021-01-29T19:33:32.537Z [INFO] TestInfoCommand.server: member joined, marking health alive: member=Node-47d839b5-3f2d-2283-24df-ca0e2b529483 writer.go:29: 2021-01-29T19:33:32.720Z [DEBUG] TestInfoCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:32.724Z [INFO] TestInfoCommand: Synced node info writer.go:29: 2021-01-29T19:33:32.725Z [DEBUG] TestInfoCommand: Node info in sync writer.go:29: 2021-01-29T19:33:32.831Z [DEBUG] TestInfoCommand.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:60740 latency=28.402937ms writer.go:29: 2021-01-29T19:33:32.838Z [INFO] TestInfoCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:32.840Z [INFO] TestInfoCommand.server: shutting down server writer.go:29: 2021-01-29T19:33:32.843Z [DEBUG] TestInfoCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.848Z [WARN] TestInfoCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:32.849Z [DEBUG] TestInfoCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:32.854Z [WARN] TestInfoCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:32.859Z [INFO] TestInfoCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:32.862Z [INFO] TestInfoCommand: consul server down writer.go:29: 2021-01-29T19:33:32.864Z [INFO] TestInfoCommand: shutdown complete writer.go:29: 2021-01-29T19:33:32.866Z [INFO] TestInfoCommand: Stopping server: protocol=DNS address=127.0.0.1:19190 network=tcp writer.go:29: 2021-01-29T19:33:32.871Z [INFO] TestInfoCommand: Stopping server: protocol=DNS address=127.0.0.1:19190 network=udp writer.go:29: 2021-01-29T19:33:32.873Z [INFO] TestInfoCommand: Stopping server: protocol=HTTP address=127.0.0.1:19191 network=tcp writer.go:29: 2021-01-29T19:33:33.375Z [INFO] TestInfoCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:33.377Z [INFO] TestInfoCommand: Endpoints down --- PASS: TestInfoCommand (1.09s) PASS ok github.com/hashicorp/consul/command/info 1.365s === RUN TestCommand_noTabs === PAUSE TestCommand_noTabs === CONT TestCommand_noTabs --- PASS: TestCommand_noTabs (0.00s) PASS ok github.com/hashicorp/consul/command/intention 0.009s === RUN TestCommand_noTabs === PAUSE TestCommand_noTabs === RUN TestCommand_Validation === PAUSE TestCommand_Validation === RUN TestCommand === PAUSE TestCommand === CONT TestCommand === CONT TestCommand_Validation [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestCommand_noTabs === RUN TestCommand_Validation/0_args --- PASS: TestCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === RUN TestCommand_Validation/1_args === RUN TestCommand_Validation/3_args --- PASS: TestCommand_Validation (0.02s) --- PASS: TestCommand_Validation/0_args (0.00s) --- PASS: TestCommand_Validation/1_args (0.00s) --- PASS: TestCommand_Validation/3_args (0.00s) === CONT TestCommand writer.go:29: 2021-01-29T19:33:36.180Z [WARN] TestCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:36.182Z [DEBUG] TestCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:36.187Z [DEBUG] TestCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:36.193Z [INFO] TestCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2bca379f-66c1-20fb-0baa-cabe290aefa2 Address:127.0.0.1:22258}]" writer.go:29: 2021-01-29T19:33:36.197Z [INFO] TestCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:22258 [Follower]" leader= writer.go:29: 2021-01-29T19:33:36.200Z [INFO] TestCommand.server.serf.wan: serf: EventMemberJoin: Node-2bca379f-66c1-20fb-0baa-cabe290aefa2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:36.204Z [INFO] TestCommand.server.serf.lan: serf: EventMemberJoin: Node-2bca379f-66c1-20fb-0baa-cabe290aefa2 127.0.0.1 writer.go:29: 2021-01-29T19:33:36.206Z [INFO] TestCommand.server: Adding LAN server: server="Node-2bca379f-66c1-20fb-0baa-cabe290aefa2 (Addr: tcp/127.0.0.1:22258) (DC: dc1)" writer.go:29: 2021-01-29T19:33:36.207Z [INFO] TestCommand.server: Handled event for server in area: event=member-join server=Node-2bca379f-66c1-20fb-0baa-cabe290aefa2.dc1 area=wan writer.go:29: 2021-01-29T19:33:36.210Z [INFO] TestCommand: Started DNS server: address=127.0.0.1:22253 network=tcp writer.go:29: 2021-01-29T19:33:36.212Z [INFO] TestCommand: Started DNS server: address=127.0.0.1:22253 network=udp writer.go:29: 2021-01-29T19:33:36.213Z [INFO] TestCommand: Started HTTP server: address=127.0.0.1:22254 network=tcp writer.go:29: 2021-01-29T19:33:36.214Z [INFO] TestCommand: started state syncer writer.go:29: 2021-01-29T19:33:36.255Z [WARN] TestCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:36.264Z [INFO] TestCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:22258 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:36.267Z [DEBUG] TestCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:36.277Z [DEBUG] TestCommand.server.raft: vote granted: from=2bca379f-66c1-20fb-0baa-cabe290aefa2 term=2 tally=1 writer.go:29: 2021-01-29T19:33:36.280Z [INFO] TestCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:36.282Z [INFO] TestCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:22258 [Leader]" writer.go:29: 2021-01-29T19:33:36.288Z [INFO] TestCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:36.289Z [INFO] TestCommand.server: New leader elected: payload=Node-2bca379f-66c1-20fb-0baa-cabe290aefa2 writer.go:29: 2021-01-29T19:33:36.290Z [DEBUG] TestCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22258 writer.go:29: 2021-01-29T19:33:36.308Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:36.343Z [DEBUG] TestCommand.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:48414 latency=6.721239ms writer.go:29: 2021-01-29T19:33:36.372Z [DEBUG] TestCommand.http: Request finished: method=GET url=/v1/connect/intentions/check?destination=db&source=foo&source-type=consul from=127.0.0.1:48416 latency=18.996408ms writer.go:29: 2021-01-29T19:33:36.392Z [DEBUG] TestCommand.http: Request finished: method=GET url=/v1/connect/intentions/check?destination=db&source=web&source-type=consul from=127.0.0.1:48418 latency=9.12758ms writer.go:29: 2021-01-29T19:33:36.392Z [INFO] TestCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:36.396Z [INFO] TestCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.399Z [DEBUG] TestCommand.server: Skipping self join check for node since the cluster is too small: node=Node-2bca379f-66c1-20fb-0baa-cabe290aefa2 writer.go:29: 2021-01-29T19:33:36.405Z [INFO] TestCommand.server: member joined, marking health alive: member=Node-2bca379f-66c1-20fb-0baa-cabe290aefa2 writer.go:29: 2021-01-29T19:33:36.400Z [INFO] TestCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:36.410Z [INFO] TestCommand.server: shutting down server writer.go:29: 2021-01-29T19:33:36.411Z [DEBUG] TestCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.413Z [WARN] TestCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:36.411Z [ERROR] TestCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:36.413Z [DEBUG] TestCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:36.416Z [WARN] TestCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:36.422Z [INFO] TestCommand: consul server down writer.go:29: 2021-01-29T19:33:36.423Z [INFO] TestCommand: shutdown complete writer.go:29: 2021-01-29T19:33:36.422Z [INFO] TestCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:36.425Z [INFO] TestCommand: Stopping server: protocol=DNS address=127.0.0.1:22253 network=tcp writer.go:29: 2021-01-29T19:33:36.431Z [INFO] TestCommand: Stopping server: protocol=DNS address=127.0.0.1:22253 network=udp writer.go:29: 2021-01-29T19:33:36.433Z [INFO] TestCommand: Stopping server: protocol=HTTP address=127.0.0.1:22254 network=tcp writer.go:29: 2021-01-29T19:33:36.934Z [INFO] TestCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:36.936Z [INFO] TestCommand: Endpoints down --- PASS: TestCommand (0.79s) PASS ok github.com/hashicorp/consul/command/intention/check 1.012s === RUN TestCommand_noTabs === PAUSE TestCommand_noTabs === RUN TestCommand_Validation === PAUSE TestCommand_Validation === RUN TestCommand === PAUSE TestCommand === RUN TestCommand_deny === PAUSE TestCommand_deny === RUN TestCommand_meta === PAUSE TestCommand_meta === RUN TestCommand_File === PAUSE TestCommand_File === RUN TestCommand_FileNoExist === PAUSE TestCommand_FileNoExist === RUN TestCommand_replace === PAUSE TestCommand_replace === CONT TestCommand_noTabs === CONT TestCommand_File --- PASS: TestCommand_noTabs (0.00s) === CONT TestCommand_meta [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestCommand_deny === CONT TestCommand [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.619Z [WARN] TestCommand_deny: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:39.621Z [DEBUG] TestCommand_deny.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:39.628Z [DEBUG] TestCommand_deny.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.637Z [WARN] TestCommand: bootstrap = true: do not enable unless necessary === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.646Z [INFO] TestCommand_deny.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e3480501-ff90-4590-1076-990f110878a2 Address:127.0.0.1:14102}]" writer.go:29: 2021-01-29T19:33:39.672Z [INFO] TestCommand_deny.server.raft: entering follower state: follower="Node at 127.0.0.1:14102 [Follower]" leader= === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.673Z [DEBUG] TestCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:39.674Z [DEBUG] TestCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.677Z [INFO] TestCommand_deny.server.serf.wan: serf: EventMemberJoin: Node-e3480501-ff90-4590-1076-990f110878a2.dc1 127.0.0.1 === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.677Z [WARN] TestCommand_File: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:39.678Z [DEBUG] TestCommand_File.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:39.679Z [DEBUG] TestCommand_File.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.680Z [WARN] TestCommand_meta: bootstrap = true: do not enable unless necessary === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.679Z [INFO] TestCommand_deny.server.serf.lan: serf: EventMemberJoin: Node-e3480501-ff90-4590-1076-990f110878a2 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.701Z [INFO] TestCommand_deny: Started DNS server: address=127.0.0.1:14097 network=udp writer.go:29: 2021-01-29T19:33:39.702Z [INFO] TestCommand_deny.server: Adding LAN server: server="Node-e3480501-ff90-4590-1076-990f110878a2 (Addr: tcp/127.0.0.1:14102) (DC: dc1)" writer.go:29: 2021-01-29T19:33:39.703Z [INFO] TestCommand_deny: Started DNS server: address=127.0.0.1:14097 network=tcp === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.698Z [DEBUG] TestCommand_meta.tlsutil: Update: version=1 === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.702Z [INFO] TestCommand_deny.server: Handled event for server in area: event=member-join server=Node-e3480501-ff90-4590-1076-990f110878a2.dc1 area=wan === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.710Z [DEBUG] TestCommand_meta.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.706Z [INFO] TestCommand_deny: Started HTTP server: address=127.0.0.1:14098 network=tcp === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.722Z [INFO] TestCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9ff0c9bc-02a0-95ce-2104-a04a8c0016c6 Address:127.0.0.1:14108}]" === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.722Z [INFO] TestCommand_File.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a42d9755-f032-7ff6-b877-38e36bc2cc05 Address:127.0.0.1:14090}]" === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.723Z [INFO] TestCommand_deny: started state syncer === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.725Z [INFO] TestCommand_File.server.serf.wan: serf: EventMemberJoin: Node-a42d9755-f032-7ff6-b877-38e36bc2cc05.dc1 127.0.0.1 === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.726Z [INFO] TestCommand.server.serf.wan: serf: EventMemberJoin: Node-9ff0c9bc-02a0-95ce-2104-a04a8c0016c6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.729Z [INFO] TestCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:14108 [Follower]" leader= === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.729Z [INFO] TestCommand_File.server.serf.lan: serf: EventMemberJoin: Node-a42d9755-f032-7ff6-b877-38e36bc2cc05 127.0.0.1 === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.738Z [WARN] TestCommand_deny.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.741Z [INFO] TestCommand_meta.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:61a6cf89-ff05-2c73-21db-7fcd2420b62b Address:127.0.0.1:14096}]" === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.746Z [INFO] TestCommand_deny.server.raft: entering candidate state: node="Node at 127.0.0.1:14102 [Candidate]" term=2 === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.776Z [INFO] TestCommand_meta.server.serf.wan: serf: EventMemberJoin: Node-61a6cf89-ff05-2c73-21db-7fcd2420b62b.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.778Z [INFO] TestCommand_meta.server.serf.lan: serf: EventMemberJoin: Node-61a6cf89-ff05-2c73-21db-7fcd2420b62b 127.0.0.1 writer.go:29: 2021-01-29T19:33:39.781Z [INFO] TestCommand_meta: Started DNS server: address=127.0.0.1:14091 network=udp === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.730Z [INFO] TestCommand.server.serf.lan: serf: EventMemberJoin: Node-9ff0c9bc-02a0-95ce-2104-a04a8c0016c6 127.0.0.1 === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.782Z [INFO] TestCommand_meta.server.raft: entering follower state: follower="Node at 127.0.0.1:14096 [Follower]" leader= === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.784Z [INFO] TestCommand: Started DNS server: address=127.0.0.1:14103 network=udp writer.go:29: 2021-01-29T19:33:39.785Z [INFO] TestCommand.server: Adding LAN server: server="Node-9ff0c9bc-02a0-95ce-2104-a04a8c0016c6 (Addr: tcp/127.0.0.1:14108) (DC: dc1)" writer.go:29: 2021-01-29T19:33:39.786Z [INFO] TestCommand.server: Handled event for server in area: event=member-join server=Node-9ff0c9bc-02a0-95ce-2104-a04a8c0016c6.dc1 area=wan === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.744Z [INFO] TestCommand_File.server.raft: entering follower state: follower="Node at 127.0.0.1:14090 [Follower]" leader= === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.786Z [DEBUG] TestCommand_deny.server.raft: votes: needed=1 === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.775Z [INFO] TestCommand_File.server: Adding LAN server: server="Node-a42d9755-f032-7ff6-b877-38e36bc2cc05 (Addr: tcp/127.0.0.1:14090) (DC: dc1)" === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.789Z [DEBUG] TestCommand_deny.server.raft: vote granted: from=e3480501-ff90-4590-1076-990f110878a2 term=2 tally=1 === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.793Z [INFO] TestCommand_meta.server: Adding LAN server: server="Node-61a6cf89-ff05-2c73-21db-7fcd2420b62b (Addr: tcp/127.0.0.1:14096) (DC: dc1)" writer.go:29: 2021-01-29T19:33:39.794Z [INFO] TestCommand_meta.server: Handled event for server in area: event=member-join server=Node-61a6cf89-ff05-2c73-21db-7fcd2420b62b.dc1 area=wan === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.776Z [INFO] TestCommand_File: Started DNS server: address=127.0.0.1:14085 network=udp === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.796Z [INFO] TestCommand_meta: Started DNS server: address=127.0.0.1:14091 network=tcp writer.go:29: 2021-01-29T19:33:39.798Z [INFO] TestCommand_meta: Started HTTP server: address=127.0.0.1:14092 network=tcp writer.go:29: 2021-01-29T19:33:39.799Z [INFO] TestCommand_meta: started state syncer === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.789Z [INFO] TestCommand: Started DNS server: address=127.0.0.1:14103 network=tcp writer.go:29: 2021-01-29T19:33:39.800Z [WARN] TestCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:39.802Z [INFO] TestCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:14108 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:39.802Z [INFO] TestCommand: Started HTTP server: address=127.0.0.1:14104 network=tcp writer.go:29: 2021-01-29T19:33:39.805Z [DEBUG] TestCommand.server.raft: votes: needed=1 === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.793Z [INFO] TestCommand_deny.server.raft: election won: tally=1 === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.807Z [DEBUG] TestCommand.server.raft: vote granted: from=9ff0c9bc-02a0-95ce-2104-a04a8c0016c6 term=2 tally=1 === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.783Z [INFO] TestCommand_File.server: Handled event for server in area: event=member-join server=Node-a42d9755-f032-7ff6-b877-38e36bc2cc05.dc1 area=wan === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.805Z [INFO] TestCommand: started state syncer writer.go:29: 2021-01-29T19:33:39.809Z [INFO] TestCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:39.812Z [INFO] TestCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:14108 [Leader]" === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.808Z [INFO] TestCommand_deny.server.raft: entering leader state: leader="Node at 127.0.0.1:14102 [Leader]" === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.801Z [INFO] TestCommand_File: Started DNS server: address=127.0.0.1:14085 network=tcp writer.go:29: 2021-01-29T19:33:39.824Z [INFO] TestCommand_File: Started HTTP server: address=127.0.0.1:14086 network=tcp writer.go:29: 2021-01-29T19:33:39.825Z [INFO] TestCommand_File: started state syncer === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.823Z [INFO] TestCommand.server: cluster leadership acquired === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.827Z [INFO] TestCommand_deny.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:39.828Z [INFO] TestCommand_deny.server: New leader elected: payload=Node-e3480501-ff90-4590-1076-990f110878a2 === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.829Z [INFO] TestCommand.server: New leader elected: payload=Node-9ff0c9bc-02a0-95ce-2104-a04a8c0016c6 writer.go:29: 2021-01-29T19:33:39.833Z [DEBUG] TestCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:14108 === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.834Z [WARN] TestCommand_meta.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.834Z [DEBUG] TestCommand.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:53592 latency=12.277555ms === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.836Z [INFO] TestCommand_meta.server.raft: entering candidate state: node="Node at 127.0.0.1:14096 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:39.839Z [DEBUG] TestCommand_meta.server.raft: votes: needed=1 === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.840Z [DEBUG] TestCommand_deny.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:14102 === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.842Z [DEBUG] TestCommand.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:53594 latency=1.288152ms === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.841Z [DEBUG] TestCommand_meta.server.raft: vote granted: from=61a6cf89-ff05-2c73-21db-7fcd2420b62b term=2 tally=1 writer.go:29: 2021-01-29T19:33:39.843Z [INFO] TestCommand_meta.server.raft: election won: tally=1 === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.845Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.845Z [INFO] TestCommand_meta.server.raft: entering leader state: leader="Node at 127.0.0.1:14096 [Leader]" === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.862Z [WARN] TestCommand_File.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:39.863Z [INFO] TestCommand_File.server.raft: entering candidate state: node="Node at 127.0.0.1:14090 [Candidate]" term=2 === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.847Z [INFO] TestCommand_meta.server: cluster leadership acquired === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.860Z [INFO] TestCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:39.876Z [INFO] TestCommand.server: shutting down server === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.862Z [INFO] TestCommand_meta.server: New leader elected: payload=Node-61a6cf89-ff05-2c73-21db-7fcd2420b62b === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.878Z [WARN] TestCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:39.880Z [ERROR] TestCommand.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.875Z [DEBUG] TestCommand_meta.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:14096 === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.885Z [DEBUG] TestCommand_File.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:39.887Z [DEBUG] TestCommand_File.server.raft: vote granted: from=a42d9755-f032-7ff6-b877-38e36bc2cc05 term=2 tally=1 writer.go:29: 2021-01-29T19:33:39.888Z [INFO] TestCommand_File.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:39.889Z [INFO] TestCommand_File.server.raft: entering leader state: leader="Node at 127.0.0.1:14090 [Leader]" === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.884Z [WARN] TestCommand.server.serf.wan: serf: Shutdown without a Leave === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.891Z [INFO] TestCommand_File.server: cluster leadership acquired === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.886Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:39.898Z [INFO] TestCommand.server.router.manager: shutting down === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.894Z [INFO] TestCommand_File.server: New leader elected: payload=Node-a42d9755-f032-7ff6-b877-38e36bc2cc05 === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.903Z [INFO] TestCommand: consul server down === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.904Z [DEBUG] TestCommand_File.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:14090 writer.go:29: 2021-01-29T19:33:39.906Z [INFO] TestCommand_File: Synced node info writer.go:29: 2021-01-29T19:33:39.908Z [DEBUG] TestCommand_File: Node info in sync === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.906Z [INFO] TestCommand: shutdown complete === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.909Z [INFO] TestCommand_meta: Synced node info === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.909Z [INFO] TestCommand: Stopping server: protocol=DNS address=127.0.0.1:14103 network=tcp === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.913Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.915Z [INFO] TestCommand: Stopping server: protocol=DNS address=127.0.0.1:14103 network=udp === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.920Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand writer.go:29: 2021-01-29T19:33:39.918Z [INFO] TestCommand: Stopping server: protocol=HTTP address=127.0.0.1:14104 network=tcp writer.go:29: 2021-01-29T19:33:39.954Z [ERROR] TestCommand.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: raft is already shutdown" === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.964Z [INFO] TestCommand_File.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.965Z [INFO] TestCommand_meta.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.965Z [INFO] TestCommand_File.leader: started routine: routine="CA root pruning" === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.966Z [INFO] TestCommand_meta.leader: started routine: routine="CA root pruning" === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.967Z [DEBUG] TestCommand_File.server: Skipping self join check for node since the cluster is too small: node=Node-a42d9755-f032-7ff6-b877-38e36bc2cc05 === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:39.967Z [DEBUG] TestCommand_meta.server: Skipping self join check for node since the cluster is too small: node=Node-61a6cf89-ff05-2c73-21db-7fcd2420b62b writer.go:29: 2021-01-29T19:33:39.968Z [INFO] TestCommand_meta.server: member joined, marking health alive: member=Node-61a6cf89-ff05-2c73-21db-7fcd2420b62b === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:39.968Z [INFO] TestCommand_File.server: member joined, marking health alive: member=Node-a42d9755-f032-7ff6-b877-38e36bc2cc05 === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:39.976Z [INFO] TestCommand_deny.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:39.979Z [INFO] TestCommand_deny.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:39.982Z [DEBUG] TestCommand_deny.server: Skipping self join check for node since the cluster is too small: node=Node-e3480501-ff90-4590-1076-990f110878a2 writer.go:29: 2021-01-29T19:33:39.984Z [INFO] TestCommand_deny.server: member joined, marking health alive: member=Node-e3480501-ff90-4590-1076-990f110878a2 writer.go:29: 2021-01-29T19:33:39.982Z [INFO] TestCommand_deny: Synced node info writer.go:29: 2021-01-29T19:33:40.043Z [DEBUG] TestCommand_deny.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:35260 latency=1.856637ms writer.go:29: 2021-01-29T19:33:40.048Z [DEBUG] TestCommand_deny.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:35262 latency=910.71µs writer.go:29: 2021-01-29T19:33:40.055Z [INFO] TestCommand_deny: Requesting shutdown writer.go:29: 2021-01-29T19:33:40.057Z [INFO] TestCommand_deny.server: shutting down server writer.go:29: 2021-01-29T19:33:40.058Z [DEBUG] TestCommand_deny.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.060Z [WARN] TestCommand_deny.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.061Z [DEBUG] TestCommand_deny.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.064Z [WARN] TestCommand_deny.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.067Z [INFO] TestCommand_deny.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:40.068Z [INFO] TestCommand_deny: consul server down writer.go:29: 2021-01-29T19:33:40.071Z [INFO] TestCommand_deny: shutdown complete writer.go:29: 2021-01-29T19:33:40.072Z [INFO] TestCommand_deny: Stopping server: protocol=DNS address=127.0.0.1:14097 network=tcp writer.go:29: 2021-01-29T19:33:40.074Z [INFO] TestCommand_deny: Stopping server: protocol=DNS address=127.0.0.1:14097 network=udp writer.go:29: 2021-01-29T19:33:40.075Z [INFO] TestCommand_deny: Stopping server: protocol=HTTP address=127.0.0.1:14098 network=tcp === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:40.082Z [DEBUG] TestCommand_File.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:52376 latency=1.437676ms writer.go:29: 2021-01-29T19:33:40.096Z [DEBUG] TestCommand_File.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:52378 latency=200.272µs writer.go:29: 2021-01-29T19:33:40.103Z [INFO] TestCommand_File: Requesting shutdown writer.go:29: 2021-01-29T19:33:40.104Z [INFO] TestCommand_File.server: shutting down server writer.go:29: 2021-01-29T19:33:40.106Z [DEBUG] TestCommand_File.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.109Z [WARN] TestCommand_File.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.109Z [DEBUG] TestCommand_File.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.114Z [WARN] TestCommand_File.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.117Z [INFO] TestCommand_File.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:40.118Z [INFO] TestCommand_File: consul server down writer.go:29: 2021-01-29T19:33:40.122Z [INFO] TestCommand_File: shutdown complete writer.go:29: 2021-01-29T19:33:40.123Z [INFO] TestCommand_File: Stopping server: protocol=DNS address=127.0.0.1:14085 network=tcp writer.go:29: 2021-01-29T19:33:40.125Z [INFO] TestCommand_File: Stopping server: protocol=DNS address=127.0.0.1:14085 network=udp writer.go:29: 2021-01-29T19:33:40.127Z [INFO] TestCommand_File: Stopping server: protocol=HTTP address=127.0.0.1:14086 network=tcp === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:40.121Z [DEBUG] TestCommand_meta.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:47752 latency=2.698331ms writer.go:29: 2021-01-29T19:33:40.135Z [DEBUG] TestCommand_meta.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:47754 latency=296.807µs writer.go:29: 2021-01-29T19:33:40.139Z [INFO] TestCommand_meta: Requesting shutdown writer.go:29: 2021-01-29T19:33:40.140Z [INFO] TestCommand_meta.server: shutting down server writer.go:29: 2021-01-29T19:33:40.142Z [DEBUG] TestCommand_meta.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.143Z [WARN] TestCommand_meta.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.143Z [DEBUG] TestCommand_meta.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.145Z [WARN] TestCommand_meta.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.169Z [INFO] TestCommand_meta.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:40.170Z [INFO] TestCommand_meta: consul server down writer.go:29: 2021-01-29T19:33:40.171Z [INFO] TestCommand_meta: shutdown complete writer.go:29: 2021-01-29T19:33:40.172Z [INFO] TestCommand_meta: Stopping server: protocol=DNS address=127.0.0.1:14091 network=tcp writer.go:29: 2021-01-29T19:33:40.174Z [INFO] TestCommand_meta: Stopping server: protocol=DNS address=127.0.0.1:14091 network=udp writer.go:29: 2021-01-29T19:33:40.176Z [INFO] TestCommand_meta: Stopping server: protocol=HTTP address=127.0.0.1:14092 network=tcp === CONT TestCommand writer.go:29: 2021-01-29T19:33:40.425Z [INFO] TestCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:40.427Z [INFO] TestCommand: Endpoints down --- PASS: TestCommand (0.86s) === CONT TestCommand_Validation === RUN TestCommand_Validation/-allow_and_-deny --- PASS: TestCommand_Validation (0.01s) --- PASS: TestCommand_Validation/-allow_and_-deny (0.01s) === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.464Z [WARN] TestCommand_replace: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:40.466Z [DEBUG] TestCommand_replace.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:40.469Z [DEBUG] TestCommand_replace.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:40.478Z [INFO] TestCommand_replace.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bd8f3e4b-add7-6b20-daba-94e64035fd62 Address:127.0.0.1:14114}]" writer.go:29: 2021-01-29T19:33:40.480Z [INFO] TestCommand_replace.server.raft: entering follower state: follower="Node at 127.0.0.1:14114 [Follower]" leader= writer.go:29: 2021-01-29T19:33:40.514Z [INFO] TestCommand_replace.server.serf.wan: serf: EventMemberJoin: Node-bd8f3e4b-add7-6b20-daba-94e64035fd62.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.519Z [INFO] TestCommand_replace.server.serf.lan: serf: EventMemberJoin: Node-bd8f3e4b-add7-6b20-daba-94e64035fd62 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.524Z [INFO] TestCommand_replace.server: Adding LAN server: server="Node-bd8f3e4b-add7-6b20-daba-94e64035fd62 (Addr: tcp/127.0.0.1:14114) (DC: dc1)" writer.go:29: 2021-01-29T19:33:40.525Z [INFO] TestCommand_replace: Started DNS server: address=127.0.0.1:14109 network=udp writer.go:29: 2021-01-29T19:33:40.534Z [INFO] TestCommand_replace: Started DNS server: address=127.0.0.1:14109 network=tcp writer.go:29: 2021-01-29T19:33:40.525Z [INFO] TestCommand_replace.server: Handled event for server in area: event=member-join server=Node-bd8f3e4b-add7-6b20-daba-94e64035fd62.dc1 area=wan writer.go:29: 2021-01-29T19:33:40.529Z [WARN] TestCommand_replace.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:40.541Z [INFO] TestCommand_replace: Started HTTP server: address=127.0.0.1:14110 network=tcp writer.go:29: 2021-01-29T19:33:40.564Z [INFO] TestCommand_replace: started state syncer writer.go:29: 2021-01-29T19:33:40.553Z [INFO] TestCommand_replace.server.raft: entering candidate state: node="Node at 127.0.0.1:14114 [Candidate]" term=2 === CONT TestCommand_deny writer.go:29: 2021-01-29T19:33:40.577Z [INFO] TestCommand_deny: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:40.578Z [INFO] TestCommand_deny: Endpoints down --- PASS: TestCommand_deny (1.01s) === CONT TestCommand_FileNoExist === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.590Z [DEBUG] TestCommand_replace.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:40.596Z [DEBUG] TestCommand_replace.server.raft: vote granted: from=bd8f3e4b-add7-6b20-daba-94e64035fd62 term=2 tally=1 writer.go:29: 2021-01-29T19:33:40.600Z [INFO] TestCommand_replace.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:40.605Z [INFO] TestCommand_replace.server.raft: entering leader state: leader="Node at 127.0.0.1:14114 [Leader]" writer.go:29: 2021-01-29T19:33:40.614Z [INFO] TestCommand_replace.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:40.615Z [INFO] TestCommand_replace.server: New leader elected: payload=Node-bd8f3e4b-add7-6b20-daba-94e64035fd62 === CONT TestCommand_FileNoExist writer.go:29: 2021-01-29T19:33:40.622Z [WARN] TestCommand_FileNoExist: bootstrap = true: do not enable unless necessary === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.618Z [DEBUG] TestCommand_replace.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:14114 === CONT TestCommand_FileNoExist writer.go:29: 2021-01-29T19:33:40.624Z [DEBUG] TestCommand_FileNoExist.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:40.626Z [DEBUG] TestCommand_FileNoExist.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.631Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_File writer.go:29: 2021-01-29T19:33:40.630Z [INFO] TestCommand_File: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:40.632Z [INFO] TestCommand_File: Endpoints down --- PASS: TestCommand_File (1.07s) === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.635Z [INFO] TestCommand_replace.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCommand_FileNoExist writer.go:29: 2021-01-29T19:33:40.634Z [INFO] TestCommand_FileNoExist.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9651305b-7299-a081-c761-9fc9173ca9f8 Address:127.0.0.1:14120}]" === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.636Z [INFO] TestCommand_replace.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.637Z [DEBUG] TestCommand_replace.server: Skipping self join check for node since the cluster is too small: node=Node-bd8f3e4b-add7-6b20-daba-94e64035fd62 writer.go:29: 2021-01-29T19:33:40.638Z [INFO] TestCommand_replace.server: member joined, marking health alive: member=Node-bd8f3e4b-add7-6b20-daba-94e64035fd62 === CONT TestCommand_FileNoExist writer.go:29: 2021-01-29T19:33:40.638Z [INFO] TestCommand_FileNoExist.server.raft: entering follower state: follower="Node at 127.0.0.1:14120 [Follower]" leader= writer.go:29: 2021-01-29T19:33:40.639Z [INFO] TestCommand_FileNoExist.server.serf.wan: serf: EventMemberJoin: Node-9651305b-7299-a081-c761-9fc9173ca9f8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.651Z [INFO] TestCommand_FileNoExist.server.serf.lan: serf: EventMemberJoin: Node-9651305b-7299-a081-c761-9fc9173ca9f8 127.0.0.1 writer.go:29: 2021-01-29T19:33:40.653Z [INFO] TestCommand_FileNoExist.server: Handled event for server in area: event=member-join server=Node-9651305b-7299-a081-c761-9fc9173ca9f8.dc1 area=wan writer.go:29: 2021-01-29T19:33:40.654Z [INFO] TestCommand_FileNoExist.server: Adding LAN server: server="Node-9651305b-7299-a081-c761-9fc9173ca9f8 (Addr: tcp/127.0.0.1:14120) (DC: dc1)" writer.go:29: 2021-01-29T19:33:40.654Z [INFO] TestCommand_FileNoExist: Started DNS server: address=127.0.0.1:14115 network=udp writer.go:29: 2021-01-29T19:33:40.657Z [INFO] TestCommand_FileNoExist: Started DNS server: address=127.0.0.1:14115 network=tcp writer.go:29: 2021-01-29T19:33:40.658Z [INFO] TestCommand_FileNoExist: Started HTTP server: address=127.0.0.1:14116 network=tcp writer.go:29: 2021-01-29T19:33:40.661Z [INFO] TestCommand_FileNoExist: started state syncer === CONT TestCommand_meta writer.go:29: 2021-01-29T19:33:40.678Z [INFO] TestCommand_meta: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:40.680Z [INFO] TestCommand_meta: Endpoints down --- PASS: TestCommand_meta (1.12s) === CONT TestCommand_FileNoExist writer.go:29: 2021-01-29T19:33:40.695Z [WARN] TestCommand_FileNoExist.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:40.697Z [INFO] TestCommand_FileNoExist.server.raft: entering candidate state: node="Node at 127.0.0.1:14120 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:40.700Z [DEBUG] TestCommand_FileNoExist.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:40.702Z [DEBUG] TestCommand_FileNoExist.server.raft: vote granted: from=9651305b-7299-a081-c761-9fc9173ca9f8 term=2 tally=1 writer.go:29: 2021-01-29T19:33:40.704Z [INFO] TestCommand_FileNoExist.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:40.707Z [INFO] TestCommand_FileNoExist.server.raft: entering leader state: leader="Node at 127.0.0.1:14120 [Leader]" writer.go:29: 2021-01-29T19:33:40.709Z [INFO] TestCommand_FileNoExist.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:40.710Z [INFO] TestCommand_FileNoExist.server: New leader elected: payload=Node-9651305b-7299-a081-c761-9fc9173ca9f8 writer.go:29: 2021-01-29T19:33:40.716Z [DEBUG] TestCommand_FileNoExist.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:14120 writer.go:29: 2021-01-29T19:33:40.721Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:40.726Z [INFO] TestCommand_FileNoExist.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:40.727Z [INFO] TestCommand_FileNoExist.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.728Z [DEBUG] TestCommand_FileNoExist.server: Skipping self join check for node since the cluster is too small: node=Node-9651305b-7299-a081-c761-9fc9173ca9f8 writer.go:29: 2021-01-29T19:33:40.729Z [INFO] TestCommand_FileNoExist.server: member joined, marking health alive: member=Node-9651305b-7299-a081-c761-9fc9173ca9f8 === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.843Z [DEBUG] TestCommand_replace.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:36338 latency=10.10722ms writer.go:29: 2021-01-29T19:33:40.854Z [DEBUG] TestCommand_replace.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:36340 latency=183.123µs writer.go:29: 2021-01-29T19:33:40.866Z [ERROR] TestCommand_replace.http: Request error: method=POST url=/v1/connect/intentions from=127.0.0.1:36342 error="duplicate intention found: ALLOW default/foo => default/bar (ID: 75fafaef-306b-b7e1-c2fe-3818ba99644b, Precedence: 9)" writer.go:29: 2021-01-29T19:33:40.869Z [DEBUG] TestCommand_replace.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:36342 latency=3.608832ms === CONT TestCommand_FileNoExist writer.go:29: 2021-01-29T19:33:40.869Z [INFO] TestCommand_FileNoExist: Requesting shutdown writer.go:29: 2021-01-29T19:33:40.887Z [INFO] TestCommand_FileNoExist.server: shutting down server writer.go:29: 2021-01-29T19:33:40.887Z [DEBUG] TestCommand_FileNoExist.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.888Z [WARN] TestCommand_FileNoExist.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.890Z [ERROR] TestCommand_FileNoExist.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:40.891Z [DEBUG] TestCommand_FileNoExist.leader: stopped routine: routine="CA root pruning" === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.886Z [DEBUG] TestCommand_replace.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:36344 latency=796.834µs writer.go:29: 2021-01-29T19:33:40.897Z [DEBUG] TestCommand_replace: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestCommand_FileNoExist writer.go:29: 2021-01-29T19:33:40.909Z [WARN] TestCommand_FileNoExist.server.serf.wan: serf: Shutdown without a Leave === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.910Z [INFO] TestCommand_replace: Synced node info === CONT TestCommand_FileNoExist writer.go:29: 2021-01-29T19:33:40.911Z [INFO] TestCommand_FileNoExist.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:40.913Z [INFO] TestCommand_FileNoExist: consul server down === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.911Z [DEBUG] TestCommand_replace.http: Request finished: method=PUT url=/v1/connect/intentions/75fafaef-306b-b7e1-c2fe-3818ba99644b from=127.0.0.1:36344 latency=16.449ms === CONT TestCommand_FileNoExist writer.go:29: 2021-01-29T19:33:40.914Z [INFO] TestCommand_FileNoExist: shutdown complete writer.go:29: 2021-01-29T19:33:40.915Z [INFO] TestCommand_FileNoExist: Stopping server: protocol=DNS address=127.0.0.1:14115 network=tcp writer.go:29: 2021-01-29T19:33:40.915Z [INFO] TestCommand_FileNoExist: Stopping server: protocol=DNS address=127.0.0.1:14115 network=udp writer.go:29: 2021-01-29T19:33:40.916Z [INFO] TestCommand_FileNoExist: Stopping server: protocol=HTTP address=127.0.0.1:14116 network=tcp === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:40.916Z [DEBUG] TestCommand_replace.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:36340 latency=158.535µs writer.go:29: 2021-01-29T19:33:40.928Z [INFO] TestCommand_replace: Requesting shutdown writer.go:29: 2021-01-29T19:33:40.929Z [INFO] TestCommand_replace.server: shutting down server writer.go:29: 2021-01-29T19:33:40.930Z [DEBUG] TestCommand_replace.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.935Z [WARN] TestCommand_replace.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.935Z [DEBUG] TestCommand_replace.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:40.937Z [WARN] TestCommand_replace.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:40.941Z [INFO] TestCommand_replace.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:40.942Z [INFO] TestCommand_replace: consul server down writer.go:29: 2021-01-29T19:33:40.943Z [INFO] TestCommand_replace: shutdown complete writer.go:29: 2021-01-29T19:33:40.944Z [INFO] TestCommand_replace: Stopping server: protocol=DNS address=127.0.0.1:14109 network=tcp writer.go:29: 2021-01-29T19:33:40.945Z [INFO] TestCommand_replace: Stopping server: protocol=DNS address=127.0.0.1:14109 network=udp writer.go:29: 2021-01-29T19:33:40.946Z [INFO] TestCommand_replace: Stopping server: protocol=HTTP address=127.0.0.1:14110 network=tcp === CONT TestCommand_FileNoExist writer.go:29: 2021-01-29T19:33:41.417Z [INFO] TestCommand_FileNoExist: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:41.421Z [INFO] TestCommand_FileNoExist: Endpoints down --- PASS: TestCommand_FileNoExist (0.84s) === CONT TestCommand_replace writer.go:29: 2021-01-29T19:33:41.447Z [INFO] TestCommand_replace: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:41.449Z [INFO] TestCommand_replace: Endpoints down --- PASS: TestCommand_replace (1.01s) PASS ok github.com/hashicorp/consul/command/intention/create 2.017s === RUN TestCommand_noTabs === PAUSE TestCommand_noTabs === RUN TestCommand_Validation === PAUSE TestCommand_Validation === RUN TestCommand === PAUSE TestCommand === CONT TestCommand_noTabs === CONT TestCommand === CONT TestCommand_Validation === RUN TestCommand_Validation/0_args [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestCommand_noTabs (0.00s) === RUN TestCommand_Validation/3_args --- PASS: TestCommand_Validation (0.01s) --- PASS: TestCommand_Validation/0_args (0.00s) --- PASS: TestCommand_Validation/3_args (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestCommand writer.go:29: 2021-01-29T19:33:44.872Z [WARN] TestCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:44.874Z [DEBUG] TestCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:44.880Z [DEBUG] TestCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:44.886Z [INFO] TestCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0024e09f-9d8d-43bb-dc50-bd45cbad6ca0 Address:127.0.0.1:26342}]" writer.go:29: 2021-01-29T19:33:44.888Z [INFO] TestCommand.server.serf.wan: serf: EventMemberJoin: Node-0024e09f-9d8d-43bb-dc50-bd45cbad6ca0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:44.888Z [INFO] TestCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:26342 [Follower]" leader= writer.go:29: 2021-01-29T19:33:44.894Z [INFO] TestCommand.server.serf.lan: serf: EventMemberJoin: Node-0024e09f-9d8d-43bb-dc50-bd45cbad6ca0 127.0.0.1 writer.go:29: 2021-01-29T19:33:44.899Z [INFO] TestCommand: Started DNS server: address=127.0.0.1:26337 network=udp writer.go:29: 2021-01-29T19:33:44.900Z [INFO] TestCommand.server: Adding LAN server: server="Node-0024e09f-9d8d-43bb-dc50-bd45cbad6ca0 (Addr: tcp/127.0.0.1:26342) (DC: dc1)" writer.go:29: 2021-01-29T19:33:44.901Z [INFO] TestCommand.server: Handled event for server in area: event=member-join server=Node-0024e09f-9d8d-43bb-dc50-bd45cbad6ca0.dc1 area=wan writer.go:29: 2021-01-29T19:33:44.902Z [INFO] TestCommand: Started DNS server: address=127.0.0.1:26337 network=tcp writer.go:29: 2021-01-29T19:33:44.912Z [INFO] TestCommand: Started HTTP server: address=127.0.0.1:26338 network=tcp writer.go:29: 2021-01-29T19:33:44.918Z [INFO] TestCommand: started state syncer writer.go:29: 2021-01-29T19:33:44.960Z [WARN] TestCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:44.962Z [INFO] TestCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:26342 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:44.968Z [DEBUG] TestCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:44.970Z [DEBUG] TestCommand.server.raft: vote granted: from=0024e09f-9d8d-43bb-dc50-bd45cbad6ca0 term=2 tally=1 writer.go:29: 2021-01-29T19:33:44.972Z [INFO] TestCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:44.973Z [INFO] TestCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:26342 [Leader]" writer.go:29: 2021-01-29T19:33:44.975Z [INFO] TestCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:44.975Z [INFO] TestCommand.server: New leader elected: payload=Node-0024e09f-9d8d-43bb-dc50-bd45cbad6ca0 writer.go:29: 2021-01-29T19:33:44.999Z [DEBUG] TestCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:26342 writer.go:29: 2021-01-29T19:33:45.005Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:45.045Z [INFO] TestCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:45.046Z [INFO] TestCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.048Z [DEBUG] TestCommand.server: Skipping self join check for node since the cluster is too small: node=Node-0024e09f-9d8d-43bb-dc50-bd45cbad6ca0 writer.go:29: 2021-01-29T19:33:45.049Z [INFO] TestCommand.server: member joined, marking health alive: member=Node-0024e09f-9d8d-43bb-dc50-bd45cbad6ca0 writer.go:29: 2021-01-29T19:33:45.164Z [DEBUG] TestCommand.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:57956 latency=33.637288ms writer.go:29: 2021-01-29T19:33:45.174Z [DEBUG] TestCommand.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:57958 latency=1.809436ms writer.go:29: 2021-01-29T19:33:45.185Z [DEBUG] TestCommand.http: Request finished: method=DELETE url=/v1/connect/intentions/105d7455-0f6d-1264-9e6c-cdcfa6d122b2 from=127.0.0.1:57958 latency=7.4114ms writer.go:29: 2021-01-29T19:33:45.190Z [DEBUG] TestCommand.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:57956 latency=456.677µs writer.go:29: 2021-01-29T19:33:45.210Z [INFO] TestCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:45.211Z [INFO] TestCommand.server: shutting down server writer.go:29: 2021-01-29T19:33:45.211Z [DEBUG] TestCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.211Z [ERROR] TestCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:45.214Z [WARN] TestCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:45.214Z [DEBUG] TestCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:45.219Z [WARN] TestCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:45.222Z [INFO] TestCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:45.222Z [INFO] TestCommand: consul server down writer.go:29: 2021-01-29T19:33:45.226Z [INFO] TestCommand: shutdown complete writer.go:29: 2021-01-29T19:33:45.227Z [INFO] TestCommand: Stopping server: protocol=DNS address=127.0.0.1:26337 network=tcp writer.go:29: 2021-01-29T19:33:45.229Z [INFO] TestCommand: Stopping server: protocol=DNS address=127.0.0.1:26337 network=udp writer.go:29: 2021-01-29T19:33:45.230Z [INFO] TestCommand: Stopping server: protocol=HTTP address=127.0.0.1:26338 network=tcp writer.go:29: 2021-01-29T19:33:45.732Z [INFO] TestCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:45.733Z [INFO] TestCommand: Endpoints down --- PASS: TestCommand (0.91s) PASS ok github.com/hashicorp/consul/command/intention/delete 1.137s === RUN TestFinder === PAUSE TestFinder === CONT TestFinder [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:33:47.439Z [WARN] TestFinder: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:47.440Z [DEBUG] TestFinder.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:47.444Z [DEBUG] TestFinder.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:47.450Z [INFO] TestFinder.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4888874b-c105-6251-a4e2-093d02dd1f15 Address:127.0.0.1:17153}]" writer.go:29: 2021-01-29T19:33:47.454Z [INFO] TestFinder.server.raft: entering follower state: follower="Node at 127.0.0.1:17153 [Follower]" leader= writer.go:29: 2021-01-29T19:33:47.458Z [INFO] TestFinder.server.serf.wan: serf: EventMemberJoin: Node-4888874b-c105-6251-a4e2-093d02dd1f15.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:47.469Z [INFO] TestFinder.server.serf.lan: serf: EventMemberJoin: Node-4888874b-c105-6251-a4e2-093d02dd1f15 127.0.0.1 writer.go:29: 2021-01-29T19:33:47.477Z [INFO] TestFinder.server: Adding LAN server: server="Node-4888874b-c105-6251-a4e2-093d02dd1f15 (Addr: tcp/127.0.0.1:17153) (DC: dc1)" writer.go:29: 2021-01-29T19:33:47.478Z [INFO] TestFinder.server: Handled event for server in area: event=member-join server=Node-4888874b-c105-6251-a4e2-093d02dd1f15.dc1 area=wan writer.go:29: 2021-01-29T19:33:47.496Z [INFO] TestFinder: Started DNS server: address=127.0.0.1:17148 network=tcp writer.go:29: 2021-01-29T19:33:47.497Z [INFO] TestFinder: Started DNS server: address=127.0.0.1:17148 network=udp writer.go:29: 2021-01-29T19:33:47.498Z [INFO] TestFinder: Started HTTP server: address=127.0.0.1:17149 network=tcp writer.go:29: 2021-01-29T19:33:47.499Z [INFO] TestFinder: started state syncer writer.go:29: 2021-01-29T19:33:47.545Z [WARN] TestFinder.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:47.546Z [INFO] TestFinder.server.raft: entering candidate state: node="Node at 127.0.0.1:17153 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:47.552Z [DEBUG] TestFinder.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:47.553Z [DEBUG] TestFinder.server.raft: vote granted: from=4888874b-c105-6251-a4e2-093d02dd1f15 term=2 tally=1 writer.go:29: 2021-01-29T19:33:47.553Z [INFO] TestFinder.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:47.555Z [INFO] TestFinder.server.raft: entering leader state: leader="Node at 127.0.0.1:17153 [Leader]" writer.go:29: 2021-01-29T19:33:47.556Z [INFO] TestFinder.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:47.558Z [INFO] TestFinder.server: New leader elected: payload=Node-4888874b-c105-6251-a4e2-093d02dd1f15 writer.go:29: 2021-01-29T19:33:47.559Z [DEBUG] TestFinder.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:17153 writer.go:29: 2021-01-29T19:33:47.568Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:47.616Z [INFO] TestFinder: Synced node info writer.go:29: 2021-01-29T19:33:47.619Z [INFO] TestFinder.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:47.620Z [INFO] TestFinder.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.621Z [DEBUG] TestFinder.server: Skipping self join check for node since the cluster is too small: node=Node-4888874b-c105-6251-a4e2-093d02dd1f15 writer.go:29: 2021-01-29T19:33:47.623Z [INFO] TestFinder.server: member joined, marking health alive: member=Node-4888874b-c105-6251-a4e2-093d02dd1f15 writer.go:29: 2021-01-29T19:33:47.927Z [DEBUG] TestFinder.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:60146 latency=4.226519ms writer.go:29: 2021-01-29T19:33:47.938Z [DEBUG] TestFinder.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:60146 latency=229.451µs writer.go:29: 2021-01-29T19:33:47.954Z [INFO] TestFinder: Requesting shutdown writer.go:29: 2021-01-29T19:33:47.956Z [INFO] TestFinder.server: shutting down server writer.go:29: 2021-01-29T19:33:47.957Z [DEBUG] TestFinder.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.958Z [WARN] TestFinder.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:47.959Z [DEBUG] TestFinder.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:47.963Z [WARN] TestFinder.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:47.968Z [INFO] TestFinder.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:47.968Z [INFO] TestFinder: consul server down writer.go:29: 2021-01-29T19:33:47.969Z [INFO] TestFinder: shutdown complete writer.go:29: 2021-01-29T19:33:47.970Z [INFO] TestFinder: Stopping server: protocol=DNS address=127.0.0.1:17148 network=tcp writer.go:29: 2021-01-29T19:33:47.971Z [INFO] TestFinder: Stopping server: protocol=DNS address=127.0.0.1:17148 network=udp writer.go:29: 2021-01-29T19:33:47.972Z [INFO] TestFinder: Stopping server: protocol=HTTP address=127.0.0.1:17149 network=tcp writer.go:29: 2021-01-29T19:33:48.473Z [INFO] TestFinder: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:48.474Z [INFO] TestFinder: Endpoints down --- PASS: TestFinder (1.07s) PASS ok github.com/hashicorp/consul/command/intention/finder 1.540s === RUN TestCommand_noTabs === PAUSE TestCommand_noTabs === RUN TestCommand_Validation === PAUSE TestCommand_Validation === RUN TestCommand_id === PAUSE TestCommand_id === RUN TestCommand_srcDst === PAUSE TestCommand_srcDst === RUN TestCommand_verticalBar === PAUSE TestCommand_verticalBar === CONT TestCommand_noTabs === CONT TestCommand_srcDst === CONT TestCommand_id === CONT TestCommand_Validation === RUN TestCommand_Validation/0_args --- PASS: TestCommand_noTabs (0.00s) === CONT TestCommand_verticalBar [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === RUN TestCommand_Validation/3_args [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestCommand_Validation (0.01s) --- PASS: TestCommand_Validation/0_args (0.00s) --- PASS: TestCommand_Validation/3_args (0.01s) === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.579Z [WARN] TestCommand_id: bootstrap = true: do not enable unless necessary === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.582Z [WARN] TestCommand_srcDst: bootstrap = true: do not enable unless necessary === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.581Z [DEBUG] TestCommand_id.tlsutil: Update: version=1 === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.583Z [DEBUG] TestCommand_srcDst.tlsutil: Update: version=1 === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.588Z [DEBUG] TestCommand_id.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.588Z [DEBUG] TestCommand_srcDst.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.601Z [INFO] TestCommand_id.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:27a95979-ef13-352e-699f-aaaacd2eb93f Address:127.0.0.1:24300}]" === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.601Z [WARN] TestCommand_verticalBar: bootstrap = true: do not enable unless necessary === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.603Z [INFO] TestCommand_id.server.serf.wan: serf: EventMemberJoin: Node-27a95979-ef13-352e-699f-aaaacd2eb93f.dc1 127.0.0.1 === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.603Z [DEBUG] TestCommand_verticalBar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:49.605Z [DEBUG] TestCommand_verticalBar.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.603Z [INFO] TestCommand_id.server.raft: entering follower state: follower="Node at 127.0.0.1:24300 [Follower]" leader= === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.608Z [INFO] TestCommand_srcDst.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:834ea607-a8c0-4bfe-2fe1-f62bd158c7d8 Address:127.0.0.1:24306}]" === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.605Z [INFO] TestCommand_id.server.serf.lan: serf: EventMemberJoin: Node-27a95979-ef13-352e-699f-aaaacd2eb93f 127.0.0.1 === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.609Z [INFO] TestCommand_verticalBar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ff553f7c-b952-30e7-bd36-6e524bbb992e Address:127.0.0.1:24312}]" === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.610Z [INFO] TestCommand_srcDst.server.serf.wan: serf: EventMemberJoin: Node-834ea607-a8c0-4bfe-2fe1-f62bd158c7d8.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:49.612Z [INFO] TestCommand_srcDst.server.serf.lan: serf: EventMemberJoin: Node-834ea607-a8c0-4bfe-2fe1-f62bd158c7d8 127.0.0.1 === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.612Z [INFO] TestCommand_id: Started DNS server: address=127.0.0.1:24295 network=udp === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.613Z [INFO] TestCommand_verticalBar.server.serf.wan: serf: EventMemberJoin: Node-ff553f7c-b952-30e7-bd36-6e524bbb992e.dc1 127.0.0.1 === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.614Z [INFO] TestCommand_id.server: Adding LAN server: server="Node-27a95979-ef13-352e-699f-aaaacd2eb93f (Addr: tcp/127.0.0.1:24300) (DC: dc1)" === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.615Z [INFO] TestCommand_srcDst: Started DNS server: address=127.0.0.1:24301 network=udp === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.616Z [INFO] TestCommand_id.server: Handled event for server in area: event=member-join server=Node-27a95979-ef13-352e-699f-aaaacd2eb93f.dc1 area=wan === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.616Z [INFO] TestCommand_verticalBar.server.serf.lan: serf: EventMemberJoin: Node-ff553f7c-b952-30e7-bd36-6e524bbb992e 127.0.0.1 === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.619Z [INFO] TestCommand_id: Started DNS server: address=127.0.0.1:24295 network=tcp === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.621Z [INFO] TestCommand_srcDst: Started DNS server: address=127.0.0.1:24301 network=tcp === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.625Z [INFO] TestCommand_id: Started HTTP server: address=127.0.0.1:24296 network=tcp === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.626Z [INFO] TestCommand_srcDst: Started HTTP server: address=127.0.0.1:24302 network=tcp writer.go:29: 2021-01-29T19:33:49.617Z [INFO] TestCommand_srcDst.server.raft: entering follower state: follower="Node at 127.0.0.1:24306 [Follower]" leader= writer.go:29: 2021-01-29T19:33:49.619Z [INFO] TestCommand_srcDst.server: Adding LAN server: server="Node-834ea607-a8c0-4bfe-2fe1-f62bd158c7d8 (Addr: tcp/127.0.0.1:24306) (DC: dc1)" === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.627Z [INFO] TestCommand_id: started state syncer === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.628Z [INFO] TestCommand_verticalBar: Started DNS server: address=127.0.0.1:24307 network=udp === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.619Z [INFO] TestCommand_srcDst.server: Handled event for server in area: event=member-join server=Node-834ea607-a8c0-4bfe-2fe1-f62bd158c7d8.dc1 area=wan writer.go:29: 2021-01-29T19:33:49.642Z [INFO] TestCommand_srcDst: started state syncer === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.619Z [INFO] TestCommand_verticalBar.server.raft: entering follower state: follower="Node at 127.0.0.1:24312 [Follower]" leader= writer.go:29: 2021-01-29T19:33:49.630Z [INFO] TestCommand_verticalBar.server: Adding LAN server: server="Node-ff553f7c-b952-30e7-bd36-6e524bbb992e (Addr: tcp/127.0.0.1:24312) (DC: dc1)" writer.go:29: 2021-01-29T19:33:49.631Z [INFO] TestCommand_verticalBar.server: Handled event for server in area: event=member-join server=Node-ff553f7c-b952-30e7-bd36-6e524bbb992e.dc1 area=wan writer.go:29: 2021-01-29T19:33:49.644Z [INFO] TestCommand_verticalBar: Started DNS server: address=127.0.0.1:24307 network=tcp writer.go:29: 2021-01-29T19:33:49.668Z [INFO] TestCommand_verticalBar: Started HTTP server: address=127.0.0.1:24308 network=tcp writer.go:29: 2021-01-29T19:33:49.669Z [INFO] TestCommand_verticalBar: started state syncer === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.671Z [WARN] TestCommand_id.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:49.683Z [INFO] TestCommand_id.server.raft: entering candidate state: node="Node at 127.0.0.1:24300 [Candidate]" term=2 === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.690Z [WARN] TestCommand_verticalBar.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.690Z [DEBUG] TestCommand_id.server.raft: votes: needed=1 === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.692Z [INFO] TestCommand_verticalBar.server.raft: entering candidate state: node="Node at 127.0.0.1:24312 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:49.695Z [DEBUG] TestCommand_verticalBar.server.raft: votes: needed=1 === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.694Z [DEBUG] TestCommand_id.server.raft: vote granted: from=27a95979-ef13-352e-699f-aaaacd2eb93f term=2 tally=1 === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.696Z [DEBUG] TestCommand_verticalBar.server.raft: vote granted: from=ff553f7c-b952-30e7-bd36-6e524bbb992e term=2 tally=1 writer.go:29: 2021-01-29T19:33:49.698Z [INFO] TestCommand_verticalBar.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:49.700Z [INFO] TestCommand_verticalBar.server.raft: entering leader state: leader="Node at 127.0.0.1:24312 [Leader]" === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.698Z [INFO] TestCommand_id.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:49.702Z [INFO] TestCommand_id.server.raft: entering leader state: leader="Node at 127.0.0.1:24300 [Leader]" === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.702Z [WARN] TestCommand_srcDst.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:49.705Z [INFO] TestCommand_srcDst.server.raft: entering candidate state: node="Node at 127.0.0.1:24306 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:49.709Z [DEBUG] TestCommand_srcDst.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:49.711Z [DEBUG] TestCommand_srcDst.server.raft: vote granted: from=834ea607-a8c0-4bfe-2fe1-f62bd158c7d8 term=2 tally=1 writer.go:29: 2021-01-29T19:33:49.712Z [INFO] TestCommand_srcDst.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:49.714Z [INFO] TestCommand_srcDst.server.raft: entering leader state: leader="Node at 127.0.0.1:24306 [Leader]" === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.705Z [INFO] TestCommand_id.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:49.709Z [INFO] TestCommand_id.server: New leader elected: payload=Node-27a95979-ef13-352e-699f-aaaacd2eb93f === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.716Z [INFO] TestCommand_srcDst.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:49.716Z [INFO] TestCommand_srcDst.server: New leader elected: payload=Node-834ea607-a8c0-4bfe-2fe1-f62bd158c7d8 === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.709Z [DEBUG] TestCommand_id.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:24300 === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.721Z [DEBUG] TestCommand_srcDst.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:24306 === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.702Z [INFO] TestCommand_verticalBar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:49.707Z [DEBUG] TestCommand_verticalBar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:24312 writer.go:29: 2021-01-29T19:33:49.708Z [INFO] TestCommand_verticalBar.server: New leader elected: payload=Node-ff553f7c-b952-30e7-bd36-6e524bbb992e === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.735Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.741Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.750Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:49.765Z [DEBUG] TestCommand_id.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:39118 latency=20.786985ms writer.go:29: 2021-01-29T19:33:49.773Z [DEBUG] TestCommand_id.http: Request finished: method=GET url=/v1/connect/intentions/f996718c-f95b-9496-b26e-0afa984dc8f0 from=127.0.0.1:39120 latency=315.003µs writer.go:29: 2021-01-29T19:33:49.779Z [INFO] TestCommand_id: Requesting shutdown writer.go:29: 2021-01-29T19:33:49.780Z [INFO] TestCommand_id.server: shutting down server writer.go:29: 2021-01-29T19:33:49.782Z [WARN] TestCommand_id.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:49.782Z [ERROR] TestCommand_id.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:33:49.784Z [WARN] TestCommand_id.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:49.798Z [INFO] TestCommand_id.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:49.798Z [INFO] TestCommand_id: consul server down writer.go:29: 2021-01-29T19:33:49.801Z [INFO] TestCommand_id: shutdown complete writer.go:29: 2021-01-29T19:33:49.802Z [INFO] TestCommand_id: Stopping server: protocol=DNS address=127.0.0.1:24295 network=tcp writer.go:29: 2021-01-29T19:33:49.802Z [INFO] TestCommand_id: Stopping server: protocol=DNS address=127.0.0.1:24295 network=udp writer.go:29: 2021-01-29T19:33:49.803Z [INFO] TestCommand_id: Stopping server: protocol=HTTP address=127.0.0.1:24296 network=tcp writer.go:29: 2021-01-29T19:33:49.806Z [ERROR] TestCommand_id.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: raft is already shutdown" writer.go:29: 2021-01-29T19:33:49.808Z [ERROR] TestCommand_id.server: failed to transfer leadership attempt, will retry: attempt=0 retry_limit=3 error="raft is already shutdown" === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.809Z [INFO] TestCommand_verticalBar.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:49.810Z [INFO] TestCommand_verticalBar.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.811Z [DEBUG] TestCommand_verticalBar.server: Skipping self join check for node since the cluster is too small: node=Node-ff553f7c-b952-30e7-bd36-6e524bbb992e writer.go:29: 2021-01-29T19:33:49.812Z [INFO] TestCommand_verticalBar.server: member joined, marking health alive: member=Node-ff553f7c-b952-30e7-bd36-6e524bbb992e === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.819Z [INFO] TestCommand_srcDst.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:49.820Z [INFO] TestCommand_srcDst.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.823Z [DEBUG] TestCommand_srcDst.server: Skipping self join check for node since the cluster is too small: node=Node-834ea607-a8c0-4bfe-2fe1-f62bd158c7d8 writer.go:29: 2021-01-29T19:33:49.825Z [INFO] TestCommand_srcDst.server: member joined, marking health alive: member=Node-834ea607-a8c0-4bfe-2fe1-f62bd158c7d8 writer.go:29: 2021-01-29T19:33:49.851Z [DEBUG] TestCommand_srcDst: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.854Z [DEBUG] TestCommand_verticalBar.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:51358 latency=3.015923ms === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.878Z [INFO] TestCommand_srcDst: Synced node info === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.923Z [DEBUG] TestCommand_verticalBar: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:49.940Z [INFO] TestCommand_verticalBar: Synced node info writer.go:29: 2021-01-29T19:33:49.965Z [DEBUG] TestCommand_verticalBar.http: Request finished: method=GET url=/v1/connect/intentions/655c4d62-c594-77e4-5e91-a177cb6394dd from=127.0.0.1:51360 latency=99.827992ms === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.967Z [DEBUG] TestCommand_srcDst.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:47398 latency=2.313934ms === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.970Z [INFO] TestCommand_verticalBar: Requesting shutdown === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.971Z [DEBUG] TestCommand_srcDst.http: Request finished: method=GET url=/v1/connect/intentions from=127.0.0.1:47400 latency=360.648µs writer.go:29: 2021-01-29T19:33:49.973Z [DEBUG] TestCommand_srcDst.http: Request finished: method=GET url=/v1/connect/intentions/3f2c7a0e-3810-6cf4-84c8-5e2e2c5c4a74 from=127.0.0.1:47400 latency=159.18µs === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.972Z [INFO] TestCommand_verticalBar.server: shutting down server writer.go:29: 2021-01-29T19:33:49.975Z [DEBUG] TestCommand_verticalBar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.976Z [WARN] TestCommand_verticalBar.server.serf.lan: serf: Shutdown without a Leave === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.977Z [INFO] TestCommand_srcDst: Requesting shutdown === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.978Z [DEBUG] TestCommand_verticalBar.leader: stopped routine: routine="CA root pruning" === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.979Z [INFO] TestCommand_srcDst.server: shutting down server === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.981Z [WARN] TestCommand_verticalBar.server.serf.wan: serf: Shutdown without a Leave === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.982Z [DEBUG] TestCommand_srcDst.leader: stopping routine: routine="CA root pruning" === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.985Z [INFO] TestCommand_verticalBar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:49.985Z [INFO] TestCommand_verticalBar: consul server down writer.go:29: 2021-01-29T19:33:49.988Z [INFO] TestCommand_verticalBar: shutdown complete writer.go:29: 2021-01-29T19:33:49.989Z [INFO] TestCommand_verticalBar: Stopping server: protocol=DNS address=127.0.0.1:24307 network=tcp === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.986Z [WARN] TestCommand_srcDst.server.serf.lan: serf: Shutdown without a Leave === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:49.990Z [INFO] TestCommand_verticalBar: Stopping server: protocol=DNS address=127.0.0.1:24307 network=udp writer.go:29: 2021-01-29T19:33:49.993Z [INFO] TestCommand_verticalBar: Stopping server: protocol=HTTP address=127.0.0.1:24308 network=tcp === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:49.986Z [DEBUG] TestCommand_srcDst.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:49.993Z [WARN] TestCommand_srcDst.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:50.001Z [INFO] TestCommand_srcDst.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:50.001Z [INFO] TestCommand_srcDst: consul server down writer.go:29: 2021-01-29T19:33:50.004Z [INFO] TestCommand_srcDst: shutdown complete writer.go:29: 2021-01-29T19:33:50.006Z [INFO] TestCommand_srcDst: Stopping server: protocol=DNS address=127.0.0.1:24301 network=tcp writer.go:29: 2021-01-29T19:33:50.007Z [INFO] TestCommand_srcDst: Stopping server: protocol=DNS address=127.0.0.1:24301 network=udp writer.go:29: 2021-01-29T19:33:50.009Z [INFO] TestCommand_srcDst: Stopping server: protocol=HTTP address=127.0.0.1:24302 network=tcp === CONT TestCommand_id writer.go:29: 2021-01-29T19:33:50.304Z [INFO] TestCommand_id: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:50.305Z [INFO] TestCommand_id: Endpoints down --- PASS: TestCommand_id (0.77s) === CONT TestCommand_verticalBar writer.go:29: 2021-01-29T19:33:50.494Z [INFO] TestCommand_verticalBar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:50.496Z [INFO] TestCommand_verticalBar: Endpoints down --- PASS: TestCommand_verticalBar (0.96s) === CONT TestCommand_srcDst writer.go:29: 2021-01-29T19:33:50.510Z [INFO] TestCommand_srcDst: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:50.511Z [INFO] TestCommand_srcDst: Endpoints down --- PASS: TestCommand_srcDst (0.98s) PASS ok github.com/hashicorp/consul/command/intention/get 1.252s === RUN TestCommand_noTabs === PAUSE TestCommand_noTabs === RUN TestCommand_Validation === PAUSE TestCommand_Validation === RUN TestCommand_matchDst === PAUSE TestCommand_matchDst === RUN TestCommand_matchSource === PAUSE TestCommand_matchSource === CONT TestCommand_noTabs --- PASS: TestCommand_noTabs (0.00s) === CONT TestCommand_matchSource === CONT TestCommand_matchDst [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestCommand_Validation === RUN TestCommand_Validation/0_args [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === RUN TestCommand_Validation/3_args === RUN TestCommand_Validation/both_source_and_dest --- PASS: TestCommand_Validation (0.01s) --- PASS: TestCommand_Validation/0_args (0.00s) --- PASS: TestCommand_Validation/3_args (0.00s) --- PASS: TestCommand_Validation/both_source_and_dest (0.00s) === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.788Z [WARN] TestCommand_matchSource: bootstrap = true: do not enable unless necessary === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.793Z [WARN] TestCommand_matchDst: bootstrap = true: do not enable unless necessary === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.790Z [DEBUG] TestCommand_matchSource.tlsutil: Update: version=1 === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.795Z [DEBUG] TestCommand_matchDst.tlsutil: Update: version=1 === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.798Z [DEBUG] TestCommand_matchSource.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.803Z [DEBUG] TestCommand_matchDst.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:54.819Z [INFO] TestCommand_matchDst.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2b17ab7f-3812-e638-6c0a-083b3f29aa64 Address:127.0.0.1:21243}]" writer.go:29: 2021-01-29T19:33:54.822Z [INFO] TestCommand_matchDst.server.raft: entering follower state: follower="Node at 127.0.0.1:21243 [Follower]" leader= === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.829Z [INFO] TestCommand_matchSource.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0cf03645-373e-1a03-89fb-b07e37e951dd Address:127.0.0.1:21237}]" writer.go:29: 2021-01-29T19:33:54.835Z [INFO] TestCommand_matchSource.server.raft: entering follower state: follower="Node at 127.0.0.1:21237 [Follower]" leader= === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.831Z [INFO] TestCommand_matchDst.server.serf.wan: serf: EventMemberJoin: Node-2b17ab7f-3812-e638-6c0a-083b3f29aa64.dc1 127.0.0.1 === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.847Z [INFO] TestCommand_matchSource.server.serf.wan: serf: EventMemberJoin: Node-0cf03645-373e-1a03-89fb-b07e37e951dd.dc1 127.0.0.1 === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.852Z [INFO] TestCommand_matchDst.server.serf.lan: serf: EventMemberJoin: Node-2b17ab7f-3812-e638-6c0a-083b3f29aa64 127.0.0.1 === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.861Z [INFO] TestCommand_matchSource.server.serf.lan: serf: EventMemberJoin: Node-0cf03645-373e-1a03-89fb-b07e37e951dd 127.0.0.1 writer.go:29: 2021-01-29T19:33:54.866Z [INFO] TestCommand_matchSource: Started DNS server: address=127.0.0.1:21232 network=udp === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.860Z [INFO] TestCommand_matchDst: Started DNS server: address=127.0.0.1:21238 network=udp writer.go:29: 2021-01-29T19:33:54.879Z [INFO] TestCommand_matchDst: Started DNS server: address=127.0.0.1:21238 network=tcp writer.go:29: 2021-01-29T19:33:54.882Z [INFO] TestCommand_matchDst.server: Adding LAN server: server="Node-2b17ab7f-3812-e638-6c0a-083b3f29aa64 (Addr: tcp/127.0.0.1:21243) (DC: dc1)" writer.go:29: 2021-01-29T19:33:54.867Z [INFO] TestCommand_matchDst.server: Handled event for server in area: event=member-join server=Node-2b17ab7f-3812-e638-6c0a-083b3f29aa64.dc1 area=wan === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.888Z [INFO] TestCommand_matchSource.server: Handled event for server in area: event=member-join server=Node-0cf03645-373e-1a03-89fb-b07e37e951dd.dc1 area=wan writer.go:29: 2021-01-29T19:33:54.889Z [INFO] TestCommand_matchSource.server: Adding LAN server: server="Node-0cf03645-373e-1a03-89fb-b07e37e951dd (Addr: tcp/127.0.0.1:21237) (DC: dc1)" === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.892Z [INFO] TestCommand_matchDst: Started HTTP server: address=127.0.0.1:21239 network=tcp writer.go:29: 2021-01-29T19:33:54.893Z [WARN] TestCommand_matchDst.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:54.897Z [INFO] TestCommand_matchDst.server.raft: entering candidate state: node="Node at 127.0.0.1:21243 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:54.896Z [INFO] TestCommand_matchDst: started state syncer === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.893Z [INFO] TestCommand_matchSource: Started DNS server: address=127.0.0.1:21232 network=tcp === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.901Z [DEBUG] TestCommand_matchDst.server.raft: votes: needed=1 === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.900Z [WARN] TestCommand_matchSource.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:54.909Z [INFO] TestCommand_matchSource.server.raft: entering candidate state: node="Node at 127.0.0.1:21237 [Candidate]" term=2 === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.908Z [DEBUG] TestCommand_matchDst.server.raft: vote granted: from=2b17ab7f-3812-e638-6c0a-083b3f29aa64 term=2 tally=1 === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.909Z [INFO] TestCommand_matchSource: Started HTTP server: address=127.0.0.1:21233 network=tcp writer.go:29: 2021-01-29T19:33:54.911Z [DEBUG] TestCommand_matchSource.server.raft: votes: needed=1 === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.911Z [INFO] TestCommand_matchDst.server.raft: election won: tally=1 === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.913Z [DEBUG] TestCommand_matchSource.server.raft: vote granted: from=0cf03645-373e-1a03-89fb-b07e37e951dd term=2 tally=1 writer.go:29: 2021-01-29T19:33:54.914Z [INFO] TestCommand_matchSource.server.raft: election won: tally=1 === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.914Z [INFO] TestCommand_matchDst.server.raft: entering leader state: leader="Node at 127.0.0.1:21243 [Leader]" === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.916Z [INFO] TestCommand_matchSource.server.raft: entering leader state: leader="Node at 127.0.0.1:21237 [Leader]" === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.916Z [INFO] TestCommand_matchDst.server: cluster leadership acquired === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.917Z [INFO] TestCommand_matchSource.server: cluster leadership acquired === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.918Z [INFO] TestCommand_matchDst.server: New leader elected: payload=Node-2b17ab7f-3812-e638-6c0a-083b3f29aa64 === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.919Z [INFO] TestCommand_matchSource.server: New leader elected: payload=Node-0cf03645-373e-1a03-89fb-b07e37e951dd writer.go:29: 2021-01-29T19:33:54.920Z [INFO] TestCommand_matchSource: started state syncer writer.go:29: 2021-01-29T19:33:54.922Z [DEBUG] TestCommand_matchSource.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:21237 === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.922Z [DEBUG] TestCommand_matchDst.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:21243 === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.933Z [INFO] TestCommand_matchSource: Synced node info writer.go:29: 2021-01-29T19:33:54.935Z [DEBUG] TestCommand_matchSource: Node info in sync writer.go:29: 2021-01-29T19:33:54.934Z [DEBUG] TestCommand_matchSource.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:43238 latency=4.969052ms writer.go:29: 2021-01-29T19:33:54.935Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:54.946Z [DEBUG] TestCommand_matchSource.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:43238 latency=3.263894ms === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:54.948Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:54.954Z [DEBUG] TestCommand_matchSource.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:43238 latency=794.934µs writer.go:29: 2021-01-29T19:33:54.960Z [DEBUG] TestCommand_matchSource.http: Request finished: method=GET url=/v1/connect/intentions/match?by=source&name=foo from=127.0.0.1:43240 latency=2.104324ms writer.go:29: 2021-01-29T19:33:54.965Z [INFO] TestCommand_matchSource: Requesting shutdown writer.go:29: 2021-01-29T19:33:54.967Z [INFO] TestCommand_matchSource.server: shutting down server writer.go:29: 2021-01-29T19:33:54.968Z [WARN] TestCommand_matchSource.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:54.970Z [WARN] TestCommand_matchSource.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:54.972Z [INFO] TestCommand_matchSource.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:54.972Z [INFO] TestCommand_matchSource: consul server down writer.go:29: 2021-01-29T19:33:54.974Z [INFO] TestCommand_matchSource: shutdown complete writer.go:29: 2021-01-29T19:33:54.975Z [INFO] TestCommand_matchSource: Stopping server: protocol=DNS address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:33:54.978Z [INFO] TestCommand_matchSource: Stopping server: protocol=DNS address=127.0.0.1:21232 network=udp writer.go:29: 2021-01-29T19:33:54.980Z [INFO] TestCommand_matchSource: Stopping server: protocol=HTTP address=127.0.0.1:21233 network=tcp === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:55.001Z [DEBUG] TestCommand_matchDst.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:48818 latency=2.62513ms === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:55.011Z [ERROR] TestCommand_matchSource.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: raft is already shutdown" writer.go:29: 2021-01-29T19:33:55.014Z [ERROR] TestCommand_matchSource.server: failed to transfer leadership attempt, will retry: attempt=0 retry_limit=3 error="raft is already shutdown" === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:55.013Z [DEBUG] TestCommand_matchDst.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:48818 latency=3.444493ms === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:55.015Z [ERROR] TestCommand_matchSource.server: failed to transfer leadership attempt, will retry: attempt=1 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:55.016Z [ERROR] TestCommand_matchSource.server: failed to transfer leadership attempt, will retry: attempt=2 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:33:55.017Z [ERROR] TestCommand_matchSource.server: failed to transfer leadership: error="failed to transfer leadership in 3 attempts" === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:55.030Z [INFO] TestCommand_matchDst: Synced node info writer.go:29: 2021-01-29T19:33:55.033Z [DEBUG] TestCommand_matchDst.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:48818 latency=15.363115ms writer.go:29: 2021-01-29T19:33:55.034Z [INFO] TestCommand_matchDst.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:55.037Z [INFO] TestCommand_matchDst.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:55.039Z [DEBUG] TestCommand_matchDst.server: Skipping self join check for node since the cluster is too small: node=Node-2b17ab7f-3812-e638-6c0a-083b3f29aa64 writer.go:29: 2021-01-29T19:33:55.041Z [INFO] TestCommand_matchDst.server: member joined, marking health alive: member=Node-2b17ab7f-3812-e638-6c0a-083b3f29aa64 writer.go:29: 2021-01-29T19:33:55.040Z [DEBUG] TestCommand_matchDst.http: Request finished: method=GET url=/v1/connect/intentions/match?by=destination&name=db from=127.0.0.1:48820 latency=135.713µs writer.go:29: 2021-01-29T19:33:55.048Z [INFO] TestCommand_matchDst: Requesting shutdown writer.go:29: 2021-01-29T19:33:55.049Z [INFO] TestCommand_matchDst.server: shutting down server writer.go:29: 2021-01-29T19:33:55.050Z [DEBUG] TestCommand_matchDst.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:55.051Z [WARN] TestCommand_matchDst.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:55.053Z [DEBUG] TestCommand_matchDst.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:55.065Z [WARN] TestCommand_matchDst.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:55.072Z [INFO] TestCommand_matchDst.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:55.080Z [INFO] TestCommand_matchDst: consul server down writer.go:29: 2021-01-29T19:33:55.083Z [INFO] TestCommand_matchDst: shutdown complete writer.go:29: 2021-01-29T19:33:55.086Z [INFO] TestCommand_matchDst: Stopping server: protocol=DNS address=127.0.0.1:21238 network=tcp writer.go:29: 2021-01-29T19:33:55.090Z [INFO] TestCommand_matchDst: Stopping server: protocol=DNS address=127.0.0.1:21238 network=udp writer.go:29: 2021-01-29T19:33:55.092Z [INFO] TestCommand_matchDst: Stopping server: protocol=HTTP address=127.0.0.1:21239 network=tcp === CONT TestCommand_matchSource writer.go:29: 2021-01-29T19:33:55.483Z [INFO] TestCommand_matchSource: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:55.485Z [INFO] TestCommand_matchSource: Endpoints down --- PASS: TestCommand_matchSource (0.75s) === CONT TestCommand_matchDst writer.go:29: 2021-01-29T19:33:55.597Z [INFO] TestCommand_matchDst: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:55.599Z [INFO] TestCommand_matchDst: Endpoints down --- PASS: TestCommand_matchDst (0.86s) PASS ok github.com/hashicorp/consul/command/intention/match 1.106s === RUN TestJoinCommand_noTabs === PAUSE TestJoinCommand_noTabs === RUN TestJoinCommandJoin_lan === PAUSE TestJoinCommandJoin_lan === RUN TestJoinCommand_wan === PAUSE TestJoinCommand_wan === RUN TestJoinCommand_noAddrs === PAUSE TestJoinCommand_noAddrs === CONT TestJoinCommand_noTabs === CONT TestJoinCommand_noAddrs --- PASS: TestJoinCommand_noAddrs (0.00s) === CONT TestJoinCommand_wan [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestJoinCommand_noTabs (0.00s) === CONT TestJoinCommandJoin_lan [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.165Z [WARN] TestJoinCommand_wan: bootstrap = true: do not enable unless necessary === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.166Z [WARN] TestJoinCommandJoin_lan: bootstrap = true: do not enable unless necessary === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.167Z [DEBUG] TestJoinCommand_wan.tlsutil: Update: version=1 === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.168Z [DEBUG] TestJoinCommandJoin_lan.tlsutil: Update: version=1 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.168Z [DEBUG] TestJoinCommand_wan.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.170Z [DEBUG] TestJoinCommandJoin_lan.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:56.180Z [INFO] TestJoinCommandJoin_lan.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:694288f5-a19c-40b6-3482-f849e2c4f719 Address:127.0.0.1:11033}]" writer.go:29: 2021-01-29T19:33:56.183Z [INFO] TestJoinCommandJoin_lan.server.raft: entering follower state: follower="Node at 127.0.0.1:11033 [Follower]" leader= === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.195Z [INFO] TestJoinCommand_wan.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:98c31cd7-4b86-8dd1-5459-f927adc3bfc0 Address:127.0.0.1:11027}]" writer.go:29: 2021-01-29T19:33:56.198Z [INFO] TestJoinCommand_wan.server.raft: entering follower state: follower="Node at 127.0.0.1:11027 [Follower]" leader= writer.go:29: 2021-01-29T19:33:56.199Z [INFO] TestJoinCommand_wan.server.serf.wan: serf: EventMemberJoin: Node-98c31cd7-4b86-8dd1-5459-f927adc3bfc0.dc1 127.0.0.1 === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.198Z [INFO] TestJoinCommandJoin_lan.server.serf.wan: serf: EventMemberJoin: Node-694288f5-a19c-40b6-3482-f849e2c4f719.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:56.204Z [INFO] TestJoinCommandJoin_lan.server.serf.lan: serf: EventMemberJoin: Node-694288f5-a19c-40b6-3482-f849e2c4f719 127.0.0.1 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.204Z [INFO] TestJoinCommand_wan.server.serf.lan: serf: EventMemberJoin: Node-98c31cd7-4b86-8dd1-5459-f927adc3bfc0 127.0.0.1 === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.206Z [INFO] TestJoinCommandJoin_lan: Started DNS server: address=127.0.0.1:11028 network=udp writer.go:29: 2021-01-29T19:33:56.208Z [INFO] TestJoinCommandJoin_lan.server: Adding LAN server: server="Node-694288f5-a19c-40b6-3482-f849e2c4f719 (Addr: tcp/127.0.0.1:11033) (DC: dc1)" === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.208Z [INFO] TestJoinCommand_wan: Started DNS server: address=127.0.0.1:11022 network=udp writer.go:29: 2021-01-29T19:33:56.210Z [INFO] TestJoinCommand_wan.server: Adding LAN server: server="Node-98c31cd7-4b86-8dd1-5459-f927adc3bfc0 (Addr: tcp/127.0.0.1:11027) (DC: dc1)" === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.208Z [INFO] TestJoinCommandJoin_lan: Started DNS server: address=127.0.0.1:11028 network=tcp === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.211Z [INFO] TestJoinCommand_wan.server: Handled event for server in area: event=member-join server=Node-98c31cd7-4b86-8dd1-5459-f927adc3bfc0.dc1 area=wan writer.go:29: 2021-01-29T19:33:56.211Z [INFO] TestJoinCommand_wan: Started DNS server: address=127.0.0.1:11022 network=tcp === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.212Z [INFO] TestJoinCommandJoin_lan.server: Handled event for server in area: event=member-join server=Node-694288f5-a19c-40b6-3482-f849e2c4f719.dc1 area=wan writer.go:29: 2021-01-29T19:33:56.212Z [INFO] TestJoinCommandJoin_lan: Started HTTP server: address=127.0.0.1:11029 network=tcp writer.go:29: 2021-01-29T19:33:56.215Z [INFO] TestJoinCommandJoin_lan: started state syncer === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.214Z [INFO] TestJoinCommand_wan: Started HTTP server: address=127.0.0.1:11023 network=tcp writer.go:29: 2021-01-29T19:33:56.218Z [INFO] TestJoinCommand_wan: started state syncer === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.223Z [WARN] TestJoinCommandJoin_lan.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:56.224Z [INFO] TestJoinCommandJoin_lan.server.raft: entering candidate state: node="Node at 127.0.0.1:11033 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:56.226Z [DEBUG] TestJoinCommandJoin_lan.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:56.227Z [DEBUG] TestJoinCommandJoin_lan.server.raft: vote granted: from=694288f5-a19c-40b6-3482-f849e2c4f719 term=2 tally=1 writer.go:29: 2021-01-29T19:33:56.228Z [INFO] TestJoinCommandJoin_lan.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:56.230Z [INFO] TestJoinCommandJoin_lan.server.raft: entering leader state: leader="Node at 127.0.0.1:11033 [Leader]" writer.go:29: 2021-01-29T19:33:56.231Z [INFO] TestJoinCommandJoin_lan.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:56.232Z [INFO] TestJoinCommandJoin_lan.server: New leader elected: payload=Node-694288f5-a19c-40b6-3482-f849e2c4f719 writer.go:29: 2021-01-29T19:33:56.235Z [DEBUG] TestJoinCommandJoin_lan.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:11033 writer.go:29: 2021-01-29T19:33:56.247Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.262Z [WARN] TestJoinCommand_wan.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:56.263Z [INFO] TestJoinCommand_wan.server.raft: entering candidate state: node="Node at 127.0.0.1:11027 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:56.265Z [DEBUG] TestJoinCommand_wan.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:56.266Z [DEBUG] TestJoinCommand_wan.server.raft: vote granted: from=98c31cd7-4b86-8dd1-5459-f927adc3bfc0 term=2 tally=1 writer.go:29: 2021-01-29T19:33:56.267Z [INFO] TestJoinCommand_wan.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:56.268Z [INFO] TestJoinCommand_wan.server.raft: entering leader state: leader="Node at 127.0.0.1:11027 [Leader]" writer.go:29: 2021-01-29T19:33:56.269Z [INFO] TestJoinCommand_wan.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:56.269Z [INFO] TestJoinCommand_wan.server: New leader elected: payload=Node-98c31cd7-4b86-8dd1-5459-f927adc3bfc0 writer.go:29: 2021-01-29T19:33:56.270Z [DEBUG] TestJoinCommand_wan.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:11027 writer.go:29: 2021-01-29T19:33:56.281Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.297Z [INFO] TestJoinCommandJoin_lan.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.297Z [INFO] TestJoinCommand_wan.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.298Z [INFO] TestJoinCommandJoin_lan.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.299Z [DEBUG] TestJoinCommandJoin_lan.server: Skipping self join check for node since the cluster is too small: node=Node-694288f5-a19c-40b6-3482-f849e2c4f719 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.299Z [INFO] TestJoinCommand_wan.leader: started routine: routine="CA root pruning" === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.300Z [INFO] TestJoinCommandJoin_lan.server: member joined, marking health alive: member=Node-694288f5-a19c-40b6-3482-f849e2c4f719 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.301Z [DEBUG] TestJoinCommand_wan.server: Skipping self join check for node since the cluster is too small: node=Node-98c31cd7-4b86-8dd1-5459-f927adc3bfc0 writer.go:29: 2021-01-29T19:33:56.305Z [INFO] TestJoinCommand_wan.server: member joined, marking health alive: member=Node-98c31cd7-4b86-8dd1-5459-f927adc3bfc0 writer.go:29: 2021-01-29T19:33:56.490Z [DEBUG] TestJoinCommand_wan: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:56.494Z [INFO] TestJoinCommand_wan: Synced node info writer.go:29: 2021-01-29T19:33:56.496Z [DEBUG] TestJoinCommand_wan: Node info in sync === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.507Z [DEBUG] TestJoinCommandJoin_lan: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:56.510Z [INFO] TestJoinCommandJoin_lan: Synced node info writer.go:29: 2021-01-29T19:33:56.511Z [DEBUG] TestJoinCommandJoin_lan: Node info in sync writer.go:29: 2021-01-29T19:33:56.538Z [WARN] TestJoinCommandJoin_lan: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:56.539Z [DEBUG] TestJoinCommandJoin_lan.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:56.551Z [DEBUG] TestJoinCommandJoin_lan.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:56.566Z [INFO] TestJoinCommandJoin_lan.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a4d463ee-8197-e77b-73ee-049e0401da82 Address:127.0.0.1:11039}]" writer.go:29: 2021-01-29T19:33:56.570Z [INFO] TestJoinCommandJoin_lan.server.raft: entering follower state: follower="Node at 127.0.0.1:11039 [Follower]" leader= writer.go:29: 2021-01-29T19:33:56.571Z [INFO] TestJoinCommandJoin_lan.server.serf.wan: serf: EventMemberJoin: Node-a4d463ee-8197-e77b-73ee-049e0401da82.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:56.577Z [INFO] TestJoinCommandJoin_lan.server.serf.lan: serf: EventMemberJoin: Node-a4d463ee-8197-e77b-73ee-049e0401da82 127.0.0.1 writer.go:29: 2021-01-29T19:33:56.580Z [INFO] TestJoinCommandJoin_lan.server: Handled event for server in area: event=member-join server=Node-a4d463ee-8197-e77b-73ee-049e0401da82.dc1 area=wan writer.go:29: 2021-01-29T19:33:56.581Z [INFO] TestJoinCommandJoin_lan: Started DNS server: address=127.0.0.1:11034 network=udp writer.go:29: 2021-01-29T19:33:56.584Z [INFO] TestJoinCommandJoin_lan: Started DNS server: address=127.0.0.1:11034 network=tcp writer.go:29: 2021-01-29T19:33:56.582Z [INFO] TestJoinCommandJoin_lan.server: Adding LAN server: server="Node-a4d463ee-8197-e77b-73ee-049e0401da82 (Addr: tcp/127.0.0.1:11039) (DC: dc1)" writer.go:29: 2021-01-29T19:33:56.586Z [INFO] TestJoinCommandJoin_lan: Started HTTP server: address=127.0.0.1:11035 network=tcp writer.go:29: 2021-01-29T19:33:56.589Z [INFO] TestJoinCommandJoin_lan: started state syncer === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.626Z [WARN] TestJoinCommand_wan: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:56.631Z [DEBUG] TestJoinCommand_wan.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:56.636Z [DEBUG] TestJoinCommand_wan.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.644Z [WARN] TestJoinCommandJoin_lan.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:56.646Z [INFO] TestJoinCommandJoin_lan.server.raft: entering candidate state: node="Node at 127.0.0.1:11039 [Candidate]" term=2 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.648Z [INFO] TestJoinCommand_wan.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:896a9c77-70b8-256a-0fc6-21a8e93b0ac4 Address:127.0.0.1:11045}]" writer.go:29: 2021-01-29T19:33:56.659Z [INFO] TestJoinCommand_wan.server.serf.wan: serf: EventMemberJoin: Node-896a9c77-70b8-256a-0fc6-21a8e93b0ac4.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:56.664Z [INFO] TestJoinCommand_wan.server.raft: entering follower state: follower="Node at 127.0.0.1:11045 [Follower]" leader= writer.go:29: 2021-01-29T19:33:56.683Z [INFO] TestJoinCommand_wan.server.serf.lan: serf: EventMemberJoin: Node-896a9c77-70b8-256a-0fc6-21a8e93b0ac4 127.0.0.1 === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.684Z [DEBUG] TestJoinCommandJoin_lan.server.raft: votes: needed=1 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.688Z [INFO] TestJoinCommand_wan: Started DNS server: address=127.0.0.1:11040 network=udp writer.go:29: 2021-01-29T19:33:56.688Z [INFO] TestJoinCommand_wan.server: Handled event for server in area: event=member-join server=Node-896a9c77-70b8-256a-0fc6-21a8e93b0ac4.dc1 area=wan writer.go:29: 2021-01-29T19:33:56.693Z [INFO] TestJoinCommand_wan: Started DNS server: address=127.0.0.1:11040 network=tcp writer.go:29: 2021-01-29T19:33:56.695Z [INFO] TestJoinCommand_wan.server: Adding LAN server: server="Node-896a9c77-70b8-256a-0fc6-21a8e93b0ac4 (Addr: tcp/127.0.0.1:11045) (DC: dc1)" === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.686Z [DEBUG] TestJoinCommandJoin_lan.server.raft: vote granted: from=a4d463ee-8197-e77b-73ee-049e0401da82 term=2 tally=1 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.696Z [INFO] TestJoinCommand_wan: Started HTTP server: address=127.0.0.1:11041 network=tcp writer.go:29: 2021-01-29T19:33:56.704Z [INFO] TestJoinCommand_wan: started state syncer === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.703Z [INFO] TestJoinCommandJoin_lan.server.raft: election won: tally=1 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.719Z [WARN] TestJoinCommand_wan.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.711Z [INFO] TestJoinCommandJoin_lan.server.raft: entering leader state: leader="Node at 127.0.0.1:11039 [Leader]" === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.721Z [INFO] TestJoinCommand_wan.server.raft: entering candidate state: node="Node at 127.0.0.1:11045 [Candidate]" term=2 === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.721Z [INFO] TestJoinCommandJoin_lan.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:56.722Z [INFO] TestJoinCommandJoin_lan.server: New leader elected: payload=Node-a4d463ee-8197-e77b-73ee-049e0401da82 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.724Z [DEBUG] TestJoinCommand_wan.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:56.726Z [DEBUG] TestJoinCommand_wan.server.raft: vote granted: from=896a9c77-70b8-256a-0fc6-21a8e93b0ac4 term=2 tally=1 === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.726Z [DEBUG] TestJoinCommandJoin_lan.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:11039 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.728Z [INFO] TestJoinCommand_wan.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:56.729Z [INFO] TestJoinCommand_wan.server.raft: entering leader state: leader="Node at 127.0.0.1:11045 [Leader]" writer.go:29: 2021-01-29T19:33:56.732Z [INFO] TestJoinCommand_wan.server: cluster leadership acquired === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.732Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.732Z [INFO] TestJoinCommand_wan.server: New leader elected: payload=Node-896a9c77-70b8-256a-0fc6-21a8e93b0ac4 writer.go:29: 2021-01-29T19:33:56.736Z [DEBUG] TestJoinCommand_wan.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:11045 === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.738Z [INFO] TestJoinCommandJoin_lan.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:56.741Z [INFO] TestJoinCommandJoin_lan.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.742Z [DEBUG] TestJoinCommandJoin_lan.server: Skipping self join check for node since the cluster is too small: node=Node-a4d463ee-8197-e77b-73ee-049e0401da82 writer.go:29: 2021-01-29T19:33:56.743Z [INFO] TestJoinCommandJoin_lan.server: member joined, marking health alive: member=Node-a4d463ee-8197-e77b-73ee-049e0401da82 === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:56.743Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:56.750Z [INFO] TestJoinCommand_wan.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:56.752Z [INFO] TestJoinCommand_wan.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:56.754Z [DEBUG] TestJoinCommand_wan.server: Skipping self join check for node since the cluster is too small: node=Node-896a9c77-70b8-256a-0fc6-21a8e93b0ac4 writer.go:29: 2021-01-29T19:33:56.755Z [INFO] TestJoinCommand_wan.server: member joined, marking health alive: member=Node-896a9c77-70b8-256a-0fc6-21a8e93b0ac4 writer.go:29: 2021-01-29T19:33:56.846Z [DEBUG] TestJoinCommand_wan: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:56.849Z [INFO] TestJoinCommand_wan: Synced node info === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:56.891Z [DEBUG] TestJoinCommandJoin_lan: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:33:56.895Z [INFO] TestJoinCommandJoin_lan: Synced node info writer.go:29: 2021-01-29T19:33:56.900Z [DEBUG] TestJoinCommandJoin_lan: Node info in sync writer.go:29: 2021-01-29T19:33:57.028Z [INFO] TestJoinCommandJoin_lan: (LAN) joining: lan_addresses=[127.0.0.1:11037] writer.go:29: 2021-01-29T19:33:57.031Z [DEBUG] TestJoinCommandJoin_lan.server.memberlist.lan: memberlist: Stream connection from=127.0.0.1:54810 writer.go:29: 2021-01-29T19:33:57.035Z [DEBUG] TestJoinCommandJoin_lan.server.memberlist.lan: memberlist: Initiating push/pull sync with: 127.0.0.1:11037 writer.go:29: 2021-01-29T19:33:57.039Z [INFO] TestJoinCommandJoin_lan.server.serf.lan: serf: EventMemberJoin: Node-694288f5-a19c-40b6-3482-f849e2c4f719 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.041Z [INFO] TestJoinCommandJoin_lan.server.serf.lan: serf: EventMemberJoin: Node-a4d463ee-8197-e77b-73ee-049e0401da82 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.041Z [INFO] TestJoinCommandJoin_lan.server: Adding LAN server: server="Node-694288f5-a19c-40b6-3482-f849e2c4f719 (Addr: tcp/127.0.0.1:11033) (DC: dc1)" writer.go:29: 2021-01-29T19:33:57.044Z [INFO] TestJoinCommandJoin_lan.server: New leader elected: payload=Node-694288f5-a19c-40b6-3482-f849e2c4f719 writer.go:29: 2021-01-29T19:33:57.045Z [ERROR] TestJoinCommandJoin_lan.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-694288f5-a19c-40b6-3482-f849e2c4f719 other=Node-a4d463ee-8197-e77b-73ee-049e0401da82 writer.go:29: 2021-01-29T19:33:57.046Z [INFO] TestJoinCommandJoin_lan.server: member joined, marking health alive: member=Node-694288f5-a19c-40b6-3482-f849e2c4f719 writer.go:29: 2021-01-29T19:33:57.043Z [INFO] TestJoinCommandJoin_lan.server: Adding LAN server: server="Node-a4d463ee-8197-e77b-73ee-049e0401da82 (Addr: tcp/127.0.0.1:11039) (DC: dc1)" writer.go:29: 2021-01-29T19:33:57.047Z [ERROR] TestJoinCommandJoin_lan.server: Two nodes are in bootstrap mode. Only one node should be in bootstrap mode, not adding Raft peer.: node_to_add=Node-a4d463ee-8197-e77b-73ee-049e0401da82 other=Node-694288f5-a19c-40b6-3482-f849e2c4f719 writer.go:29: 2021-01-29T19:33:57.048Z [INFO] TestJoinCommandJoin_lan.server: member joined, marking health alive: member=Node-a4d463ee-8197-e77b-73ee-049e0401da82 writer.go:29: 2021-01-29T19:33:57.051Z [INFO] TestJoinCommandJoin_lan: (LAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:33:57.055Z [WARN] TestJoinCommandJoin_lan.server.memberlist.wan: memberlist: Failed to resolve Node-694288f5-a19c-40b6-3482-f849e2c4f719.dc1/127.0.0.1:11032: lookup Node-694288f5-a19c-40b6-3482-f849e2c4f719.dc1/127.0.0.1: no such host writer.go:29: 2021-01-29T19:33:57.057Z [DEBUG] TestJoinCommandJoin_lan.server: Failed to flood-join server at address: server=Node-694288f5-a19c-40b6-3482-f849e2c4f719.dc1 address=127.0.0.1:11032 error="1 error occurred: * Failed to resolve Node-694288f5-a19c-40b6-3482-f849e2c4f719.dc1/127.0.0.1:11032: lookup Node-694288f5-a19c-40b6-3482-f849e2c4f719.dc1/127.0.0.1: no such host " writer.go:29: 2021-01-29T19:33:57.057Z [DEBUG] TestJoinCommandJoin_lan: systemd notify failed: error="No socket" writer.go:29: 2021-01-29T19:33:57.060Z [DEBUG] TestJoinCommandJoin_lan.http: Request finished: method=PUT url=/v1/agent/join/127.0.0.1:11037 from=127.0.0.1:39270 latency=32.249882ms writer.go:29: 2021-01-29T19:33:57.057Z [WARN] TestJoinCommandJoin_lan.server.memberlist.wan: memberlist: Failed to resolve Node-a4d463ee-8197-e77b-73ee-049e0401da82.dc1/127.0.0.1:11038: lookup Node-a4d463ee-8197-e77b-73ee-049e0401da82.dc1/127.0.0.1: no such host writer.go:29: 2021-01-29T19:33:57.064Z [DEBUG] TestJoinCommandJoin_lan.server: Failed to flood-join server at address: server=Node-a4d463ee-8197-e77b-73ee-049e0401da82.dc1 address=127.0.0.1:11038 error="1 error occurred: * Failed to resolve Node-a4d463ee-8197-e77b-73ee-049e0401da82.dc1/127.0.0.1:11038: lookup Node-a4d463ee-8197-e77b-73ee-049e0401da82.dc1/127.0.0.1: no such host " writer.go:29: 2021-01-29T19:33:57.065Z [INFO] TestJoinCommandJoin_lan: Requesting shutdown writer.go:29: 2021-01-29T19:33:57.066Z [INFO] TestJoinCommandJoin_lan.server: shutting down server writer.go:29: 2021-01-29T19:33:57.067Z [DEBUG] TestJoinCommandJoin_lan.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.068Z [WARN] TestJoinCommandJoin_lan.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.069Z [DEBUG] TestJoinCommandJoin_lan.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.071Z [WARN] TestJoinCommandJoin_lan.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.074Z [INFO] TestJoinCommandJoin_lan.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:57.075Z [INFO] TestJoinCommandJoin_lan: consul server down writer.go:29: 2021-01-29T19:33:57.078Z [INFO] TestJoinCommandJoin_lan: shutdown complete writer.go:29: 2021-01-29T19:33:57.079Z [INFO] TestJoinCommandJoin_lan: Stopping server: protocol=DNS address=127.0.0.1:11034 network=tcp writer.go:29: 2021-01-29T19:33:57.081Z [INFO] TestJoinCommandJoin_lan: Stopping server: protocol=DNS address=127.0.0.1:11034 network=udp writer.go:29: 2021-01-29T19:33:57.083Z [INFO] TestJoinCommandJoin_lan: Stopping server: protocol=HTTP address=127.0.0.1:11035 network=tcp === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:57.133Z [INFO] TestJoinCommand_wan: (WAN) joining: wan_addresses=[127.0.0.1:11044] writer.go:29: 2021-01-29T19:33:57.139Z [DEBUG] TestJoinCommand_wan.server.memberlist.wan: memberlist: Stream connection from=127.0.0.1:54978 writer.go:29: 2021-01-29T19:33:57.152Z [DEBUG] TestJoinCommand_wan.server.memberlist.wan: memberlist: Initiating push/pull sync with: 127.0.0.1:11044 writer.go:29: 2021-01-29T19:33:57.155Z [INFO] TestJoinCommand_wan.server.serf.wan: serf: EventMemberJoin: Node-98c31cd7-4b86-8dd1-5459-f927adc3bfc0.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.158Z [INFO] TestJoinCommand_wan.server: Handled event for server in area: event=member-join server=Node-98c31cd7-4b86-8dd1-5459-f927adc3bfc0.dc1 area=wan writer.go:29: 2021-01-29T19:33:57.155Z [INFO] TestJoinCommand_wan.server.serf.wan: serf: EventMemberJoin: Node-896a9c77-70b8-256a-0fc6-21a8e93b0ac4.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:57.162Z [INFO] TestJoinCommand_wan.server: Handled event for server in area: event=member-join server=Node-896a9c77-70b8-256a-0fc6-21a8e93b0ac4.dc1 area=wan writer.go:29: 2021-01-29T19:33:57.164Z [INFO] TestJoinCommand_wan: (WAN) joined: number_of_nodes=1 writer.go:29: 2021-01-29T19:33:57.168Z [DEBUG] TestJoinCommand_wan.http: Request finished: method=PUT url=/v1/agent/join/127.0.0.1:11044?wan=1 from=127.0.0.1:54566 latency=35.36445ms writer.go:29: 2021-01-29T19:33:57.172Z [INFO] TestJoinCommand_wan: Requesting shutdown writer.go:29: 2021-01-29T19:33:57.174Z [INFO] TestJoinCommand_wan.server: shutting down server writer.go:29: 2021-01-29T19:33:57.175Z [DEBUG] TestJoinCommand_wan.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.177Z [WARN] TestJoinCommand_wan.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.177Z [DEBUG] TestJoinCommand_wan.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.180Z [WARN] TestJoinCommand_wan.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.183Z [INFO] TestJoinCommand_wan.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:57.184Z [INFO] TestJoinCommand_wan: consul server down writer.go:29: 2021-01-29T19:33:57.187Z [INFO] TestJoinCommand_wan: shutdown complete writer.go:29: 2021-01-29T19:33:57.189Z [INFO] TestJoinCommand_wan: Stopping server: protocol=DNS address=127.0.0.1:11040 network=tcp writer.go:29: 2021-01-29T19:33:57.191Z [INFO] TestJoinCommand_wan: Stopping server: protocol=DNS address=127.0.0.1:11040 network=udp writer.go:29: 2021-01-29T19:33:57.194Z [INFO] TestJoinCommand_wan: Stopping server: protocol=HTTP address=127.0.0.1:11041 network=tcp === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:57.586Z [INFO] TestJoinCommandJoin_lan: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:57.589Z [INFO] TestJoinCommandJoin_lan: Endpoints down writer.go:29: 2021-01-29T19:33:57.590Z [INFO] TestJoinCommandJoin_lan: Requesting shutdown writer.go:29: 2021-01-29T19:33:57.592Z [INFO] TestJoinCommandJoin_lan.server: shutting down server writer.go:29: 2021-01-29T19:33:57.594Z [DEBUG] TestJoinCommandJoin_lan.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.595Z [WARN] TestJoinCommandJoin_lan.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.595Z [DEBUG] TestJoinCommandJoin_lan.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.599Z [WARN] TestJoinCommandJoin_lan.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.601Z [INFO] TestJoinCommandJoin_lan.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:57.602Z [INFO] TestJoinCommandJoin_lan: consul server down writer.go:29: 2021-01-29T19:33:57.606Z [INFO] TestJoinCommandJoin_lan: shutdown complete writer.go:29: 2021-01-29T19:33:57.607Z [INFO] TestJoinCommandJoin_lan: Stopping server: protocol=DNS address=127.0.0.1:11028 network=tcp writer.go:29: 2021-01-29T19:33:57.609Z [INFO] TestJoinCommandJoin_lan: Stopping server: protocol=DNS address=127.0.0.1:11028 network=udp writer.go:29: 2021-01-29T19:33:57.611Z [INFO] TestJoinCommandJoin_lan: Stopping server: protocol=HTTP address=127.0.0.1:11029 network=tcp === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:57.697Z [INFO] TestJoinCommand_wan: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:57.698Z [INFO] TestJoinCommand_wan: Endpoints down writer.go:29: 2021-01-29T19:33:57.700Z [INFO] TestJoinCommand_wan: Requesting shutdown writer.go:29: 2021-01-29T19:33:57.701Z [INFO] TestJoinCommand_wan.server: shutting down server writer.go:29: 2021-01-29T19:33:57.703Z [DEBUG] TestJoinCommand_wan.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.704Z [WARN] TestJoinCommand_wan.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:57.704Z [DEBUG] TestJoinCommand_wan.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:57.712Z [WARN] TestJoinCommand_wan.server.serf.wan: serf: Shutdown without a Leave === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:57.714Z [DEBUG] TestJoinCommandJoin_lan.server.memberlist.lan: memberlist: Failed ping: Node-a4d463ee-8197-e77b-73ee-049e0401da82 (timeout reached) === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:57.715Z [INFO] TestJoinCommand_wan.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:57.715Z [INFO] TestJoinCommand_wan: consul server down writer.go:29: 2021-01-29T19:33:57.719Z [INFO] TestJoinCommand_wan: shutdown complete writer.go:29: 2021-01-29T19:33:57.720Z [INFO] TestJoinCommand_wan: Stopping server: protocol=DNS address=127.0.0.1:11022 network=tcp writer.go:29: 2021-01-29T19:33:57.722Z [INFO] TestJoinCommand_wan: Stopping server: protocol=DNS address=127.0.0.1:11022 network=udp writer.go:29: 2021-01-29T19:33:57.723Z [INFO] TestJoinCommand_wan: Stopping server: protocol=HTTP address=127.0.0.1:11023 network=tcp === CONT TestJoinCommandJoin_lan writer.go:29: 2021-01-29T19:33:58.112Z [INFO] TestJoinCommandJoin_lan: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:58.113Z [INFO] TestJoinCommandJoin_lan: Endpoints down --- PASS: TestJoinCommandJoin_lan (2.00s) === CONT TestJoinCommand_wan writer.go:29: 2021-01-29T19:33:58.230Z [INFO] TestJoinCommand_wan: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:58.231Z [INFO] TestJoinCommand_wan: Endpoints down --- PASS: TestJoinCommand_wan (2.12s) PASS ok github.com/hashicorp/consul/command/join 2.311s === RUN TestKeygenCommand_noTabs === PAUSE TestKeygenCommand_noTabs === RUN TestKeygenCommand === PAUSE TestKeygenCommand === CONT TestKeygenCommand_noTabs --- PASS: TestKeygenCommand_noTabs (0.00s) === CONT TestKeygenCommand --- PASS: TestKeygenCommand (0.00s) PASS ok github.com/hashicorp/consul/command/keygen 0.022s === RUN TestKeyringCommand_noTabs === PAUSE TestKeyringCommand_noTabs === RUN TestKeyringCommand === PAUSE TestKeyringCommand === RUN TestKeyringCommand_help === PAUSE TestKeyringCommand_help === RUN TestKeyringCommand_failedConnection === PAUSE TestKeyringCommand_failedConnection === RUN TestKeyringCommand_invalidLocalOnly === PAUSE TestKeyringCommand_invalidLocalOnly === RUN TestKeyringCommand_invalidRelayFactor === PAUSE TestKeyringCommand_invalidRelayFactor === CONT TestKeyringCommand_noTabs === CONT TestKeyringCommand_invalidLocalOnly === CONT TestKeyringCommand_invalidRelayFactor === CONT TestKeyringCommand_help --- PASS: TestKeyringCommand_noTabs (0.00s) === CONT TestKeyringCommand_failedConnection === CONT TestKeyringCommand --- PASS: TestKeyringCommand_invalidLocalOnly (0.00s) --- PASS: TestKeyringCommand_invalidRelayFactor (0.00s) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestKeyringCommand_failedConnection (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestKeyringCommand_help (0.00s) === CONT TestKeyringCommand writer.go:29: 2021-01-29T19:33:58.635Z [WARN] TestKeyringCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:33:58.637Z [DEBUG] TestKeyringCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:33:58.638Z [DEBUG] TestKeyringCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:58.644Z [INFO] TestKeyringCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e182a0ea-0f64-00f2-8e71-c58949127384 Address:127.0.0.1:20216}]" writer.go:29: 2021-01-29T19:33:58.645Z [INFO] TestKeyringCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:20216 [Follower]" leader= writer.go:29: 2021-01-29T19:33:58.648Z [INFO] TestKeyringCommand.server.serf.wan: serf: EventMemberJoin: Node-e182a0ea-0f64-00f2-8e71-c58949127384.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:33:58.650Z [INFO] TestKeyringCommand.server.serf.lan: serf: EventMemberJoin: Node-e182a0ea-0f64-00f2-8e71-c58949127384 127.0.0.1 writer.go:29: 2021-01-29T19:33:58.653Z [INFO] TestKeyringCommand.server: Adding LAN server: server="Node-e182a0ea-0f64-00f2-8e71-c58949127384 (Addr: tcp/127.0.0.1:20216) (DC: dc1)" writer.go:29: 2021-01-29T19:33:58.654Z [INFO] TestKeyringCommand.server: Handled event for server in area: event=member-join server=Node-e182a0ea-0f64-00f2-8e71-c58949127384.dc1 area=wan writer.go:29: 2021-01-29T19:33:58.654Z [INFO] TestKeyringCommand: Started DNS server: address=127.0.0.1:20211 network=udp writer.go:29: 2021-01-29T19:33:58.658Z [INFO] TestKeyringCommand: Started DNS server: address=127.0.0.1:20211 network=tcp writer.go:29: 2021-01-29T19:33:58.661Z [INFO] TestKeyringCommand: Started HTTP server: address=127.0.0.1:20212 network=tcp writer.go:29: 2021-01-29T19:33:58.663Z [INFO] TestKeyringCommand: started state syncer writer.go:29: 2021-01-29T19:33:58.704Z [WARN] TestKeyringCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:33:58.707Z [INFO] TestKeyringCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:20216 [Candidate]" term=2 writer.go:29: 2021-01-29T19:33:58.710Z [DEBUG] TestKeyringCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:33:58.713Z [DEBUG] TestKeyringCommand.server.raft: vote granted: from=e182a0ea-0f64-00f2-8e71-c58949127384 term=2 tally=1 writer.go:29: 2021-01-29T19:33:58.714Z [INFO] TestKeyringCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:33:58.716Z [INFO] TestKeyringCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:20216 [Leader]" writer.go:29: 2021-01-29T19:33:58.718Z [INFO] TestKeyringCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:33:58.718Z [INFO] TestKeyringCommand.server: New leader elected: payload=Node-e182a0ea-0f64-00f2-8e71-c58949127384 writer.go:29: 2021-01-29T19:33:58.726Z [DEBUG] TestKeyringCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:20216 writer.go:29: 2021-01-29T19:33:58.731Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:33:58.764Z [INFO] TestKeyringCommand: Synced node info writer.go:29: 2021-01-29T19:33:58.767Z [DEBUG] TestKeyringCommand: Node info in sync writer.go:29: 2021-01-29T19:33:58.816Z [INFO] TestKeyringCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:33:58.817Z [INFO] TestKeyringCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:58.819Z [DEBUG] TestKeyringCommand.server: Skipping self join check for node since the cluster is too small: node=Node-e182a0ea-0f64-00f2-8e71-c58949127384 writer.go:29: 2021-01-29T19:33:58.821Z [INFO] TestKeyringCommand.server: member joined, marking health alive: member=Node-e182a0ea-0f64-00f2-8e71-c58949127384 writer.go:29: 2021-01-29T19:33:58.858Z [INFO] TestKeyringCommand.server.serf.wan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:58.864Z [DEBUG] TestKeyringCommand.server.serf.wan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384.dc1 writer.go:29: 2021-01-29T19:33:58.876Z [DEBUG] TestKeyringCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:33:58.898Z [INFO] TestKeyringCommand.server.serf.lan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:58.917Z [DEBUG] TestKeyringCommand.server.serf.lan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384 writer.go:29: 2021-01-29T19:33:58.922Z [DEBUG] TestKeyringCommand.http: Request finished: method=GET url=/v1/operator/keyring from=127.0.0.1:44834 latency=67.609984ms writer.go:29: 2021-01-29T19:33:58.934Z [INFO] TestKeyringCommand.server.serf.wan: serf: Received install-key query writer.go:29: 2021-01-29T19:33:58.951Z [DEBUG] TestKeyringCommand.server.serf.wan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384.dc1 writer.go:29: 2021-01-29T19:33:58.954Z [INFO] TestKeyringCommand.server.serf.lan: serf: Received install-key query writer.go:29: 2021-01-29T19:33:58.956Z [DEBUG] TestKeyringCommand.server.serf.lan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384 writer.go:29: 2021-01-29T19:33:58.958Z [DEBUG] TestKeyringCommand.http: Request finished: method=POST url=/v1/operator/keyring from=127.0.0.1:44836 latency=25.134299ms writer.go:29: 2021-01-29T19:33:58.966Z [INFO] TestKeyringCommand.server.serf.wan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:58.969Z [DEBUG] TestKeyringCommand.server.serf.wan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384.dc1 writer.go:29: 2021-01-29T19:33:58.971Z [INFO] TestKeyringCommand.server.serf.lan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:58.975Z [DEBUG] TestKeyringCommand.server.serf.lan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384 writer.go:29: 2021-01-29T19:33:58.982Z [DEBUG] TestKeyringCommand.http: Request finished: method=GET url=/v1/operator/keyring from=127.0.0.1:44838 latency=17.015256ms writer.go:29: 2021-01-29T19:33:58.996Z [INFO] TestKeyringCommand.server.serf.wan: serf: Received use-key query writer.go:29: 2021-01-29T19:33:59.017Z [DEBUG] TestKeyringCommand.server.serf.wan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384.dc1 writer.go:29: 2021-01-29T19:33:59.026Z [INFO] TestKeyringCommand.server.serf.lan: serf: Received use-key query writer.go:29: 2021-01-29T19:33:59.030Z [DEBUG] TestKeyringCommand.server.serf.lan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384 writer.go:29: 2021-01-29T19:33:59.035Z [DEBUG] TestKeyringCommand.http: Request finished: method=PUT url=/v1/operator/keyring from=127.0.0.1:44840 latency=41.280566ms writer.go:29: 2021-01-29T19:33:59.043Z [INFO] TestKeyringCommand.server.serf.wan: serf: Received remove-key query writer.go:29: 2021-01-29T19:33:59.048Z [DEBUG] TestKeyringCommand.server.serf.wan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384.dc1 writer.go:29: 2021-01-29T19:33:59.052Z [INFO] TestKeyringCommand.server.serf.lan: serf: Received remove-key query writer.go:29: 2021-01-29T19:33:59.063Z [DEBUG] TestKeyringCommand.server.serf.lan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384 writer.go:29: 2021-01-29T19:33:59.066Z [DEBUG] TestKeyringCommand.http: Request finished: method=DELETE url=/v1/operator/keyring from=127.0.0.1:44842 latency=22.872614ms writer.go:29: 2021-01-29T19:33:59.076Z [INFO] TestKeyringCommand.server.serf.wan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:59.083Z [DEBUG] TestKeyringCommand.server.serf.wan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384.dc1 writer.go:29: 2021-01-29T19:33:59.085Z [INFO] TestKeyringCommand.server.serf.lan: serf: Received list-keys query writer.go:29: 2021-01-29T19:33:59.092Z [DEBUG] TestKeyringCommand.server.serf.lan: serf: messageQueryResponseType: Node-e182a0ea-0f64-00f2-8e71-c58949127384 writer.go:29: 2021-01-29T19:33:59.095Z [DEBUG] TestKeyringCommand.http: Request finished: method=GET url=/v1/operator/keyring from=127.0.0.1:44844 latency=22.190039ms writer.go:29: 2021-01-29T19:33:59.097Z [INFO] TestKeyringCommand: Requesting shutdown writer.go:29: 2021-01-29T19:33:59.099Z [INFO] TestKeyringCommand.server: shutting down server writer.go:29: 2021-01-29T19:33:59.102Z [DEBUG] TestKeyringCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.108Z [WARN] TestKeyringCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:59.108Z [DEBUG] TestKeyringCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:33:59.111Z [WARN] TestKeyringCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:33:59.131Z [INFO] TestKeyringCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:33:59.134Z [INFO] TestKeyringCommand: consul server down writer.go:29: 2021-01-29T19:33:59.136Z [INFO] TestKeyringCommand: shutdown complete writer.go:29: 2021-01-29T19:33:59.137Z [INFO] TestKeyringCommand: Stopping server: protocol=DNS address=127.0.0.1:20211 network=tcp writer.go:29: 2021-01-29T19:33:59.138Z [INFO] TestKeyringCommand: Stopping server: protocol=DNS address=127.0.0.1:20211 network=udp writer.go:29: 2021-01-29T19:33:59.140Z [INFO] TestKeyringCommand: Stopping server: protocol=HTTP address=127.0.0.1:20212 network=tcp writer.go:29: 2021-01-29T19:33:59.645Z [INFO] TestKeyringCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:33:59.646Z [INFO] TestKeyringCommand: Endpoints down --- PASS: TestKeyringCommand (1.05s) PASS ok github.com/hashicorp/consul/command/keyring 1.271s === RUN TestKVCommand_noTabs === PAUSE TestKVCommand_noTabs === CONT TestKVCommand_noTabs --- PASS: TestKVCommand_noTabs (0.00s) PASS ok github.com/hashicorp/consul/command/kv 0.011s === RUN TestKVDeleteCommand_noTabs === PAUSE TestKVDeleteCommand_noTabs === RUN TestKVDeleteCommand_Validation === PAUSE TestKVDeleteCommand_Validation === RUN TestKVDeleteCommand === PAUSE TestKVDeleteCommand === RUN TestKVDeleteCommand_Recurse === PAUSE TestKVDeleteCommand_Recurse === RUN TestKVDeleteCommand_CAS === PAUSE TestKVDeleteCommand_CAS === CONT TestKVDeleteCommand_Recurse === CONT TestKVDeleteCommand_Validation === CONT TestKVDeleteCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestKVDeleteCommand_CAS [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestKVDeleteCommand_Validation (0.01s) === CONT TestKVDeleteCommand_noTabs --- PASS: TestKVDeleteCommand_noTabs (0.00s) === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:04.953Z [WARN] TestKVDeleteCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:04.960Z [DEBUG] TestKVDeleteCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:04.966Z [DEBUG] TestKVDeleteCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:04.971Z [INFO] TestKVDeleteCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5fc8e4a8-76ef-9325-af27-5f1ebb2c182f Address:127.0.0.1:28390}]" === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:04.971Z [WARN] TestKVDeleteCommand_CAS: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:04.974Z [DEBUG] TestKVDeleteCommand_CAS.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:04.976Z [DEBUG] TestKVDeleteCommand_CAS.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:04.993Z [WARN] TestKVDeleteCommand_Recurse: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:05.015Z [DEBUG] TestKVDeleteCommand_Recurse.tlsutil: Update: version=1 === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:04.973Z [INFO] TestKVDeleteCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:28390 [Follower]" leader= === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.027Z [INFO] TestKVDeleteCommand_CAS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:9d2cfcd0-ad3b-2560-9dae-2b078515afe1 Address:127.0.0.1:28396}]" === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.008Z [INFO] TestKVDeleteCommand.server.serf.wan: serf: EventMemberJoin: Node-5fc8e4a8-76ef-9325-af27-5f1ebb2c182f.dc1 127.0.0.1 === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.029Z [INFO] TestKVDeleteCommand_CAS.server.serf.wan: serf: EventMemberJoin: Node-9d2cfcd0-ad3b-2560-9dae-2b078515afe1.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:05.031Z [INFO] TestKVDeleteCommand_CAS.server.serf.lan: serf: EventMemberJoin: Node-9d2cfcd0-ad3b-2560-9dae-2b078515afe1 127.0.0.1 writer.go:29: 2021-01-29T19:34:05.034Z [INFO] TestKVDeleteCommand_CAS: Started DNS server: address=127.0.0.1:28391 network=udp === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.033Z [INFO] TestKVDeleteCommand.server.serf.lan: serf: EventMemberJoin: Node-5fc8e4a8-76ef-9325-af27-5f1ebb2c182f 127.0.0.1 === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.035Z [INFO] TestKVDeleteCommand_CAS.server.raft: entering follower state: follower="Node at 127.0.0.1:28396 [Follower]" leader= === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.052Z [DEBUG] TestKVDeleteCommand_Recurse.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.053Z [INFO] TestKVDeleteCommand: Started DNS server: address=127.0.0.1:28385 network=udp writer.go:29: 2021-01-29T19:34:05.055Z [INFO] TestKVDeleteCommand: Started DNS server: address=127.0.0.1:28385 network=tcp === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.052Z [INFO] TestKVDeleteCommand_CAS.server: Adding LAN server: server="Node-9d2cfcd0-ad3b-2560-9dae-2b078515afe1 (Addr: tcp/127.0.0.1:28396) (DC: dc1)" === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.059Z [INFO] TestKVDeleteCommand_Recurse.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:79fb055c-a834-36c1-0040-a723621ea20c Address:127.0.0.1:28384}]" === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.052Z [INFO] TestKVDeleteCommand_CAS.server: Handled event for server in area: event=member-join server=Node-9d2cfcd0-ad3b-2560-9dae-2b078515afe1.dc1 area=wan writer.go:29: 2021-01-29T19:34:05.055Z [INFO] TestKVDeleteCommand_CAS: Started DNS server: address=127.0.0.1:28391 network=tcp === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.061Z [INFO] TestKVDeleteCommand_Recurse.server.serf.wan: serf: EventMemberJoin: Node-79fb055c-a834-36c1-0040-a723621ea20c.dc1 127.0.0.1 === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.063Z [INFO] TestKVDeleteCommand_CAS: Started HTTP server: address=127.0.0.1:28392 network=tcp === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.055Z [INFO] TestKVDeleteCommand.server: Handled event for server in area: event=member-join server=Node-5fc8e4a8-76ef-9325-af27-5f1ebb2c182f.dc1 area=wan === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.065Z [INFO] TestKVDeleteCommand_CAS: started state syncer === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.056Z [INFO] TestKVDeleteCommand.server: Adding LAN server: server="Node-5fc8e4a8-76ef-9325-af27-5f1ebb2c182f (Addr: tcp/127.0.0.1:28390) (DC: dc1)" === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.067Z [INFO] TestKVDeleteCommand_Recurse.server.serf.lan: serf: EventMemberJoin: Node-79fb055c-a834-36c1-0040-a723621ea20c 127.0.0.1 === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.058Z [INFO] TestKVDeleteCommand: Started HTTP server: address=127.0.0.1:28386 network=tcp writer.go:29: 2021-01-29T19:34:05.072Z [INFO] TestKVDeleteCommand: started state syncer writer.go:29: 2021-01-29T19:34:05.071Z [WARN] TestKVDeleteCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:05.088Z [INFO] TestKVDeleteCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:28390 [Candidate]" term=2 === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.069Z [INFO] TestKVDeleteCommand_Recurse.server.raft: entering follower state: follower="Node at 127.0.0.1:28384 [Follower]" leader= writer.go:29: 2021-01-29T19:34:05.077Z [INFO] TestKVDeleteCommand_Recurse: Started DNS server: address=127.0.0.1:28379 network=udp === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.090Z [DEBUG] TestKVDeleteCommand.server.raft: votes: needed=1 === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.090Z [INFO] TestKVDeleteCommand_Recurse: Started DNS server: address=127.0.0.1:28379 network=tcp === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.091Z [DEBUG] TestKVDeleteCommand.server.raft: vote granted: from=5fc8e4a8-76ef-9325-af27-5f1ebb2c182f term=2 tally=1 writer.go:29: 2021-01-29T19:34:05.093Z [INFO] TestKVDeleteCommand.server.raft: election won: tally=1 === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.094Z [WARN] TestKVDeleteCommand_CAS.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.094Z [INFO] TestKVDeleteCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:28390 [Leader]" === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.095Z [INFO] TestKVDeleteCommand_CAS.server.raft: entering candidate state: node="Node at 127.0.0.1:28396 [Candidate]" term=2 === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.079Z [INFO] TestKVDeleteCommand_Recurse.server: Adding LAN server: server="Node-79fb055c-a834-36c1-0040-a723621ea20c (Addr: tcp/127.0.0.1:28384) (DC: dc1)" === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.096Z [INFO] TestKVDeleteCommand.server: cluster leadership acquired === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.098Z [DEBUG] TestKVDeleteCommand_CAS.server.raft: votes: needed=1 === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.080Z [INFO] TestKVDeleteCommand_Recurse.server: Handled event for server in area: event=member-join server=Node-79fb055c-a834-36c1-0040-a723621ea20c.dc1 area=wan === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.097Z [INFO] TestKVDeleteCommand.server: New leader elected: payload=Node-5fc8e4a8-76ef-9325-af27-5f1ebb2c182f === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.099Z [DEBUG] TestKVDeleteCommand_CAS.server.raft: vote granted: from=9d2cfcd0-ad3b-2560-9dae-2b078515afe1 term=2 tally=1 === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.092Z [INFO] TestKVDeleteCommand_Recurse: Started HTTP server: address=127.0.0.1:28380 network=tcp === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.101Z [INFO] TestKVDeleteCommand_CAS.server.raft: election won: tally=1 === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.097Z [DEBUG] TestKVDeleteCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:28390 === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.102Z [INFO] TestKVDeleteCommand_CAS.server.raft: entering leader state: leader="Node at 127.0.0.1:28396 [Leader]" writer.go:29: 2021-01-29T19:34:05.103Z [INFO] TestKVDeleteCommand_CAS.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:05.105Z [DEBUG] TestKVDeleteCommand_CAS.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:28396 === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.102Z [INFO] TestKVDeleteCommand_Recurse: started state syncer === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.105Z [INFO] TestKVDeleteCommand_CAS.server: New leader elected: payload=Node-9d2cfcd0-ad3b-2560-9dae-2b078515afe1 writer.go:29: 2021-01-29T19:34:05.110Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.112Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.150Z [INFO] TestKVDeleteCommand_CAS.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:05.153Z [INFO] TestKVDeleteCommand_CAS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.156Z [DEBUG] TestKVDeleteCommand_CAS.server: Skipping self join check for node since the cluster is too small: node=Node-9d2cfcd0-ad3b-2560-9dae-2b078515afe1 === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.158Z [WARN] TestKVDeleteCommand_Recurse.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:05.159Z [INFO] TestKVDeleteCommand_Recurse.server.raft: entering candidate state: node="Node at 127.0.0.1:28384 [Candidate]" term=2 === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.158Z [INFO] TestKVDeleteCommand_CAS.server: member joined, marking health alive: member=Node-9d2cfcd0-ad3b-2560-9dae-2b078515afe1 === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.149Z [INFO] TestKVDeleteCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:05.163Z [INFO] TestKVDeleteCommand.leader: started routine: routine="CA root pruning" === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.159Z [INFO] TestKVDeleteCommand_CAS: Synced node info === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.165Z [DEBUG] TestKVDeleteCommand.server: Skipping self join check for node since the cluster is too small: node=Node-5fc8e4a8-76ef-9325-af27-5f1ebb2c182f writer.go:29: 2021-01-29T19:34:05.166Z [INFO] TestKVDeleteCommand.server: member joined, marking health alive: member=Node-5fc8e4a8-76ef-9325-af27-5f1ebb2c182f === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.216Z [DEBUG] TestKVDeleteCommand_Recurse.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:05.217Z [DEBUG] TestKVDeleteCommand_Recurse.server.raft: vote granted: from=79fb055c-a834-36c1-0040-a723621ea20c term=2 tally=1 writer.go:29: 2021-01-29T19:34:05.218Z [INFO] TestKVDeleteCommand_Recurse.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:05.219Z [INFO] TestKVDeleteCommand_Recurse.server.raft: entering leader state: leader="Node at 127.0.0.1:28384 [Leader]" writer.go:29: 2021-01-29T19:34:05.220Z [INFO] TestKVDeleteCommand_Recurse.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:05.225Z [INFO] TestKVDeleteCommand_Recurse.server: New leader elected: payload=Node-79fb055c-a834-36c1-0040-a723621ea20c writer.go:29: 2021-01-29T19:34:05.250Z [DEBUG] TestKVDeleteCommand_Recurse.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:28384 writer.go:29: 2021-01-29T19:34:05.259Z [INFO] TestKVDeleteCommand_Recurse: Synced node info writer.go:29: 2021-01-29T19:34:05.261Z [DEBUG] TestKVDeleteCommand_Recurse: Node info in sync writer.go:29: 2021-01-29T19:34:05.267Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:05.275Z [INFO] TestKVDeleteCommand_Recurse.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:05.277Z [INFO] TestKVDeleteCommand_Recurse.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.279Z [DEBUG] TestKVDeleteCommand_Recurse.server: Skipping self join check for node since the cluster is too small: node=Node-79fb055c-a834-36c1-0040-a723621ea20c writer.go:29: 2021-01-29T19:34:05.282Z [INFO] TestKVDeleteCommand_Recurse.server: member joined, marking health alive: member=Node-79fb055c-a834-36c1-0040-a723621ea20c === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.320Z [DEBUG] TestKVDeleteCommand_CAS.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:52660 latency=12.079465ms writer.go:29: 2021-01-29T19:34:05.327Z [DEBUG] TestKVDeleteCommand_CAS.http: Request finished: method=DELETE url=/v1/kv/foo?cas=1 from=127.0.0.1:52662 latency=573.499µs writer.go:29: 2021-01-29T19:34:05.329Z [DEBUG] TestKVDeleteCommand_CAS.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:52660 latency=170.328µs writer.go:29: 2021-01-29T19:34:05.335Z [DEBUG] TestKVDeleteCommand_CAS.http: Request finished: method=DELETE url=/v1/kv/foo?cas=12 from=127.0.0.1:52664 latency=741.126µs writer.go:29: 2021-01-29T19:34:05.337Z [DEBUG] TestKVDeleteCommand_CAS.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:52660 latency=64.123µs writer.go:29: 2021-01-29T19:34:05.339Z [INFO] TestKVDeleteCommand_CAS: Requesting shutdown writer.go:29: 2021-01-29T19:34:05.341Z [INFO] TestKVDeleteCommand_CAS.server: shutting down server writer.go:29: 2021-01-29T19:34:05.342Z [DEBUG] TestKVDeleteCommand_CAS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.344Z [WARN] TestKVDeleteCommand_CAS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.344Z [DEBUG] TestKVDeleteCommand_CAS.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.347Z [WARN] TestKVDeleteCommand_CAS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.351Z [INFO] TestKVDeleteCommand_CAS: consul server down writer.go:29: 2021-01-29T19:34:05.352Z [INFO] TestKVDeleteCommand_CAS: shutdown complete writer.go:29: 2021-01-29T19:34:05.353Z [INFO] TestKVDeleteCommand_CAS: Stopping server: protocol=DNS address=127.0.0.1:28391 network=tcp writer.go:29: 2021-01-29T19:34:05.355Z [INFO] TestKVDeleteCommand_CAS: Stopping server: protocol=DNS address=127.0.0.1:28391 network=udp writer.go:29: 2021-01-29T19:34:05.357Z [INFO] TestKVDeleteCommand_CAS: Stopping server: protocol=HTTP address=127.0.0.1:28392 network=tcp writer.go:29: 2021-01-29T19:34:05.358Z [INFO] TestKVDeleteCommand_CAS.server.router.manager: shutting down === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.377Z [DEBUG] TestKVDeleteCommand.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:46498 latency=1.393171ms writer.go:29: 2021-01-29T19:34:05.382Z [DEBUG] TestKVDeleteCommand.http: Request finished: method=DELETE url=/v1/kv/foo from=127.0.0.1:46500 latency=701.475µs writer.go:29: 2021-01-29T19:34:05.384Z [DEBUG] TestKVDeleteCommand.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:46498 latency=35.2µs writer.go:29: 2021-01-29T19:34:05.385Z [INFO] TestKVDeleteCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:05.387Z [INFO] TestKVDeleteCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:05.389Z [DEBUG] TestKVDeleteCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.390Z [ERROR] TestKVDeleteCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:05.392Z [WARN] TestKVDeleteCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.392Z [DEBUG] TestKVDeleteCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.395Z [WARN] TestKVDeleteCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.399Z [INFO] TestKVDeleteCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:05.399Z [INFO] TestKVDeleteCommand: consul server down writer.go:29: 2021-01-29T19:34:05.406Z [INFO] TestKVDeleteCommand: shutdown complete writer.go:29: 2021-01-29T19:34:05.408Z [INFO] TestKVDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:28385 network=tcp writer.go:29: 2021-01-29T19:34:05.410Z [INFO] TestKVDeleteCommand: Stopping server: protocol=DNS address=127.0.0.1:28385 network=udp writer.go:29: 2021-01-29T19:34:05.411Z [INFO] TestKVDeleteCommand: Stopping server: protocol=HTTP address=127.0.0.1:28386 network=tcp === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:05.617Z [DEBUG] TestKVDeleteCommand_Recurse.http: Request finished: method=PUT url=/v1/kv/foo/a from=127.0.0.1:59938 latency=1.722354ms writer.go:29: 2021-01-29T19:34:05.622Z [DEBUG] TestKVDeleteCommand_Recurse.http: Request finished: method=PUT url=/v1/kv/foo/b from=127.0.0.1:59938 latency=757.353µs writer.go:29: 2021-01-29T19:34:05.624Z [DEBUG] TestKVDeleteCommand_Recurse.http: Request finished: method=PUT url=/v1/kv/food from=127.0.0.1:59938 latency=551.791µs writer.go:29: 2021-01-29T19:34:05.634Z [DEBUG] TestKVDeleteCommand_Recurse.http: Request finished: method=DELETE url=/v1/kv/foo?recurse= from=127.0.0.1:59940 latency=671.827µs writer.go:29: 2021-01-29T19:34:05.636Z [DEBUG] TestKVDeleteCommand_Recurse.http: Request finished: method=GET url=/v1/kv/foo/a from=127.0.0.1:59938 latency=23.866µs writer.go:29: 2021-01-29T19:34:05.638Z [DEBUG] TestKVDeleteCommand_Recurse.http: Request finished: method=GET url=/v1/kv/foo/b from=127.0.0.1:59938 latency=29.817µs writer.go:29: 2021-01-29T19:34:05.641Z [DEBUG] TestKVDeleteCommand_Recurse.http: Request finished: method=GET url=/v1/kv/food from=127.0.0.1:59938 latency=47.989µs writer.go:29: 2021-01-29T19:34:05.646Z [INFO] TestKVDeleteCommand_Recurse: Requesting shutdown writer.go:29: 2021-01-29T19:34:05.647Z [INFO] TestKVDeleteCommand_Recurse.server: shutting down server writer.go:29: 2021-01-29T19:34:05.648Z [DEBUG] TestKVDeleteCommand_Recurse.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.649Z [WARN] TestKVDeleteCommand_Recurse.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.649Z [DEBUG] TestKVDeleteCommand_Recurse.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:05.651Z [WARN] TestKVDeleteCommand_Recurse.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:05.654Z [INFO] TestKVDeleteCommand_Recurse.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:05.654Z [INFO] TestKVDeleteCommand_Recurse: consul server down writer.go:29: 2021-01-29T19:34:05.657Z [INFO] TestKVDeleteCommand_Recurse: shutdown complete writer.go:29: 2021-01-29T19:34:05.658Z [INFO] TestKVDeleteCommand_Recurse: Stopping server: protocol=DNS address=127.0.0.1:28379 network=tcp writer.go:29: 2021-01-29T19:34:05.659Z [INFO] TestKVDeleteCommand_Recurse: Stopping server: protocol=DNS address=127.0.0.1:28379 network=udp writer.go:29: 2021-01-29T19:34:05.660Z [INFO] TestKVDeleteCommand_Recurse: Stopping server: protocol=HTTP address=127.0.0.1:28380 network=tcp === CONT TestKVDeleteCommand_CAS writer.go:29: 2021-01-29T19:34:05.858Z [INFO] TestKVDeleteCommand_CAS: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:05.859Z [INFO] TestKVDeleteCommand_CAS: Endpoints down --- PASS: TestKVDeleteCommand_CAS (0.95s) === CONT TestKVDeleteCommand writer.go:29: 2021-01-29T19:34:05.913Z [INFO] TestKVDeleteCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:05.915Z [INFO] TestKVDeleteCommand: Endpoints down --- PASS: TestKVDeleteCommand (1.01s) === CONT TestKVDeleteCommand_Recurse writer.go:29: 2021-01-29T19:34:06.162Z [INFO] TestKVDeleteCommand_Recurse: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:06.164Z [INFO] TestKVDeleteCommand_Recurse: Endpoints down --- PASS: TestKVDeleteCommand_Recurse (1.25s) PASS ok github.com/hashicorp/consul/command/kv/del 1.449s === RUN TestKVExportCommand_noTabs === PAUSE TestKVExportCommand_noTabs === RUN TestKVExportCommand === PAUSE TestKVExportCommand === CONT TestKVExportCommand_noTabs === CONT TestKVExportCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestKVExportCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestKVExportCommand writer.go:29: 2021-01-29T19:34:06.662Z [WARN] TestKVExportCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:06.664Z [DEBUG] TestKVExportCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:06.668Z [DEBUG] TestKVExportCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:06.693Z [INFO] TestKVExportCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3f3635a5-2bdd-bd30-0da9-22f18ad86f15 Address:127.0.0.1:25321}]" writer.go:29: 2021-01-29T19:34:06.697Z [INFO] TestKVExportCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:25321 [Follower]" leader= writer.go:29: 2021-01-29T19:34:06.704Z [INFO] TestKVExportCommand.server.serf.wan: serf: EventMemberJoin: Node-3f3635a5-2bdd-bd30-0da9-22f18ad86f15.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:06.728Z [INFO] TestKVExportCommand.server.serf.lan: serf: EventMemberJoin: Node-3f3635a5-2bdd-bd30-0da9-22f18ad86f15 127.0.0.1 writer.go:29: 2021-01-29T19:34:06.742Z [WARN] TestKVExportCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:06.761Z [INFO] TestKVExportCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:25321 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:06.762Z [WARN] TestKVExportCommand.server.raft: unable to get address for sever, using fallback address: id=3f3635a5-2bdd-bd30-0da9-22f18ad86f15 fallback=127.0.0.1:25321 error="Could not find address for server id 3f3635a5-2bdd-bd30-0da9-22f18ad86f15" writer.go:29: 2021-01-29T19:34:06.764Z [DEBUG] TestKVExportCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:06.764Z [DEBUG] TestKVExportCommand.server.raft: vote granted: from=3f3635a5-2bdd-bd30-0da9-22f18ad86f15 term=2 tally=1 writer.go:29: 2021-01-29T19:34:06.765Z [INFO] TestKVExportCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:06.766Z [INFO] TestKVExportCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:25321 [Leader]" writer.go:29: 2021-01-29T19:34:06.766Z [INFO] TestKVExportCommand.server: Adding LAN server: server="Node-3f3635a5-2bdd-bd30-0da9-22f18ad86f15 (Addr: tcp/127.0.0.1:25321) (DC: dc1)" writer.go:29: 2021-01-29T19:34:06.769Z [INFO] TestKVExportCommand.server: New leader elected: payload=Node-3f3635a5-2bdd-bd30-0da9-22f18ad86f15 writer.go:29: 2021-01-29T19:34:06.766Z [INFO] TestKVExportCommand.server: Handled event for server in area: event=member-join server=Node-3f3635a5-2bdd-bd30-0da9-22f18ad86f15.dc1 area=wan writer.go:29: 2021-01-29T19:34:06.768Z [DEBUG] TestKVExportCommand.server: Cannot upgrade to new ACLs: leaderMode=3 mode=1 found=false leader=127.0.0.1:25321 writer.go:29: 2021-01-29T19:34:06.768Z [INFO] TestKVExportCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:06.770Z [INFO] TestKVExportCommand: Started DNS server: address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:34:06.787Z [INFO] TestKVExportCommand: Started DNS server: address=127.0.0.1:25316 network=tcp writer.go:29: 2021-01-29T19:34:06.789Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:06.800Z [INFO] TestKVExportCommand: Started HTTP server: address=127.0.0.1:25317 network=tcp writer.go:29: 2021-01-29T19:34:06.802Z [INFO] TestKVExportCommand: started state syncer writer.go:29: 2021-01-29T19:34:06.809Z [INFO] TestKVExportCommand: Synced node info writer.go:29: 2021-01-29T19:34:06.811Z [DEBUG] TestKVExportCommand: Node info in sync writer.go:29: 2021-01-29T19:34:06.832Z [DEBUG] TestKVExportCommand.http: Request finished: method=PUT url=/v1/kv/foo/b from=127.0.0.1:55302 latency=22.695062ms writer.go:29: 2021-01-29T19:34:06.855Z [DEBUG] TestKVExportCommand.http: Request finished: method=PUT url=/v1/kv/foo/c from=127.0.0.1:55302 latency=19.451081ms writer.go:29: 2021-01-29T19:34:06.876Z [DEBUG] TestKVExportCommand.http: Request finished: method=PUT url=/v1/kv/bar from=127.0.0.1:55302 latency=6.467002ms writer.go:29: 2021-01-29T19:34:06.881Z [DEBUG] TestKVExportCommand.http: Request finished: method=PUT url=/v1/kv/foo/a from=127.0.0.1:55302 latency=851.887µs writer.go:29: 2021-01-29T19:34:06.886Z [INFO] TestKVExportCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:06.888Z [INFO] TestKVExportCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.886Z [DEBUG] TestKVExportCommand.http: Request finished: method=GET url=/v1/kv/foo?recurse= from=127.0.0.1:55304 latency=531.329µs writer.go:29: 2021-01-29T19:34:06.889Z [DEBUG] TestKVExportCommand.server: Skipping self join check for node since the cluster is too small: node=Node-3f3635a5-2bdd-bd30-0da9-22f18ad86f15 writer.go:29: 2021-01-29T19:34:06.895Z [INFO] TestKVExportCommand.server: member joined, marking health alive: member=Node-3f3635a5-2bdd-bd30-0da9-22f18ad86f15 writer.go:29: 2021-01-29T19:34:06.893Z [INFO] TestKVExportCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:06.898Z [DEBUG] TestKVExportCommand.server: Skipping self join check for node since the cluster is too small: node=Node-3f3635a5-2bdd-bd30-0da9-22f18ad86f15 writer.go:29: 2021-01-29T19:34:06.898Z [INFO] TestKVExportCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:06.900Z [DEBUG] TestKVExportCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.900Z [WARN] TestKVExportCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:06.902Z [DEBUG] TestKVExportCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:06.905Z [WARN] TestKVExportCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:06.908Z [INFO] TestKVExportCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:06.908Z [INFO] TestKVExportCommand: consul server down writer.go:29: 2021-01-29T19:34:06.912Z [INFO] TestKVExportCommand: shutdown complete writer.go:29: 2021-01-29T19:34:06.914Z [INFO] TestKVExportCommand: Stopping server: protocol=DNS address=127.0.0.1:25316 network=tcp writer.go:29: 2021-01-29T19:34:06.915Z [INFO] TestKVExportCommand: Stopping server: protocol=DNS address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:34:06.917Z [INFO] TestKVExportCommand: Stopping server: protocol=HTTP address=127.0.0.1:25317 network=tcp writer.go:29: 2021-01-29T19:34:07.418Z [INFO] TestKVExportCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:07.420Z [INFO] TestKVExportCommand: Endpoints down --- PASS: TestKVExportCommand (0.85s) PASS ok github.com/hashicorp/consul/command/kv/exp 1.247s === RUN TestKVGetCommand_noTabs === PAUSE TestKVGetCommand_noTabs === RUN TestKVGetCommand_Validation === PAUSE TestKVGetCommand_Validation === RUN TestKVGetCommand === PAUSE TestKVGetCommand === RUN TestKVGetCommand_Base64 === PAUSE TestKVGetCommand_Base64 === RUN TestKVGetCommand_Missing === PAUSE TestKVGetCommand_Missing === RUN TestKVGetCommand_Empty === PAUSE TestKVGetCommand_Empty === RUN TestKVGetCommand_Detailed === PAUSE TestKVGetCommand_Detailed === RUN TestKVGetCommand_Keys === PAUSE TestKVGetCommand_Keys === RUN TestKVGetCommand_Recurse === PAUSE TestKVGetCommand_Recurse === RUN TestKVGetCommand_RecurseBase64 === PAUSE TestKVGetCommand_RecurseBase64 === RUN TestKVGetCommand_DetailedBase64 kv_get_test.go:338: DM-skipped --- SKIP: TestKVGetCommand_DetailedBase64 (0.00s) === CONT TestKVGetCommand_noTabs === CONT TestKVGetCommand_Detailed === CONT TestKVGetCommand_Base64 === CONT TestKVGetCommand_RecurseBase64 [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestKVGetCommand_noTabs (0.00s) === CONT TestKVGetCommand [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.536Z [WARN] TestKVGetCommand_RecurseBase64: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:09.540Z [DEBUG] TestKVGetCommand_RecurseBase64.tlsutil: Update: version=1 === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.546Z [WARN] TestKVGetCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:09.554Z [DEBUG] TestKVGetCommand.tlsutil: Update: version=1 === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.557Z [DEBUG] TestKVGetCommand_RecurseBase64.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.561Z [DEBUG] TestKVGetCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.578Z [INFO] TestKVGetCommand_RecurseBase64.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5c2e96a2-9e6e-a0b3-c97c-54d19a09f064 Address:127.0.0.1:15117}]" === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.580Z [WARN] TestKVGetCommand_Base64: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:09.581Z [DEBUG] TestKVGetCommand_Base64.tlsutil: Update: version=1 === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:09.585Z [WARN] TestKVGetCommand_Detailed: bootstrap = true: do not enable unless necessary === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.585Z [DEBUG] TestKVGetCommand_Base64.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:09.587Z [DEBUG] TestKVGetCommand_Detailed.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:09.589Z [DEBUG] TestKVGetCommand_Detailed.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.591Z [INFO] TestKVGetCommand_Base64.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1dd194ee-2ff7-cdf1-4b18-5d62bd912949 Address:127.0.0.1:15111}]" === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.592Z [INFO] TestKVGetCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:874f7768-b458-a323-456a-8be93e7eedbc Address:127.0.0.1:15129}]" === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.594Z [INFO] TestKVGetCommand_Base64.server.raft: entering follower state: follower="Node at 127.0.0.1:15111 [Follower]" leader= === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.581Z [INFO] TestKVGetCommand_RecurseBase64.server.serf.wan: serf: EventMemberJoin: Node-5c2e96a2-9e6e-a0b3-c97c-54d19a09f064.dc1 127.0.0.1 === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.597Z [INFO] TestKVGetCommand.server.serf.wan: serf: EventMemberJoin: Node-874f7768-b458-a323-456a-8be93e7eedbc.dc1 127.0.0.1 === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.596Z [INFO] TestKVGetCommand_Base64.server.serf.wan: serf: EventMemberJoin: Node-1dd194ee-2ff7-cdf1-4b18-5d62bd912949.dc1 127.0.0.1 === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.600Z [INFO] TestKVGetCommand_RecurseBase64.server.serf.lan: serf: EventMemberJoin: Node-5c2e96a2-9e6e-a0b3-c97c-54d19a09f064 127.0.0.1 writer.go:29: 2021-01-29T19:34:09.604Z [INFO] TestKVGetCommand_RecurseBase64: Started DNS server: address=127.0.0.1:15112 network=udp writer.go:29: 2021-01-29T19:34:09.582Z [INFO] TestKVGetCommand_RecurseBase64.server.raft: entering follower state: follower="Node at 127.0.0.1:15117 [Follower]" leader= === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.604Z [INFO] TestKVGetCommand_Base64.server.serf.lan: serf: EventMemberJoin: Node-1dd194ee-2ff7-cdf1-4b18-5d62bd912949 127.0.0.1 === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.602Z [INFO] TestKVGetCommand.server.serf.lan: serf: EventMemberJoin: Node-874f7768-b458-a323-456a-8be93e7eedbc 127.0.0.1 === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.623Z [INFO] TestKVGetCommand_Base64: Started DNS server: address=127.0.0.1:15106 network=udp === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.619Z [INFO] TestKVGetCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:15129 [Follower]" leader= === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.635Z [INFO] TestKVGetCommand_Base64.server: Handled event for server in area: event=member-join server=Node-1dd194ee-2ff7-cdf1-4b18-5d62bd912949.dc1 area=wan === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.638Z [INFO] TestKVGetCommand_RecurseBase64.server: Adding LAN server: server="Node-5c2e96a2-9e6e-a0b3-c97c-54d19a09f064 (Addr: tcp/127.0.0.1:15117) (DC: dc1)" writer.go:29: 2021-01-29T19:34:09.639Z [INFO] TestKVGetCommand_RecurseBase64.server: Handled event for server in area: event=member-join server=Node-5c2e96a2-9e6e-a0b3-c97c-54d19a09f064.dc1 area=wan === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.639Z [INFO] TestKVGetCommand.server: Adding LAN server: server="Node-874f7768-b458-a323-456a-8be93e7eedbc (Addr: tcp/127.0.0.1:15129) (DC: dc1)" === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.637Z [WARN] TestKVGetCommand_Base64.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.641Z [INFO] TestKVGetCommand.server: Handled event for server in area: event=member-join server=Node-874f7768-b458-a323-456a-8be93e7eedbc.dc1 area=wan === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.641Z [INFO] TestKVGetCommand_Base64.server.raft: entering candidate state: node="Node at 127.0.0.1:15111 [Candidate]" term=2 === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.639Z [INFO] TestKVGetCommand_RecurseBase64: Started DNS server: address=127.0.0.1:15112 network=tcp === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.642Z [INFO] TestKVGetCommand: Started DNS server: address=127.0.0.1:15124 network=udp === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.638Z [INFO] TestKVGetCommand_Base64: Started DNS server: address=127.0.0.1:15106 network=tcp writer.go:29: 2021-01-29T19:34:09.639Z [INFO] TestKVGetCommand_Base64.server: Adding LAN server: server="Node-1dd194ee-2ff7-cdf1-4b18-5d62bd912949 (Addr: tcp/127.0.0.1:15111) (DC: dc1)" === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.644Z [WARN] TestKVGetCommand_RecurseBase64.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:09.646Z [INFO] TestKVGetCommand_RecurseBase64.server.raft: entering candidate state: node="Node at 127.0.0.1:15117 [Candidate]" term=2 === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.646Z [INFO] TestKVGetCommand: Started DNS server: address=127.0.0.1:15124 network=tcp === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.645Z [INFO] TestKVGetCommand_RecurseBase64: Started HTTP server: address=127.0.0.1:15113 network=tcp === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.647Z [INFO] TestKVGetCommand_Base64: Started HTTP server: address=127.0.0.1:15107 network=tcp === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.648Z [INFO] TestKVGetCommand_RecurseBase64: started state syncer === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.649Z [INFO] TestKVGetCommand_Base64: started state syncer === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.649Z [INFO] TestKVGetCommand: Started HTTP server: address=127.0.0.1:15125 network=tcp === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.651Z [DEBUG] TestKVGetCommand_RecurseBase64.server.raft: votes: needed=1 === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.651Z [INFO] TestKVGetCommand: started state syncer === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.653Z [DEBUG] TestKVGetCommand_RecurseBase64.server.raft: vote granted: from=5c2e96a2-9e6e-a0b3-c97c-54d19a09f064 term=2 tally=1 writer.go:29: 2021-01-29T19:34:09.654Z [INFO] TestKVGetCommand_RecurseBase64.server.raft: election won: tally=1 === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:09.654Z [INFO] TestKVGetCommand_Detailed.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0bd715c1-e3cb-1712-f930-728d213d07d3 Address:127.0.0.1:15123}]" === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.655Z [INFO] TestKVGetCommand_RecurseBase64.server.raft: entering leader state: leader="Node at 127.0.0.1:15117 [Leader]" writer.go:29: 2021-01-29T19:34:09.662Z [INFO] TestKVGetCommand_RecurseBase64.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:09.664Z [INFO] TestKVGetCommand_RecurseBase64.server: New leader elected: payload=Node-5c2e96a2-9e6e-a0b3-c97c-54d19a09f064 === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.660Z [DEBUG] TestKVGetCommand_Base64.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:09.667Z [DEBUG] TestKVGetCommand_Base64.server.raft: vote granted: from=1dd194ee-2ff7-cdf1-4b18-5d62bd912949 term=2 tally=1 === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.664Z [DEBUG] TestKVGetCommand_RecurseBase64.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15117 === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:09.657Z [INFO] TestKVGetCommand_Detailed.server.serf.wan: serf: EventMemberJoin: Node-0bd715c1-e3cb-1712-f930-728d213d07d3.dc1 127.0.0.1 === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.673Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:09.662Z [INFO] TestKVGetCommand_Detailed.server.raft: entering follower state: follower="Node at 127.0.0.1:15123 [Follower]" leader= === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.677Z [WARN] TestKVGetCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:09.679Z [INFO] TestKVGetCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:15129 [Candidate]" term=2 === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.669Z [INFO] TestKVGetCommand_Base64.server.raft: election won: tally=1 === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.681Z [DEBUG] TestKVGetCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:09.682Z [DEBUG] TestKVGetCommand.server.raft: vote granted: from=874f7768-b458-a323-456a-8be93e7eedbc term=2 tally=1 writer.go:29: 2021-01-29T19:34:09.684Z [INFO] TestKVGetCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:09.685Z [INFO] TestKVGetCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:15129 [Leader]" === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:09.685Z [INFO] TestKVGetCommand_Detailed.server.serf.lan: serf: EventMemberJoin: Node-0bd715c1-e3cb-1712-f930-728d213d07d3 127.0.0.1 writer.go:29: 2021-01-29T19:34:09.689Z [INFO] TestKVGetCommand_Detailed: Started DNS server: address=127.0.0.1:15118 network=udp === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.682Z [INFO] TestKVGetCommand_Base64.server.raft: entering leader state: leader="Node at 127.0.0.1:15111 [Leader]" === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.693Z [INFO] TestKVGetCommand.server: cluster leadership acquired === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.695Z [INFO] TestKVGetCommand_Base64.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:09.697Z [INFO] TestKVGetCommand_Base64.server: New leader elected: payload=Node-1dd194ee-2ff7-cdf1-4b18-5d62bd912949 === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:09.691Z [INFO] TestKVGetCommand_Detailed.server: Adding LAN server: server="Node-0bd715c1-e3cb-1712-f930-728d213d07d3 (Addr: tcp/127.0.0.1:15123) (DC: dc1)" === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.698Z [INFO] TestKVGetCommand.server: New leader elected: payload=Node-874f7768-b458-a323-456a-8be93e7eedbc === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.700Z [DEBUG] TestKVGetCommand_Base64.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15111 === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:09.703Z [INFO] TestKVGetCommand_Detailed.server: Handled event for server in area: event=member-join server=Node-0bd715c1-e3cb-1712-f930-728d213d07d3.dc1 area=wan writer.go:29: 2021-01-29T19:34:09.704Z [INFO] TestKVGetCommand_Detailed: Started DNS server: address=127.0.0.1:15118 network=tcp writer.go:29: 2021-01-29T19:34:09.706Z [INFO] TestKVGetCommand_Detailed: Started HTTP server: address=127.0.0.1:15119 network=tcp writer.go:29: 2021-01-29T19:34:09.707Z [INFO] TestKVGetCommand_Detailed: started state syncer === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.715Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.700Z [DEBUG] TestKVGetCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15129 writer.go:29: 2021-01-29T19:34:09.726Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:09.742Z [WARN] TestKVGetCommand_Detailed.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:09.743Z [INFO] TestKVGetCommand_Detailed.server.raft: entering candidate state: node="Node at 127.0.0.1:15123 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:09.745Z [DEBUG] TestKVGetCommand_Detailed.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:09.746Z [DEBUG] TestKVGetCommand_Detailed.server.raft: vote granted: from=0bd715c1-e3cb-1712-f930-728d213d07d3 term=2 tally=1 writer.go:29: 2021-01-29T19:34:09.746Z [INFO] TestKVGetCommand_Detailed.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:09.747Z [INFO] TestKVGetCommand_Detailed.server.raft: entering leader state: leader="Node at 127.0.0.1:15123 [Leader]" writer.go:29: 2021-01-29T19:34:09.748Z [INFO] TestKVGetCommand_Detailed.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:09.749Z [INFO] TestKVGetCommand_Detailed.server: New leader elected: payload=Node-0bd715c1-e3cb-1712-f930-728d213d07d3 writer.go:29: 2021-01-29T19:34:09.749Z [DEBUG] TestKVGetCommand_Detailed.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15123 writer.go:29: 2021-01-29T19:34:09.754Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.771Z [INFO] TestKVGetCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:09.772Z [INFO] TestKVGetCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.774Z [DEBUG] TestKVGetCommand.server: Skipping self join check for node since the cluster is too small: node=Node-874f7768-b458-a323-456a-8be93e7eedbc === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:09.774Z [INFO] TestKVGetCommand_Detailed.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:09.776Z [INFO] TestKVGetCommand_Detailed.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.778Z [DEBUG] TestKVGetCommand_Detailed.server: Skipping self join check for node since the cluster is too small: node=Node-0bd715c1-e3cb-1712-f930-728d213d07d3 writer.go:29: 2021-01-29T19:34:09.779Z [INFO] TestKVGetCommand_Detailed.server: member joined, marking health alive: member=Node-0bd715c1-e3cb-1712-f930-728d213d07d3 === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.774Z [INFO] TestKVGetCommand.server: member joined, marking health alive: member=Node-874f7768-b458-a323-456a-8be93e7eedbc === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:09.784Z [INFO] TestKVGetCommand_RecurseBase64.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:09.785Z [INFO] TestKVGetCommand_RecurseBase64.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.787Z [DEBUG] TestKVGetCommand_RecurseBase64.server: Skipping self join check for node since the cluster is too small: node=Node-5c2e96a2-9e6e-a0b3-c97c-54d19a09f064 writer.go:29: 2021-01-29T19:34:09.788Z [INFO] TestKVGetCommand_RecurseBase64.server: member joined, marking health alive: member=Node-5c2e96a2-9e6e-a0b3-c97c-54d19a09f064 === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.793Z [INFO] TestKVGetCommand_Base64.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:09.800Z [INFO] TestKVGetCommand_Base64.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.803Z [DEBUG] TestKVGetCommand_Base64.server: Skipping self join check for node since the cluster is too small: node=Node-1dd194ee-2ff7-cdf1-4b18-5d62bd912949 === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.809Z [DEBUG] TestKVGetCommand.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:32850 latency=19.508272ms === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.809Z [INFO] TestKVGetCommand_Base64.server: member joined, marking health alive: member=Node-1dd194ee-2ff7-cdf1-4b18-5d62bd912949 writer.go:29: 2021-01-29T19:34:09.808Z [INFO] TestKVGetCommand_Base64: Synced node info writer.go:29: 2021-01-29T19:34:09.815Z [DEBUG] TestKVGetCommand_Base64: Node info in sync === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:09.826Z [DEBUG] TestKVGetCommand.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:32852 latency=1.047212ms writer.go:29: 2021-01-29T19:34:09.833Z [INFO] TestKVGetCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:09.835Z [INFO] TestKVGetCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:09.837Z [DEBUG] TestKVGetCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.839Z [WARN] TestKVGetCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:09.846Z [ERROR] TestKVGetCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:09.846Z [DEBUG] TestKVGetCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.860Z [WARN] TestKVGetCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:09.863Z [INFO] TestKVGetCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:09.863Z [INFO] TestKVGetCommand: consul server down writer.go:29: 2021-01-29T19:34:09.866Z [INFO] TestKVGetCommand: shutdown complete writer.go:29: 2021-01-29T19:34:09.867Z [INFO] TestKVGetCommand: Stopping server: protocol=DNS address=127.0.0.1:15124 network=tcp writer.go:29: 2021-01-29T19:34:09.868Z [INFO] TestKVGetCommand: Stopping server: protocol=DNS address=127.0.0.1:15124 network=udp writer.go:29: 2021-01-29T19:34:09.870Z [INFO] TestKVGetCommand: Stopping server: protocol=HTTP address=127.0.0.1:15125 network=tcp === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:09.967Z [DEBUG] TestKVGetCommand_Base64.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:37554 latency=2.117693ms writer.go:29: 2021-01-29T19:34:09.972Z [DEBUG] TestKVGetCommand_Base64.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:37556 latency=1.372882ms writer.go:29: 2021-01-29T19:34:09.975Z [INFO] TestKVGetCommand_Base64: Requesting shutdown writer.go:29: 2021-01-29T19:34:09.976Z [INFO] TestKVGetCommand_Base64.server: shutting down server writer.go:29: 2021-01-29T19:34:09.984Z [DEBUG] TestKVGetCommand_Base64.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.986Z [DEBUG] TestKVGetCommand_Base64.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:09.986Z [WARN] TestKVGetCommand_Base64.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:09.991Z [WARN] TestKVGetCommand_Base64.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:09.995Z [INFO] TestKVGetCommand_Base64.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:09.996Z [INFO] TestKVGetCommand_Base64: consul server down writer.go:29: 2021-01-29T19:34:10.003Z [INFO] TestKVGetCommand_Base64: shutdown complete writer.go:29: 2021-01-29T19:34:10.004Z [INFO] TestKVGetCommand_Base64: Stopping server: protocol=DNS address=127.0.0.1:15106 network=tcp writer.go:29: 2021-01-29T19:34:10.006Z [INFO] TestKVGetCommand_Base64: Stopping server: protocol=DNS address=127.0.0.1:15106 network=udp writer.go:29: 2021-01-29T19:34:10.008Z [INFO] TestKVGetCommand_Base64: Stopping server: protocol=HTTP address=127.0.0.1:15107 network=tcp === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:10.013Z [DEBUG] TestKVGetCommand_RecurseBase64.http: Request finished: method=PUT url=/v1/kv/foo/b from=127.0.0.1:39014 latency=1.198356ms writer.go:29: 2021-01-29T19:34:10.016Z [DEBUG] TestKVGetCommand_RecurseBase64.http: Request finished: method=PUT url=/v1/kv/foo/c from=127.0.0.1:39014 latency=721.81µs writer.go:29: 2021-01-29T19:34:10.018Z [DEBUG] TestKVGetCommand_RecurseBase64.http: Request finished: method=PUT url=/v1/kv/foo/a from=127.0.0.1:39014 latency=884.261µs writer.go:29: 2021-01-29T19:34:10.023Z [DEBUG] TestKVGetCommand_RecurseBase64.http: Request finished: method=GET url=/v1/kv/foo?recurse= from=127.0.0.1:39016 latency=104.432µs writer.go:29: 2021-01-29T19:34:10.024Z [INFO] TestKVGetCommand_RecurseBase64: Requesting shutdown writer.go:29: 2021-01-29T19:34:10.025Z [INFO] TestKVGetCommand_RecurseBase64.server: shutting down server writer.go:29: 2021-01-29T19:34:10.026Z [DEBUG] TestKVGetCommand_RecurseBase64.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.026Z [WARN] TestKVGetCommand_RecurseBase64.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.026Z [ERROR] TestKVGetCommand_RecurseBase64.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:10.026Z [DEBUG] TestKVGetCommand_RecurseBase64.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.030Z [WARN] TestKVGetCommand_RecurseBase64.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.032Z [INFO] TestKVGetCommand_RecurseBase64.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:10.033Z [INFO] TestKVGetCommand_RecurseBase64: consul server down writer.go:29: 2021-01-29T19:34:10.037Z [INFO] TestKVGetCommand_RecurseBase64: shutdown complete writer.go:29: 2021-01-29T19:34:10.039Z [INFO] TestKVGetCommand_RecurseBase64: Stopping server: protocol=DNS address=127.0.0.1:15112 network=tcp writer.go:29: 2021-01-29T19:34:10.041Z [INFO] TestKVGetCommand_RecurseBase64: Stopping server: protocol=DNS address=127.0.0.1:15112 network=udp writer.go:29: 2021-01-29T19:34:10.043Z [INFO] TestKVGetCommand_RecurseBase64: Stopping server: protocol=HTTP address=127.0.0.1:15113 network=tcp === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:10.053Z [DEBUG] TestKVGetCommand_Detailed: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:10.058Z [INFO] TestKVGetCommand_Detailed: Synced node info writer.go:29: 2021-01-29T19:34:10.107Z [DEBUG] TestKVGetCommand_Detailed.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:32870 latency=11.372587ms writer.go:29: 2021-01-29T19:34:10.117Z [DEBUG] TestKVGetCommand_Detailed.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:32872 latency=145.509µs writer.go:29: 2021-01-29T19:34:10.119Z [INFO] TestKVGetCommand_Detailed: Requesting shutdown writer.go:29: 2021-01-29T19:34:10.121Z [INFO] TestKVGetCommand_Detailed.server: shutting down server writer.go:29: 2021-01-29T19:34:10.122Z [DEBUG] TestKVGetCommand_Detailed.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.123Z [WARN] TestKVGetCommand_Detailed.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.123Z [DEBUG] TestKVGetCommand_Detailed.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.127Z [WARN] TestKVGetCommand_Detailed.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.129Z [INFO] TestKVGetCommand_Detailed.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:10.130Z [INFO] TestKVGetCommand_Detailed: consul server down writer.go:29: 2021-01-29T19:34:10.134Z [INFO] TestKVGetCommand_Detailed: shutdown complete writer.go:29: 2021-01-29T19:34:10.136Z [INFO] TestKVGetCommand_Detailed: Stopping server: protocol=DNS address=127.0.0.1:15118 network=tcp writer.go:29: 2021-01-29T19:34:10.138Z [INFO] TestKVGetCommand_Detailed: Stopping server: protocol=DNS address=127.0.0.1:15118 network=udp writer.go:29: 2021-01-29T19:34:10.140Z [INFO] TestKVGetCommand_Detailed: Stopping server: protocol=HTTP address=127.0.0.1:15119 network=tcp === CONT TestKVGetCommand writer.go:29: 2021-01-29T19:34:10.371Z [INFO] TestKVGetCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:10.373Z [INFO] TestKVGetCommand: Endpoints down --- PASS: TestKVGetCommand (0.89s) === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.409Z [WARN] TestKVGetCommand_Empty: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:10.412Z [DEBUG] TestKVGetCommand_Empty.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:10.414Z [DEBUG] TestKVGetCommand_Empty.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:10.418Z [INFO] TestKVGetCommand_Empty.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:0fc6ec08-ab03-54ce-470a-3df853b74e93 Address:127.0.0.1:15135}]" writer.go:29: 2021-01-29T19:34:10.421Z [INFO] TestKVGetCommand_Empty.server.raft: entering follower state: follower="Node at 127.0.0.1:15135 [Follower]" leader= writer.go:29: 2021-01-29T19:34:10.421Z [INFO] TestKVGetCommand_Empty.server.serf.wan: serf: EventMemberJoin: Node-0fc6ec08-ab03-54ce-470a-3df853b74e93.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.427Z [INFO] TestKVGetCommand_Empty.server.serf.lan: serf: EventMemberJoin: Node-0fc6ec08-ab03-54ce-470a-3df853b74e93 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.430Z [INFO] TestKVGetCommand_Empty.server: Adding LAN server: server="Node-0fc6ec08-ab03-54ce-470a-3df853b74e93 (Addr: tcp/127.0.0.1:15135) (DC: dc1)" writer.go:29: 2021-01-29T19:34:10.430Z [INFO] TestKVGetCommand_Empty: Started DNS server: address=127.0.0.1:15130 network=udp writer.go:29: 2021-01-29T19:34:10.433Z [INFO] TestKVGetCommand_Empty: Started DNS server: address=127.0.0.1:15130 network=tcp writer.go:29: 2021-01-29T19:34:10.436Z [INFO] TestKVGetCommand_Empty: Started HTTP server: address=127.0.0.1:15131 network=tcp writer.go:29: 2021-01-29T19:34:10.437Z [INFO] TestKVGetCommand_Empty: started state syncer writer.go:29: 2021-01-29T19:34:10.430Z [INFO] TestKVGetCommand_Empty.server: Handled event for server in area: event=member-join server=Node-0fc6ec08-ab03-54ce-470a-3df853b74e93.dc1 area=wan writer.go:29: 2021-01-29T19:34:10.465Z [WARN] TestKVGetCommand_Empty.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:10.467Z [INFO] TestKVGetCommand_Empty.server.raft: entering candidate state: node="Node at 127.0.0.1:15135 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:10.495Z [DEBUG] TestKVGetCommand_Empty.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:10.497Z [DEBUG] TestKVGetCommand_Empty.server.raft: vote granted: from=0fc6ec08-ab03-54ce-470a-3df853b74e93 term=2 tally=1 writer.go:29: 2021-01-29T19:34:10.498Z [INFO] TestKVGetCommand_Empty.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:10.499Z [INFO] TestKVGetCommand_Empty.server.raft: entering leader state: leader="Node at 127.0.0.1:15135 [Leader]" writer.go:29: 2021-01-29T19:34:10.500Z [INFO] TestKVGetCommand_Empty.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:10.502Z [INFO] TestKVGetCommand_Empty.server: New leader elected: payload=Node-0fc6ec08-ab03-54ce-470a-3df853b74e93 writer.go:29: 2021-01-29T19:34:10.505Z [DEBUG] TestKVGetCommand_Empty.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15135 === CONT TestKVGetCommand_Base64 writer.go:29: 2021-01-29T19:34:10.509Z [INFO] TestKVGetCommand_Base64: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:10.511Z [INFO] TestKVGetCommand_Base64: Endpoints down --- PASS: TestKVGetCommand_Base64 (1.03s) === CONT TestKVGetCommand_Missing === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.518Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:10.526Z [INFO] TestKVGetCommand_Empty.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:10.527Z [INFO] TestKVGetCommand_Empty.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.528Z [DEBUG] TestKVGetCommand_Empty.server: Skipping self join check for node since the cluster is too small: node=Node-0fc6ec08-ab03-54ce-470a-3df853b74e93 writer.go:29: 2021-01-29T19:34:10.529Z [INFO] TestKVGetCommand_Empty.server: member joined, marking health alive: member=Node-0fc6ec08-ab03-54ce-470a-3df853b74e93 === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.534Z [WARN] TestKVGetCommand_Missing: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:10.536Z [DEBUG] TestKVGetCommand_Missing.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:10.538Z [DEBUG] TestKVGetCommand_Missing.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVGetCommand_RecurseBase64 writer.go:29: 2021-01-29T19:34:10.544Z [INFO] TestKVGetCommand_RecurseBase64: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:10.544Z [INFO] TestKVGetCommand_RecurseBase64: Endpoints down --- PASS: TestKVGetCommand_RecurseBase64 (1.07s) === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.552Z [WARN] TestKVGetCommand_Recurse: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:10.553Z [DEBUG] TestKVGetCommand_Recurse.tlsutil: Update: version=1 === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.554Z [INFO] TestKVGetCommand_Missing.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f54580de-499e-f5bb-573e-ac4f88c3fd3e Address:127.0.0.1:15141}]" === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.555Z [DEBUG] TestKVGetCommand_Recurse.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.555Z [INFO] TestKVGetCommand_Missing.server.raft: entering follower state: follower="Node at 127.0.0.1:15141 [Follower]" leader= writer.go:29: 2021-01-29T19:34:10.555Z [INFO] TestKVGetCommand_Missing.server.serf.wan: serf: EventMemberJoin: Node-f54580de-499e-f5bb-573e-ac4f88c3fd3e.dc1 127.0.0.1 === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.558Z [INFO] TestKVGetCommand_Recurse.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d6178bf0-d804-cf5b-ed29-ba84e567808e Address:127.0.0.1:15147}]" === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.559Z [INFO] TestKVGetCommand_Missing.server.serf.lan: serf: EventMemberJoin: Node-f54580de-499e-f5bb-573e-ac4f88c3fd3e 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.561Z [INFO] TestKVGetCommand_Missing: Started DNS server: address=127.0.0.1:15136 network=udp === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.561Z [INFO] TestKVGetCommand_Recurse.server.serf.wan: serf: EventMemberJoin: Node-d6178bf0-d804-cf5b-ed29-ba84e567808e.dc1 127.0.0.1 === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.563Z [INFO] TestKVGetCommand_Missing.server: Adding LAN server: server="Node-f54580de-499e-f5bb-573e-ac4f88c3fd3e (Addr: tcp/127.0.0.1:15141) (DC: dc1)" writer.go:29: 2021-01-29T19:34:10.564Z [INFO] TestKVGetCommand_Missing.server: Handled event for server in area: event=member-join server=Node-f54580de-499e-f5bb-573e-ac4f88c3fd3e.dc1 area=wan writer.go:29: 2021-01-29T19:34:10.565Z [INFO] TestKVGetCommand_Missing: Started DNS server: address=127.0.0.1:15136 network=tcp writer.go:29: 2021-01-29T19:34:10.566Z [INFO] TestKVGetCommand_Missing: Started HTTP server: address=127.0.0.1:15137 network=tcp === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.562Z [INFO] TestKVGetCommand_Recurse.server.raft: entering follower state: follower="Node at 127.0.0.1:15147 [Follower]" leader= === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.580Z [INFO] TestKVGetCommand_Missing: started state syncer === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.564Z [INFO] TestKVGetCommand_Recurse.server.serf.lan: serf: EventMemberJoin: Node-d6178bf0-d804-cf5b-ed29-ba84e567808e 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.596Z [INFO] TestKVGetCommand_Recurse.server: Adding LAN server: server="Node-d6178bf0-d804-cf5b-ed29-ba84e567808e (Addr: tcp/127.0.0.1:15147) (DC: dc1)" writer.go:29: 2021-01-29T19:34:10.598Z [INFO] TestKVGetCommand_Recurse.server: Handled event for server in area: event=member-join server=Node-d6178bf0-d804-cf5b-ed29-ba84e567808e.dc1 area=wan writer.go:29: 2021-01-29T19:34:10.600Z [INFO] TestKVGetCommand_Recurse: Started DNS server: address=127.0.0.1:15142 network=udp writer.go:29: 2021-01-29T19:34:10.602Z [INFO] TestKVGetCommand_Recurse: Started DNS server: address=127.0.0.1:15142 network=tcp === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.622Z [WARN] TestKVGetCommand_Missing.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.625Z [INFO] TestKVGetCommand_Recurse: Started HTTP server: address=127.0.0.1:15143 network=tcp === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.625Z [INFO] TestKVGetCommand_Missing.server.raft: entering candidate state: node="Node at 127.0.0.1:15141 [Candidate]" term=2 === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.634Z [WARN] TestKVGetCommand_Recurse.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.630Z [DEBUG] TestKVGetCommand_Missing.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:10.647Z [DEBUG] TestKVGetCommand_Missing.server.raft: vote granted: from=f54580de-499e-f5bb-573e-ac4f88c3fd3e term=2 tally=1 === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.636Z [INFO] TestKVGetCommand_Recurse.server.raft: entering candidate state: node="Node at 127.0.0.1:15147 [Candidate]" term=2 === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.657Z [INFO] TestKVGetCommand_Missing.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:10.659Z [INFO] TestKVGetCommand_Missing.server.raft: entering leader state: leader="Node at 127.0.0.1:15141 [Leader]" writer.go:29: 2021-01-29T19:34:10.662Z [INFO] TestKVGetCommand_Missing.server: cluster leadership acquired === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.663Z [INFO] TestKVGetCommand_Recurse: started state syncer === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.666Z [INFO] TestKVGetCommand_Missing.server: New leader elected: payload=Node-f54580de-499e-f5bb-573e-ac4f88c3fd3e === CONT TestKVGetCommand_Detailed writer.go:29: 2021-01-29T19:34:10.669Z [INFO] TestKVGetCommand_Detailed: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:10.693Z [INFO] TestKVGetCommand_Detailed: Endpoints down --- PASS: TestKVGetCommand_Detailed (1.23s) === CONT TestKVGetCommand_Keys === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.724Z [DEBUG] TestKVGetCommand_Missing.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15141 === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.725Z [DEBUG] TestKVGetCommand_Recurse.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:10.738Z [DEBUG] TestKVGetCommand_Recurse.server.raft: vote granted: from=d6178bf0-d804-cf5b-ed29-ba84e567808e term=2 tally=1 writer.go:29: 2021-01-29T19:34:10.740Z [INFO] TestKVGetCommand_Recurse.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:10.742Z [INFO] TestKVGetCommand_Recurse.server.raft: entering leader state: leader="Node at 127.0.0.1:15147 [Leader]" === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.764Z [DEBUG] TestKVGetCommand_Empty.http: Request finished: method=PUT url=/v1/kv/empty from=127.0.0.1:42370 latency=107.369768ms === CONT TestKVGetCommand_Keys writer.go:29: 2021-01-29T19:34:10.765Z [WARN] TestKVGetCommand_Keys: bootstrap = true: do not enable unless necessary === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.765Z [INFO] TestKVGetCommand_Recurse.server: cluster leadership acquired === CONT TestKVGetCommand_Keys writer.go:29: 2021-01-29T19:34:10.766Z [DEBUG] TestKVGetCommand_Keys.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:10.767Z [DEBUG] TestKVGetCommand_Keys.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.767Z [DEBUG] TestKVGetCommand_Missing.http: Request finished: method=GET url=/v1/kv/not-a-real-key from=127.0.0.1:45054 latency=57.518µs === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.768Z [INFO] TestKVGetCommand_Recurse.server: New leader elected: payload=Node-d6178bf0-d804-cf5b-ed29-ba84e567808e === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.770Z [INFO] TestKVGetCommand_Missing: Requesting shutdown writer.go:29: 2021-01-29T19:34:10.771Z [INFO] TestKVGetCommand_Missing.server: shutting down server === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.770Z [DEBUG] TestKVGetCommand_Empty: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.772Z [WARN] TestKVGetCommand_Missing.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.777Z [ERROR] TestKVGetCommand_Missing.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.778Z [DEBUG] TestKVGetCommand_Empty.http: Request finished: method=GET url=/v1/kv/empty from=127.0.0.1:42374 latency=394.766µs writer.go:29: 2021-01-29T19:34:10.780Z [INFO] TestKVGetCommand_Empty: Synced node info === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.785Z [WARN] TestKVGetCommand_Missing.server.serf.wan: serf: Shutdown without a Leave === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.783Z [INFO] TestKVGetCommand_Empty: Requesting shutdown === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.788Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.784Z [DEBUG] TestKVGetCommand_Recurse.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15147 === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.792Z [INFO] TestKVGetCommand_Missing.server.router.manager: shutting down === CONT TestKVGetCommand_Keys writer.go:29: 2021-01-29T19:34:10.794Z [INFO] TestKVGetCommand_Keys.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e290dd3e-0fd1-6d97-9017-00c97977d459 Address:127.0.0.1:15153}]" === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.788Z [INFO] TestKVGetCommand_Empty.server: shutting down server === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.792Z [ERROR] TestKVGetCommand_Missing.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: leadership lost while committing log" === CONT TestKVGetCommand_Keys writer.go:29: 2021-01-29T19:34:10.797Z [INFO] TestKVGetCommand_Keys.server.raft: entering follower state: follower="Node at 127.0.0.1:15153 [Follower]" leader= === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.792Z [INFO] TestKVGetCommand_Missing: consul server down writer.go:29: 2021-01-29T19:34:10.804Z [INFO] TestKVGetCommand_Missing: shutdown complete === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.798Z [DEBUG] TestKVGetCommand_Empty.leader: stopping routine: routine="CA root pruning" === CONT TestKVGetCommand_Keys writer.go:29: 2021-01-29T19:34:10.814Z [INFO] TestKVGetCommand_Keys.server.serf.wan: serf: EventMemberJoin: Node-e290dd3e-0fd1-6d97-9017-00c97977d459.dc1 127.0.0.1 === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.811Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.812Z [INFO] TestKVGetCommand_Missing: Stopping server: protocol=DNS address=127.0.0.1:15136 network=tcp === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.813Z [WARN] TestKVGetCommand_Empty.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.818Z [DEBUG] TestKVGetCommand_Empty.leader: stopped routine: routine="CA root pruning" === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:10.819Z [INFO] TestKVGetCommand_Missing: Stopping server: protocol=DNS address=127.0.0.1:15136 network=udp writer.go:29: 2021-01-29T19:34:10.830Z [INFO] TestKVGetCommand_Missing: Stopping server: protocol=HTTP address=127.0.0.1:15137 network=tcp === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.821Z [WARN] TestKVGetCommand_Empty.server.serf.wan: serf: Shutdown without a Leave === CONT TestKVGetCommand_Keys writer.go:29: 2021-01-29T19:34:10.833Z [INFO] TestKVGetCommand_Keys.server.serf.lan: serf: EventMemberJoin: Node-e290dd3e-0fd1-6d97-9017-00c97977d459 127.0.0.1 writer.go:29: 2021-01-29T19:34:10.837Z [INFO] TestKVGetCommand_Keys.server: Adding LAN server: server="Node-e290dd3e-0fd1-6d97-9017-00c97977d459 (Addr: tcp/127.0.0.1:15153) (DC: dc1)" writer.go:29: 2021-01-29T19:34:10.838Z [INFO] TestKVGetCommand_Keys: Started DNS server: address=127.0.0.1:15148 network=udp writer.go:29: 2021-01-29T19:34:10.838Z [INFO] TestKVGetCommand_Keys.server: Handled event for server in area: event=member-join server=Node-e290dd3e-0fd1-6d97-9017-00c97977d459.dc1 area=wan === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.839Z [INFO] TestKVGetCommand_Empty.server.router.manager: shutting down === CONT TestKVGetCommand_Keys writer.go:29: 2021-01-29T19:34:10.839Z [INFO] TestKVGetCommand_Keys: Started DNS server: address=127.0.0.1:15148 network=tcp === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.841Z [INFO] TestKVGetCommand_Empty: consul server down writer.go:29: 2021-01-29T19:34:10.842Z [INFO] TestKVGetCommand_Empty: shutdown complete === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.841Z [INFO] TestKVGetCommand_Recurse.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:10.843Z [INFO] TestKVGetCommand_Empty: Stopping server: protocol=DNS address=127.0.0.1:15130 network=tcp writer.go:29: 2021-01-29T19:34:10.847Z [INFO] TestKVGetCommand_Empty: Stopping server: protocol=DNS address=127.0.0.1:15130 network=udp writer.go:29: 2021-01-29T19:34:10.847Z [INFO] TestKVGetCommand_Empty: Stopping server: protocol=HTTP address=127.0.0.1:15131 network=tcp === CONT TestKVGetCommand_Keys writer.go:29: 2021-01-29T19:34:10.849Z [WARN] TestKVGetCommand_Keys.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:10.850Z [INFO] TestKVGetCommand_Keys.server.raft: entering candidate state: node="Node at 127.0.0.1:15153 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:10.852Z [INFO] TestKVGetCommand_Keys: Started HTTP server: address=127.0.0.1:15149 network=tcp writer.go:29: 2021-01-29T19:34:10.853Z [INFO] TestKVGetCommand_Keys: started state syncer === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:10.844Z [INFO] TestKVGetCommand_Recurse.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.861Z [DEBUG] TestKVGetCommand_Recurse.server: Skipping self join check for node since the cluster is too small: node=Node-d6178bf0-d804-cf5b-ed29-ba84e567808e writer.go:29: 2021-01-29T19:34:10.868Z [INFO] TestKVGetCommand_Recurse.server: member joined, marking health alive: member=Node-d6178bf0-d804-cf5b-ed29-ba84e567808e === CONT TestKVGetCommand_Keys writer.go:29: 2021-01-29T19:34:10.876Z [DEBUG] TestKVGetCommand_Keys.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:10.878Z [DEBUG] TestKVGetCommand_Keys.server.raft: vote granted: from=e290dd3e-0fd1-6d97-9017-00c97977d459 term=2 tally=1 writer.go:29: 2021-01-29T19:34:10.880Z [INFO] TestKVGetCommand_Keys.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:10.882Z [INFO] TestKVGetCommand_Keys.server.raft: entering leader state: leader="Node at 127.0.0.1:15153 [Leader]" writer.go:29: 2021-01-29T19:34:10.885Z [INFO] TestKVGetCommand_Keys.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:10.886Z [INFO] TestKVGetCommand_Keys.server: New leader elected: payload=Node-e290dd3e-0fd1-6d97-9017-00c97977d459 writer.go:29: 2021-01-29T19:34:10.890Z [DEBUG] TestKVGetCommand_Keys.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15153 writer.go:29: 2021-01-29T19:34:10.906Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:10.925Z [INFO] TestKVGetCommand_Keys.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:10.927Z [INFO] TestKVGetCommand_Keys.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.929Z [DEBUG] TestKVGetCommand_Keys.server: Skipping self join check for node since the cluster is too small: node=Node-e290dd3e-0fd1-6d97-9017-00c97977d459 writer.go:29: 2021-01-29T19:34:10.928Z [DEBUG] TestKVGetCommand_Keys.http: Request finished: method=PUT url=/v1/kv/foo/bar from=127.0.0.1:51952 latency=5.257611ms writer.go:29: 2021-01-29T19:34:10.934Z [INFO] TestKVGetCommand_Keys.server: member joined, marking health alive: member=Node-e290dd3e-0fd1-6d97-9017-00c97977d459 writer.go:29: 2021-01-29T19:34:10.936Z [DEBUG] TestKVGetCommand_Keys.http: Request finished: method=PUT url=/v1/kv/foo/baz from=127.0.0.1:51952 latency=950.323µs writer.go:29: 2021-01-29T19:34:10.945Z [DEBUG] TestKVGetCommand_Keys.http: Request finished: method=PUT url=/v1/kv/foo/zip from=127.0.0.1:51952 latency=2.929775ms writer.go:29: 2021-01-29T19:34:10.957Z [DEBUG] TestKVGetCommand_Keys.http: Request finished: method=GET url=/v1/kv/foo/?keys=&separator=%2F from=127.0.0.1:51954 latency=403.856µs writer.go:29: 2021-01-29T19:34:10.965Z [INFO] TestKVGetCommand_Keys: Requesting shutdown writer.go:29: 2021-01-29T19:34:10.972Z [INFO] TestKVGetCommand_Keys.server: shutting down server writer.go:29: 2021-01-29T19:34:10.979Z [DEBUG] TestKVGetCommand_Keys.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.980Z [WARN] TestKVGetCommand_Keys.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:10.981Z [ERROR] TestKVGetCommand_Keys.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:10.982Z [DEBUG] TestKVGetCommand_Keys.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:10.985Z [WARN] TestKVGetCommand_Keys.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:11.006Z [INFO] TestKVGetCommand_Keys.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:11.007Z [INFO] TestKVGetCommand_Keys: consul server down writer.go:29: 2021-01-29T19:34:11.011Z [INFO] TestKVGetCommand_Keys: shutdown complete writer.go:29: 2021-01-29T19:34:11.013Z [INFO] TestKVGetCommand_Keys: Stopping server: protocol=DNS address=127.0.0.1:15148 network=tcp writer.go:29: 2021-01-29T19:34:11.016Z [INFO] TestKVGetCommand_Keys: Stopping server: protocol=DNS address=127.0.0.1:15148 network=udp writer.go:29: 2021-01-29T19:34:11.018Z [INFO] TestKVGetCommand_Keys: Stopping server: protocol=HTTP address=127.0.0.1:15149 network=tcp === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:11.012Z [DEBUG] TestKVGetCommand_Recurse.http: Request finished: method=PUT url=/v1/kv/foo/a from=127.0.0.1:44698 latency=12.200514ms writer.go:29: 2021-01-29T19:34:11.028Z [DEBUG] TestKVGetCommand_Recurse.http: Request finished: method=PUT url=/v1/kv/foo/b from=127.0.0.1:44698 latency=1.166633ms writer.go:29: 2021-01-29T19:34:11.032Z [DEBUG] TestKVGetCommand_Recurse.http: Request finished: method=PUT url=/v1/kv/foo/c from=127.0.0.1:44698 latency=1.58987ms writer.go:29: 2021-01-29T19:34:11.040Z [DEBUG] TestKVGetCommand_Recurse.http: Request finished: method=GET url=/v1/kv/foo?recurse= from=127.0.0.1:44700 latency=101.804µs writer.go:29: 2021-01-29T19:34:11.044Z [INFO] TestKVGetCommand_Recurse: Requesting shutdown writer.go:29: 2021-01-29T19:34:11.046Z [INFO] TestKVGetCommand_Recurse.server: shutting down server writer.go:29: 2021-01-29T19:34:11.048Z [DEBUG] TestKVGetCommand_Recurse.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:11.050Z [WARN] TestKVGetCommand_Recurse.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:11.050Z [ERROR] TestKVGetCommand_Recurse.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:11.051Z [DEBUG] TestKVGetCommand_Recurse.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:11.059Z [WARN] TestKVGetCommand_Recurse.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:11.065Z [INFO] TestKVGetCommand_Recurse.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:11.068Z [INFO] TestKVGetCommand_Recurse: consul server down writer.go:29: 2021-01-29T19:34:11.070Z [INFO] TestKVGetCommand_Recurse: shutdown complete writer.go:29: 2021-01-29T19:34:11.073Z [INFO] TestKVGetCommand_Recurse: Stopping server: protocol=DNS address=127.0.0.1:15142 network=tcp writer.go:29: 2021-01-29T19:34:11.089Z [INFO] TestKVGetCommand_Recurse: Stopping server: protocol=DNS address=127.0.0.1:15142 network=udp writer.go:29: 2021-01-29T19:34:11.096Z [INFO] TestKVGetCommand_Recurse: Stopping server: protocol=HTTP address=127.0.0.1:15143 network=tcp === CONT TestKVGetCommand_Missing writer.go:29: 2021-01-29T19:34:11.331Z [INFO] TestKVGetCommand_Missing: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:11.333Z [INFO] TestKVGetCommand_Missing: Endpoints down --- PASS: TestKVGetCommand_Missing (0.82s) === CONT TestKVGetCommand_Validation --- PASS: TestKVGetCommand_Validation (0.01s) === CONT TestKVGetCommand_Empty writer.go:29: 2021-01-29T19:34:11.349Z [INFO] TestKVGetCommand_Empty: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:11.351Z [INFO] TestKVGetCommand_Empty: Endpoints down --- PASS: TestKVGetCommand_Empty (0.98s) === CONT TestKVGetCommand_Keys writer.go:29: 2021-01-29T19:34:11.521Z [INFO] TestKVGetCommand_Keys: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:11.522Z [INFO] TestKVGetCommand_Keys: Endpoints down --- PASS: TestKVGetCommand_Keys (0.81s) === CONT TestKVGetCommand_Recurse writer.go:29: 2021-01-29T19:34:11.598Z [INFO] TestKVGetCommand_Recurse: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:11.599Z [INFO] TestKVGetCommand_Recurse: Endpoints down --- PASS: TestKVGetCommand_Recurse (1.06s) PASS ok github.com/hashicorp/consul/command/kv/get 2.525s === RUN TestKVImportCommand_noTabs === PAUSE TestKVImportCommand_noTabs === RUN TestKVImportCommand === PAUSE TestKVImportCommand === CONT TestKVImportCommand_noTabs === CONT TestKVImportCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestKVImportCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestKVImportCommand writer.go:29: 2021-01-29T19:34:14.523Z [WARN] TestKVImportCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:14.525Z [DEBUG] TestKVImportCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:14.528Z [DEBUG] TestKVImportCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:14.545Z [INFO] TestKVImportCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:358b73e3-430b-e696-e451-15cc3ebb78e2 Address:127.0.0.1:26342}]" writer.go:29: 2021-01-29T19:34:14.548Z [INFO] TestKVImportCommand.server.serf.wan: serf: EventMemberJoin: Node-358b73e3-430b-e696-e451-15cc3ebb78e2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:14.549Z [INFO] TestKVImportCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:26342 [Follower]" leader= writer.go:29: 2021-01-29T19:34:14.550Z [INFO] TestKVImportCommand.server.serf.lan: serf: EventMemberJoin: Node-358b73e3-430b-e696-e451-15cc3ebb78e2 127.0.0.1 writer.go:29: 2021-01-29T19:34:14.572Z [INFO] TestKVImportCommand: Started DNS server: address=127.0.0.1:26337 network=udp writer.go:29: 2021-01-29T19:34:14.574Z [INFO] TestKVImportCommand.server: Handled event for server in area: event=member-join server=Node-358b73e3-430b-e696-e451-15cc3ebb78e2.dc1 area=wan writer.go:29: 2021-01-29T19:34:14.574Z [INFO] TestKVImportCommand.server: Adding LAN server: server="Node-358b73e3-430b-e696-e451-15cc3ebb78e2 (Addr: tcp/127.0.0.1:26342) (DC: dc1)" writer.go:29: 2021-01-29T19:34:14.576Z [INFO] TestKVImportCommand: Started DNS server: address=127.0.0.1:26337 network=tcp writer.go:29: 2021-01-29T19:34:14.585Z [INFO] TestKVImportCommand: Started HTTP server: address=127.0.0.1:26338 network=tcp writer.go:29: 2021-01-29T19:34:14.587Z [INFO] TestKVImportCommand: started state syncer writer.go:29: 2021-01-29T19:34:14.605Z [WARN] TestKVImportCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:14.607Z [INFO] TestKVImportCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:26342 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:14.611Z [DEBUG] TestKVImportCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:14.615Z [DEBUG] TestKVImportCommand.server.raft: vote granted: from=358b73e3-430b-e696-e451-15cc3ebb78e2 term=2 tally=1 writer.go:29: 2021-01-29T19:34:14.620Z [INFO] TestKVImportCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:14.627Z [INFO] TestKVImportCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:26342 [Leader]" writer.go:29: 2021-01-29T19:34:14.632Z [INFO] TestKVImportCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:14.634Z [INFO] TestKVImportCommand.server: New leader elected: payload=Node-358b73e3-430b-e696-e451-15cc3ebb78e2 writer.go:29: 2021-01-29T19:34:14.637Z [DEBUG] TestKVImportCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:26342 writer.go:29: 2021-01-29T19:34:14.660Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:14.739Z [INFO] TestKVImportCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:14.740Z [INFO] TestKVImportCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.743Z [DEBUG] TestKVImportCommand.server: Skipping self join check for node since the cluster is too small: node=Node-358b73e3-430b-e696-e451-15cc3ebb78e2 writer.go:29: 2021-01-29T19:34:14.745Z [INFO] TestKVImportCommand.server: member joined, marking health alive: member=Node-358b73e3-430b-e696-e451-15cc3ebb78e2 writer.go:29: 2021-01-29T19:34:14.777Z [DEBUG] TestKVImportCommand.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:58068 latency=4.149249ms writer.go:29: 2021-01-29T19:34:14.782Z [DEBUG] TestKVImportCommand.http: Request finished: method=PUT url=/v1/kv/foo/a from=127.0.0.1:58068 latency=2.048859ms writer.go:29: 2021-01-29T19:34:14.790Z [DEBUG] TestKVImportCommand.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:58070 latency=3.316894ms writer.go:29: 2021-01-29T19:34:14.800Z [DEBUG] TestKVImportCommand.http: Request finished: method=GET url=/v1/kv/foo/a from=127.0.0.1:58070 latency=337.976µs writer.go:29: 2021-01-29T19:34:14.823Z [INFO] TestKVImportCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:14.824Z [INFO] TestKVImportCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:14.826Z [DEBUG] TestKVImportCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.826Z [ERROR] TestKVImportCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:14.827Z [WARN] TestKVImportCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:14.828Z [DEBUG] TestKVImportCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:14.836Z [WARN] TestKVImportCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:14.839Z [INFO] TestKVImportCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:14.841Z [INFO] TestKVImportCommand: consul server down writer.go:29: 2021-01-29T19:34:14.843Z [INFO] TestKVImportCommand: shutdown complete writer.go:29: 2021-01-29T19:34:14.844Z [INFO] TestKVImportCommand: Stopping server: protocol=DNS address=127.0.0.1:26337 network=tcp writer.go:29: 2021-01-29T19:34:14.845Z [INFO] TestKVImportCommand: Stopping server: protocol=DNS address=127.0.0.1:26337 network=udp writer.go:29: 2021-01-29T19:34:14.847Z [INFO] TestKVImportCommand: Stopping server: protocol=HTTP address=127.0.0.1:26338 network=tcp writer.go:29: 2021-01-29T19:34:15.348Z [INFO] TestKVImportCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:15.350Z [INFO] TestKVImportCommand: Endpoints down --- PASS: TestKVImportCommand (0.93s) PASS ok github.com/hashicorp/consul/command/kv/imp 1.048s ? github.com/hashicorp/consul/command/kv/impexp [no test files] === RUN TestKVPutCommand_noTabs === PAUSE TestKVPutCommand_noTabs === RUN TestKVPutCommand_Validation === PAUSE TestKVPutCommand_Validation === RUN TestKVPutCommand === PAUSE TestKVPutCommand === RUN TestKVPutCommand_EmptyDataQuoted kv_put_test.go:108: DM-skipped --- SKIP: TestKVPutCommand_EmptyDataQuoted (0.00s) === RUN TestKVPutCommand_Base64 === PAUSE TestKVPutCommand_Base64 === RUN TestKVPutCommand_File === PAUSE TestKVPutCommand_File === RUN TestKVPutCommand_FileNoExist === PAUSE TestKVPutCommand_FileNoExist === RUN TestKVPutCommand_Stdin === PAUSE TestKVPutCommand_Stdin === RUN TestKVPutCommand_NegativeVal === PAUSE TestKVPutCommand_NegativeVal === RUN TestKVPutCommand_Flags === PAUSE TestKVPutCommand_Flags === RUN TestKVPutCommand_CAS === PAUSE TestKVPutCommand_CAS === CONT TestKVPutCommand_noTabs --- PASS: TestKVPutCommand_noTabs (0.00s) === CONT TestKVPutCommand_CAS === CONT TestKVPutCommand_FileNoExist === CONT TestKVPutCommand_Base64 === CONT TestKVPutCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestKVPutCommand_FileNoExist (0.00s) === CONT TestKVPutCommand_File [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:34:15.274Z [WARN] TestKVPutCommand_File: bootstrap = true: do not enable unless necessary === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.275Z [WARN] TestKVPutCommand_CAS: bootstrap = true: do not enable unless necessary === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.276Z [WARN] TestKVPutCommand_Base64: bootstrap = true: do not enable unless necessary === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.276Z [DEBUG] TestKVPutCommand_File.tlsutil: Update: version=1 === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.277Z [DEBUG] TestKVPutCommand_CAS.tlsutil: Update: version=1 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.278Z [DEBUG] TestKVPutCommand_Base64.tlsutil: Update: version=1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.278Z [WARN] TestKVPutCommand: bootstrap = true: do not enable unless necessary === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.281Z [DEBUG] TestKVPutCommand_CAS.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.281Z [DEBUG] TestKVPutCommand.tlsutil: Update: version=1 === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.285Z [DEBUG] TestKVPutCommand_File.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.285Z [DEBUG] TestKVPutCommand_Base64.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.284Z [DEBUG] TestKVPutCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.307Z [INFO] TestKVPutCommand_CAS.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:083c0cfc-2a83-c1aa-9144-39e21855a865 Address:127.0.0.1:23279}]" writer.go:29: 2021-01-29T19:34:15.310Z [INFO] TestKVPutCommand_CAS.server.raft: entering follower state: follower="Node at 127.0.0.1:23279 [Follower]" leader= === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.310Z [INFO] TestKVPutCommand_Base64.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:98c7ff0d-6716-3bea-c5c2-6fd3108d7c91 Address:127.0.0.1:23285}]" writer.go:29: 2021-01-29T19:34:15.316Z [INFO] TestKVPutCommand_Base64.server.serf.wan: serf: EventMemberJoin: Node-98c7ff0d-6716-3bea-c5c2-6fd3108d7c91.dc1 127.0.0.1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.315Z [INFO] TestKVPutCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:522f76ef-762d-58de-70d3-07ac44c963c6 Address:127.0.0.1:23291}]" === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.315Z [INFO] TestKVPutCommand_CAS.server.serf.wan: serf: EventMemberJoin: Node-083c0cfc-2a83-c1aa-9144-39e21855a865.dc1 127.0.0.1 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.318Z [INFO] TestKVPutCommand_Base64.server.serf.lan: serf: EventMemberJoin: Node-98c7ff0d-6716-3bea-c5c2-6fd3108d7c91 127.0.0.1 === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.319Z [INFO] TestKVPutCommand_CAS.server.serf.lan: serf: EventMemberJoin: Node-083c0cfc-2a83-c1aa-9144-39e21855a865 127.0.0.1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.319Z [INFO] TestKVPutCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:23291 [Follower]" leader= === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.322Z [INFO] TestKVPutCommand_Base64: Started DNS server: address=127.0.0.1:23280 network=udp === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.322Z [INFO] TestKVPutCommand.server.serf.wan: serf: EventMemberJoin: Node-522f76ef-762d-58de-70d3-07ac44c963c6.dc1 127.0.0.1 === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.324Z [INFO] TestKVPutCommand_File.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:64797916-d4a3-2dbb-2ebb-65db5696b831 Address:127.0.0.1:23297}]" === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.324Z [INFO] TestKVPutCommand_Base64.server.raft: entering follower state: follower="Node at 127.0.0.1:23285 [Follower]" leader= === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.326Z [INFO] TestKVPutCommand.server.serf.lan: serf: EventMemberJoin: Node-522f76ef-762d-58de-70d3-07ac44c963c6 127.0.0.1 === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.323Z [INFO] TestKVPutCommand_CAS: Started DNS server: address=127.0.0.1:23274 network=udp writer.go:29: 2021-01-29T19:34:15.330Z [INFO] TestKVPutCommand_CAS: Started DNS server: address=127.0.0.1:23274 network=tcp === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.332Z [INFO] TestKVPutCommand.server: Adding LAN server: server="Node-522f76ef-762d-58de-70d3-07ac44c963c6 (Addr: tcp/127.0.0.1:23291) (DC: dc1)" === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.327Z [INFO] TestKVPutCommand_File.server.serf.wan: serf: EventMemberJoin: Node-64797916-d4a3-2dbb-2ebb-65db5696b831.dc1 127.0.0.1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.333Z [INFO] TestKVPutCommand.server: Handled event for server in area: event=member-join server=Node-522f76ef-762d-58de-70d3-07ac44c963c6.dc1 area=wan === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.334Z [INFO] TestKVPutCommand_CAS: Started HTTP server: address=127.0.0.1:23275 network=tcp === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.337Z [INFO] TestKVPutCommand: Started DNS server: address=127.0.0.1:23286 network=udp === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.337Z [INFO] TestKVPutCommand_CAS: started state syncer === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.335Z [INFO] TestKVPutCommand_File.server.raft: entering follower state: follower="Node at 127.0.0.1:23297 [Follower]" leader= === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.340Z [INFO] TestKVPutCommand: Started DNS server: address=127.0.0.1:23286 network=tcp === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.328Z [INFO] TestKVPutCommand_CAS.server: Adding LAN server: server="Node-083c0cfc-2a83-c1aa-9144-39e21855a865 (Addr: tcp/127.0.0.1:23279) (DC: dc1)" === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.342Z [INFO] TestKVPutCommand_Base64.server: Handled event for server in area: event=member-join server=Node-98c7ff0d-6716-3bea-c5c2-6fd3108d7c91.dc1 area=wan === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.343Z [INFO] TestKVPutCommand_File.server.serf.lan: serf: EventMemberJoin: Node-64797916-d4a3-2dbb-2ebb-65db5696b831 127.0.0.1 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.345Z [INFO] TestKVPutCommand_Base64.server: Adding LAN server: server="Node-98c7ff0d-6716-3bea-c5c2-6fd3108d7c91 (Addr: tcp/127.0.0.1:23285) (DC: dc1)" === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.344Z [INFO] TestKVPutCommand: Started HTTP server: address=127.0.0.1:23287 network=tcp === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.328Z [INFO] TestKVPutCommand_CAS.server: Handled event for server in area: event=member-join server=Node-083c0cfc-2a83-c1aa-9144-39e21855a865.dc1 area=wan === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.347Z [INFO] TestKVPutCommand: started state syncer === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.346Z [INFO] TestKVPutCommand_Base64: Started DNS server: address=127.0.0.1:23280 network=tcp === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.347Z [INFO] TestKVPutCommand_File: Started DNS server: address=127.0.0.1:23292 network=udp writer.go:29: 2021-01-29T19:34:15.349Z [INFO] TestKVPutCommand_File.server: Adding LAN server: server="Node-64797916-d4a3-2dbb-2ebb-65db5696b831 (Addr: tcp/127.0.0.1:23297) (DC: dc1)" writer.go:29: 2021-01-29T19:34:15.349Z [INFO] TestKVPutCommand_File.server: Handled event for server in area: event=member-join server=Node-64797916-d4a3-2dbb-2ebb-65db5696b831.dc1 area=wan === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.357Z [INFO] TestKVPutCommand_Base64: Started HTTP server: address=127.0.0.1:23281 network=tcp writer.go:29: 2021-01-29T19:34:15.358Z [INFO] TestKVPutCommand_Base64: started state syncer === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.359Z [INFO] TestKVPutCommand_File: Started DNS server: address=127.0.0.1:23292 network=tcp writer.go:29: 2021-01-29T19:34:15.368Z [INFO] TestKVPutCommand_File: Started HTTP server: address=127.0.0.1:23293 network=tcp writer.go:29: 2021-01-29T19:34:15.370Z [INFO] TestKVPutCommand_File: started state syncer === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.383Z [WARN] TestKVPutCommand_Base64.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.383Z [WARN] TestKVPutCommand_CAS.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.385Z [INFO] TestKVPutCommand_Base64.server.raft: entering candidate state: node="Node at 127.0.0.1:23285 [Candidate]" term=2 === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.387Z [INFO] TestKVPutCommand_CAS.server.raft: entering candidate state: node="Node at 127.0.0.1:23279 [Candidate]" term=2 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.389Z [DEBUG] TestKVPutCommand_Base64.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:15.391Z [DEBUG] TestKVPutCommand_Base64.server.raft: vote granted: from=98c7ff0d-6716-3bea-c5c2-6fd3108d7c91 term=2 tally=1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.391Z [WARN] TestKVPutCommand.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.392Z [INFO] TestKVPutCommand_Base64.server.raft: election won: tally=1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.393Z [INFO] TestKVPutCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:23291 [Candidate]" term=2 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.394Z [INFO] TestKVPutCommand_Base64.server.raft: entering leader state: leader="Node at 127.0.0.1:23285 [Leader]" === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.396Z [WARN] TestKVPutCommand_File.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:15.397Z [INFO] TestKVPutCommand_File.server.raft: entering candidate state: node="Node at 127.0.0.1:23297 [Candidate]" term=2 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.397Z [INFO] TestKVPutCommand_Base64.server: cluster leadership acquired === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.400Z [DEBUG] TestKVPutCommand.server.raft: votes: needed=1 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.400Z [INFO] TestKVPutCommand_Base64.server: New leader elected: payload=Node-98c7ff0d-6716-3bea-c5c2-6fd3108d7c91 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.401Z [DEBUG] TestKVPutCommand.server.raft: vote granted: from=522f76ef-762d-58de-70d3-07ac44c963c6 term=2 tally=1 writer.go:29: 2021-01-29T19:34:15.403Z [INFO] TestKVPutCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:15.405Z [INFO] TestKVPutCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:23291 [Leader]" writer.go:29: 2021-01-29T19:34:15.407Z [INFO] TestKVPutCommand.server: cluster leadership acquired === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.407Z [DEBUG] TestKVPutCommand_CAS.server.raft: votes: needed=1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.410Z [INFO] TestKVPutCommand.server: New leader elected: payload=Node-522f76ef-762d-58de-70d3-07ac44c963c6 === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.410Z [DEBUG] TestKVPutCommand_CAS.server.raft: vote granted: from=083c0cfc-2a83-c1aa-9144-39e21855a865 term=2 tally=1 writer.go:29: 2021-01-29T19:34:15.411Z [INFO] TestKVPutCommand_CAS.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:15.413Z [INFO] TestKVPutCommand_CAS.server.raft: entering leader state: leader="Node at 127.0.0.1:23279 [Leader]" writer.go:29: 2021-01-29T19:34:15.415Z [INFO] TestKVPutCommand_CAS.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:15.418Z [INFO] TestKVPutCommand_CAS.server: New leader elected: payload=Node-083c0cfc-2a83-c1aa-9144-39e21855a865 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.414Z [DEBUG] TestKVPutCommand_Base64.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23285 === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.420Z [DEBUG] TestKVPutCommand_File.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:15.421Z [DEBUG] TestKVPutCommand_File.server.raft: vote granted: from=64797916-d4a3-2dbb-2ebb-65db5696b831 term=2 tally=1 writer.go:29: 2021-01-29T19:34:15.423Z [INFO] TestKVPutCommand_File.server.raft: election won: tally=1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.422Z [DEBUG] TestKVPutCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23291 === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.425Z [INFO] TestKVPutCommand_File.server.raft: entering leader state: leader="Node at 127.0.0.1:23297 [Leader]" writer.go:29: 2021-01-29T19:34:15.427Z [INFO] TestKVPutCommand_File.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:15.430Z [INFO] TestKVPutCommand_File.server: New leader elected: payload=Node-64797916-d4a3-2dbb-2ebb-65db5696b831 === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.424Z [DEBUG] TestKVPutCommand_CAS.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23279 === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.441Z [DEBUG] TestKVPutCommand_File.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23297 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.444Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.463Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.472Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.490Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:15.499Z [DEBUG] TestKVPutCommand_File.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:49080 latency=16.889438ms === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.517Z [DEBUG] TestKVPutCommand.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:51754 latency=10.703189ms === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.513Z [DEBUG] TestKVPutCommand_File.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:49084 latency=4.101867ms writer.go:29: 2021-01-29T19:34:15.520Z [INFO] TestKVPutCommand_File: Requesting shutdown === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.521Z [DEBUG] TestKVPutCommand.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:51758 latency=259.317µs writer.go:29: 2021-01-29T19:34:15.522Z [INFO] TestKVPutCommand: Requesting shutdown === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.521Z [INFO] TestKVPutCommand_File.server: shutting down server === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.523Z [INFO] TestKVPutCommand.server: shutting down server === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.523Z [WARN] TestKVPutCommand_File.server.serf.lan: serf: Shutdown without a Leave === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.525Z [INFO] TestKVPutCommand_Base64: Synced node info === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.523Z [ERROR] TestKVPutCommand_File.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:15.533Z [WARN] TestKVPutCommand_File.server.serf.wan: serf: Shutdown without a Leave === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.524Z [WARN] TestKVPutCommand.server.serf.lan: serf: Shutdown without a Leave === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.536Z [INFO] TestKVPutCommand_File.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:15.537Z [INFO] TestKVPutCommand_File: consul server down === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.524Z [ERROR] TestKVPutCommand.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.538Z [INFO] TestKVPutCommand_File: shutdown complete writer.go:29: 2021-01-29T19:34:15.539Z [INFO] TestKVPutCommand_File: Stopping server: protocol=DNS address=127.0.0.1:23292 network=tcp === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.540Z [WARN] TestKVPutCommand.server.serf.wan: serf: Shutdown without a Leave === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.540Z [ERROR] TestKVPutCommand_File.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: raft is already shutdown" writer.go:29: 2021-01-29T19:34:15.545Z [INFO] TestKVPutCommand_File: Stopping server: protocol=DNS address=127.0.0.1:23292 network=udp === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.543Z [INFO] TestKVPutCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:15.546Z [INFO] TestKVPutCommand_File: Stopping server: protocol=HTTP address=127.0.0.1:23293 network=tcp === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.549Z [INFO] TestKVPutCommand_Base64.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.553Z [INFO] TestKVPutCommand_CAS.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:15.555Z [INFO] TestKVPutCommand_CAS.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.556Z [DEBUG] TestKVPutCommand_CAS.server: Skipping self join check for node since the cluster is too small: node=Node-083c0cfc-2a83-c1aa-9144-39e21855a865 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.554Z [INFO] TestKVPutCommand_Base64.leader: started routine: routine="CA root pruning" === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.558Z [INFO] TestKVPutCommand_CAS.server: member joined, marking health alive: member=Node-083c0cfc-2a83-c1aa-9144-39e21855a865 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.560Z [DEBUG] TestKVPutCommand_Base64.server: Skipping self join check for node since the cluster is too small: node=Node-98c7ff0d-6716-3bea-c5c2-6fd3108d7c91 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.546Z [INFO] TestKVPutCommand.leader: started routine: routine="CA root pruning" === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.570Z [INFO] TestKVPutCommand_Base64.server: member joined, marking health alive: member=Node-98c7ff0d-6716-3bea-c5c2-6fd3108d7c91 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.572Z [DEBUG] TestKVPutCommand.server: Skipping self join check for node since the cluster is too small: node=Node-522f76ef-762d-58de-70d3-07ac44c963c6 writer.go:29: 2021-01-29T19:34:15.546Z [INFO] TestKVPutCommand: consul server down === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.581Z [DEBUG] TestKVPutCommand_CAS.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:35854 latency=1.102959ms === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:15.546Z [INFO] TestKVPutCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:15.582Z [INFO] TestKVPutCommand.server: member joined, marking health alive: member=Node-522f76ef-762d-58de-70d3-07ac44c963c6 writer.go:29: 2021-01-29T19:34:15.584Z [INFO] TestKVPutCommand: shutdown complete writer.go:29: 2021-01-29T19:34:15.590Z [INFO] TestKVPutCommand: Stopping server: protocol=DNS address=127.0.0.1:23286 network=tcp writer.go:29: 2021-01-29T19:34:15.589Z [ERROR] TestKVPutCommand.server: failed to reconcile member: member="{Node-522f76ef-762d-58de-70d3-07ac44c963c6 127.0.0.1 23289 map[acls:0 bootstrap:1 build:1.7.4: dc:dc1 id:522f76ef-762d-58de-70d3-07ac44c963c6 port:23291 raft_vsn:3 role:consul segment: vsn:2 vsn_max:3 vsn_min:2 wan_join_port:23290] alive 1 5 2 2 5 4}" error="raft is already shutdown" writer.go:29: 2021-01-29T19:34:15.593Z [DEBUG] TestKVPutCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.591Z [INFO] TestKVPutCommand: Stopping server: protocol=DNS address=127.0.0.1:23286 network=udp writer.go:29: 2021-01-29T19:34:15.595Z [DEBUG] TestKVPutCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.597Z [INFO] TestKVPutCommand: Stopping server: protocol=HTTP address=127.0.0.1:23287 network=tcp === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.609Z [DEBUG] TestKVPutCommand_CAS.http: Request finished: method=PUT url=/v1/kv/foo?cas=123 from=127.0.0.1:35856 latency=1.919419ms writer.go:29: 2021-01-29T19:34:15.617Z [DEBUG] TestKVPutCommand_CAS.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:35854 latency=638.419µs writer.go:29: 2021-01-29T19:34:15.635Z [DEBUG] TestKVPutCommand_CAS: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:15.636Z [DEBUG] TestKVPutCommand_CAS.http: Request finished: method=PUT url=/v1/kv/foo?cas=11 from=127.0.0.1:35858 latency=2.991253ms writer.go:29: 2021-01-29T19:34:15.636Z [INFO] TestKVPutCommand_CAS: Synced node info writer.go:29: 2021-01-29T19:34:15.641Z [DEBUG] TestKVPutCommand_CAS: Node info in sync writer.go:29: 2021-01-29T19:34:15.639Z [DEBUG] TestKVPutCommand_CAS.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:35854 latency=141.06µs writer.go:29: 2021-01-29T19:34:15.657Z [INFO] TestKVPutCommand_CAS: Requesting shutdown writer.go:29: 2021-01-29T19:34:15.659Z [INFO] TestKVPutCommand_CAS.server: shutting down server writer.go:29: 2021-01-29T19:34:15.660Z [DEBUG] TestKVPutCommand_CAS.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.661Z [WARN] TestKVPutCommand_CAS.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:15.664Z [DEBUG] TestKVPutCommand_CAS.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.667Z [WARN] TestKVPutCommand_CAS.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:15.670Z [INFO] TestKVPutCommand_CAS.server.router.manager: shutting down === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.680Z [DEBUG] TestKVPutCommand_Base64.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:57312 latency=4.227182ms === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.670Z [INFO] TestKVPutCommand_CAS: consul server down writer.go:29: 2021-01-29T19:34:15.684Z [INFO] TestKVPutCommand_CAS: shutdown complete === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.685Z [DEBUG] TestKVPutCommand_Base64.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:57314 latency=148.242µs === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:15.685Z [INFO] TestKVPutCommand_CAS: Stopping server: protocol=DNS address=127.0.0.1:23274 network=tcp writer.go:29: 2021-01-29T19:34:15.687Z [INFO] TestKVPutCommand_CAS: Stopping server: protocol=DNS address=127.0.0.1:23274 network=udp writer.go:29: 2021-01-29T19:34:15.688Z [INFO] TestKVPutCommand_CAS: Stopping server: protocol=HTTP address=127.0.0.1:23275 network=tcp === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:15.687Z [INFO] TestKVPutCommand_Base64: Requesting shutdown writer.go:29: 2021-01-29T19:34:15.696Z [INFO] TestKVPutCommand_Base64.server: shutting down server writer.go:29: 2021-01-29T19:34:15.714Z [DEBUG] TestKVPutCommand_Base64.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.719Z [DEBUG] TestKVPutCommand_Base64.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:15.721Z [WARN] TestKVPutCommand_Base64.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:15.739Z [WARN] TestKVPutCommand_Base64.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:15.743Z [INFO] TestKVPutCommand_Base64.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:15.743Z [INFO] TestKVPutCommand_Base64: consul server down writer.go:29: 2021-01-29T19:34:15.746Z [INFO] TestKVPutCommand_Base64: shutdown complete writer.go:29: 2021-01-29T19:34:15.747Z [INFO] TestKVPutCommand_Base64: Stopping server: protocol=DNS address=127.0.0.1:23280 network=tcp writer.go:29: 2021-01-29T19:34:15.747Z [INFO] TestKVPutCommand_Base64: Stopping server: protocol=DNS address=127.0.0.1:23280 network=udp writer.go:29: 2021-01-29T19:34:15.749Z [INFO] TestKVPutCommand_Base64: Stopping server: protocol=HTTP address=127.0.0.1:23281 network=tcp === CONT TestKVPutCommand_File writer.go:29: 2021-01-29T19:34:16.047Z [INFO] TestKVPutCommand_File: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:16.049Z [INFO] TestKVPutCommand_File: Endpoints down --- PASS: TestKVPutCommand_File (0.83s) === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.059Z [WARN] TestKVPutCommand_NegativeVal: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:16.060Z [DEBUG] TestKVPutCommand_NegativeVal.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:16.093Z [DEBUG] TestKVPutCommand_NegativeVal.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestKVPutCommand writer.go:29: 2021-01-29T19:34:16.101Z [INFO] TestKVPutCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:16.103Z [INFO] TestKVPutCommand: Endpoints down --- PASS: TestKVPutCommand (0.88s) === CONT TestKVPutCommand_Flags === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.099Z [INFO] TestKVPutCommand_NegativeVal.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ad6bd3da-c648-8dc3-1ba9-2121ef0e9d59 Address:127.0.0.1:23303}]" writer.go:29: 2021-01-29T19:34:16.112Z [INFO] TestKVPutCommand_NegativeVal.server.raft: entering follower state: follower="Node at 127.0.0.1:23303 [Follower]" leader= writer.go:29: 2021-01-29T19:34:16.116Z [INFO] TestKVPutCommand_NegativeVal.server.serf.wan: serf: EventMemberJoin: Node-ad6bd3da-c648-8dc3-1ba9-2121ef0e9d59.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:16.119Z [INFO] TestKVPutCommand_NegativeVal.server.serf.lan: serf: EventMemberJoin: Node-ad6bd3da-c648-8dc3-1ba9-2121ef0e9d59 127.0.0.1 writer.go:29: 2021-01-29T19:34:16.122Z [INFO] TestKVPutCommand_NegativeVal.server: Handled event for server in area: event=member-join server=Node-ad6bd3da-c648-8dc3-1ba9-2121ef0e9d59.dc1 area=wan writer.go:29: 2021-01-29T19:34:16.122Z [INFO] TestKVPutCommand_NegativeVal.server: Adding LAN server: server="Node-ad6bd3da-c648-8dc3-1ba9-2121ef0e9d59 (Addr: tcp/127.0.0.1:23303) (DC: dc1)" writer.go:29: 2021-01-29T19:34:16.123Z [INFO] TestKVPutCommand_NegativeVal: Started DNS server: address=127.0.0.1:23298 network=udp writer.go:29: 2021-01-29T19:34:16.125Z [INFO] TestKVPutCommand_NegativeVal: Started DNS server: address=127.0.0.1:23298 network=tcp writer.go:29: 2021-01-29T19:34:16.128Z [INFO] TestKVPutCommand_NegativeVal: Started HTTP server: address=127.0.0.1:23299 network=tcp writer.go:29: 2021-01-29T19:34:16.128Z [INFO] TestKVPutCommand_NegativeVal: started state syncer === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.130Z [WARN] TestKVPutCommand_Flags: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:16.131Z [DEBUG] TestKVPutCommand_Flags.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:16.133Z [DEBUG] TestKVPutCommand_Flags.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:16.138Z [INFO] TestKVPutCommand_Flags.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:140ead72-740d-8c01-0129-4c98185e32bb Address:127.0.0.1:23309}]" writer.go:29: 2021-01-29T19:34:16.140Z [INFO] TestKVPutCommand_Flags.server.raft: entering follower state: follower="Node at 127.0.0.1:23309 [Follower]" leader= writer.go:29: 2021-01-29T19:34:16.140Z [INFO] TestKVPutCommand_Flags.server.serf.wan: serf: EventMemberJoin: Node-140ead72-740d-8c01-0129-4c98185e32bb.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:16.146Z [INFO] TestKVPutCommand_Flags.server.serf.lan: serf: EventMemberJoin: Node-140ead72-740d-8c01-0129-4c98185e32bb 127.0.0.1 writer.go:29: 2021-01-29T19:34:16.148Z [INFO] TestKVPutCommand_Flags.server: Handled event for server in area: event=member-join server=Node-140ead72-740d-8c01-0129-4c98185e32bb.dc1 area=wan writer.go:29: 2021-01-29T19:34:16.149Z [INFO] TestKVPutCommand_Flags.server: Adding LAN server: server="Node-140ead72-740d-8c01-0129-4c98185e32bb (Addr: tcp/127.0.0.1:23309) (DC: dc1)" writer.go:29: 2021-01-29T19:34:16.150Z [INFO] TestKVPutCommand_Flags: Started DNS server: address=127.0.0.1:23304 network=udp writer.go:29: 2021-01-29T19:34:16.152Z [INFO] TestKVPutCommand_Flags: Started DNS server: address=127.0.0.1:23304 network=tcp writer.go:29: 2021-01-29T19:34:16.153Z [INFO] TestKVPutCommand_Flags: Started HTTP server: address=127.0.0.1:23305 network=tcp writer.go:29: 2021-01-29T19:34:16.154Z [INFO] TestKVPutCommand_Flags: started state syncer === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.168Z [WARN] TestKVPutCommand_NegativeVal.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:16.169Z [INFO] TestKVPutCommand_NegativeVal.server.raft: entering candidate state: node="Node at 127.0.0.1:23303 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:16.172Z [DEBUG] TestKVPutCommand_NegativeVal.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:16.173Z [DEBUG] TestKVPutCommand_NegativeVal.server.raft: vote granted: from=ad6bd3da-c648-8dc3-1ba9-2121ef0e9d59 term=2 tally=1 writer.go:29: 2021-01-29T19:34:16.174Z [INFO] TestKVPutCommand_NegativeVal.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:16.175Z [INFO] TestKVPutCommand_NegativeVal.server.raft: entering leader state: leader="Node at 127.0.0.1:23303 [Leader]" writer.go:29: 2021-01-29T19:34:16.176Z [INFO] TestKVPutCommand_NegativeVal.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:16.177Z [INFO] TestKVPutCommand_NegativeVal.server: New leader elected: payload=Node-ad6bd3da-c648-8dc3-1ba9-2121ef0e9d59 writer.go:29: 2021-01-29T19:34:16.178Z [DEBUG] TestKVPutCommand_NegativeVal.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23303 === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.184Z [WARN] TestKVPutCommand_Flags.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:16.185Z [INFO] TestKVPutCommand_Flags.server.raft: entering candidate state: node="Node at 127.0.0.1:23309 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:16.187Z [DEBUG] TestKVPutCommand_Flags.server.raft: votes: needed=1 === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.187Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.188Z [DEBUG] TestKVPutCommand_Flags.server.raft: vote granted: from=140ead72-740d-8c01-0129-4c98185e32bb term=2 tally=1 writer.go:29: 2021-01-29T19:34:16.188Z [INFO] TestKVPutCommand_Flags.server.raft: election won: tally=1 === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:16.189Z [INFO] TestKVPutCommand_CAS: Waiting for endpoints to shut down === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.189Z [INFO] TestKVPutCommand_Flags.server.raft: entering leader state: leader="Node at 127.0.0.1:23309 [Leader]" === CONT TestKVPutCommand_CAS writer.go:29: 2021-01-29T19:34:16.190Z [INFO] TestKVPutCommand_CAS: Endpoints down --- PASS: TestKVPutCommand_CAS (0.97s) === CONT TestKVPutCommand_Validation === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.190Z [INFO] TestKVPutCommand_Flags.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:16.191Z [INFO] TestKVPutCommand_Flags.server: New leader elected: payload=Node-140ead72-740d-8c01-0129-4c98185e32bb writer.go:29: 2021-01-29T19:34:16.193Z [DEBUG] TestKVPutCommand_Flags.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23309 === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.193Z [INFO] TestKVPutCommand_NegativeVal.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:16.197Z [INFO] TestKVPutCommand_NegativeVal.leader: started routine: routine="CA root pruning" --- PASS: TestKVPutCommand_Validation (0.01s) === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.198Z [DEBUG] TestKVPutCommand_NegativeVal.server: Skipping self join check for node since the cluster is too small: node=Node-ad6bd3da-c648-8dc3-1ba9-2121ef0e9d59 writer.go:29: 2021-01-29T19:34:16.199Z [INFO] TestKVPutCommand_NegativeVal.server: member joined, marking health alive: member=Node-ad6bd3da-c648-8dc3-1ba9-2121ef0e9d59 === CONT TestKVPutCommand_Stdin === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.201Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:16.236Z [INFO] TestKVPutCommand_Flags.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:16.238Z [INFO] TestKVPutCommand_Flags.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:16.239Z [DEBUG] TestKVPutCommand_Flags.server: Skipping self join check for node since the cluster is too small: node=Node-140ead72-740d-8c01-0129-4c98185e32bb writer.go:29: 2021-01-29T19:34:16.240Z [INFO] TestKVPutCommand_Flags.server: member joined, marking health alive: member=Node-140ead72-740d-8c01-0129-4c98185e32bb === CONT TestKVPutCommand_Stdin writer.go:29: 2021-01-29T19:34:16.240Z [WARN] TestKVPutCommand_Stdin: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:16.241Z [DEBUG] TestKVPutCommand_Stdin.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:16.243Z [DEBUG] TestKVPutCommand_Stdin.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:16.248Z [INFO] TestKVPutCommand_Stdin.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:74fda25c-998f-276d-14aa-00b394c9c5dc Address:127.0.0.1:23315}]" writer.go:29: 2021-01-29T19:34:16.250Z [INFO] TestKVPutCommand_Stdin.server.raft: entering follower state: follower="Node at 127.0.0.1:23315 [Follower]" leader= writer.go:29: 2021-01-29T19:34:16.251Z [INFO] TestKVPutCommand_Stdin.server.serf.wan: serf: EventMemberJoin: Node-74fda25c-998f-276d-14aa-00b394c9c5dc.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:16.254Z [INFO] TestKVPutCommand_Stdin.server.serf.lan: serf: EventMemberJoin: Node-74fda25c-998f-276d-14aa-00b394c9c5dc 127.0.0.1 === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:16.250Z [INFO] TestKVPutCommand_Base64: Waiting for endpoints to shut down === CONT TestKVPutCommand_Stdin writer.go:29: 2021-01-29T19:34:16.256Z [INFO] TestKVPutCommand_Stdin.server: Handled event for server in area: event=member-join server=Node-74fda25c-998f-276d-14aa-00b394c9c5dc.dc1 area=wan === CONT TestKVPutCommand_Base64 writer.go:29: 2021-01-29T19:34:16.257Z [INFO] TestKVPutCommand_Base64: Endpoints down --- PASS: TestKVPutCommand_Base64 (1.04s) === CONT TestKVPutCommand_Stdin writer.go:29: 2021-01-29T19:34:16.256Z [INFO] TestKVPutCommand_Stdin.server: Adding LAN server: server="Node-74fda25c-998f-276d-14aa-00b394c9c5dc (Addr: tcp/127.0.0.1:23315) (DC: dc1)" writer.go:29: 2021-01-29T19:34:16.256Z [INFO] TestKVPutCommand_Stdin: Started DNS server: address=127.0.0.1:23310 network=udp writer.go:29: 2021-01-29T19:34:16.260Z [INFO] TestKVPutCommand_Stdin: Started DNS server: address=127.0.0.1:23310 network=tcp writer.go:29: 2021-01-29T19:34:16.261Z [INFO] TestKVPutCommand_Stdin: Started HTTP server: address=127.0.0.1:23311 network=tcp writer.go:29: 2021-01-29T19:34:16.262Z [INFO] TestKVPutCommand_Stdin: started state syncer === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.269Z [DEBUG] TestKVPutCommand_Flags: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:16.272Z [INFO] TestKVPutCommand_Flags: Synced node info writer.go:29: 2021-01-29T19:34:16.274Z [DEBUG] TestKVPutCommand_Flags: Node info in sync === CONT TestKVPutCommand_Stdin writer.go:29: 2021-01-29T19:34:16.322Z [WARN] TestKVPutCommand_Stdin.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:16.323Z [INFO] TestKVPutCommand_Stdin.server.raft: entering candidate state: node="Node at 127.0.0.1:23315 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:16.330Z [DEBUG] TestKVPutCommand_Stdin.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:16.333Z [DEBUG] TestKVPutCommand_Stdin.server.raft: vote granted: from=74fda25c-998f-276d-14aa-00b394c9c5dc term=2 tally=1 writer.go:29: 2021-01-29T19:34:16.335Z [INFO] TestKVPutCommand_Stdin.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:16.337Z [INFO] TestKVPutCommand_Stdin.server.raft: entering leader state: leader="Node at 127.0.0.1:23315 [Leader]" writer.go:29: 2021-01-29T19:34:16.339Z [INFO] TestKVPutCommand_Stdin.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:16.340Z [INFO] TestKVPutCommand_Stdin.server: New leader elected: payload=Node-74fda25c-998f-276d-14aa-00b394c9c5dc writer.go:29: 2021-01-29T19:34:16.340Z [DEBUG] TestKVPutCommand_Stdin.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23315 writer.go:29: 2021-01-29T19:34:16.372Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:16.393Z [INFO] TestKVPutCommand_Stdin.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.412Z [DEBUG] TestKVPutCommand_Flags.http: Request finished: method=PUT url=/v1/kv/foo?flags=12345 from=127.0.0.1:34414 latency=4.617701ms === CONT TestKVPutCommand_Stdin writer.go:29: 2021-01-29T19:34:16.410Z [INFO] TestKVPutCommand_Stdin.leader: started routine: routine="CA root pruning" === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.415Z [DEBUG] TestKVPutCommand_NegativeVal.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:44024 latency=4.049381ms === CONT TestKVPutCommand_Stdin writer.go:29: 2021-01-29T19:34:16.418Z [DEBUG] TestKVPutCommand_Stdin.server: Skipping self join check for node since the cluster is too small: node=Node-74fda25c-998f-276d-14aa-00b394c9c5dc === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.420Z [DEBUG] TestKVPutCommand_Flags.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:34418 latency=3.883982ms writer.go:29: 2021-01-29T19:34:16.423Z [INFO] TestKVPutCommand_Flags: Requesting shutdown writer.go:29: 2021-01-29T19:34:16.425Z [INFO] TestKVPutCommand_Flags.server: shutting down server writer.go:29: 2021-01-29T19:34:16.426Z [DEBUG] TestKVPutCommand_Flags.leader: stopping routine: routine="CA root pruning" === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.422Z [DEBUG] TestKVPutCommand_NegativeVal.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:44028 latency=62.781µs === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.428Z [WARN] TestKVPutCommand_Flags.server.serf.lan: serf: Shutdown without a Leave === CONT TestKVPutCommand_Stdin writer.go:29: 2021-01-29T19:34:16.421Z [INFO] TestKVPutCommand_Stdin.server: member joined, marking health alive: member=Node-74fda25c-998f-276d-14aa-00b394c9c5dc === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.432Z [DEBUG] TestKVPutCommand_Flags.leader: stopped routine: routine="CA root pruning" === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.430Z [INFO] TestKVPutCommand_NegativeVal: Requesting shutdown === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.434Z [WARN] TestKVPutCommand_Flags.server.serf.wan: serf: Shutdown without a Leave === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.434Z [INFO] TestKVPutCommand_NegativeVal.server: shutting down server writer.go:29: 2021-01-29T19:34:16.438Z [DEBUG] TestKVPutCommand_NegativeVal.leader: stopping routine: routine="CA root pruning" === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.440Z [INFO] TestKVPutCommand_Flags.server.router.manager: shutting down === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.441Z [WARN] TestKVPutCommand_NegativeVal.server.serf.lan: serf: Shutdown without a Leave === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.442Z [INFO] TestKVPutCommand_Flags: consul server down writer.go:29: 2021-01-29T19:34:16.444Z [INFO] TestKVPutCommand_Flags: shutdown complete === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.439Z [ERROR] TestKVPutCommand_NegativeVal.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.445Z [INFO] TestKVPutCommand_Flags: Stopping server: protocol=DNS address=127.0.0.1:23304 network=tcp writer.go:29: 2021-01-29T19:34:16.447Z [INFO] TestKVPutCommand_Flags: Stopping server: protocol=DNS address=127.0.0.1:23304 network=udp === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.441Z [DEBUG] TestKVPutCommand_NegativeVal.leader: stopped routine: routine="CA root pruning" === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.448Z [INFO] TestKVPutCommand_Flags: Stopping server: protocol=HTTP address=127.0.0.1:23305 network=tcp === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.445Z [WARN] TestKVPutCommand_NegativeVal.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:16.451Z [INFO] TestKVPutCommand_NegativeVal.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:16.452Z [INFO] TestKVPutCommand_NegativeVal: consul server down writer.go:29: 2021-01-29T19:34:16.454Z [INFO] TestKVPutCommand_NegativeVal: shutdown complete writer.go:29: 2021-01-29T19:34:16.456Z [INFO] TestKVPutCommand_NegativeVal: Stopping server: protocol=DNS address=127.0.0.1:23298 network=tcp writer.go:29: 2021-01-29T19:34:16.457Z [INFO] TestKVPutCommand_NegativeVal: Stopping server: protocol=DNS address=127.0.0.1:23298 network=udp writer.go:29: 2021-01-29T19:34:16.459Z [INFO] TestKVPutCommand_NegativeVal: Stopping server: protocol=HTTP address=127.0.0.1:23299 network=tcp === CONT TestKVPutCommand_Stdin writer.go:29: 2021-01-29T19:34:16.532Z [DEBUG] TestKVPutCommand_Stdin.http: Request finished: method=PUT url=/v1/kv/foo from=127.0.0.1:50542 latency=1.085467ms writer.go:29: 2021-01-29T19:34:16.538Z [DEBUG] TestKVPutCommand_Stdin.http: Request finished: method=GET url=/v1/kv/foo from=127.0.0.1:50544 latency=1.276188ms writer.go:29: 2021-01-29T19:34:16.539Z [INFO] TestKVPutCommand_Stdin: Requesting shutdown writer.go:29: 2021-01-29T19:34:16.541Z [INFO] TestKVPutCommand_Stdin.server: shutting down server writer.go:29: 2021-01-29T19:34:16.543Z [DEBUG] TestKVPutCommand_Stdin.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:16.545Z [WARN] TestKVPutCommand_Stdin.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:16.545Z [ERROR] TestKVPutCommand_Stdin.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:16.546Z [DEBUG] TestKVPutCommand_Stdin.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:16.567Z [WARN] TestKVPutCommand_Stdin.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:16.576Z [INFO] TestKVPutCommand_Stdin.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:16.576Z [INFO] TestKVPutCommand_Stdin: consul server down writer.go:29: 2021-01-29T19:34:16.578Z [INFO] TestKVPutCommand_Stdin: shutdown complete writer.go:29: 2021-01-29T19:34:16.579Z [INFO] TestKVPutCommand_Stdin: Stopping server: protocol=DNS address=127.0.0.1:23310 network=tcp writer.go:29: 2021-01-29T19:34:16.580Z [INFO] TestKVPutCommand_Stdin: Stopping server: protocol=DNS address=127.0.0.1:23310 network=udp writer.go:29: 2021-01-29T19:34:16.580Z [INFO] TestKVPutCommand_Stdin: Stopping server: protocol=HTTP address=127.0.0.1:23311 network=tcp === CONT TestKVPutCommand_Flags writer.go:29: 2021-01-29T19:34:16.949Z [INFO] TestKVPutCommand_Flags: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:16.950Z [INFO] TestKVPutCommand_Flags: Endpoints down --- PASS: TestKVPutCommand_Flags (0.85s) === CONT TestKVPutCommand_NegativeVal writer.go:29: 2021-01-29T19:34:16.966Z [INFO] TestKVPutCommand_NegativeVal: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:16.967Z [INFO] TestKVPutCommand_NegativeVal: Endpoints down --- PASS: TestKVPutCommand_NegativeVal (0.92s) === CONT TestKVPutCommand_Stdin writer.go:29: 2021-01-29T19:34:17.082Z [INFO] TestKVPutCommand_Stdin: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:17.082Z [INFO] TestKVPutCommand_Stdin: Endpoints down --- PASS: TestKVPutCommand_Stdin (0.88s) PASS ok github.com/hashicorp/consul/command/kv/put 2.031s === RUN TestLeaveCommand_noTabs === PAUSE TestLeaveCommand_noTabs === RUN TestLeaveCommand === PAUSE TestLeaveCommand === RUN TestLeaveCommand_FailOnNonFlagArgs === PAUSE TestLeaveCommand_FailOnNonFlagArgs === CONT TestLeaveCommand_noTabs === CONT TestLeaveCommand_FailOnNonFlagArgs === CONT TestLeaveCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestLeaveCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestLeaveCommand_FailOnNonFlagArgs writer.go:29: 2021-01-29T19:34:18.427Z [WARN] TestLeaveCommand_FailOnNonFlagArgs: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:18.428Z [DEBUG] TestLeaveCommand_FailOnNonFlagArgs.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:18.430Z [DEBUG] TestLeaveCommand_FailOnNonFlagArgs.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLeaveCommand writer.go:29: 2021-01-29T19:34:18.432Z [WARN] TestLeaveCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:18.434Z [DEBUG] TestLeaveCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:18.436Z [DEBUG] TestLeaveCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLeaveCommand_FailOnNonFlagArgs writer.go:29: 2021-01-29T19:34:18.439Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:afa4e545-5c7e-10af-f1b0-a4973baa0ed6 Address:127.0.0.1:21243}]" writer.go:29: 2021-01-29T19:34:18.441Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server.serf.wan: serf: EventMemberJoin: Node-afa4e545-5c7e-10af-f1b0-a4973baa0ed6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.443Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server.serf.lan: serf: EventMemberJoin: Node-afa4e545-5c7e-10af-f1b0-a4973baa0ed6 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.444Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server.raft: entering follower state: follower="Node at 127.0.0.1:21243 [Follower]" leader= writer.go:29: 2021-01-29T19:34:18.451Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: Started DNS server: address=127.0.0.1:21238 network=udp writer.go:29: 2021-01-29T19:34:18.455Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server: Adding LAN server: server="Node-afa4e545-5c7e-10af-f1b0-a4973baa0ed6 (Addr: tcp/127.0.0.1:21243) (DC: dc1)" writer.go:29: 2021-01-29T19:34:18.455Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: Started DNS server: address=127.0.0.1:21238 network=tcp writer.go:29: 2021-01-29T19:34:18.458Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: Started HTTP server: address=127.0.0.1:21239 network=tcp writer.go:29: 2021-01-29T19:34:18.459Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: started state syncer writer.go:29: 2021-01-29T19:34:18.455Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server: Handled event for server in area: event=member-join server=Node-afa4e545-5c7e-10af-f1b0-a4973baa0ed6.dc1 area=wan === CONT TestLeaveCommand writer.go:29: 2021-01-29T19:34:18.469Z [INFO] TestLeaveCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b1bbbd82-a655-f12e-7667-30e892126594 Address:127.0.0.1:21237}]" writer.go:29: 2021-01-29T19:34:18.473Z [INFO] TestLeaveCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:21237 [Follower]" leader= === CONT TestLeaveCommand_FailOnNonFlagArgs writer.go:29: 2021-01-29T19:34:18.499Z [WARN] TestLeaveCommand_FailOnNonFlagArgs.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestLeaveCommand writer.go:29: 2021-01-29T19:34:18.475Z [INFO] TestLeaveCommand.server.serf.wan: serf: EventMemberJoin: Node-b1bbbd82-a655-f12e-7667-30e892126594.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.502Z [INFO] TestLeaveCommand.server.serf.lan: serf: EventMemberJoin: Node-b1bbbd82-a655-f12e-7667-30e892126594 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.505Z [INFO] TestLeaveCommand: Started DNS server: address=127.0.0.1:21232 network=udp writer.go:29: 2021-01-29T19:34:18.506Z [INFO] TestLeaveCommand.server: Adding LAN server: server="Node-b1bbbd82-a655-f12e-7667-30e892126594 (Addr: tcp/127.0.0.1:21237) (DC: dc1)" === CONT TestLeaveCommand_FailOnNonFlagArgs writer.go:29: 2021-01-29T19:34:18.500Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server.raft: entering candidate state: node="Node at 127.0.0.1:21243 [Candidate]" term=2 === CONT TestLeaveCommand writer.go:29: 2021-01-29T19:34:18.506Z [INFO] TestLeaveCommand.server: Handled event for server in area: event=member-join server=Node-b1bbbd82-a655-f12e-7667-30e892126594.dc1 area=wan writer.go:29: 2021-01-29T19:34:18.507Z [INFO] TestLeaveCommand: Started DNS server: address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:34:18.511Z [INFO] TestLeaveCommand: Started HTTP server: address=127.0.0.1:21233 network=tcp writer.go:29: 2021-01-29T19:34:18.512Z [INFO] TestLeaveCommand: started state syncer === CONT TestLeaveCommand_FailOnNonFlagArgs writer.go:29: 2021-01-29T19:34:18.508Z [DEBUG] TestLeaveCommand_FailOnNonFlagArgs.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:18.514Z [DEBUG] TestLeaveCommand_FailOnNonFlagArgs.server.raft: vote granted: from=afa4e545-5c7e-10af-f1b0-a4973baa0ed6 term=2 tally=1 writer.go:29: 2021-01-29T19:34:18.515Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:18.516Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server.raft: entering leader state: leader="Node at 127.0.0.1:21243 [Leader]" writer.go:29: 2021-01-29T19:34:18.518Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:18.520Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server: New leader elected: payload=Node-afa4e545-5c7e-10af-f1b0-a4973baa0ed6 === CONT TestLeaveCommand writer.go:29: 2021-01-29T19:34:18.524Z [WARN] TestLeaveCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:18.525Z [INFO] TestLeaveCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:21237 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:18.527Z [DEBUG] TestLeaveCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:18.528Z [DEBUG] TestLeaveCommand.server.raft: vote granted: from=b1bbbd82-a655-f12e-7667-30e892126594 term=2 tally=1 writer.go:29: 2021-01-29T19:34:18.529Z [INFO] TestLeaveCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:18.530Z [INFO] TestLeaveCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:21237 [Leader]" === CONT TestLeaveCommand_FailOnNonFlagArgs writer.go:29: 2021-01-29T19:34:18.527Z [DEBUG] TestLeaveCommand_FailOnNonFlagArgs.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:21243 === CONT TestLeaveCommand writer.go:29: 2021-01-29T19:34:18.532Z [INFO] TestLeaveCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:18.532Z [INFO] TestLeaveCommand.server: New leader elected: payload=Node-b1bbbd82-a655-f12e-7667-30e892126594 === CONT TestLeaveCommand_FailOnNonFlagArgs writer.go:29: 2021-01-29T19:34:18.539Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLeaveCommand writer.go:29: 2021-01-29T19:34:18.533Z [DEBUG] TestLeaveCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:21237 writer.go:29: 2021-01-29T19:34:18.565Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLeaveCommand_FailOnNonFlagArgs writer.go:29: 2021-01-29T19:34:18.594Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:18.595Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:18.596Z [DEBUG] TestLeaveCommand_FailOnNonFlagArgs.server: Skipping self join check for node since the cluster is too small: node=Node-afa4e545-5c7e-10af-f1b0-a4973baa0ed6 writer.go:29: 2021-01-29T19:34:18.598Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server: member joined, marking health alive: member=Node-afa4e545-5c7e-10af-f1b0-a4973baa0ed6 === CONT TestLeaveCommand writer.go:29: 2021-01-29T19:34:18.596Z [INFO] TestLeaveCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:18.607Z [INFO] TestLeaveCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:18.613Z [DEBUG] TestLeaveCommand.server: Skipping self join check for node since the cluster is too small: node=Node-b1bbbd82-a655-f12e-7667-30e892126594 writer.go:29: 2021-01-29T19:34:18.623Z [INFO] TestLeaveCommand.server: member joined, marking health alive: member=Node-b1bbbd82-a655-f12e-7667-30e892126594 writer.go:29: 2021-01-29T19:34:18.625Z [INFO] TestLeaveCommand.server: server starting leave writer.go:29: 2021-01-29T19:34:18.630Z [INFO] TestLeaveCommand.server.serf.wan: serf: EventMemberLeave: Node-b1bbbd82-a655-f12e-7667-30e892126594.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:18.631Z [INFO] TestLeaveCommand.server: Handled event for server in area: event=member-leave server=Node-b1bbbd82-a655-f12e-7667-30e892126594.dc1 area=wan writer.go:29: 2021-01-29T19:34:18.631Z [INFO] TestLeaveCommand.server.router.manager: shutting down === CONT TestLeaveCommand_FailOnNonFlagArgs writer.go:29: 2021-01-29T19:34:18.693Z [DEBUG] TestLeaveCommand_FailOnNonFlagArgs: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:18.695Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: Synced node info writer.go:29: 2021-01-29T19:34:18.696Z [DEBUG] TestLeaveCommand_FailOnNonFlagArgs: Node info in sync writer.go:29: 2021-01-29T19:34:18.712Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: Requesting shutdown writer.go:29: 2021-01-29T19:34:18.714Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server: shutting down server writer.go:29: 2021-01-29T19:34:18.715Z [DEBUG] TestLeaveCommand_FailOnNonFlagArgs.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:18.717Z [WARN] TestLeaveCommand_FailOnNonFlagArgs.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:18.717Z [DEBUG] TestLeaveCommand_FailOnNonFlagArgs.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:18.720Z [WARN] TestLeaveCommand_FailOnNonFlagArgs.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:18.723Z [INFO] TestLeaveCommand_FailOnNonFlagArgs.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:18.725Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: consul server down writer.go:29: 2021-01-29T19:34:18.726Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: shutdown complete writer.go:29: 2021-01-29T19:34:18.727Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: Stopping server: protocol=DNS address=127.0.0.1:21238 network=tcp writer.go:29: 2021-01-29T19:34:18.728Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: Stopping server: protocol=DNS address=127.0.0.1:21238 network=udp writer.go:29: 2021-01-29T19:34:18.728Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: Stopping server: protocol=HTTP address=127.0.0.1:21239 network=tcp === CONT TestLeaveCommand writer.go:29: 2021-01-29T19:34:18.938Z [DEBUG] TestLeaveCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:18.941Z [INFO] TestLeaveCommand: Synced node info === CONT TestLeaveCommand_FailOnNonFlagArgs writer.go:29: 2021-01-29T19:34:19.229Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:19.232Z [INFO] TestLeaveCommand_FailOnNonFlagArgs: Endpoints down --- PASS: TestLeaveCommand_FailOnNonFlagArgs (0.84s) === CONT TestLeaveCommand writer.go:29: 2021-01-29T19:34:19.793Z [DEBUG] TestLeaveCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:19.795Z [DEBUG] TestLeaveCommand: Node info in sync writer.go:29: 2021-01-29T19:34:19.797Z [DEBUG] TestLeaveCommand: Node info in sync writer.go:29: 2021-01-29T19:34:20.554Z [DEBUG] TestLeaveCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:21.632Z [INFO] TestLeaveCommand.server.serf.lan: serf: EventMemberLeave: Node-b1bbbd82-a655-f12e-7667-30e892126594 127.0.0.1 writer.go:29: 2021-01-29T19:34:21.634Z [INFO] TestLeaveCommand.server: Removing LAN server: server="Node-b1bbbd82-a655-f12e-7667-30e892126594 (Addr: tcp/127.0.0.1:21237) (DC: dc1)" writer.go:29: 2021-01-29T19:34:21.634Z [WARN] TestLeaveCommand.server: deregistering self should be done by follower: name=Node-b1bbbd82-a655-f12e-7667-30e892126594 writer.go:29: 2021-01-29T19:34:22.553Z [ERROR] TestLeaveCommand.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" writer.go:29: 2021-01-29T19:34:24.552Z [ERROR] TestLeaveCommand.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" writer.go:29: 2021-01-29T19:34:24.634Z [INFO] TestLeaveCommand.server: Waiting to drain RPC traffic: drain_time=5s writer.go:29: 2021-01-29T19:34:26.552Z [ERROR] TestLeaveCommand.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" writer.go:29: 2021-01-29T19:34:28.552Z [ERROR] TestLeaveCommand.server.autopilot: Error updating cluster health: error="error getting server raft protocol versions: No servers found" writer.go:29: 2021-01-29T19:34:28.553Z [ERROR] TestLeaveCommand.server.autopilot: Error promoting servers: error="error getting server raft protocol versions: No servers found" writer.go:29: 2021-01-29T19:34:29.636Z [INFO] TestLeaveCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:29.639Z [INFO] TestLeaveCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:29.642Z [DEBUG] TestLeaveCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:29.644Z [DEBUG] TestLeaveCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:29.648Z [INFO] TestLeaveCommand: consul server down writer.go:29: 2021-01-29T19:34:29.650Z [INFO] TestLeaveCommand: shutdown complete writer.go:29: 2021-01-29T19:34:29.652Z [DEBUG] TestLeaveCommand.http: Request finished: method=PUT url=/v1/agent/leave from=127.0.0.1:43368 latency=11.027491891s writer.go:29: 2021-01-29T19:34:29.654Z [INFO] TestLeaveCommand: Stopping server: protocol=DNS address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:34:29.655Z [INFO] TestLeaveCommand: Stopping server: protocol=DNS address=127.0.0.1:21232 network=udp writer.go:29: 2021-01-29T19:34:29.656Z [INFO] TestLeaveCommand: Stopping server: protocol=HTTP address=127.0.0.1:21233 network=tcp writer.go:29: 2021-01-29T19:34:30.158Z [INFO] TestLeaveCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:30.160Z [INFO] TestLeaveCommand: Endpoints down --- PASS: TestLeaveCommand (11.76s) PASS ok github.com/hashicorp/consul/command/leave 11.963s === RUN TestLockCommand_noTabs === PAUSE TestLockCommand_noTabs === RUN TestLockCommand_BadArgs === PAUSE TestLockCommand_BadArgs === RUN TestLockCommand === PAUSE TestLockCommand === RUN TestLockCommand_NoShell === PAUSE TestLockCommand_NoShell === RUN TestLockCommand_TryLock === PAUSE TestLockCommand_TryLock === RUN TestLockCommand_TrySemaphore === PAUSE TestLockCommand_TrySemaphore === RUN TestLockCommand_MonitorRetry_Lock_Default === PAUSE TestLockCommand_MonitorRetry_Lock_Default === RUN TestLockCommand_MonitorRetry_Semaphore_Default === PAUSE TestLockCommand_MonitorRetry_Semaphore_Default === RUN TestLockCommand_MonitorRetry_Lock_Arg === PAUSE TestLockCommand_MonitorRetry_Lock_Arg === RUN TestLockCommand_MonitorRetry_Semaphore_Arg === PAUSE TestLockCommand_MonitorRetry_Semaphore_Arg === RUN TestLockCommand_ChildExitCode lock_test.go:302: DM-skipped --- SKIP: TestLockCommand_ChildExitCode (0.00s) === CONT TestLockCommand_noTabs --- PASS: TestLockCommand_noTabs (0.00s) === CONT TestLockCommand_MonitorRetry_Semaphore_Default [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestLockCommand_MonitorRetry_Lock_Default === CONT TestLockCommand_MonitorRetry_Semaphore_Arg === CONT TestLockCommand_MonitorRetry_Lock_Arg [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:21.979Z [WARN] TestLockCommand_MonitorRetry_Semaphore_Default: bootstrap = true: do not enable unless necessary === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:21.980Z [WARN] TestLockCommand_MonitorRetry_Semaphore_Arg: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:21.983Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.tlsutil: Update: version=1 === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:21.981Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.tlsutil: Update: version=1 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:21.987Z [WARN] TestLockCommand_MonitorRetry_Lock_Arg: bootstrap = true: do not enable unless necessary === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:21.987Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:21.989Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.tlsutil: Update: version=1 === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:21.991Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:21.994Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:22.002Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b35adb5b-e58b-c619-d284-3621e5182a90 Address:127.0.0.1:22276}]" === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.002Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:6e461530-03d7-69ce-f5d4-244b0ef36041 Address:127.0.0.1:22270}]" === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.001Z [WARN] TestLockCommand_MonitorRetry_Lock_Default: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:22.005Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:22.007Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.003Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:73d917f6-155e-d707-516e-f3194f71e311 Address:127.0.0.1:22258}]" writer.go:29: 2021-01-29T19:34:22.013Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server.serf.wan: serf: EventMemberJoin: Node-73d917f6-155e-d707-516e-f3194f71e311.dc1 127.0.0.1 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.012Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server.raft: entering follower state: follower="Node at 127.0.0.1:22276 [Follower]" leader= === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.015Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server.serf.lan: serf: EventMemberJoin: Node-73d917f6-155e-d707-516e-f3194f71e311 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.019Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: Started DNS server: address=127.0.0.1:22253 network=udp writer.go:29: 2021-01-29T19:34:22.020Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server.raft: entering follower state: follower="Node at 127.0.0.1:22258 [Follower]" leader= === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.007Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server.serf.wan: serf: EventMemberJoin: Node-6e461530-03d7-69ce-f5d4-244b0ef36041.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.026Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server.serf.lan: serf: EventMemberJoin: Node-6e461530-03d7-69ce-f5d4-244b0ef36041 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.018Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server.raft: entering follower state: follower="Node at 127.0.0.1:22270 [Follower]" leader= === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.020Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server.serf.wan: serf: EventMemberJoin: Node-b35adb5b-e58b-c619-d284-3621e5182a90.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.032Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server.serf.lan: serf: EventMemberJoin: Node-b35adb5b-e58b-c619-d284-3621e5182a90 127.0.0.1 === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.037Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: Started DNS server: address=127.0.0.1:22265 network=tcp === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.037Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server: Adding LAN server: server="Node-73d917f6-155e-d707-516e-f3194f71e311 (Addr: tcp/127.0.0.1:22258) (DC: dc1)" writer.go:29: 2021-01-29T19:34:22.039Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server: Handled event for server in area: event=member-join server=Node-73d917f6-155e-d707-516e-f3194f71e311.dc1 area=wan === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.039Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server: Adding LAN server: server="Node-b35adb5b-e58b-c619-d284-3621e5182a90 (Addr: tcp/127.0.0.1:22276) (DC: dc1)" === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.044Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: Started DNS server: address=127.0.0.1:22253 network=tcp === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.042Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: Started DNS server: address=127.0.0.1:22271 network=udp writer.go:29: 2021-01-29T19:34:22.047Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server: Handled event for server in area: event=member-join server=Node-b35adb5b-e58b-c619-d284-3621e5182a90.dc1 area=wan === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.046Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: Started HTTP server: address=127.0.0.1:22254 network=tcp === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.049Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: Started DNS server: address=127.0.0.1:22271 network=tcp writer.go:29: 2021-01-29T19:34:22.050Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: Started HTTP server: address=127.0.0.1:22272 network=tcp writer.go:29: 2021-01-29T19:34:22.051Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: started state syncer === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.049Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: started state syncer === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.041Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server: Adding LAN server: server="Node-6e461530-03d7-69ce-f5d4-244b0ef36041 (Addr: tcp/127.0.0.1:22270) (DC: dc1)" === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.053Z [WARN] TestLockCommand_MonitorRetry_Lock_Arg.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.053Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server: Handled event for server in area: event=member-join server=Node-6e461530-03d7-69ce-f5d4-244b0ef36041.dc1 area=wan === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.055Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server.raft: entering candidate state: node="Node at 127.0.0.1:22276 [Candidate]" term=2 === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.053Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: Started DNS server: address=127.0.0.1:22265 network=udp writer.go:29: 2021-01-29T19:34:22.058Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: Started HTTP server: address=127.0.0.1:22266 network=tcp writer.go:29: 2021-01-29T19:34:22.058Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: started state syncer === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.057Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.server.raft: votes: needed=1 === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.060Z [WARN] TestLockCommand_MonitorRetry_Semaphore_Default.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:22.061Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server.raft: entering candidate state: node="Node at 127.0.0.1:22258 [Candidate]" term=2 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.061Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.server.raft: vote granted: from=b35adb5b-e58b-c619-d284-3621e5182a90 term=2 tally=1 writer.go:29: 2021-01-29T19:34:22.062Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server.raft: election won: tally=1 === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.062Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:637b8c5d-8d0a-f21f-4ec6-6942c6132102 Address:127.0.0.1:22264}]" === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.063Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:22.064Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.server.raft: vote granted: from=73d917f6-155e-d707-516e-f3194f71e311 term=2 tally=1 writer.go:29: 2021-01-29T19:34:22.065Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:22.066Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server.raft: entering leader state: leader="Node at 127.0.0.1:22258 [Leader]" === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.065Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server.raft: entering follower state: follower="Node at 127.0.0.1:22264 [Follower]" leader= === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.064Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server.raft: entering leader state: leader="Node at 127.0.0.1:22276 [Leader]" === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.067Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server: cluster leadership acquired === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.069Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server: cluster leadership acquired === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.069Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server: New leader elected: payload=Node-73d917f6-155e-d707-516e-f3194f71e311 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.071Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server: New leader elected: payload=Node-b35adb5b-e58b-c619-d284-3621e5182a90 === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.066Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server.serf.wan: serf: EventMemberJoin: Node-637b8c5d-8d0a-f21f-4ec6-6942c6132102.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.074Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server.serf.lan: serf: EventMemberJoin: Node-637b8c5d-8d0a-f21f-4ec6-6942c6132102 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.080Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: Started DNS server: address=127.0.0.1:22259 network=udp writer.go:29: 2021-01-29T19:34:22.082Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server: Handled event for server in area: event=member-join server=Node-637b8c5d-8d0a-f21f-4ec6-6942c6132102.dc1 area=wan writer.go:29: 2021-01-29T19:34:22.083Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: Started DNS server: address=127.0.0.1:22259 network=tcp writer.go:29: 2021-01-29T19:34:22.085Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: Started HTTP server: address=127.0.0.1:22260 network=tcp writer.go:29: 2021-01-29T19:34:22.086Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: started state syncer === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.084Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22258 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.087Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22276 === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.090Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.089Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server: Adding LAN server: server="Node-637b8c5d-8d0a-f21f-4ec6-6942c6132102 (Addr: tcp/127.0.0.1:22264) (DC: dc1)" === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.099Z [WARN] TestLockCommand_MonitorRetry_Semaphore_Arg.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:22.102Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server.raft: entering candidate state: node="Node at 127.0.0.1:22270 [Candidate]" term=2 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.106Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.106Z [WARN] TestLockCommand_MonitorRetry_Lock_Default.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:22.108Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server.raft: entering candidate state: node="Node at 127.0.0.1:22264 [Candidate]" term=2 === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.109Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:22.111Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.server.raft: vote granted: from=6e461530-03d7-69ce-f5d4-244b0ef36041 term=2 tally=1 writer.go:29: 2021-01-29T19:34:22.112Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server.raft: election won: tally=1 === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.111Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.server.raft: votes: needed=1 === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.113Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server.raft: entering leader state: leader="Node at 127.0.0.1:22270 [Leader]" writer.go:29: 2021-01-29T19:34:22.115Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server: cluster leadership acquired === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.115Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.server.raft: vote granted: from=637b8c5d-8d0a-f21f-4ec6-6942c6132102 term=2 tally=1 === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.117Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server: New leader elected: payload=Node-6e461530-03d7-69ce-f5d4-244b0ef36041 === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.117Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:22.119Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server.raft: entering leader state: leader="Node at 127.0.0.1:22264 [Leader]" writer.go:29: 2021-01-29T19:34:22.120Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server: cluster leadership acquired === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.120Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22270 === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.134Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22264 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.135Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:22.138Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.leader: started routine: routine="CA root pruning" === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.140Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server: New leader elected: payload=Node-637b8c5d-8d0a-f21f-4ec6-6942c6132102 === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.143Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.146Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.146Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.leader: started routine: routine="CA root pruning" === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.140Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.server: Skipping self join check for node since the cluster is too small: node=Node-b35adb5b-e58b-c619-d284-3621e5182a90 === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.149Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.server: Skipping self join check for node since the cluster is too small: node=Node-73d917f6-155e-d707-516e-f3194f71e311 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.150Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server: member joined, marking health alive: member=Node-b35adb5b-e58b-c619-d284-3621e5182a90 === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.151Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server: member joined, marking health alive: member=Node-73d917f6-155e-d707-516e-f3194f71e311 === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.159Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.157Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: Synced node info writer.go:29: 2021-01-29T19:34:22.161Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg: Node info in sync writer.go:29: 2021-01-29T19:34:22.165Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:22.166Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.168Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.server: Skipping self join check for node since the cluster is too small: node=Node-6e461530-03d7-69ce-f5d4-244b0ef36041 writer.go:29: 2021-01-29T19:34:22.169Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server: member joined, marking health alive: member=Node-6e461530-03d7-69ce-f5d4-244b0ef36041 === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.170Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:22.171Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.172Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.server: Skipping self join check for node since the cluster is too small: node=Node-637b8c5d-8d0a-f21f-4ec6-6942c6132102 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.163Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: Synced node info === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.174Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server: member joined, marking health alive: member=Node-637b8c5d-8d0a-f21f-4ec6-6942c6132102 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.175Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg: Node info in sync === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.271Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:37104 latency=2.225873ms writer.go:29: 2021-01-29T19:34:22.277Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:37104 latency=1.938272ms writer.go:29: 2021-01-29T19:34:22.280Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=PUT url=/v1/kv/test/prefix/ec0cfc9b-725e-c85b-8a4a-1e8b9a0a29bf?acquire=ec0cfc9b-725e-c85b-8a4a-1e8b9a0a29bf&flags=16210313421097356768 from=127.0.0.1:37104 latency=1.1349ms writer.go:29: 2021-01-29T19:34:22.282Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=GET url=/v1/kv/test/prefix?recurse=&wait=15000ms from=127.0.0.1:37104 latency=472.222µs writer.go:29: 2021-01-29T19:34:22.285Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?cas=0&flags=16210313421097356768 from=127.0.0.1:37104 latency=823.899µs writer.go:29: 2021-01-29T19:34:22.291Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=GET url=/v1/kv/test/prefix?consistent=&recurse= from=127.0.0.1:37104 latency=279.421µs writer.go:29: 2021-01-29T19:34:22.293Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock from=127.0.0.1:37106 latency=125.769µs writer.go:29: 2021-01-29T19:34:22.295Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?cas=14&flags=16210313421097356768 from=127.0.0.1:37106 latency=787.829µs writer.go:29: 2021-01-29T19:34:22.295Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=GET url=/v1/kv/test/prefix?consistent=&index=14&recurse= from=127.0.0.1:37104 latency=2.025314ms writer.go:29: 2021-01-29T19:34:22.300Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=DELETE url=/v1/kv/test/prefix/ec0cfc9b-725e-c85b-8a4a-1e8b9a0a29bf from=127.0.0.1:37106 latency=3.601399ms writer.go:29: 2021-01-29T19:34:22.304Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=GET url=/v1/kv/test/prefix?recurse= from=127.0.0.1:37106 latency=1.268987ms === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.332Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.334Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.338Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: Synced node info === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.340Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: Synced node info === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.341Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=DELETE url=/v1/kv/test/prefix/.lock?cas=15 from=127.0.0.1:37106 latency=32.96807ms writer.go:29: 2021-01-29T19:34:22.343Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.http: Request finished: method=PUT url=/v1/session/destroy/ec0cfc9b-725e-c85b-8a4a-1e8b9a0a29bf from=127.0.0.1:37104 latency=35.45693ms writer.go:29: 2021-01-29T19:34:22.345Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: Requesting shutdown writer.go:29: 2021-01-29T19:34:22.346Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server: shutting down server writer.go:29: 2021-01-29T19:34:22.347Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.347Z [WARN] TestLockCommand_MonitorRetry_Semaphore_Arg.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:22.349Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Arg.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.351Z [WARN] TestLockCommand_MonitorRetry_Semaphore_Arg.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:22.359Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:22.361Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: consul server down writer.go:29: 2021-01-29T19:34:22.362Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: shutdown complete writer.go:29: 2021-01-29T19:34:22.364Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: Stopping server: protocol=DNS address=127.0.0.1:22265 network=tcp === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.364Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:48604 latency=15.621549ms === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.365Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: Stopping server: protocol=DNS address=127.0.0.1:22265 network=udp writer.go:29: 2021-01-29T19:34:22.366Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: Stopping server: protocol=HTTP address=127.0.0.1:22266 network=tcp === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.368Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:48604 latency=1.109985ms writer.go:29: 2021-01-29T19:34:22.374Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=PUT url=/v1/kv/test/prefix/cc002b69-ba49-0135-19f1-8896957f1dc8?acquire=cc002b69-ba49-0135-19f1-8896957f1dc8&flags=16210313421097356768 from=127.0.0.1:48604 latency=4.213815ms writer.go:29: 2021-01-29T19:34:22.375Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=GET url=/v1/kv/test/prefix?recurse=&wait=15000ms from=127.0.0.1:48604 latency=56.297µs writer.go:29: 2021-01-29T19:34:22.378Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?cas=0&flags=16210313421097356768 from=127.0.0.1:48604 latency=482.485µs writer.go:29: 2021-01-29T19:34:22.381Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=GET url=/v1/kv/test/prefix?consistent=&recurse= from=127.0.0.1:48604 latency=721.057µs === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.385Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:35862 latency=692.762µs === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.385Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock from=127.0.0.1:48608 latency=139.17µs === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.388Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:35862 latency=1.257371ms === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.388Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?cas=14&flags=16210313421097356768 from=127.0.0.1:48608 latency=991.892µs === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.389Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?wait=15000ms from=127.0.0.1:35862 latency=18.533µs === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.388Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=GET url=/v1/kv/test/prefix?consistent=&index=14&recurse= from=127.0.0.1:48604 latency=4.340399ms writer.go:29: 2021-01-29T19:34:22.392Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=DELETE url=/v1/kv/test/prefix/cc002b69-ba49-0135-19f1-8896957f1dc8 from=127.0.0.1:48608 latency=2.384721ms writer.go:29: 2021-01-29T19:34:22.394Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=GET url=/v1/kv/test/prefix?recurse= from=127.0.0.1:48608 latency=93.824µs writer.go:29: 2021-01-29T19:34:22.397Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=DELETE url=/v1/kv/test/prefix/.lock?cas=15 from=127.0.0.1:48608 latency=652.264µs writer.go:29: 2021-01-29T19:34:22.399Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: Requesting shutdown writer.go:29: 2021-01-29T19:34:22.400Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server: shutting down server writer.go:29: 2021-01-29T19:34:22.401Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.402Z [WARN] TestLockCommand_MonitorRetry_Semaphore_Default.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:22.402Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.404Z [WARN] TestLockCommand_MonitorRetry_Semaphore_Default.server.serf.wan: serf: Shutdown without a Leave === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.405Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?acquire=6aead2a1-61a3-bc03-76fc-b6d8d47a5986&flags=3304740253564472344 from=127.0.0.1:35862 latency=13.331668ms === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.405Z [DEBUG] TestLockCommand_MonitorRetry_Semaphore_Default.http: Request finished: method=PUT url=/v1/session/destroy/cc002b69-ba49-0135-19f1-8896957f1dc8 from=127.0.0.1:48604 latency=6.117993ms writer.go:29: 2021-01-29T19:34:22.407Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default.server.router.manager: shutting down === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.409Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?consistent= from=127.0.0.1:35862 latency=83.314µs === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.407Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: consul server down writer.go:29: 2021-01-29T19:34:22.410Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: shutdown complete writer.go:29: 2021-01-29T19:34:22.412Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: Stopping server: protocol=DNS address=127.0.0.1:22253 network=tcp writer.go:29: 2021-01-29T19:34:22.413Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: Stopping server: protocol=DNS address=127.0.0.1:22253 network=udp === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.413Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?flags=3304740253564472344&release=6aead2a1-61a3-bc03-76fc-b6d8d47a5986 from=127.0.0.1:35862 latency=2.397596ms === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.414Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: Stopping server: protocol=HTTP address=127.0.0.1:22254 network=tcp === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.417Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?consistent=&index=13 from=127.0.0.1:35866 latency=4.195729ms writer.go:29: 2021-01-29T19:34:22.419Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock from=127.0.0.1:35862 latency=70.5µs writer.go:29: 2021-01-29T19:34:22.420Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.http: Request finished: method=PUT url=/v1/session/destroy/6aead2a1-61a3-bc03-76fc-b6d8d47a5986 from=127.0.0.1:35866 latency=947.8µs writer.go:29: 2021-01-29T19:34:22.422Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.http: Request finished: method=DELETE url=/v1/kv/test/prefix/.lock?cas=14 from=127.0.0.1:35862 latency=1.013187ms writer.go:29: 2021-01-29T19:34:22.425Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: Requesting shutdown writer.go:29: 2021-01-29T19:34:22.426Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server: shutting down server writer.go:29: 2021-01-29T19:34:22.428Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.429Z [WARN] TestLockCommand_MonitorRetry_Lock_Arg.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:22.431Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Arg.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.434Z [WARN] TestLockCommand_MonitorRetry_Lock_Arg.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:22.437Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:22.437Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: consul server down writer.go:29: 2021-01-29T19:34:22.439Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: shutdown complete writer.go:29: 2021-01-29T19:34:22.441Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: Stopping server: protocol=DNS address=127.0.0.1:22271 network=tcp writer.go:29: 2021-01-29T19:34:22.442Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: Stopping server: protocol=DNS address=127.0.0.1:22271 network=udp writer.go:29: 2021-01-29T19:34:22.443Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: Stopping server: protocol=HTTP address=127.0.0.1:22272 network=tcp === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:22.464Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:53828 latency=526.3µs writer.go:29: 2021-01-29T19:34:22.466Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:53828 latency=591.419µs writer.go:29: 2021-01-29T19:34:22.468Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?wait=15000ms from=127.0.0.1:53828 latency=52.91µs writer.go:29: 2021-01-29T19:34:22.470Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?acquire=3715f485-6a91-d25d-1db0-91ae431c1148&flags=3304740253564472344 from=127.0.0.1:53828 latency=591.952µs writer.go:29: 2021-01-29T19:34:22.475Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?consistent= from=127.0.0.1:53828 latency=153.882µs writer.go:29: 2021-01-29T19:34:22.479Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?flags=3304740253564472344&release=3715f485-6a91-d25d-1db0-91ae431c1148 from=127.0.0.1:53828 latency=506.119µs writer.go:29: 2021-01-29T19:34:22.481Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock from=127.0.0.1:53828 latency=54.37µs writer.go:29: 2021-01-29T19:34:22.481Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?consistent=&index=13 from=127.0.0.1:53830 latency=192.367µs writer.go:29: 2021-01-29T19:34:22.483Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.http: Request finished: method=DELETE url=/v1/kv/test/prefix/.lock?cas=14 from=127.0.0.1:53828 latency=461.555µs writer.go:29: 2021-01-29T19:34:22.485Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.http: Request finished: method=PUT url=/v1/session/destroy/3715f485-6a91-d25d-1db0-91ae431c1148 from=127.0.0.1:53830 latency=832.482µs writer.go:29: 2021-01-29T19:34:22.487Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: Requesting shutdown writer.go:29: 2021-01-29T19:34:22.487Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server: shutting down server writer.go:29: 2021-01-29T19:34:22.488Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.489Z [WARN] TestLockCommand_MonitorRetry_Lock_Default.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:22.489Z [DEBUG] TestLockCommand_MonitorRetry_Lock_Default.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:22.491Z [WARN] TestLockCommand_MonitorRetry_Lock_Default.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:22.493Z [INFO] TestLockCommand_MonitorRetry_Lock_Default.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:22.493Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: consul server down writer.go:29: 2021-01-29T19:34:22.495Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: shutdown complete writer.go:29: 2021-01-29T19:34:22.495Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: Stopping server: protocol=DNS address=127.0.0.1:22259 network=tcp writer.go:29: 2021-01-29T19:34:22.496Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: Stopping server: protocol=DNS address=127.0.0.1:22259 network=udp writer.go:29: 2021-01-29T19:34:22.497Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: Stopping server: protocol=HTTP address=127.0.0.1:22260 network=tcp === CONT TestLockCommand_MonitorRetry_Semaphore_Arg writer.go:29: 2021-01-29T19:34:22.867Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:22.868Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Arg: Endpoints down --- PASS: TestLockCommand_MonitorRetry_Semaphore_Arg (0.95s) === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:22.878Z [WARN] TestLockCommand_NoShell: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:22.879Z [DEBUG] TestLockCommand_NoShell.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:22.881Z [DEBUG] TestLockCommand_NoShell.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:22.886Z [INFO] TestLockCommand_NoShell.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b3789db6-c2e9-f6c2-d683-e84af7506184 Address:127.0.0.1:22282}]" writer.go:29: 2021-01-29T19:34:22.888Z [INFO] TestLockCommand_NoShell.server.raft: entering follower state: follower="Node at 127.0.0.1:22282 [Follower]" leader= writer.go:29: 2021-01-29T19:34:22.889Z [INFO] TestLockCommand_NoShell.server.serf.wan: serf: EventMemberJoin: Node-b3789db6-c2e9-f6c2-d683-e84af7506184.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.900Z [INFO] TestLockCommand_NoShell.server.serf.lan: serf: EventMemberJoin: Node-b3789db6-c2e9-f6c2-d683-e84af7506184 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.907Z [INFO] TestLockCommand_NoShell: Started DNS server: address=127.0.0.1:22277 network=udp writer.go:29: 2021-01-29T19:34:22.907Z [INFO] TestLockCommand_NoShell.server: Adding LAN server: server="Node-b3789db6-c2e9-f6c2-d683-e84af7506184 (Addr: tcp/127.0.0.1:22282) (DC: dc1)" writer.go:29: 2021-01-29T19:34:22.907Z [INFO] TestLockCommand_NoShell.server: Handled event for server in area: event=member-join server=Node-b3789db6-c2e9-f6c2-d683-e84af7506184.dc1 area=wan writer.go:29: 2021-01-29T19:34:22.909Z [INFO] TestLockCommand_NoShell: Started DNS server: address=127.0.0.1:22277 network=tcp writer.go:29: 2021-01-29T19:34:22.912Z [INFO] TestLockCommand_NoShell: Started HTTP server: address=127.0.0.1:22278 network=tcp === CONT TestLockCommand_MonitorRetry_Semaphore_Default writer.go:29: 2021-01-29T19:34:22.916Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:22.917Z [INFO] TestLockCommand_MonitorRetry_Semaphore_Default: Endpoints down --- PASS: TestLockCommand_MonitorRetry_Semaphore_Default (1.01s) === CONT TestLockCommand_TrySemaphore === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:22.913Z [INFO] TestLockCommand_NoShell: started state syncer === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:22.926Z [WARN] TestLockCommand_TrySemaphore: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:22.927Z [DEBUG] TestLockCommand_TrySemaphore.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:22.928Z [DEBUG] TestLockCommand_TrySemaphore.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:22.939Z [WARN] TestLockCommand_NoShell.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:22.940Z [INFO] TestLockCommand_NoShell.server.raft: entering candidate state: node="Node at 127.0.0.1:22282 [Candidate]" term=2 === CONT TestLockCommand_MonitorRetry_Lock_Arg writer.go:29: 2021-01-29T19:34:22.944Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:22.945Z [INFO] TestLockCommand_MonitorRetry_Lock_Arg: Endpoints down --- PASS: TestLockCommand_MonitorRetry_Lock_Arg (1.03s) === CONT TestLockCommand_TryLock === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:22.951Z [DEBUG] TestLockCommand_NoShell.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:22.952Z [DEBUG] TestLockCommand_NoShell.server.raft: vote granted: from=b3789db6-c2e9-f6c2-d683-e84af7506184 term=2 tally=1 writer.go:29: 2021-01-29T19:34:22.952Z [INFO] TestLockCommand_NoShell.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:22.953Z [INFO] TestLockCommand_NoShell.server.raft: entering leader state: leader="Node at 127.0.0.1:22282 [Leader]" writer.go:29: 2021-01-29T19:34:22.955Z [INFO] TestLockCommand_NoShell.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:22.956Z [INFO] TestLockCommand_NoShell.server: New leader elected: payload=Node-b3789db6-c2e9-f6c2-d683-e84af7506184 === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:22.958Z [INFO] TestLockCommand_TrySemaphore.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:7bd83a8e-3bdd-7ae6-b099-5c060bedea42 Address:127.0.0.1:22288}]" writer.go:29: 2021-01-29T19:34:22.960Z [INFO] TestLockCommand_TrySemaphore.server.raft: entering follower state: follower="Node at 127.0.0.1:22288 [Follower]" leader= === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:22.977Z [WARN] TestLockCommand_TryLock: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:22.978Z [DEBUG] TestLockCommand_TryLock.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:22.979Z [DEBUG] TestLockCommand_TryLock.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:22.985Z [DEBUG] TestLockCommand_NoShell.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22282 === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:22.993Z [INFO] TestLockCommand_TryLock.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:fa79b7a8-2228-9e5a-519c-56b0b3ac1c7f Address:127.0.0.1:22294}]" writer.go:29: 2021-01-29T19:34:22.995Z [INFO] TestLockCommand_TryLock.server.serf.wan: serf: EventMemberJoin: Node-fa79b7a8-2228-9e5a-519c-56b0b3ac1c7f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:22.998Z [INFO] TestLockCommand_TryLock.server.serf.lan: serf: EventMemberJoin: Node-fa79b7a8-2228-9e5a-519c-56b0b3ac1c7f 127.0.0.1 === CONT TestLockCommand_MonitorRetry_Lock_Default writer.go:29: 2021-01-29T19:34:23.000Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:23.004Z [INFO] TestLockCommand_MonitorRetry_Lock_Default: Endpoints down --- PASS: TestLockCommand_MonitorRetry_Lock_Default (1.09s) === CONT TestLockCommand === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.004Z [INFO] TestLockCommand_TryLock: Started DNS server: address=127.0.0.1:22289 network=udp writer.go:29: 2021-01-29T19:34:23.001Z [INFO] TestLockCommand_TryLock.server.raft: entering follower state: follower="Node at 127.0.0.1:22294 [Follower]" leader= writer.go:29: 2021-01-29T19:34:23.013Z [INFO] TestLockCommand_TryLock.server: Handled event for server in area: event=member-join server=Node-fa79b7a8-2228-9e5a-519c-56b0b3ac1c7f.dc1 area=wan === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.011Z [INFO] TestLockCommand_TrySemaphore.server.serf.wan: serf: EventMemberJoin: Node-7bd83a8e-3bdd-7ae6-b099-5c060bedea42.dc1 127.0.0.1 === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.015Z [INFO] TestLockCommand_TryLock.server: Adding LAN server: server="Node-fa79b7a8-2228-9e5a-519c-56b0b3ac1c7f (Addr: tcp/127.0.0.1:22294) (DC: dc1)" writer.go:29: 2021-01-29T19:34:23.016Z [INFO] TestLockCommand_TryLock: Started DNS server: address=127.0.0.1:22289 network=tcp === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.020Z [INFO] TestLockCommand_TrySemaphore.server.serf.lan: serf: EventMemberJoin: Node-7bd83a8e-3bdd-7ae6-b099-5c060bedea42 127.0.0.1 === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.021Z [WARN] TestLockCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:23.023Z [DEBUG] TestLockCommand.tlsutil: Update: version=1 === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.024Z [INFO] TestLockCommand_TrySemaphore: Started DNS server: address=127.0.0.1:22283 network=udp === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.021Z [INFO] TestLockCommand_TryLock: Started HTTP server: address=127.0.0.1:22290 network=tcp === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.025Z [DEBUG] TestLockCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.025Z [INFO] TestLockCommand_TryLock: started state syncer === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.026Z [WARN] TestLockCommand_TrySemaphore.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:23.027Z [INFO] TestLockCommand_TrySemaphore.server.raft: entering candidate state: node="Node at 127.0.0.1:22288 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:23.029Z [WARN] TestLockCommand_TrySemaphore.server.raft: unable to get address for sever, using fallback address: id=7bd83a8e-3bdd-7ae6-b099-5c060bedea42 fallback=127.0.0.1:22288 error="Could not find address for server id 7bd83a8e-3bdd-7ae6-b099-5c060bedea42" === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.022Z [INFO] TestLockCommand_NoShell: Synced node info === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.030Z [INFO] TestLockCommand_TrySemaphore: Started DNS server: address=127.0.0.1:22283 network=tcp === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.030Z [DEBUG] TestLockCommand_NoShell: Node info in sync writer.go:29: 2021-01-29T19:34:23.028Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.033Z [INFO] TestLockCommand_TrySemaphore.server: Handled event for server in area: event=member-join server=Node-7bd83a8e-3bdd-7ae6-b099-5c060bedea42.dc1 area=wan writer.go:29: 2021-01-29T19:34:23.034Z [DEBUG] TestLockCommand_TrySemaphore.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:23.035Z [DEBUG] TestLockCommand_TrySemaphore.server.raft: vote granted: from=7bd83a8e-3bdd-7ae6-b099-5c060bedea42 term=2 tally=1 === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.035Z [INFO] TestLockCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:976b8050-d7e2-bffa-f9cf-bd0bdc7b2b85 Address:127.0.0.1:22300}]" === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.036Z [INFO] TestLockCommand_TrySemaphore.server.raft: election won: tally=1 === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.036Z [INFO] TestLockCommand.server.serf.wan: serf: EventMemberJoin: Node-976b8050-d7e2-bffa-f9cf-bd0bdc7b2b85.dc1 127.0.0.1 === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.037Z [INFO] TestLockCommand_TrySemaphore.server.raft: entering leader state: leader="Node at 127.0.0.1:22288 [Leader]" writer.go:29: 2021-01-29T19:34:23.038Z [INFO] TestLockCommand_TrySemaphore.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:23.035Z [INFO] TestLockCommand_TrySemaphore: Started HTTP server: address=127.0.0.1:22284 network=tcp === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.038Z [INFO] TestLockCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:22300 [Follower]" leader= === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.040Z [INFO] TestLockCommand_TrySemaphore: started state syncer === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.038Z [INFO] TestLockCommand.server.serf.lan: serf: EventMemberJoin: Node-976b8050-d7e2-bffa-f9cf-bd0bdc7b2b85 127.0.0.1 === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.041Z [DEBUG] TestLockCommand_TrySemaphore.server: Cannot upgrade to new ACLs: leaderMode=3 mode=1 found=false leader=127.0.0.1:22288 === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.045Z [INFO] TestLockCommand: Started DNS server: address=127.0.0.1:22295 network=udp writer.go:29: 2021-01-29T19:34:23.047Z [INFO] TestLockCommand.server: Adding LAN server: server="Node-976b8050-d7e2-bffa-f9cf-bd0bdc7b2b85 (Addr: tcp/127.0.0.1:22300) (DC: dc1)" === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.036Z [INFO] TestLockCommand_TrySemaphore.server: Adding LAN server: server="Node-7bd83a8e-3bdd-7ae6-b099-5c060bedea42 (Addr: tcp/127.0.0.1:22288) (DC: dc1)" === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.049Z [INFO] TestLockCommand.server: Handled event for server in area: event=member-join server=Node-976b8050-d7e2-bffa-f9cf-bd0bdc7b2b85.dc1 area=wan === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.045Z [INFO] TestLockCommand_TrySemaphore: Synced node info writer.go:29: 2021-01-29T19:34:23.053Z [DEBUG] TestLockCommand_TrySemaphore: Node info in sync writer.go:29: 2021-01-29T19:34:23.052Z [INFO] TestLockCommand_TrySemaphore.server: New leader elected: payload=Node-7bd83a8e-3bdd-7ae6-b099-5c060bedea42 === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.055Z [INFO] TestLockCommand: Started DNS server: address=127.0.0.1:22295 network=tcp === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.096Z [INFO] TestLockCommand_NoShell.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.097Z [INFO] TestLockCommand: Started HTTP server: address=127.0.0.1:22296 network=tcp === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.097Z [INFO] TestLockCommand_NoShell.leader: started routine: routine="CA root pruning" === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.098Z [INFO] TestLockCommand: started state syncer === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.098Z [DEBUG] TestLockCommand_NoShell.server: Skipping self join check for node since the cluster is too small: node=Node-b3789db6-c2e9-f6c2-d683-e84af7506184 writer.go:29: 2021-01-29T19:34:23.101Z [INFO] TestLockCommand_NoShell.server: member joined, marking health alive: member=Node-b3789db6-c2e9-f6c2-d683-e84af7506184 === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.103Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.111Z [WARN] TestLockCommand.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.092Z [WARN] TestLockCommand_TryLock.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.114Z [INFO] TestLockCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:22300 [Candidate]" term=2 === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.114Z [INFO] TestLockCommand_TryLock.server.raft: entering candidate state: node="Node at 127.0.0.1:22294 [Candidate]" term=2 === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.116Z [INFO] TestLockCommand_TrySemaphore.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.116Z [DEBUG] TestLockCommand_TryLock.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:23.117Z [DEBUG] TestLockCommand_TryLock.server.raft: vote granted: from=fa79b7a8-2228-9e5a-519c-56b0b3ac1c7f term=2 tally=1 === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.117Z [INFO] TestLockCommand_TrySemaphore.leader: started routine: routine="CA root pruning" === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.118Z [INFO] TestLockCommand_TryLock.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:23.118Z [INFO] TestLockCommand_TryLock.server.raft: entering leader state: leader="Node at 127.0.0.1:22294 [Leader]" === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.118Z [DEBUG] TestLockCommand_TrySemaphore.server: Skipping self join check for node since the cluster is too small: node=Node-7bd83a8e-3bdd-7ae6-b099-5c060bedea42 writer.go:29: 2021-01-29T19:34:23.120Z [INFO] TestLockCommand_TrySemaphore.server: member joined, marking health alive: member=Node-7bd83a8e-3bdd-7ae6-b099-5c060bedea42 === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.119Z [INFO] TestLockCommand_TryLock.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:23.120Z [INFO] TestLockCommand_TryLock.server: New leader elected: payload=Node-fa79b7a8-2228-9e5a-519c-56b0b3ac1c7f === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.122Z [DEBUG] TestLockCommand_TrySemaphore.server: Skipping self join check for node since the cluster is too small: node=Node-7bd83a8e-3bdd-7ae6-b099-5c060bedea42 === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.124Z [DEBUG] TestLockCommand.server.raft: votes: needed=1 === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.122Z [DEBUG] TestLockCommand_TryLock.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22294 === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.125Z [DEBUG] TestLockCommand.server.raft: vote granted: from=976b8050-d7e2-bffa-f9cf-bd0bdc7b2b85 term=2 tally=1 writer.go:29: 2021-01-29T19:34:23.127Z [INFO] TestLockCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:23.128Z [INFO] TestLockCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:22300 [Leader]" === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.129Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.129Z [INFO] TestLockCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:23.132Z [INFO] TestLockCommand.server: New leader elected: payload=Node-976b8050-d7e2-bffa-f9cf-bd0bdc7b2b85 writer.go:29: 2021-01-29T19:34:23.135Z [DEBUG] TestLockCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22300 === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.138Z [INFO] TestLockCommand_TryLock.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:23.139Z [INFO] TestLockCommand_TryLock.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.141Z [DEBUG] TestLockCommand_TryLock.server: Skipping self join check for node since the cluster is too small: node=Node-fa79b7a8-2228-9e5a-519c-56b0b3ac1c7f === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.140Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:57650 latency=2.681894ms === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.143Z [INFO] TestLockCommand_TryLock.server: member joined, marking health alive: member=Node-fa79b7a8-2228-9e5a-519c-56b0b3ac1c7f === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.150Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.151Z [DEBUG] TestLockCommand_NoShell.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:34164 latency=3.696975ms === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.156Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:57650 latency=5.718672ms === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.167Z [INFO] TestLockCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.167Z [DEBUG] TestLockCommand_NoShell.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:34164 latency=1.381849ms === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.168Z [INFO] TestLockCommand.leader: started routine: routine="CA root pruning" === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.169Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=PUT url=/v1/kv/test/prefix/77be0d48-a109-1418-4480-06bcbc239dba?acquire=77be0d48-a109-1418-4480-06bcbc239dba&flags=16210313421097356768 from=127.0.0.1:57650 latency=3.080088ms === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.170Z [DEBUG] TestLockCommand.server: Skipping self join check for node since the cluster is too small: node=Node-976b8050-d7e2-bffa-f9cf-bd0bdc7b2b85 === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.185Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=GET url=/v1/kv/test/prefix?recurse=&wait=10000ms from=127.0.0.1:57650 latency=13.564492ms writer.go:29: 2021-01-29T19:34:23.195Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?cas=0&flags=16210313421097356768 from=127.0.0.1:57650 latency=3.555466ms === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.171Z [INFO] TestLockCommand.server: member joined, marking health alive: member=Node-976b8050-d7e2-bffa-f9cf-bd0bdc7b2b85 === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.186Z [DEBUG] TestLockCommand_NoShell.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?wait=15000ms from=127.0.0.1:34164 latency=111.195µs === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.203Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=GET url=/v1/kv/test/prefix?consistent=&recurse= from=127.0.0.1:57650 latency=140.103µs === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.206Z [DEBUG] TestLockCommand_NoShell.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?acquire=975c73c7-ec6c-44d1-1e87-0c1171054229&flags=3304740253564472344 from=127.0.0.1:34164 latency=3.970711ms writer.go:29: 2021-01-29T19:34:23.217Z [DEBUG] TestLockCommand_NoShell.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?consistent= from=127.0.0.1:34164 latency=209.545µs === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.221Z [DEBUG] TestLockCommand: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.222Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock from=127.0.0.1:57654 latency=129.324µs === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.225Z [INFO] TestLockCommand: Synced node info === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.226Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?cas=14&flags=16210313421097356768 from=127.0.0.1:57654 latency=2.213633ms writer.go:29: 2021-01-29T19:34:23.226Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=GET url=/v1/kv/test/prefix?consistent=&index=14&recurse= from=127.0.0.1:57650 latency=14.669234ms === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.230Z [DEBUG] TestLockCommand_NoShell.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?flags=3304740253564472344&release=975c73c7-ec6c-44d1-1e87-0c1171054229 from=127.0.0.1:34168 latency=1.257546ms writer.go:29: 2021-01-29T19:34:23.231Z [DEBUG] TestLockCommand_NoShell.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?consistent=&index=13 from=127.0.0.1:34164 latency=11.526539ms writer.go:29: 2021-01-29T19:34:23.235Z [DEBUG] TestLockCommand_NoShell.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock from=127.0.0.1:34168 latency=172.738µs writer.go:29: 2021-01-29T19:34:23.238Z [DEBUG] TestLockCommand_NoShell.http: Request finished: method=DELETE url=/v1/kv/test/prefix/.lock?cas=14 from=127.0.0.1:34168 latency=835.967µs writer.go:29: 2021-01-29T19:34:23.242Z [INFO] TestLockCommand_NoShell: Requesting shutdown === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.232Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=DELETE url=/v1/kv/test/prefix/77be0d48-a109-1418-4480-06bcbc239dba from=127.0.0.1:57654 latency=3.445796ms === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.239Z [DEBUG] TestLockCommand_NoShell.http: Request finished: method=PUT url=/v1/session/destroy/975c73c7-ec6c-44d1-1e87-0c1171054229 from=127.0.0.1:34164 latency=1.156457ms writer.go:29: 2021-01-29T19:34:23.249Z [INFO] TestLockCommand_NoShell.server: shutting down server writer.go:29: 2021-01-29T19:34:23.253Z [DEBUG] TestLockCommand_NoShell.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.255Z [WARN] TestLockCommand_NoShell.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.257Z [DEBUG] TestLockCommand_NoShell.leader: stopped routine: routine="CA root pruning" === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.254Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=GET url=/v1/kv/test/prefix?recurse= from=127.0.0.1:57654 latency=120.703µs === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.258Z [WARN] TestLockCommand_NoShell.server.serf.wan: serf: Shutdown without a Leave === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.257Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=PUT url=/v1/session/destroy/77be0d48-a109-1418-4480-06bcbc239dba from=127.0.0.1:57650 latency=1.328377ms writer.go:29: 2021-01-29T19:34:23.261Z [DEBUG] TestLockCommand_TrySemaphore.http: Request finished: method=DELETE url=/v1/kv/test/prefix/.lock?cas=15 from=127.0.0.1:57654 latency=1.429302ms writer.go:29: 2021-01-29T19:34:23.267Z [INFO] TestLockCommand_TrySemaphore: Requesting shutdown === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.261Z [INFO] TestLockCommand_NoShell.server.router.manager: shutting down === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.269Z [INFO] TestLockCommand_TrySemaphore.server: shutting down server === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.268Z [INFO] TestLockCommand_NoShell: consul server down writer.go:29: 2021-01-29T19:34:23.270Z [INFO] TestLockCommand_NoShell: shutdown complete === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.270Z [DEBUG] TestLockCommand_TrySemaphore.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.275Z [WARN] TestLockCommand_TrySemaphore.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.279Z [DEBUG] TestLockCommand_TrySemaphore.leader: stopped routine: routine="CA root pruning" === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.273Z [INFO] TestLockCommand_NoShell: Stopping server: protocol=DNS address=127.0.0.1:22277 network=tcp writer.go:29: 2021-01-29T19:34:23.283Z [INFO] TestLockCommand_NoShell: Stopping server: protocol=DNS address=127.0.0.1:22277 network=udp === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.287Z [WARN] TestLockCommand_TrySemaphore.server.serf.wan: serf: Shutdown without a Leave === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.284Z [INFO] TestLockCommand_NoShell: Stopping server: protocol=HTTP address=127.0.0.1:22278 network=tcp === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.305Z [INFO] TestLockCommand_TrySemaphore.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:23.306Z [INFO] TestLockCommand_TrySemaphore: consul server down writer.go:29: 2021-01-29T19:34:23.307Z [INFO] TestLockCommand_TrySemaphore: shutdown complete writer.go:29: 2021-01-29T19:34:23.308Z [INFO] TestLockCommand_TrySemaphore: Stopping server: protocol=DNS address=127.0.0.1:22283 network=tcp writer.go:29: 2021-01-29T19:34:23.309Z [INFO] TestLockCommand_TrySemaphore: Stopping server: protocol=DNS address=127.0.0.1:22283 network=udp writer.go:29: 2021-01-29T19:34:23.310Z [INFO] TestLockCommand_TrySemaphore: Stopping server: protocol=HTTP address=127.0.0.1:22284 network=tcp === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.330Z [DEBUG] TestLockCommand.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:57840 latency=5.824687ms writer.go:29: 2021-01-29T19:34:23.336Z [DEBUG] TestLockCommand.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:57840 latency=1.656098ms writer.go:29: 2021-01-29T19:34:23.337Z [DEBUG] TestLockCommand.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?wait=15000ms from=127.0.0.1:57840 latency=31.771µs writer.go:29: 2021-01-29T19:34:23.340Z [DEBUG] TestLockCommand.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?acquire=af6463a8-6dd7-43a8-5c48-fb01857a710d&flags=3304740253564472344 from=127.0.0.1:57840 latency=994.28µs writer.go:29: 2021-01-29T19:34:23.345Z [DEBUG] TestLockCommand.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?consistent= from=127.0.0.1:57840 latency=210.541µs writer.go:29: 2021-01-29T19:34:23.348Z [DEBUG] TestLockCommand.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?flags=3304740253564472344&release=af6463a8-6dd7-43a8-5c48-fb01857a710d from=127.0.0.1:57842 latency=540.891µs writer.go:29: 2021-01-29T19:34:23.350Z [DEBUG] TestLockCommand.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?consistent=&index=13 from=127.0.0.1:57840 latency=2.129957ms writer.go:29: 2021-01-29T19:34:23.350Z [DEBUG] TestLockCommand.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock from=127.0.0.1:57842 latency=86.564µs writer.go:29: 2021-01-29T19:34:23.352Z [DEBUG] TestLockCommand.http: Request finished: method=PUT url=/v1/session/destroy/af6463a8-6dd7-43a8-5c48-fb01857a710d from=127.0.0.1:57840 latency=630.389µs writer.go:29: 2021-01-29T19:34:23.355Z [DEBUG] TestLockCommand.http: Request finished: method=DELETE url=/v1/kv/test/prefix/.lock?cas=14 from=127.0.0.1:57842 latency=1.075913ms writer.go:29: 2021-01-29T19:34:23.358Z [INFO] TestLockCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:23.359Z [INFO] TestLockCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:23.360Z [DEBUG] TestLockCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.362Z [WARN] TestLockCommand.server.serf.lan: serf: Shutdown without a Leave === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.365Z [DEBUG] TestLockCommand_TryLock: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:23.369Z [INFO] TestLockCommand_TryLock: Synced node info writer.go:29: 2021-01-29T19:34:23.370Z [DEBUG] TestLockCommand_TryLock: Node info in sync === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.364Z [DEBUG] TestLockCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.366Z [WARN] TestLockCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.375Z [INFO] TestLockCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:23.375Z [INFO] TestLockCommand: consul server down writer.go:29: 2021-01-29T19:34:23.377Z [INFO] TestLockCommand: shutdown complete writer.go:29: 2021-01-29T19:34:23.377Z [INFO] TestLockCommand: Stopping server: protocol=DNS address=127.0.0.1:22295 network=tcp writer.go:29: 2021-01-29T19:34:23.378Z [INFO] TestLockCommand: Stopping server: protocol=DNS address=127.0.0.1:22295 network=udp writer.go:29: 2021-01-29T19:34:23.379Z [INFO] TestLockCommand: Stopping server: protocol=HTTP address=127.0.0.1:22296 network=tcp === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:23.506Z [DEBUG] TestLockCommand_TryLock.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:45156 latency=745.49µs writer.go:29: 2021-01-29T19:34:23.510Z [DEBUG] TestLockCommand_TryLock.http: Request finished: method=PUT url=/v1/session/create from=127.0.0.1:45156 latency=797.532µs writer.go:29: 2021-01-29T19:34:23.512Z [DEBUG] TestLockCommand_TryLock.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?wait=10000ms from=127.0.0.1:45156 latency=24.076µs writer.go:29: 2021-01-29T19:34:23.513Z [DEBUG] TestLockCommand_TryLock.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?acquire=b6528218-6c15-192e-933d-004c9809bd73&flags=3304740253564472344 from=127.0.0.1:45156 latency=412.468µs writer.go:29: 2021-01-29T19:34:23.516Z [DEBUG] TestLockCommand_TryLock.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?consistent= from=127.0.0.1:45156 latency=143.113µs writer.go:29: 2021-01-29T19:34:23.526Z [DEBUG] TestLockCommand_TryLock.http: Request finished: method=PUT url=/v1/kv/test/prefix/.lock?flags=3304740253564472344&release=b6528218-6c15-192e-933d-004c9809bd73 from=127.0.0.1:45158 latency=1.986834ms writer.go:29: 2021-01-29T19:34:23.526Z [DEBUG] TestLockCommand_TryLock.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock?consistent=&index=13 from=127.0.0.1:45156 latency=6.934182ms writer.go:29: 2021-01-29T19:34:23.529Z [DEBUG] TestLockCommand_TryLock.http: Request finished: method=GET url=/v1/kv/test/prefix/.lock from=127.0.0.1:45158 latency=136.426µs writer.go:29: 2021-01-29T19:34:23.532Z [DEBUG] TestLockCommand_TryLock.http: Request finished: method=PUT url=/v1/session/destroy/b6528218-6c15-192e-933d-004c9809bd73 from=127.0.0.1:45156 latency=613.648µs writer.go:29: 2021-01-29T19:34:23.532Z [DEBUG] TestLockCommand_TryLock.http: Request finished: method=DELETE url=/v1/kv/test/prefix/.lock?cas=14 from=127.0.0.1:45158 latency=990.602µs writer.go:29: 2021-01-29T19:34:23.535Z [INFO] TestLockCommand_TryLock: Requesting shutdown writer.go:29: 2021-01-29T19:34:23.536Z [INFO] TestLockCommand_TryLock.server: shutting down server writer.go:29: 2021-01-29T19:34:23.538Z [DEBUG] TestLockCommand_TryLock.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.540Z [WARN] TestLockCommand_TryLock.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.540Z [DEBUG] TestLockCommand_TryLock.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.542Z [WARN] TestLockCommand_TryLock.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.545Z [INFO] TestLockCommand_TryLock.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:23.545Z [INFO] TestLockCommand_TryLock: consul server down writer.go:29: 2021-01-29T19:34:23.548Z [INFO] TestLockCommand_TryLock: shutdown complete writer.go:29: 2021-01-29T19:34:23.549Z [INFO] TestLockCommand_TryLock: Stopping server: protocol=DNS address=127.0.0.1:22289 network=tcp writer.go:29: 2021-01-29T19:34:23.551Z [INFO] TestLockCommand_TryLock: Stopping server: protocol=DNS address=127.0.0.1:22289 network=udp writer.go:29: 2021-01-29T19:34:23.555Z [INFO] TestLockCommand_TryLock: Stopping server: protocol=HTTP address=127.0.0.1:22290 network=tcp === CONT TestLockCommand_NoShell writer.go:29: 2021-01-29T19:34:23.792Z [INFO] TestLockCommand_NoShell: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:23.793Z [INFO] TestLockCommand_NoShell: Endpoints down --- PASS: TestLockCommand_NoShell (0.92s) === CONT TestLockCommand_BadArgs --- PASS: TestLockCommand_BadArgs (0.00s) === CONT TestLockCommand_TrySemaphore writer.go:29: 2021-01-29T19:34:23.811Z [INFO] TestLockCommand_TrySemaphore: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:23.812Z [INFO] TestLockCommand_TrySemaphore: Endpoints down --- PASS: TestLockCommand_TrySemaphore (0.90s) === CONT TestLockCommand writer.go:29: 2021-01-29T19:34:23.880Z [INFO] TestLockCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:23.881Z [INFO] TestLockCommand: Endpoints down --- PASS: TestLockCommand (0.88s) === CONT TestLockCommand_TryLock writer.go:29: 2021-01-29T19:34:24.058Z [INFO] TestLockCommand_TryLock: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:24.059Z [INFO] TestLockCommand_TryLock: Endpoints down --- PASS: TestLockCommand_TryLock (1.11s) PASS ok github.com/hashicorp/consul/command/lock 2.294s === RUN TestLoginCommand_noTabs === PAUSE TestLoginCommand_noTabs === RUN TestLoginCommand === PAUSE TestLoginCommand === RUN TestLoginCommand_k8s === PAUSE TestLoginCommand_k8s === CONT TestLoginCommand_noTabs === CONT TestLoginCommand_k8s --- PASS: TestLoginCommand_noTabs (0.00s) === CONT TestLoginCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:34:23.445Z [WARN] TestLoginCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:23.447Z [DEBUG] TestLoginCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:23.450Z [DEBUG] TestLoginCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.451Z [WARN] TestLoginCommand_k8s: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:23.452Z [DEBUG] TestLoginCommand_k8s.tlsutil: Update: version=1 === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.453Z [INFO] TestLoginCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4f239d6f-a927-d953-8e20-1efed9e77b60 Address:127.0.0.1:25327}]" === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.454Z [DEBUG] TestLoginCommand_k8s.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.456Z [INFO] TestLoginCommand.server.serf.wan: serf: EventMemberJoin: Node-4f239d6f-a927-d953-8e20-1efed9e77b60.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:23.456Z [INFO] TestLoginCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:25327 [Follower]" leader= writer.go:29: 2021-01-29T19:34:23.464Z [INFO] TestLoginCommand.server.serf.lan: serf: EventMemberJoin: Node-4f239d6f-a927-d953-8e20-1efed9e77b60 127.0.0.1 === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.466Z [INFO] TestLoginCommand_k8s.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:f0f374b9-6b2d-4ba4-2688-8590fd885d6b Address:127.0.0.1:25321}]" writer.go:29: 2021-01-29T19:34:23.468Z [INFO] TestLoginCommand_k8s.server.serf.wan: serf: EventMemberJoin: Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b.dc1 127.0.0.1 === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.469Z [INFO] TestLoginCommand: Started DNS server: address=127.0.0.1:25322 network=udp writer.go:29: 2021-01-29T19:34:23.470Z [INFO] TestLoginCommand.server: Handled event for server in area: event=member-join server=Node-4f239d6f-a927-d953-8e20-1efed9e77b60.dc1 area=wan === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.471Z [INFO] TestLoginCommand_k8s.server.serf.lan: serf: EventMemberJoin: Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b 127.0.0.1 === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.470Z [INFO] TestLoginCommand.server: Adding LAN server: server="Node-4f239d6f-a927-d953-8e20-1efed9e77b60 (Addr: tcp/127.0.0.1:25327) (DC: dc1)" === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.471Z [INFO] TestLoginCommand_k8s.server.raft: entering follower state: follower="Node at 127.0.0.1:25321 [Follower]" leader= === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.471Z [INFO] TestLoginCommand: Started DNS server: address=127.0.0.1:25322 network=tcp === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.477Z [INFO] TestLoginCommand_k8s: Started DNS server: address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:34:23.481Z [INFO] TestLoginCommand_k8s: Started DNS server: address=127.0.0.1:25316 network=tcp === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.479Z [INFO] TestLoginCommand: Started HTTP server: address=127.0.0.1:25323 network=tcp === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.484Z [INFO] TestLoginCommand_k8s: Started HTTP server: address=127.0.0.1:25317 network=tcp === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.485Z [INFO] TestLoginCommand: started state syncer === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.486Z [INFO] TestLoginCommand_k8s: started state syncer writer.go:29: 2021-01-29T19:34:23.478Z [INFO] TestLoginCommand_k8s.server: Adding LAN server: server="Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b (Addr: tcp/127.0.0.1:25321) (DC: dc1)" writer.go:29: 2021-01-29T19:34:23.479Z [INFO] TestLoginCommand_k8s.server: Handled event for server in area: event=member-join server=Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b.dc1 area=wan === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.501Z [WARN] TestLoginCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:23.503Z [INFO] TestLoginCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:25327 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:23.507Z [DEBUG] TestLoginCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:23.508Z [DEBUG] TestLoginCommand.server.raft: vote granted: from=4f239d6f-a927-d953-8e20-1efed9e77b60 term=2 tally=1 writer.go:29: 2021-01-29T19:34:23.509Z [INFO] TestLoginCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:23.514Z [INFO] TestLoginCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:25327 [Leader]" writer.go:29: 2021-01-29T19:34:23.516Z [INFO] TestLoginCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:23.518Z [INFO] TestLoginCommand.server: New leader elected: payload=Node-4f239d6f-a927-d953-8e20-1efed9e77b60 writer.go:29: 2021-01-29T19:34:23.519Z [INFO] TestLoginCommand.server: initializing acls === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.530Z [DEBUG] TestLoginCommand_k8s.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.520Z [INFO] TestLoginCommand.server: initializing acls writer.go:29: 2021-01-29T19:34:23.534Z [WARN] TestLoginCommand.server: Configuring a non-UUID master token is deprecated === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.536Z [WARN] TestLoginCommand_k8s.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.527Z [INFO] TestLoginCommand.server: Created ACL 'global-management' policy === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.537Z [INFO] TestLoginCommand_k8s.server.raft: entering candidate state: node="Node at 127.0.0.1:25321 [Candidate]" term=2 === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.538Z [WARN] TestLoginCommand.server: Configuring a non-UUID master token is deprecated === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.540Z [DEBUG] TestLoginCommand_k8s.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:23.541Z [DEBUG] TestLoginCommand_k8s.server.raft: vote granted: from=f0f374b9-6b2d-4ba4-2688-8590fd885d6b term=2 tally=1 === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.553Z [INFO] TestLoginCommand.server: Bootstrapped ACL master token from configuration === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.543Z [INFO] TestLoginCommand_k8s.server.raft: election won: tally=1 === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.553Z [INFO] TestLoginCommand.server: Bootstrapped ACL master token from configuration writer.go:29: 2021-01-29T19:34:23.556Z [INFO] TestLoginCommand.leader: started routine: routine="legacy ACL token upgrade" === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.554Z [INFO] TestLoginCommand_k8s.server.raft: entering leader state: leader="Node at 127.0.0.1:25321 [Leader]" === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.557Z [INFO] TestLoginCommand.leader: started routine: routine="acl token reaping" === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.558Z [INFO] TestLoginCommand_k8s.server: cluster leadership acquired === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.555Z [INFO] TestLoginCommand.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:23.560Z [DEBUG] TestLoginCommand.server: transitioning out of legacy ACL mode === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.559Z [INFO] TestLoginCommand_k8s.server: New leader elected: payload=Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.559Z [INFO] TestLoginCommand.server.serf.lan: serf: EventMemberUpdate: Node-4f239d6f-a927-d953-8e20-1efed9e77b60 === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.560Z [INFO] TestLoginCommand_k8s.server: initializing acls === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.562Z [INFO] TestLoginCommand.server.serf.wan: serf: EventMemberUpdate: Node-4f239d6f-a927-d953-8e20-1efed9e77b60.dc1 === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.565Z [INFO] TestLoginCommand_k8s.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:23.566Z [WARN] TestLoginCommand_k8s.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:23.569Z [INFO] TestLoginCommand_k8s.server: Bootstrapped ACL master token from configuration === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.562Z [INFO] TestLoginCommand.server.serf.lan: serf: EventMemberUpdate: Node-4f239d6f-a927-d953-8e20-1efed9e77b60 === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.573Z [INFO] TestLoginCommand_k8s.server: Created ACL anonymous token from configuration === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.562Z [INFO] TestLoginCommand.server: Updating LAN server: server="Node-4f239d6f-a927-d953-8e20-1efed9e77b60 (Addr: tcp/127.0.0.1:25327) (DC: dc1)" === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.574Z [INFO] TestLoginCommand_k8s.leader: started routine: routine="legacy ACL token upgrade" === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.575Z [INFO] TestLoginCommand.server: Updating LAN server: server="Node-4f239d6f-a927-d953-8e20-1efed9e77b60 (Addr: tcp/127.0.0.1:25327) (DC: dc1)" writer.go:29: 2021-01-29T19:34:23.564Z [INFO] TestLoginCommand.server: Handled event for server in area: event=member-update server=Node-4f239d6f-a927-d953-8e20-1efed9e77b60.dc1 area=wan === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.576Z [INFO] TestLoginCommand_k8s.leader: started routine: routine="acl token reaping" === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.565Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.578Z [INFO] TestLoginCommand_k8s.server.serf.lan: serf: EventMemberUpdate: Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b writer.go:29: 2021-01-29T19:34:23.581Z [INFO] TestLoginCommand_k8s.server.serf.wan: serf: EventMemberUpdate: Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b.dc1 === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.581Z [INFO] TestLoginCommand.server.serf.wan: serf: EventMemberUpdate: Node-4f239d6f-a927-d953-8e20-1efed9e77b60.dc1 === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.581Z [INFO] TestLoginCommand_k8s.server: Updating LAN server: server="Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b (Addr: tcp/127.0.0.1:25321) (DC: dc1)" === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.584Z [INFO] TestLoginCommand.server: Handled event for server in area: event=member-update server=Node-4f239d6f-a927-d953-8e20-1efed9e77b60.dc1 area=wan === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.584Z [INFO] TestLoginCommand_k8s.server: Handled event for server in area: event=member-update server=Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b.dc1 area=wan writer.go:29: 2021-01-29T19:34:23.586Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:23.615Z [INFO] TestLoginCommand_k8s.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:23.616Z [INFO] TestLoginCommand_k8s.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.617Z [DEBUG] TestLoginCommand_k8s.server: Skipping self join check for node since the cluster is too small: node=Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b writer.go:29: 2021-01-29T19:34:23.618Z [INFO] TestLoginCommand_k8s.server: member joined, marking health alive: member=Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.619Z [INFO] TestLoginCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.621Z [DEBUG] TestLoginCommand_k8s.server: Skipping self join check for node since the cluster is too small: node=Node-f0f374b9-6b2d-4ba4-2688-8590fd885d6b === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.620Z [INFO] TestLoginCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.622Z [DEBUG] TestLoginCommand.server: Skipping self join check for node since the cluster is too small: node=Node-4f239d6f-a927-d953-8e20-1efed9e77b60 writer.go:29: 2021-01-29T19:34:23.625Z [INFO] TestLoginCommand.server: member joined, marking health alive: member=Node-4f239d6f-a927-d953-8e20-1efed9e77b60 writer.go:29: 2021-01-29T19:34:23.629Z [DEBUG] TestLoginCommand.server: Skipping self join check for node since the cluster is too small: node=Node-4f239d6f-a927-d953-8e20-1efed9e77b60 writer.go:29: 2021-01-29T19:34:23.631Z [DEBUG] TestLoginCommand.server: Skipping self join check for node since the cluster is too small: node=Node-4f239d6f-a927-d953-8e20-1efed9e77b60 === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.635Z [DEBUG] TestLoginCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:55408 latency=6.200495ms writer.go:29: 2021-01-29T19:34:23.639Z [DEBUG] TestLoginCommand_k8s.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:55408 latency=1.644922ms === RUN TestLoginCommand_k8s/try_login_with_method_configured_and_binding_rules === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.701Z [DEBUG] TestLoginCommand_k8s.http: Request finished: method=POST url=/v1/acl/login from=127.0.0.1:55410 latency=54.198843ms writer.go:29: 2021-01-29T19:34:23.708Z [INFO] TestLoginCommand_k8s: Requesting shutdown writer.go:29: 2021-01-29T19:34:23.709Z [INFO] TestLoginCommand_k8s.server: shutting down server writer.go:29: 2021-01-29T19:34:23.710Z [DEBUG] TestLoginCommand_k8s.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.711Z [DEBUG] TestLoginCommand_k8s.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:23.710Z [ERROR] TestLoginCommand_k8s.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:23.711Z [DEBUG] TestLoginCommand_k8s.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.712Z [DEBUG] TestLoginCommand_k8s.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:23.712Z [DEBUG] TestLoginCommand_k8s.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:23.716Z [WARN] TestLoginCommand_k8s.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.716Z [DEBUG] TestLoginCommand_k8s.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:23.721Z [WARN] TestLoginCommand_k8s.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.737Z [INFO] TestLoginCommand_k8s.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:23.738Z [INFO] TestLoginCommand_k8s: consul server down writer.go:29: 2021-01-29T19:34:23.740Z [INFO] TestLoginCommand_k8s: shutdown complete === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.741Z [DEBUG] TestLoginCommand: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:23.742Z [INFO] TestLoginCommand_k8s: Stopping server: protocol=DNS address=127.0.0.1:25316 network=tcp writer.go:29: 2021-01-29T19:34:23.743Z [INFO] TestLoginCommand_k8s: Stopping server: protocol=DNS address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:34:23.744Z [INFO] TestLoginCommand_k8s: Stopping server: protocol=HTTP address=127.0.0.1:25317 network=tcp === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.745Z [INFO] TestLoginCommand: Synced node info writer.go:29: 2021-01-29T19:34:23.747Z [DEBUG] TestLoginCommand: Node info in sync === RUN TestLoginCommand/method_is_required === RUN TestLoginCommand/token-sink-file_is_required === RUN TestLoginCommand/bearer-token-file_is_required === RUN TestLoginCommand/bearer-token-file_is_empty === RUN TestLoginCommand/try_login_with_no_method_configured === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.920Z [ERROR] TestLoginCommand.http: Request error: method=POST url=/v1/acl/login from=127.0.0.1:34636 error="ACL not found" writer.go:29: 2021-01-29T19:34:23.922Z [DEBUG] TestLoginCommand.http: Request finished: method=POST url=/v1/acl/login from=127.0.0.1:34636 latency=2.177687ms writer.go:29: 2021-01-29T19:34:23.935Z [DEBUG] TestLoginCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:34638 latency=3.286701ms === RUN TestLoginCommand/try_login_with_method_configured_but_no_binding_rules === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.945Z [ERROR] TestLoginCommand.http: Request error: method=POST url=/v1/acl/login from=127.0.0.1:34640 error="Permission denied" writer.go:29: 2021-01-29T19:34:23.947Z [DEBUG] TestLoginCommand.http: Request finished: method=POST url=/v1/acl/login from=127.0.0.1:34640 latency=3.818565ms writer.go:29: 2021-01-29T19:34:23.970Z [DEBUG] TestLoginCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:34638 latency=19.494548ms === RUN TestLoginCommand/try_login_with_method_configured_and_binding_rules === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:23.989Z [DEBUG] TestLoginCommand.http: Request finished: method=POST url=/v1/acl/login from=127.0.0.1:34642 latency=9.335769ms writer.go:29: 2021-01-29T19:34:23.994Z [INFO] TestLoginCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:23.995Z [INFO] TestLoginCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:23.996Z [DEBUG] TestLoginCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:23.997Z [DEBUG] TestLoginCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:23.998Z [DEBUG] TestLoginCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.999Z [WARN] TestLoginCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:23.998Z [DEBUG] TestLoginCommand.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:23.999Z [DEBUG] TestLoginCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:23.999Z [DEBUG] TestLoginCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:24.002Z [WARN] TestLoginCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:24.005Z [INFO] TestLoginCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:24.007Z [INFO] TestLoginCommand: consul server down writer.go:29: 2021-01-29T19:34:24.009Z [INFO] TestLoginCommand: shutdown complete writer.go:29: 2021-01-29T19:34:24.011Z [INFO] TestLoginCommand: Stopping server: protocol=DNS address=127.0.0.1:25322 network=tcp writer.go:29: 2021-01-29T19:34:24.012Z [INFO] TestLoginCommand: Stopping server: protocol=DNS address=127.0.0.1:25322 network=udp writer.go:29: 2021-01-29T19:34:24.014Z [INFO] TestLoginCommand: Stopping server: protocol=HTTP address=127.0.0.1:25323 network=tcp === CONT TestLoginCommand_k8s writer.go:29: 2021-01-29T19:34:24.246Z [INFO] TestLoginCommand_k8s: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:24.248Z [INFO] TestLoginCommand_k8s: Endpoints down --- PASS: TestLoginCommand_k8s (0.84s) --- PASS: TestLoginCommand_k8s/try_login_with_method_configured_and_binding_rules (0.06s) === CONT TestLoginCommand writer.go:29: 2021-01-29T19:34:24.523Z [INFO] TestLoginCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:24.525Z [INFO] TestLoginCommand: Endpoints down --- PASS: TestLoginCommand (1.12s) --- PASS: TestLoginCommand/method_is_required (0.00s) --- PASS: TestLoginCommand/token-sink-file_is_required (0.00s) --- PASS: TestLoginCommand/bearer-token-file_is_required (0.00s) --- PASS: TestLoginCommand/bearer-token-file_is_empty (0.00s) --- PASS: TestLoginCommand/try_login_with_no_method_configured (0.01s) --- PASS: TestLoginCommand/try_login_with_method_configured_but_no_binding_rules (0.01s) --- PASS: TestLoginCommand/try_login_with_method_configured_and_binding_rules (0.02s) PASS ok github.com/hashicorp/consul/command/login 1.395s === RUN TestLogout_noTabs === PAUSE TestLogout_noTabs === RUN TestLogoutCommand === PAUSE TestLogoutCommand === RUN TestLogoutCommand_k8s === PAUSE TestLogoutCommand_k8s === CONT TestLogout_noTabs --- PASS: TestLogout_noTabs (0.00s) === CONT TestLogoutCommand_k8s === CONT TestLogoutCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:34:30.627Z [WARN] TestLogoutCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:30.628Z [DEBUG] TestLogoutCommand.tlsutil: Update: version=1 === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.629Z [WARN] TestLogoutCommand_k8s: bootstrap = true: do not enable unless necessary === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.630Z [DEBUG] TestLogoutCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.631Z [DEBUG] TestLogoutCommand_k8s.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:30.633Z [DEBUG] TestLogoutCommand_k8s.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.639Z [INFO] TestLogoutCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:bd7f851c-5434-ae6b-d89c-eebf16010130 Address:127.0.0.1:15117}]" writer.go:29: 2021-01-29T19:34:30.641Z [INFO] TestLogoutCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:15117 [Follower]" leader= === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.640Z [INFO] TestLogoutCommand_k8s.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b64874a9-0cc0-db25-6da0-2185f7e80bc7 Address:127.0.0.1:15111}]" === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.642Z [INFO] TestLogoutCommand.server.serf.wan: serf: EventMemberJoin: Node-bd7f851c-5434-ae6b-d89c-eebf16010130.dc1 127.0.0.1 === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.647Z [INFO] TestLogoutCommand_k8s.server.raft: entering follower state: follower="Node at 127.0.0.1:15111 [Follower]" leader= === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.648Z [INFO] TestLogoutCommand.server.serf.lan: serf: EventMemberJoin: Node-bd7f851c-5434-ae6b-d89c-eebf16010130 127.0.0.1 writer.go:29: 2021-01-29T19:34:30.655Z [INFO] TestLogoutCommand: Started DNS server: address=127.0.0.1:15112 network=udp writer.go:29: 2021-01-29T19:34:30.657Z [INFO] TestLogoutCommand.server: Adding LAN server: server="Node-bd7f851c-5434-ae6b-d89c-eebf16010130 (Addr: tcp/127.0.0.1:15117) (DC: dc1)" writer.go:29: 2021-01-29T19:34:30.659Z [INFO] TestLogoutCommand.server: Handled event for server in area: event=member-join server=Node-bd7f851c-5434-ae6b-d89c-eebf16010130.dc1 area=wan writer.go:29: 2021-01-29T19:34:30.664Z [INFO] TestLogoutCommand: Started DNS server: address=127.0.0.1:15112 network=tcp === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.663Z [INFO] TestLogoutCommand_k8s.server.serf.wan: serf: EventMemberJoin: Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7.dc1 127.0.0.1 === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.668Z [INFO] TestLogoutCommand: Started HTTP server: address=127.0.0.1:15113 network=tcp writer.go:29: 2021-01-29T19:34:30.670Z [INFO] TestLogoutCommand: started state syncer === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.670Z [INFO] TestLogoutCommand_k8s.server.serf.lan: serf: EventMemberJoin: Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7 127.0.0.1 writer.go:29: 2021-01-29T19:34:30.676Z [INFO] TestLogoutCommand_k8s.server: Adding LAN server: server="Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7 (Addr: tcp/127.0.0.1:15111) (DC: dc1)" writer.go:29: 2021-01-29T19:34:30.678Z [INFO] TestLogoutCommand_k8s.server: Handled event for server in area: event=member-join server=Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7.dc1 area=wan writer.go:29: 2021-01-29T19:34:30.679Z [INFO] TestLogoutCommand_k8s: Started DNS server: address=127.0.0.1:15106 network=tcp writer.go:29: 2021-01-29T19:34:30.680Z [INFO] TestLogoutCommand_k8s: Started DNS server: address=127.0.0.1:15106 network=udp writer.go:29: 2021-01-29T19:34:30.681Z [INFO] TestLogoutCommand_k8s: Started HTTP server: address=127.0.0.1:15107 network=tcp writer.go:29: 2021-01-29T19:34:30.683Z [INFO] TestLogoutCommand_k8s: started state syncer === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.704Z [WARN] TestLogoutCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:30.705Z [INFO] TestLogoutCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:15117 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:30.707Z [DEBUG] TestLogoutCommand.server.raft: votes: needed=1 === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.707Z [WARN] TestLogoutCommand_k8s.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.708Z [DEBUG] TestLogoutCommand.server.raft: vote granted: from=bd7f851c-5434-ae6b-d89c-eebf16010130 term=2 tally=1 === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.709Z [INFO] TestLogoutCommand_k8s.server.raft: entering candidate state: node="Node at 127.0.0.1:15111 [Candidate]" term=2 === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.710Z [INFO] TestLogoutCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:30.711Z [INFO] TestLogoutCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:15117 [Leader]" writer.go:29: 2021-01-29T19:34:30.712Z [DEBUG] TestLogoutCommand.server: Cannot upgrade to new ACLs: leaderMode=2 mode=2 found=true leader=127.0.0.1:15117 writer.go:29: 2021-01-29T19:34:30.713Z [INFO] TestLogoutCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:30.715Z [INFO] TestLogoutCommand.server: New leader elected: payload=Node-bd7f851c-5434-ae6b-d89c-eebf16010130 writer.go:29: 2021-01-29T19:34:30.716Z [INFO] TestLogoutCommand.server: initializing acls writer.go:29: 2021-01-29T19:34:30.719Z [INFO] TestLogoutCommand.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:30.721Z [WARN] TestLogoutCommand.server: Configuring a non-UUID master token is deprecated === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.727Z [DEBUG] TestLogoutCommand_k8s.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:30.729Z [DEBUG] TestLogoutCommand_k8s.server.raft: vote granted: from=b64874a9-0cc0-db25-6da0-2185f7e80bc7 term=2 tally=1 writer.go:29: 2021-01-29T19:34:30.730Z [INFO] TestLogoutCommand_k8s.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:30.731Z [INFO] TestLogoutCommand_k8s.server.raft: entering leader state: leader="Node at 127.0.0.1:15111 [Leader]" writer.go:29: 2021-01-29T19:34:30.729Z [DEBUG] TestLogoutCommand_k8s.server: Cannot upgrade to new ACLs: leaderMode=3 mode=2 found=true leader= writer.go:29: 2021-01-29T19:34:30.733Z [INFO] TestLogoutCommand_k8s.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:30.734Z [INFO] TestLogoutCommand_k8s.server: initializing acls writer.go:29: 2021-01-29T19:34:30.734Z [INFO] TestLogoutCommand_k8s.server: New leader elected: payload=Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7 writer.go:29: 2021-01-29T19:34:30.738Z [INFO] TestLogoutCommand_k8s.server: Created ACL 'global-management' policy writer.go:29: 2021-01-29T19:34:30.738Z [WARN] TestLogoutCommand_k8s.server: Configuring a non-UUID master token is deprecated writer.go:29: 2021-01-29T19:34:30.740Z [INFO] TestLogoutCommand_k8s.server: Bootstrapped ACL master token from configuration === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.736Z [INFO] TestLogoutCommand.server: Bootstrapped ACL master token from configuration === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.743Z [INFO] TestLogoutCommand_k8s.server: Created ACL anonymous token from configuration writer.go:29: 2021-01-29T19:34:30.744Z [INFO] TestLogoutCommand_k8s.leader: started routine: routine="legacy ACL token upgrade" === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.746Z [INFO] TestLogoutCommand.server: Created ACL anonymous token from configuration === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.746Z [INFO] TestLogoutCommand_k8s.leader: started routine: routine="acl token reaping" === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.748Z [INFO] TestLogoutCommand.leader: started routine: routine="legacy ACL token upgrade" === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.751Z [INFO] TestLogoutCommand_k8s.server.serf.lan: serf: EventMemberUpdate: Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7 === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.753Z [INFO] TestLogoutCommand.leader: started routine: routine="acl token reaping" === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.755Z [INFO] TestLogoutCommand_k8s.server.serf.wan: serf: EventMemberUpdate: Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7.dc1 === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.756Z [INFO] TestLogoutCommand.server.serf.lan: serf: EventMemberUpdate: Node-bd7f851c-5434-ae6b-d89c-eebf16010130 writer.go:29: 2021-01-29T19:34:30.760Z [INFO] TestLogoutCommand.server: Updating LAN server: server="Node-bd7f851c-5434-ae6b-d89c-eebf16010130 (Addr: tcp/127.0.0.1:15117) (DC: dc1)" writer.go:29: 2021-01-29T19:34:30.760Z [INFO] TestLogoutCommand.server.serf.wan: serf: EventMemberUpdate: Node-bd7f851c-5434-ae6b-d89c-eebf16010130.dc1 === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.757Z [INFO] TestLogoutCommand_k8s.server: Updating LAN server: server="Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7 (Addr: tcp/127.0.0.1:15111) (DC: dc1)" === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.763Z [INFO] TestLogoutCommand.server: Handled event for server in area: event=member-update server=Node-bd7f851c-5434-ae6b-d89c-eebf16010130.dc1 area=wan writer.go:29: 2021-01-29T19:34:30.765Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.760Z [INFO] TestLogoutCommand_k8s.server: Handled event for server in area: event=member-update server=Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7.dc1 area=wan writer.go:29: 2021-01-29T19:34:30.761Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:30.814Z [INFO] TestLogoutCommand_k8s.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:30.816Z [INFO] TestLogoutCommand_k8s.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:30.818Z [DEBUG] TestLogoutCommand_k8s.server: Skipping self join check for node since the cluster is too small: node=Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7 writer.go:29: 2021-01-29T19:34:30.821Z [INFO] TestLogoutCommand_k8s.server: member joined, marking health alive: member=Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7 === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.823Z [INFO] TestLogoutCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.823Z [DEBUG] TestLogoutCommand_k8s.server: Skipping self join check for node since the cluster is too small: node=Node-b64874a9-0cc0-db25-6da0-2185f7e80bc7 === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:30.824Z [INFO] TestLogoutCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:30.826Z [DEBUG] TestLogoutCommand.server: Skipping self join check for node since the cluster is too small: node=Node-bd7f851c-5434-ae6b-d89c-eebf16010130 writer.go:29: 2021-01-29T19:34:30.827Z [INFO] TestLogoutCommand.server: member joined, marking health alive: member=Node-bd7f851c-5434-ae6b-d89c-eebf16010130 writer.go:29: 2021-01-29T19:34:30.828Z [DEBUG] TestLogoutCommand.server: Skipping self join check for node since the cluster is too small: node=Node-bd7f851c-5434-ae6b-d89c-eebf16010130 writer.go:29: 2021-01-29T19:34:30.864Z [DEBUG] TestLogoutCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:30.869Z [INFO] TestLogoutCommand: Synced node info === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.904Z [DEBUG] TestLogoutCommand_k8s: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:30.915Z [INFO] TestLogoutCommand_k8s: Synced node info writer.go:29: 2021-01-29T19:34:30.963Z [DEBUG] TestLogoutCommand_k8s: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:30.965Z [DEBUG] TestLogoutCommand_k8s: Node info in sync writer.go:29: 2021-01-29T19:34:30.967Z [DEBUG] TestLogoutCommand_k8s: Node info in sync === RUN TestLogoutCommand_k8s/no_token_specified === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:30.979Z [ERROR] TestLogoutCommand_k8s.http: Request error: method=POST url=/v1/acl/logout from=127.0.0.1:37682 error="ACL not found" writer.go:29: 2021-01-29T19:34:30.981Z [DEBUG] TestLogoutCommand_k8s.http: Request finished: method=POST url=/v1/acl/logout from=127.0.0.1:37682 latency=2.824892ms === RUN TestLogoutCommand_k8s/logout_of_deleted_token === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:31.015Z [ERROR] TestLogoutCommand_k8s.http: Request error: method=POST url=/v1/acl/logout from=127.0.0.1:37684 error="ACL not found" writer.go:29: 2021-01-29T19:34:31.044Z [DEBUG] TestLogoutCommand_k8s.http: Request finished: method=POST url=/v1/acl/logout from=127.0.0.1:37684 latency=28.250997ms === RUN TestLogoutCommand/no_token_specified === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:31.091Z [ERROR] TestLogoutCommand.http: Request error: method=POST url=/v1/acl/logout from=127.0.0.1:39144 error="ACL not found" === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:31.093Z [DEBUG] TestLogoutCommand_k8s.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:37686 latency=12.540761ms === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:31.094Z [DEBUG] TestLogoutCommand.http: Request finished: method=POST url=/v1/acl/logout from=127.0.0.1:39144 latency=3.154441ms === RUN TestLogoutCommand_k8s/logout_of_ordinary_token === RUN TestLogoutCommand/logout_of_deleted_token === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:31.104Z [ERROR] TestLogoutCommand.http: Request error: method=POST url=/v1/acl/logout from=127.0.0.1:39148 error="ACL not found" writer.go:29: 2021-01-29T19:34:31.115Z [DEBUG] TestLogoutCommand.http: Request finished: method=POST url=/v1/acl/logout from=127.0.0.1:39148 latency=10.74615ms === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:31.123Z [ERROR] TestLogoutCommand_k8s.http: Request error: method=POST url=/v1/acl/logout from=127.0.0.1:37690 error="Permission denied" writer.go:29: 2021-01-29T19:34:31.129Z [DEBUG] TestLogoutCommand_k8s.http: Request finished: method=POST url=/v1/acl/logout from=127.0.0.1:37690 latency=6.001949ms writer.go:29: 2021-01-29T19:34:31.164Z [DEBUG] TestLogoutCommand_k8s.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:37686 latency=22.957031ms === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:31.173Z [DEBUG] TestLogoutCommand.http: Request finished: method=PUT url=/v1/acl/token from=127.0.0.1:39150 latency=35.844729ms === RUN TestLogoutCommand/logout_of_ordinary_token === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:31.179Z [ERROR] TestLogoutCommand.http: Request error: method=POST url=/v1/acl/logout from=127.0.0.1:39152 error="Permission denied" writer.go:29: 2021-01-29T19:34:31.181Z [DEBUG] TestLogoutCommand.http: Request finished: method=POST url=/v1/acl/logout from=127.0.0.1:39152 latency=1.547374ms === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:31.185Z [DEBUG] TestLogoutCommand_k8s.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:37686 latency=16.4226ms === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:31.185Z [DEBUG] TestLogoutCommand.http: Request finished: method=PUT url=/v1/acl/auth-method from=127.0.0.1:39150 latency=1.790604ms writer.go:29: 2021-01-29T19:34:31.190Z [DEBUG] TestLogoutCommand.http: Request finished: method=PUT url=/v1/acl/binding-rule from=127.0.0.1:39150 latency=738.276µs writer.go:29: 2021-01-29T19:34:31.200Z [DEBUG] TestLogoutCommand.http: Request finished: method=POST url=/v1/acl/login from=127.0.0.1:39150 latency=4.894956ms === RUN TestLogoutCommand/logout_of_login_token === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:31.209Z [DEBUG] TestLogoutCommand_k8s.http: Request finished: method=POST url=/v1/acl/login from=127.0.0.1:37686 latency=21.826001ms === RUN TestLogoutCommand_k8s/logout_of_login_token === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:31.210Z [DEBUG] TestLogoutCommand.http: Request finished: method=POST url=/v1/acl/logout from=127.0.0.1:39156 latency=711.507µs writer.go:29: 2021-01-29T19:34:31.215Z [INFO] TestLogoutCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:31.216Z [INFO] TestLogoutCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:31.217Z [DEBUG] TestLogoutCommand.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:31.218Z [DEBUG] TestLogoutCommand.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:31.219Z [DEBUG] TestLogoutCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:31.220Z [WARN] TestLogoutCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:31.219Z [DEBUG] TestLogoutCommand.leader: stopped routine: routine="legacy ACL token upgrade" === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:31.223Z [DEBUG] TestLogoutCommand_k8s.http: Request finished: method=POST url=/v1/acl/logout from=127.0.0.1:37702 latency=5.82887ms writer.go:29: 2021-01-29T19:34:31.227Z [INFO] TestLogoutCommand_k8s: Requesting shutdown === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:31.219Z [DEBUG] TestLogoutCommand.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:31.220Z [DEBUG] TestLogoutCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:31.223Z [WARN] TestLogoutCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:31.236Z [INFO] TestLogoutCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:31.237Z [INFO] TestLogoutCommand: consul server down writer.go:29: 2021-01-29T19:34:31.239Z [INFO] TestLogoutCommand: shutdown complete writer.go:29: 2021-01-29T19:34:31.240Z [INFO] TestLogoutCommand: Stopping server: protocol=DNS address=127.0.0.1:15112 network=tcp writer.go:29: 2021-01-29T19:34:31.240Z [INFO] TestLogoutCommand: Stopping server: protocol=DNS address=127.0.0.1:15112 network=udp === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:31.229Z [INFO] TestLogoutCommand_k8s.server: shutting down server === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:31.242Z [INFO] TestLogoutCommand: Stopping server: protocol=HTTP address=127.0.0.1:15113 network=tcp === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:31.242Z [DEBUG] TestLogoutCommand_k8s.leader: stopping routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:31.246Z [DEBUG] TestLogoutCommand_k8s.leader: stopping routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:31.250Z [DEBUG] TestLogoutCommand_k8s.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:31.246Z [DEBUG] TestLogoutCommand_k8s.leader: stopped routine: routine="legacy ACL token upgrade" writer.go:29: 2021-01-29T19:34:31.250Z [DEBUG] TestLogoutCommand_k8s.leader: stopped routine: routine="acl token reaping" writer.go:29: 2021-01-29T19:34:31.253Z [WARN] TestLogoutCommand_k8s.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:31.253Z [DEBUG] TestLogoutCommand_k8s.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:31.256Z [WARN] TestLogoutCommand_k8s.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:31.269Z [INFO] TestLogoutCommand_k8s.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:31.272Z [INFO] TestLogoutCommand_k8s: consul server down writer.go:29: 2021-01-29T19:34:31.274Z [INFO] TestLogoutCommand_k8s: shutdown complete writer.go:29: 2021-01-29T19:34:31.275Z [INFO] TestLogoutCommand_k8s: Stopping server: protocol=DNS address=127.0.0.1:15106 network=tcp writer.go:29: 2021-01-29T19:34:31.277Z [INFO] TestLogoutCommand_k8s: Stopping server: protocol=DNS address=127.0.0.1:15106 network=udp writer.go:29: 2021-01-29T19:34:31.278Z [INFO] TestLogoutCommand_k8s: Stopping server: protocol=HTTP address=127.0.0.1:15107 network=tcp === CONT TestLogoutCommand writer.go:29: 2021-01-29T19:34:31.745Z [INFO] TestLogoutCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:31.747Z [INFO] TestLogoutCommand: Endpoints down --- PASS: TestLogoutCommand (1.17s) --- PASS: TestLogoutCommand/no_token_specified (0.02s) --- PASS: TestLogoutCommand/logout_of_deleted_token (0.03s) --- PASS: TestLogoutCommand/logout_of_ordinary_token (0.01s) --- PASS: TestLogoutCommand/logout_of_login_token (0.01s) === CONT TestLogoutCommand_k8s writer.go:29: 2021-01-29T19:34:31.780Z [INFO] TestLogoutCommand_k8s: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:31.781Z [INFO] TestLogoutCommand_k8s: Endpoints down --- PASS: TestLogoutCommand_k8s (1.21s) --- PASS: TestLogoutCommand_k8s/no_token_specified (0.02s) --- PASS: TestLogoutCommand_k8s/logout_of_deleted_token (0.07s) --- PASS: TestLogoutCommand_k8s/logout_of_ordinary_token (0.04s) --- PASS: TestLogoutCommand_k8s/logout_of_login_token (0.01s) PASS ok github.com/hashicorp/consul/command/logout 1.464s === RUN TestMaintCommand_noTabs === PAUSE TestMaintCommand_noTabs === RUN TestMaintCommand_ConflictingArgs === PAUSE TestMaintCommand_ConflictingArgs === RUN TestMaintCommand_NoArgs === PAUSE TestMaintCommand_NoArgs === RUN TestMaintCommand_EnableNodeMaintenance === PAUSE TestMaintCommand_EnableNodeMaintenance === RUN TestMaintCommand_DisableNodeMaintenance === PAUSE TestMaintCommand_DisableNodeMaintenance === RUN TestMaintCommand_EnableServiceMaintenance === PAUSE TestMaintCommand_EnableServiceMaintenance === RUN TestMaintCommand_DisableServiceMaintenance === PAUSE TestMaintCommand_DisableServiceMaintenance === RUN TestMaintCommand_ServiceMaintenance_NoService === PAUSE TestMaintCommand_ServiceMaintenance_NoService === CONT TestMaintCommand_noTabs === CONT TestMaintCommand_EnableServiceMaintenance === CONT TestMaintCommand_EnableNodeMaintenance === CONT TestMaintCommand_ServiceMaintenance_NoService [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestMaintCommand_noTabs (0.03s) === CONT TestMaintCommand_DisableServiceMaintenance === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.913Z [WARN] TestMaintCommand_EnableServiceMaintenance: bootstrap = true: do not enable unless necessary === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.916Z [WARN] TestMaintCommand_DisableServiceMaintenance: bootstrap = true: do not enable unless necessary === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.932Z [DEBUG] TestMaintCommand_EnableServiceMaintenance.tlsutil: Update: version=1 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.933Z [DEBUG] TestMaintCommand_DisableServiceMaintenance.tlsutil: Update: version=1 === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.934Z [DEBUG] TestMaintCommand_EnableServiceMaintenance.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.935Z [DEBUG] TestMaintCommand_DisableServiceMaintenance.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:31.935Z [WARN] TestMaintCommand_ServiceMaintenance_NoService: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:31.936Z [DEBUG] TestMaintCommand_ServiceMaintenance_NoService.tlsutil: Update: version=1 === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:31.932Z [WARN] TestMaintCommand_EnableNodeMaintenance: bootstrap = true: do not enable unless necessary === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:31.938Z [DEBUG] TestMaintCommand_ServiceMaintenance_NoService.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:31.939Z [DEBUG] TestMaintCommand_EnableNodeMaintenance.tlsutil: Update: version=1 === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.940Z [INFO] TestMaintCommand_EnableServiceMaintenance.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a316a981-eef5-5562-65ce-dbcd578c3915 Address:127.0.0.1:19195}]" === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:31.940Z [DEBUG] TestMaintCommand_EnableNodeMaintenance.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.944Z [INFO] TestMaintCommand_EnableServiceMaintenance.server.serf.wan: serf: EventMemberJoin: Node-a316a981-eef5-5562-65ce-dbcd578c3915.dc1 127.0.0.1 === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:31.962Z [INFO] TestMaintCommand_EnableNodeMaintenance.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:48a0aee3-30a7-2609-87fb-288107bf02a2 Address:127.0.0.1:19201}]" writer.go:29: 2021-01-29T19:34:31.965Z [INFO] TestMaintCommand_EnableNodeMaintenance.server.serf.wan: serf: EventMemberJoin: Node-48a0aee3-30a7-2609-87fb-288107bf02a2.dc1 127.0.0.1 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.968Z [INFO] TestMaintCommand_DisableServiceMaintenance.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:04542bf3-2bc0-79ef-7391-322f719af73c Address:127.0.0.1:19213}]" === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:31.971Z [INFO] TestMaintCommand_EnableNodeMaintenance.server.serf.lan: serf: EventMemberJoin: Node-48a0aee3-30a7-2609-87fb-288107bf02a2 127.0.0.1 writer.go:29: 2021-01-29T19:34:31.974Z [INFO] TestMaintCommand_EnableNodeMaintenance.server.raft: entering follower state: follower="Node at 127.0.0.1:19201 [Follower]" leader= writer.go:29: 2021-01-29T19:34:31.981Z [INFO] TestMaintCommand_EnableNodeMaintenance: Started DNS server: address=127.0.0.1:19196 network=udp === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.981Z [INFO] TestMaintCommand_DisableServiceMaintenance.server.serf.wan: serf: EventMemberJoin: Node-04542bf3-2bc0-79ef-7391-322f719af73c.dc1 127.0.0.1 === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.960Z [INFO] TestMaintCommand_EnableServiceMaintenance.server.serf.lan: serf: EventMemberJoin: Node-a316a981-eef5-5562-65ce-dbcd578c3915 127.0.0.1 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.985Z [INFO] TestMaintCommand_DisableServiceMaintenance.server.serf.lan: serf: EventMemberJoin: Node-04542bf3-2bc0-79ef-7391-322f719af73c 127.0.0.1 writer.go:29: 2021-01-29T19:34:32.003Z [INFO] TestMaintCommand_DisableServiceMaintenance.server.raft: entering follower state: follower="Node at 127.0.0.1:19213 [Follower]" leader= === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.006Z [INFO] TestMaintCommand_EnableNodeMaintenance.server: Adding LAN server: server="Node-48a0aee3-30a7-2609-87fb-288107bf02a2 (Addr: tcp/127.0.0.1:19201) (DC: dc1)" === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.004Z [INFO] TestMaintCommand_DisableServiceMaintenance.server: Adding LAN server: server="Node-04542bf3-2bc0-79ef-7391-322f719af73c (Addr: tcp/127.0.0.1:19213) (DC: dc1)" === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.987Z [INFO] TestMaintCommand_EnableServiceMaintenance: Started DNS server: address=127.0.0.1:19190 network=udp === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.010Z [INFO] TestMaintCommand_EnableNodeMaintenance.server: Handled event for server in area: event=member-join server=Node-48a0aee3-30a7-2609-87fb-288107bf02a2.dc1 area=wan === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.010Z [INFO] TestMaintCommand_DisableServiceMaintenance.server: Handled event for server in area: event=member-join server=Node-04542bf3-2bc0-79ef-7391-322f719af73c.dc1 area=wan === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.012Z [INFO] TestMaintCommand_EnableNodeMaintenance: Started DNS server: address=127.0.0.1:19196 network=tcp === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:31.961Z [INFO] TestMaintCommand_EnableServiceMaintenance.server.raft: entering follower state: follower="Node at 127.0.0.1:19195 [Follower]" leader= === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.006Z [INFO] TestMaintCommand_DisableServiceMaintenance: Started DNS server: address=127.0.0.1:19208 network=tcp === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.034Z [INFO] TestMaintCommand_EnableNodeMaintenance: Started HTTP server: address=127.0.0.1:19197 network=tcp === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.036Z [INFO] TestMaintCommand_DisableServiceMaintenance: Started DNS server: address=127.0.0.1:19208 network=udp writer.go:29: 2021-01-29T19:34:32.038Z [INFO] TestMaintCommand_DisableServiceMaintenance: Started HTTP server: address=127.0.0.1:19209 network=tcp writer.go:29: 2021-01-29T19:34:32.039Z [INFO] TestMaintCommand_DisableServiceMaintenance: started state syncer === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.037Z [INFO] TestMaintCommand_EnableNodeMaintenance: started state syncer === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.003Z [INFO] TestMaintCommand_EnableServiceMaintenance.server: Handled event for server in area: event=member-join server=Node-a316a981-eef5-5562-65ce-dbcd578c3915.dc1 area=wan === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.043Z [WARN] TestMaintCommand_EnableNodeMaintenance.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.045Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8d539d94-8ebe-ad59-3d33-16bd7a413e5a Address:127.0.0.1:19207}]" === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.045Z [INFO] TestMaintCommand_EnableNodeMaintenance.server.raft: entering candidate state: node="Node at 127.0.0.1:19201 [Candidate]" term=2 === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.012Z [INFO] TestMaintCommand_EnableServiceMaintenance.server: Adding LAN server: server="Node-a316a981-eef5-5562-65ce-dbcd578c3915 (Addr: tcp/127.0.0.1:19195) (DC: dc1)" === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.047Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server.raft: entering follower state: follower="Node at 127.0.0.1:19207 [Follower]" leader= === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.025Z [INFO] TestMaintCommand_EnableServiceMaintenance: Started DNS server: address=127.0.0.1:19190 network=tcp === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.048Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server.serf.wan: serf: EventMemberJoin: Node-8d539d94-8ebe-ad59-3d33-16bd7a413e5a.dc1 127.0.0.1 === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.050Z [DEBUG] TestMaintCommand_EnableNodeMaintenance.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:32.052Z [DEBUG] TestMaintCommand_EnableNodeMaintenance.server.raft: vote granted: from=48a0aee3-30a7-2609-87fb-288107bf02a2 term=2 tally=1 writer.go:29: 2021-01-29T19:34:32.055Z [INFO] TestMaintCommand_EnableNodeMaintenance.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:32.058Z [INFO] TestMaintCommand_EnableNodeMaintenance.server.raft: entering leader state: leader="Node at 127.0.0.1:19201 [Leader]" === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.055Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server.serf.lan: serf: EventMemberJoin: Node-8d539d94-8ebe-ad59-3d33-16bd7a413e5a 127.0.0.1 === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.058Z [INFO] TestMaintCommand_EnableServiceMaintenance: Started HTTP server: address=127.0.0.1:19191 network=tcp writer.go:29: 2021-01-29T19:34:32.061Z [INFO] TestMaintCommand_EnableServiceMaintenance: started state syncer === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.062Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: Started DNS server: address=127.0.0.1:19202 network=udp === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.060Z [WARN] TestMaintCommand_DisableServiceMaintenance.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.060Z [INFO] TestMaintCommand_EnableNodeMaintenance.server: cluster leadership acquired === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.063Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server: Adding LAN server: server="Node-8d539d94-8ebe-ad59-3d33-16bd7a413e5a (Addr: tcp/127.0.0.1:19207) (DC: dc1)" === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.063Z [INFO] TestMaintCommand_EnableNodeMaintenance.server: New leader elected: payload=Node-48a0aee3-30a7-2609-87fb-288107bf02a2 === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.066Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server: Handled event for server in area: event=member-join server=Node-8d539d94-8ebe-ad59-3d33-16bd7a413e5a.dc1 area=wan === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.065Z [DEBUG] TestMaintCommand_EnableNodeMaintenance.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19201 === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.067Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: Started DNS server: address=127.0.0.1:19202 network=tcp === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.064Z [INFO] TestMaintCommand_DisableServiceMaintenance.server.raft: entering candidate state: node="Node at 127.0.0.1:19213 [Candidate]" term=2 === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.069Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: Started HTTP server: address=127.0.0.1:19203 network=tcp === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.071Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.070Z [DEBUG] TestMaintCommand_DisableServiceMaintenance.server.raft: votes: needed=1 === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.071Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: started state syncer === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.072Z [DEBUG] TestMaintCommand_DisableServiceMaintenance.server.raft: vote granted: from=04542bf3-2bc0-79ef-7391-322f719af73c term=2 tally=1 === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.074Z [WARN] TestMaintCommand_EnableServiceMaintenance.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:32.075Z [INFO] TestMaintCommand_EnableServiceMaintenance.server.raft: entering candidate state: node="Node at 127.0.0.1:19195 [Candidate]" term=2 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.074Z [INFO] TestMaintCommand_DisableServiceMaintenance.server.raft: election won: tally=1 === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.076Z [DEBUG] TestMaintCommand_EnableServiceMaintenance.server.raft: votes: needed=1 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.076Z [INFO] TestMaintCommand_DisableServiceMaintenance.server.raft: entering leader state: leader="Node at 127.0.0.1:19213 [Leader]" writer.go:29: 2021-01-29T19:34:32.078Z [INFO] TestMaintCommand_DisableServiceMaintenance.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:32.079Z [INFO] TestMaintCommand_DisableServiceMaintenance.server: New leader elected: payload=Node-04542bf3-2bc0-79ef-7391-322f719af73c writer.go:29: 2021-01-29T19:34:32.079Z [DEBUG] TestMaintCommand_DisableServiceMaintenance.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19213 === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.077Z [DEBUG] TestMaintCommand_EnableServiceMaintenance.server.raft: vote granted: from=a316a981-eef5-5562-65ce-dbcd578c3915 term=2 tally=1 writer.go:29: 2021-01-29T19:34:32.082Z [INFO] TestMaintCommand_EnableServiceMaintenance.server.raft: election won: tally=1 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.084Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.084Z [INFO] TestMaintCommand_EnableServiceMaintenance.server.raft: entering leader state: leader="Node at 127.0.0.1:19195 [Leader]" writer.go:29: 2021-01-29T19:34:32.086Z [INFO] TestMaintCommand_EnableServiceMaintenance.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:32.087Z [INFO] TestMaintCommand_EnableServiceMaintenance.server: New leader elected: payload=Node-a316a981-eef5-5562-65ce-dbcd578c3915 writer.go:29: 2021-01-29T19:34:32.088Z [DEBUG] TestMaintCommand_EnableServiceMaintenance.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19195 writer.go:29: 2021-01-29T19:34:32.093Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.117Z [DEBUG] TestMaintCommand_DisableServiceMaintenance.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:43708 latency=15.526108ms === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.128Z [WARN] TestMaintCommand_ServiceMaintenance_NoService.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:32.139Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server.raft: entering candidate state: node="Node at 127.0.0.1:19207 [Candidate]" term=2 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.140Z [INFO] TestMaintCommand_DisableServiceMaintenance: Synced node info === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.147Z [INFO] TestMaintCommand_EnableNodeMaintenance.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:32.149Z [INFO] TestMaintCommand_EnableNodeMaintenance.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.150Z [DEBUG] TestMaintCommand_EnableNodeMaintenance.server: Skipping self join check for node since the cluster is too small: node=Node-48a0aee3-30a7-2609-87fb-288107bf02a2 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.154Z [INFO] TestMaintCommand_DisableServiceMaintenance: Synced service: service=test === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.152Z [INFO] TestMaintCommand_EnableNodeMaintenance.server: member joined, marking health alive: member=Node-48a0aee3-30a7-2609-87fb-288107bf02a2 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.157Z [DEBUG] TestMaintCommand_DisableServiceMaintenance.http: Request finished: method=PUT url=/v1/agent/service/maintenance/test?enable=false from=127.0.0.1:43708 latency=25.6965ms writer.go:29: 2021-01-29T19:34:32.160Z [INFO] TestMaintCommand_DisableServiceMaintenance: Requesting shutdown writer.go:29: 2021-01-29T19:34:32.161Z [INFO] TestMaintCommand_DisableServiceMaintenance.server: shutting down server writer.go:29: 2021-01-29T19:34:32.163Z [WARN] TestMaintCommand_DisableServiceMaintenance.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.165Z [ERROR] TestMaintCommand_DisableServiceMaintenance.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:32.173Z [WARN] TestMaintCommand_DisableServiceMaintenance.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.176Z [INFO] TestMaintCommand_DisableServiceMaintenance.server.router.manager: shutting down === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.176Z [DEBUG] TestMaintCommand_ServiceMaintenance_NoService.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:32.178Z [DEBUG] TestMaintCommand_ServiceMaintenance_NoService.server.raft: vote granted: from=8d539d94-8ebe-ad59-3d33-16bd7a413e5a term=2 tally=1 writer.go:29: 2021-01-29T19:34:32.179Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:32.181Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server.raft: entering leader state: leader="Node at 127.0.0.1:19207 [Leader]" === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.183Z [INFO] TestMaintCommand_DisableServiceMaintenance: consul server down writer.go:29: 2021-01-29T19:34:32.184Z [INFO] TestMaintCommand_DisableServiceMaintenance: shutdown complete writer.go:29: 2021-01-29T19:34:32.186Z [INFO] TestMaintCommand_DisableServiceMaintenance: Stopping server: protocol=DNS address=127.0.0.1:19208 network=tcp writer.go:29: 2021-01-29T19:34:32.189Z [ERROR] TestMaintCommand_DisableServiceMaintenance.server.connect: Raft apply failed: error="leadership lost while committing log" writer.go:29: 2021-01-29T19:34:32.191Z [ERROR] TestMaintCommand_DisableServiceMaintenance.server: failed to establish leadership: error="leadership lost while committing log" === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.190Z [INFO] TestMaintCommand_EnableServiceMaintenance.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.192Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server: cluster leadership acquired === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.194Z [INFO] TestMaintCommand_EnableServiceMaintenance.leader: started routine: routine="CA root pruning" === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.190Z [INFO] TestMaintCommand_DisableServiceMaintenance: Stopping server: protocol=DNS address=127.0.0.1:19208 network=udp === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.195Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server: New leader elected: payload=Node-8d539d94-8ebe-ad59-3d33-16bd7a413e5a === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.195Z [DEBUG] TestMaintCommand_EnableServiceMaintenance.server: Skipping self join check for node since the cluster is too small: node=Node-a316a981-eef5-5562-65ce-dbcd578c3915 === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.196Z [INFO] TestMaintCommand_DisableServiceMaintenance: Stopping server: protocol=HTTP address=127.0.0.1:19209 network=tcp === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.197Z [INFO] TestMaintCommand_EnableServiceMaintenance.server: member joined, marking health alive: member=Node-a316a981-eef5-5562-65ce-dbcd578c3915 === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.199Z [DEBUG] TestMaintCommand_ServiceMaintenance_NoService.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19207 === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.198Z [INFO] TestMaintCommand_EnableServiceMaintenance: Synced node info === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.202Z [DEBUG] TestMaintCommand_EnableNodeMaintenance.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:57978 latency=2.651447ms === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.208Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.216Z [INFO] TestMaintCommand_EnableNodeMaintenance: Node entered maintenance mode === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.220Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.220Z [INFO] TestMaintCommand_EnableNodeMaintenance: Synced node info === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.221Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.223Z [DEBUG] TestMaintCommand_ServiceMaintenance_NoService.server: Skipping self join check for node since the cluster is too small: node=Node-8d539d94-8ebe-ad59-3d33-16bd7a413e5a writer.go:29: 2021-01-29T19:34:32.224Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server: member joined, marking health alive: member=Node-8d539d94-8ebe-ad59-3d33-16bd7a413e5a === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.227Z [INFO] TestMaintCommand_EnableNodeMaintenance: Synced check: check=_node_maintenance writer.go:29: 2021-01-29T19:34:32.230Z [DEBUG] TestMaintCommand_EnableNodeMaintenance.http: Request finished: method=PUT url=/v1/agent/maintenance?enable=true&reason=broken from=127.0.0.1:57978 latency=16.898269ms writer.go:29: 2021-01-29T19:34:32.232Z [INFO] TestMaintCommand_EnableNodeMaintenance: Requesting shutdown writer.go:29: 2021-01-29T19:34:32.233Z [INFO] TestMaintCommand_EnableNodeMaintenance.server: shutting down server writer.go:29: 2021-01-29T19:34:32.235Z [DEBUG] TestMaintCommand_EnableNodeMaintenance.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.236Z [WARN] TestMaintCommand_EnableNodeMaintenance.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.235Z [ERROR] TestMaintCommand_EnableNodeMaintenance.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:32.236Z [DEBUG] TestMaintCommand_EnableNodeMaintenance.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.239Z [WARN] TestMaintCommand_EnableNodeMaintenance.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.245Z [INFO] TestMaintCommand_EnableNodeMaintenance.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:32.245Z [INFO] TestMaintCommand_EnableNodeMaintenance: consul server down writer.go:29: 2021-01-29T19:34:32.249Z [INFO] TestMaintCommand_EnableNodeMaintenance: shutdown complete writer.go:29: 2021-01-29T19:34:32.250Z [INFO] TestMaintCommand_EnableNodeMaintenance: Stopping server: protocol=DNS address=127.0.0.1:19196 network=tcp writer.go:29: 2021-01-29T19:34:32.252Z [INFO] TestMaintCommand_EnableNodeMaintenance: Stopping server: protocol=DNS address=127.0.0.1:19196 network=udp writer.go:29: 2021-01-29T19:34:32.253Z [INFO] TestMaintCommand_EnableNodeMaintenance: Stopping server: protocol=HTTP address=127.0.0.1:19197 network=tcp === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.283Z [DEBUG] TestMaintCommand_ServiceMaintenance_NoService.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:33988 latency=725.737µs writer.go:29: 2021-01-29T19:34:32.289Z [DEBUG] TestMaintCommand_ServiceMaintenance_NoService.http: Request finished: method=PUT url=/v1/agent/service/maintenance/redis?enable=true&reason=broken from=127.0.0.1:33988 latency=1.550105ms writer.go:29: 2021-01-29T19:34:32.292Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: Requesting shutdown writer.go:29: 2021-01-29T19:34:32.294Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server: shutting down server writer.go:29: 2021-01-29T19:34:32.296Z [DEBUG] TestMaintCommand_ServiceMaintenance_NoService.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.297Z [WARN] TestMaintCommand_ServiceMaintenance_NoService.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.297Z [DEBUG] TestMaintCommand_ServiceMaintenance_NoService.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.296Z [ERROR] TestMaintCommand_ServiceMaintenance_NoService.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:32.300Z [WARN] TestMaintCommand_ServiceMaintenance_NoService.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.305Z [INFO] TestMaintCommand_ServiceMaintenance_NoService.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:32.305Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: consul server down writer.go:29: 2021-01-29T19:34:32.309Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: shutdown complete writer.go:29: 2021-01-29T19:34:32.311Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: Stopping server: protocol=DNS address=127.0.0.1:19202 network=tcp writer.go:29: 2021-01-29T19:34:32.313Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: Stopping server: protocol=DNS address=127.0.0.1:19202 network=udp writer.go:29: 2021-01-29T19:34:32.314Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: Stopping server: protocol=HTTP address=127.0.0.1:19203 network=tcp === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.499Z [DEBUG] TestMaintCommand_EnableServiceMaintenance.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:32794 latency=1.443497ms writer.go:29: 2021-01-29T19:34:32.503Z [INFO] TestMaintCommand_EnableServiceMaintenance: Service entered maintenance mode: service=test writer.go:29: 2021-01-29T19:34:32.504Z [DEBUG] TestMaintCommand_EnableServiceMaintenance: Node info in sync writer.go:29: 2021-01-29T19:34:32.522Z [INFO] TestMaintCommand_EnableServiceMaintenance: Synced service: service=test writer.go:29: 2021-01-29T19:34:32.523Z [DEBUG] TestMaintCommand_EnableServiceMaintenance: Check in sync: check=_service_maintenance:test writer.go:29: 2021-01-29T19:34:32.524Z [DEBUG] TestMaintCommand_EnableServiceMaintenance.http: Request finished: method=PUT url=/v1/agent/service/maintenance/test?enable=true&reason=broken from=127.0.0.1:32794 latency=22.263471ms writer.go:29: 2021-01-29T19:34:32.526Z [INFO] TestMaintCommand_EnableServiceMaintenance: Requesting shutdown writer.go:29: 2021-01-29T19:34:32.528Z [INFO] TestMaintCommand_EnableServiceMaintenance.server: shutting down server writer.go:29: 2021-01-29T19:34:32.530Z [DEBUG] TestMaintCommand_EnableServiceMaintenance.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.534Z [DEBUG] TestMaintCommand_EnableServiceMaintenance.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.534Z [WARN] TestMaintCommand_EnableServiceMaintenance.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.537Z [WARN] TestMaintCommand_EnableServiceMaintenance.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:32.540Z [INFO] TestMaintCommand_EnableServiceMaintenance.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:32.541Z [INFO] TestMaintCommand_EnableServiceMaintenance: consul server down writer.go:29: 2021-01-29T19:34:32.546Z [INFO] TestMaintCommand_EnableServiceMaintenance: shutdown complete writer.go:29: 2021-01-29T19:34:32.548Z [INFO] TestMaintCommand_EnableServiceMaintenance: Stopping server: protocol=DNS address=127.0.0.1:19190 network=tcp writer.go:29: 2021-01-29T19:34:32.552Z [INFO] TestMaintCommand_EnableServiceMaintenance: Stopping server: protocol=DNS address=127.0.0.1:19190 network=udp writer.go:29: 2021-01-29T19:34:32.557Z [INFO] TestMaintCommand_EnableServiceMaintenance: Stopping server: protocol=HTTP address=127.0.0.1:19191 network=tcp === CONT TestMaintCommand_DisableServiceMaintenance writer.go:29: 2021-01-29T19:34:32.698Z [INFO] TestMaintCommand_DisableServiceMaintenance: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:32.699Z [INFO] TestMaintCommand_DisableServiceMaintenance: Endpoints down --- PASS: TestMaintCommand_DisableServiceMaintenance (0.85s) === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.709Z [WARN] TestMaintCommand_DisableNodeMaintenance: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:32.722Z [DEBUG] TestMaintCommand_DisableNodeMaintenance.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:32.724Z [DEBUG] TestMaintCommand_DisableNodeMaintenance.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:32.743Z [INFO] TestMaintCommand_DisableNodeMaintenance.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4b75fc47-8415-0415-3407-851f24e9488a Address:127.0.0.1:19219}]" writer.go:29: 2021-01-29T19:34:32.744Z [INFO] TestMaintCommand_DisableNodeMaintenance.server.raft: entering follower state: follower="Node at 127.0.0.1:19219 [Follower]" leader= writer.go:29: 2021-01-29T19:34:32.744Z [INFO] TestMaintCommand_DisableNodeMaintenance.server.serf.wan: serf: EventMemberJoin: Node-4b75fc47-8415-0415-3407-851f24e9488a.dc1 127.0.0.1 === CONT TestMaintCommand_EnableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.755Z [INFO] TestMaintCommand_EnableNodeMaintenance: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:32.756Z [INFO] TestMaintCommand_EnableNodeMaintenance: Endpoints down --- PASS: TestMaintCommand_EnableNodeMaintenance (0.94s) === CONT TestMaintCommand_NoArgs === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.755Z [INFO] TestMaintCommand_DisableNodeMaintenance.server.serf.lan: serf: EventMemberJoin: Node-4b75fc47-8415-0415-3407-851f24e9488a 127.0.0.1 writer.go:29: 2021-01-29T19:34:32.771Z [INFO] TestMaintCommand_DisableNodeMaintenance.server: Handled event for server in area: event=member-join server=Node-4b75fc47-8415-0415-3407-851f24e9488a.dc1 area=wan writer.go:29: 2021-01-29T19:34:32.771Z [INFO] TestMaintCommand_DisableNodeMaintenance.server: Adding LAN server: server="Node-4b75fc47-8415-0415-3407-851f24e9488a (Addr: tcp/127.0.0.1:19219) (DC: dc1)" === CONT TestMaintCommand_NoArgs writer.go:29: 2021-01-29T19:34:32.773Z [WARN] TestMaintCommand_NoArgs: bootstrap = true: do not enable unless necessary === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.774Z [INFO] TestMaintCommand_DisableNodeMaintenance: Started DNS server: address=127.0.0.1:19214 network=udp writer.go:29: 2021-01-29T19:34:32.774Z [INFO] TestMaintCommand_DisableNodeMaintenance: Started DNS server: address=127.0.0.1:19214 network=tcp === CONT TestMaintCommand_NoArgs writer.go:29: 2021-01-29T19:34:32.774Z [DEBUG] TestMaintCommand_NoArgs.tlsutil: Update: version=1 === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.776Z [INFO] TestMaintCommand_DisableNodeMaintenance: Started HTTP server: address=127.0.0.1:19215 network=tcp === CONT TestMaintCommand_NoArgs writer.go:29: 2021-01-29T19:34:32.776Z [DEBUG] TestMaintCommand_NoArgs.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.777Z [INFO] TestMaintCommand_DisableNodeMaintenance: started state syncer === CONT TestMaintCommand_NoArgs writer.go:29: 2021-01-29T19:34:32.785Z [INFO] TestMaintCommand_NoArgs.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5914939c-c62c-6f69-385a-94e4ef639625 Address:127.0.0.1:19225}]" writer.go:29: 2021-01-29T19:34:32.787Z [INFO] TestMaintCommand_NoArgs.server.raft: entering follower state: follower="Node at 127.0.0.1:19225 [Follower]" leader= writer.go:29: 2021-01-29T19:34:32.788Z [INFO] TestMaintCommand_NoArgs.server.serf.wan: serf: EventMemberJoin: Node-5914939c-c62c-6f69-385a-94e4ef639625.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:32.791Z [INFO] TestMaintCommand_NoArgs.server.serf.lan: serf: EventMemberJoin: Node-5914939c-c62c-6f69-385a-94e4ef639625 127.0.0.1 writer.go:29: 2021-01-29T19:34:32.793Z [INFO] TestMaintCommand_NoArgs.server: Handled event for server in area: event=member-join server=Node-5914939c-c62c-6f69-385a-94e4ef639625.dc1 area=wan writer.go:29: 2021-01-29T19:34:32.793Z [INFO] TestMaintCommand_NoArgs: Started DNS server: address=127.0.0.1:19220 network=udp writer.go:29: 2021-01-29T19:34:32.795Z [INFO] TestMaintCommand_NoArgs: Started DNS server: address=127.0.0.1:19220 network=tcp writer.go:29: 2021-01-29T19:34:32.797Z [INFO] TestMaintCommand_NoArgs: Started HTTP server: address=127.0.0.1:19221 network=tcp writer.go:29: 2021-01-29T19:34:32.798Z [INFO] TestMaintCommand_NoArgs: started state syncer writer.go:29: 2021-01-29T19:34:32.794Z [INFO] TestMaintCommand_NoArgs.server: Adding LAN server: server="Node-5914939c-c62c-6f69-385a-94e4ef639625 (Addr: tcp/127.0.0.1:19225) (DC: dc1)" === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.807Z [WARN] TestMaintCommand_DisableNodeMaintenance.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:32.808Z [INFO] TestMaintCommand_DisableNodeMaintenance.server.raft: entering candidate state: node="Node at 127.0.0.1:19219 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:32.815Z [DEBUG] TestMaintCommand_DisableNodeMaintenance.server.raft: votes: needed=1 === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.816Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: Waiting for endpoints to shut down === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.816Z [DEBUG] TestMaintCommand_DisableNodeMaintenance.server.raft: vote granted: from=4b75fc47-8415-0415-3407-851f24e9488a term=2 tally=1 === CONT TestMaintCommand_ServiceMaintenance_NoService writer.go:29: 2021-01-29T19:34:32.818Z [INFO] TestMaintCommand_ServiceMaintenance_NoService: Endpoints down --- PASS: TestMaintCommand_ServiceMaintenance_NoService (1.00s) === CONT TestMaintCommand_ConflictingArgs === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.818Z [INFO] TestMaintCommand_DisableNodeMaintenance.server.raft: election won: tally=1 --- PASS: TestMaintCommand_ConflictingArgs (0.00s) === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:32.820Z [INFO] TestMaintCommand_DisableNodeMaintenance.server.raft: entering leader state: leader="Node at 127.0.0.1:19219 [Leader]" writer.go:29: 2021-01-29T19:34:32.822Z [INFO] TestMaintCommand_DisableNodeMaintenance.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:32.825Z [INFO] TestMaintCommand_DisableNodeMaintenance.server: New leader elected: payload=Node-4b75fc47-8415-0415-3407-851f24e9488a writer.go:29: 2021-01-29T19:34:32.825Z [DEBUG] TestMaintCommand_DisableNodeMaintenance.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19219 writer.go:29: 2021-01-29T19:34:32.834Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:32.840Z [INFO] TestMaintCommand_DisableNodeMaintenance.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:32.842Z [INFO] TestMaintCommand_DisableNodeMaintenance.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.844Z [DEBUG] TestMaintCommand_DisableNodeMaintenance.server: Skipping self join check for node since the cluster is too small: node=Node-4b75fc47-8415-0415-3407-851f24e9488a writer.go:29: 2021-01-29T19:34:32.845Z [INFO] TestMaintCommand_DisableNodeMaintenance.server: member joined, marking health alive: member=Node-4b75fc47-8415-0415-3407-851f24e9488a === CONT TestMaintCommand_NoArgs writer.go:29: 2021-01-29T19:34:32.853Z [WARN] TestMaintCommand_NoArgs.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:32.854Z [INFO] TestMaintCommand_NoArgs.server.raft: entering candidate state: node="Node at 127.0.0.1:19225 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:32.856Z [DEBUG] TestMaintCommand_NoArgs.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:32.857Z [DEBUG] TestMaintCommand_NoArgs.server.raft: vote granted: from=5914939c-c62c-6f69-385a-94e4ef639625 term=2 tally=1 writer.go:29: 2021-01-29T19:34:32.858Z [INFO] TestMaintCommand_NoArgs.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:32.859Z [INFO] TestMaintCommand_NoArgs.server.raft: entering leader state: leader="Node at 127.0.0.1:19225 [Leader]" writer.go:29: 2021-01-29T19:34:32.860Z [INFO] TestMaintCommand_NoArgs.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:32.861Z [INFO] TestMaintCommand_NoArgs.server: New leader elected: payload=Node-5914939c-c62c-6f69-385a-94e4ef639625 writer.go:29: 2021-01-29T19:34:32.862Z [DEBUG] TestMaintCommand_NoArgs.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:19225 writer.go:29: 2021-01-29T19:34:32.870Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:32.885Z [INFO] TestMaintCommand_NoArgs.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:32.886Z [INFO] TestMaintCommand_NoArgs.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:32.887Z [DEBUG] TestMaintCommand_NoArgs.server: Skipping self join check for node since the cluster is too small: node=Node-5914939c-c62c-6f69-385a-94e4ef639625 writer.go:29: 2021-01-29T19:34:32.888Z [INFO] TestMaintCommand_NoArgs.server: member joined, marking health alive: member=Node-5914939c-c62c-6f69-385a-94e4ef639625 writer.go:29: 2021-01-29T19:34:33.059Z [INFO] TestMaintCommand_NoArgs: Service entered maintenance mode: service=test writer.go:29: 2021-01-29T19:34:33.061Z [INFO] TestMaintCommand_NoArgs: Node entered maintenance mode === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:33.058Z [INFO] TestMaintCommand_EnableServiceMaintenance: Waiting for endpoints to shut down === CONT TestMaintCommand_NoArgs writer.go:29: 2021-01-29T19:34:33.066Z [DEBUG] TestMaintCommand_NoArgs.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:34302 latency=646.708µs writer.go:29: 2021-01-29T19:34:33.068Z [DEBUG] TestMaintCommand_NoArgs.http: Request finished: method=GET url=/v1/agent/checks from=127.0.0.1:34302 latency=74.402µs === CONT TestMaintCommand_EnableServiceMaintenance writer.go:29: 2021-01-29T19:34:33.066Z [INFO] TestMaintCommand_EnableServiceMaintenance: Endpoints down --- PASS: TestMaintCommand_EnableServiceMaintenance (1.25s) === CONT TestMaintCommand_NoArgs writer.go:29: 2021-01-29T19:34:33.071Z [INFO] TestMaintCommand_NoArgs: Requesting shutdown writer.go:29: 2021-01-29T19:34:33.073Z [INFO] TestMaintCommand_NoArgs.server: shutting down server writer.go:29: 2021-01-29T19:34:33.081Z [DEBUG] TestMaintCommand_NoArgs.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:33.081Z [ERROR] TestMaintCommand_NoArgs.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:33.085Z [WARN] TestMaintCommand_NoArgs.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:33.085Z [DEBUG] TestMaintCommand_NoArgs.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:33.092Z [WARN] TestMaintCommand_NoArgs.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:33.095Z [INFO] TestMaintCommand_NoArgs.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:33.095Z [INFO] TestMaintCommand_NoArgs: consul server down writer.go:29: 2021-01-29T19:34:33.097Z [INFO] TestMaintCommand_NoArgs: shutdown complete writer.go:29: 2021-01-29T19:34:33.098Z [INFO] TestMaintCommand_NoArgs: Stopping server: protocol=DNS address=127.0.0.1:19220 network=tcp writer.go:29: 2021-01-29T19:34:33.098Z [INFO] TestMaintCommand_NoArgs: Stopping server: protocol=DNS address=127.0.0.1:19220 network=udp writer.go:29: 2021-01-29T19:34:33.099Z [INFO] TestMaintCommand_NoArgs: Stopping server: protocol=HTTP address=127.0.0.1:19221 network=tcp === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:33.135Z [DEBUG] TestMaintCommand_DisableNodeMaintenance.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:56414 latency=1.351526ms writer.go:29: 2021-01-29T19:34:33.154Z [INFO] TestMaintCommand_DisableNodeMaintenance: Synced node info writer.go:29: 2021-01-29T19:34:33.156Z [DEBUG] TestMaintCommand_DisableNodeMaintenance.http: Request finished: method=PUT url=/v1/agent/maintenance?enable=false from=127.0.0.1:56414 latency=11.018591ms writer.go:29: 2021-01-29T19:34:33.158Z [INFO] TestMaintCommand_DisableNodeMaintenance: Requesting shutdown writer.go:29: 2021-01-29T19:34:33.160Z [INFO] TestMaintCommand_DisableNodeMaintenance.server: shutting down server writer.go:29: 2021-01-29T19:34:33.162Z [DEBUG] TestMaintCommand_DisableNodeMaintenance.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:33.164Z [WARN] TestMaintCommand_DisableNodeMaintenance.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:33.162Z [ERROR] TestMaintCommand_DisableNodeMaintenance.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:33.164Z [DEBUG] TestMaintCommand_DisableNodeMaintenance.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:33.168Z [WARN] TestMaintCommand_DisableNodeMaintenance.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:33.176Z [INFO] TestMaintCommand_DisableNodeMaintenance.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:33.176Z [INFO] TestMaintCommand_DisableNodeMaintenance: consul server down writer.go:29: 2021-01-29T19:34:33.181Z [INFO] TestMaintCommand_DisableNodeMaintenance: shutdown complete writer.go:29: 2021-01-29T19:34:33.182Z [INFO] TestMaintCommand_DisableNodeMaintenance: Stopping server: protocol=DNS address=127.0.0.1:19214 network=tcp writer.go:29: 2021-01-29T19:34:33.184Z [INFO] TestMaintCommand_DisableNodeMaintenance: Stopping server: protocol=DNS address=127.0.0.1:19214 network=udp writer.go:29: 2021-01-29T19:34:33.185Z [INFO] TestMaintCommand_DisableNodeMaintenance: Stopping server: protocol=HTTP address=127.0.0.1:19215 network=tcp === CONT TestMaintCommand_NoArgs writer.go:29: 2021-01-29T19:34:33.600Z [INFO] TestMaintCommand_NoArgs: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:33.602Z [INFO] TestMaintCommand_NoArgs: Endpoints down --- PASS: TestMaintCommand_NoArgs (0.85s) === CONT TestMaintCommand_DisableNodeMaintenance writer.go:29: 2021-01-29T19:34:33.687Z [INFO] TestMaintCommand_DisableNodeMaintenance: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:33.688Z [INFO] TestMaintCommand_DisableNodeMaintenance: Endpoints down --- PASS: TestMaintCommand_DisableNodeMaintenance (0.99s) PASS ok github.com/hashicorp/consul/command/maint 2.128s === RUN TestMembersCommand_noTabs === PAUSE TestMembersCommand_noTabs === RUN TestMembersCommand === PAUSE TestMembersCommand === RUN TestMembersCommand_WAN === PAUSE TestMembersCommand_WAN === RUN TestMembersCommand_statusFilter === PAUSE TestMembersCommand_statusFilter === RUN TestMembersCommand_statusFilter_failed === PAUSE TestMembersCommand_statusFilter_failed === RUN TestMembersCommand_verticalBar === PAUSE TestMembersCommand_verticalBar === CONT TestMembersCommand_noTabs === CONT TestMembersCommand_statusFilter_failed === CONT TestMembersCommand_verticalBar === CONT TestMembersCommand_WAN [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestMembersCommand_noTabs (0.01s) === CONT TestMembersCommand_statusFilter [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:34:36.886Z [WARN] TestMembersCommand_statusFilter: bootstrap = true: do not enable unless necessary === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:36.888Z [WARN] TestMembersCommand_WAN: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:36.889Z [DEBUG] TestMembersCommand_WAN.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:36.893Z [DEBUG] TestMembersCommand_WAN.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:36.887Z [DEBUG] TestMembersCommand_statusFilter.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:36.915Z [DEBUG] TestMembersCommand_statusFilter.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:36.919Z [INFO] TestMembersCommand_WAN.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3b234eb8-b9c9-4944-f924-3abc4825b4f4 Address:127.0.0.1:25327}]" writer.go:29: 2021-01-29T19:34:36.923Z [INFO] TestMembersCommand_WAN.server.raft: entering follower state: follower="Node at 127.0.0.1:25327 [Follower]" leader= === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:36.928Z [WARN] TestMembersCommand_statusFilter_failed: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:36.929Z [DEBUG] TestMembersCommand_statusFilter_failed.tlsutil: Update: version=1 === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:36.928Z [INFO] TestMembersCommand_WAN.server.serf.wan: serf: EventMemberJoin: Node-3b234eb8-b9c9-4944-f924-3abc4825b4f4.dc1 127.0.0.1 === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:36.930Z [DEBUG] TestMembersCommand_statusFilter_failed.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:36.932Z [INFO] TestMembersCommand_WAN.server.serf.lan: serf: EventMemberJoin: Node-3b234eb8-b9c9-4944-f924-3abc4825b4f4 127.0.0.1 writer.go:29: 2021-01-29T19:34:36.936Z [INFO] TestMembersCommand_WAN: Started DNS server: address=127.0.0.1:25322 network=udp writer.go:29: 2021-01-29T19:34:36.940Z [INFO] TestMembersCommand_WAN.server: Adding LAN server: server="Node-3b234eb8-b9c9-4944-f924-3abc4825b4f4 (Addr: tcp/127.0.0.1:25327) (DC: dc1)" === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:36.951Z [WARN] TestMembersCommand_verticalBar: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:36.952Z [WARN] TestMembersCommand_verticalBar: Node name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: node_name=name|with|bars === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:36.961Z [INFO] TestMembersCommand_statusFilter_failed.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d985e87b-c997-00d1-3b15-7f3d2bba8a72 Address:127.0.0.1:25321}]" writer.go:29: 2021-01-29T19:34:36.964Z [INFO] TestMembersCommand_statusFilter_failed.server.serf.wan: serf: EventMemberJoin: Node-d985e87b-c997-00d1-3b15-7f3d2bba8a72.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:36.966Z [INFO] TestMembersCommand_statusFilter_failed.server.serf.lan: serf: EventMemberJoin: Node-d985e87b-c997-00d1-3b15-7f3d2bba8a72 127.0.0.1 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:36.954Z [DEBUG] TestMembersCommand_verticalBar.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:36.972Z [DEBUG] TestMembersCommand_verticalBar.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:36.944Z [INFO] TestMembersCommand_WAN: Started DNS server: address=127.0.0.1:25322 network=tcp === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:36.967Z [INFO] TestMembersCommand_statusFilter_failed.server.raft: entering follower state: follower="Node at 127.0.0.1:25321 [Follower]" leader= writer.go:29: 2021-01-29T19:34:36.969Z [INFO] TestMembersCommand_statusFilter_failed: Started DNS server: address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:34:36.998Z [INFO] TestMembersCommand_statusFilter_failed.server: Adding LAN server: server="Node-d985e87b-c997-00d1-3b15-7f3d2bba8a72 (Addr: tcp/127.0.0.1:25321) (DC: dc1)" === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:36.946Z [INFO] TestMembersCommand_WAN.server: Handled event for server in area: event=member-join server=Node-3b234eb8-b9c9-4944-f924-3abc4825b4f4.dc1 area=wan writer.go:29: 2021-01-29T19:34:36.974Z [WARN] TestMembersCommand_WAN.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:37.003Z [INFO] TestMembersCommand_WAN.server.raft: entering candidate state: node="Node at 127.0.0.1:25327 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:36.998Z [INFO] TestMembersCommand_WAN: Started HTTP server: address=127.0.0.1:25323 network=tcp writer.go:29: 2021-01-29T19:34:37.005Z [INFO] TestMembersCommand_WAN: started state syncer === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:37.009Z [INFO] TestMembersCommand_statusFilter_failed: Started DNS server: address=127.0.0.1:25316 network=tcp === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:37.008Z [DEBUG] TestMembersCommand_WAN.server.raft: votes: needed=1 === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:37.010Z [INFO] TestMembersCommand_statusFilter_failed: Started HTTP server: address=127.0.0.1:25317 network=tcp === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:37.010Z [DEBUG] TestMembersCommand_WAN.server.raft: vote granted: from=3b234eb8-b9c9-4944-f924-3abc4825b4f4 term=2 tally=1 === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:37.011Z [INFO] TestMembersCommand_statusFilter_failed: started state syncer === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:37.011Z [INFO] TestMembersCommand_WAN.server.raft: election won: tally=1 === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:37.013Z [INFO] TestMembersCommand_statusFilter_failed.server: Handled event for server in area: event=member-join server=Node-d985e87b-c997-00d1-3b15-7f3d2bba8a72.dc1 area=wan === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:37.013Z [INFO] TestMembersCommand_WAN.server.raft: entering leader state: leader="Node at 127.0.0.1:25327 [Leader]" writer.go:29: 2021-01-29T19:34:37.014Z [INFO] TestMembersCommand_WAN.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:37.015Z [INFO] TestMembersCommand_WAN.server: New leader elected: payload=Node-3b234eb8-b9c9-4944-f924-3abc4825b4f4 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.017Z [INFO] TestMembersCommand_verticalBar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:47223be1-7d72-f7de-b191-cb4d500ecbb6 Address:127.0.0.1:25333}]" === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:37.016Z [DEBUG] TestMembersCommand_WAN.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25327 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.018Z [INFO] TestMembersCommand_verticalBar.server.serf.wan: serf: EventMemberJoin: name|with|bars.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:37.020Z [INFO] TestMembersCommand_verticalBar.server.raft: entering follower state: follower="Node at 127.0.0.1:25333 [Follower]" leader= writer.go:29: 2021-01-29T19:34:37.022Z [INFO] TestMembersCommand_verticalBar.server.serf.lan: serf: EventMemberJoin: name|with|bars 127.0.0.1 writer.go:29: 2021-01-29T19:34:37.025Z [INFO] TestMembersCommand_verticalBar: Started DNS server: address=127.0.0.1:25328 network=udp === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.015Z [INFO] TestMembersCommand_statusFilter.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8968bb3b-1e85-f0aa-975d-16adbcd69827 Address:127.0.0.1:25339}]" === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.027Z [INFO] TestMembersCommand_verticalBar.server: Adding LAN server: server="name|with|bars (Addr: tcp/127.0.0.1:25333) (DC: dc1)" writer.go:29: 2021-01-29T19:34:37.028Z [INFO] TestMembersCommand_verticalBar.server: Handled event for server in area: event=member-join server=name|with|bars.dc1 area=wan === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.029Z [INFO] TestMembersCommand_statusFilter.server.serf.wan: serf: EventMemberJoin: Node-8968bb3b-1e85-f0aa-975d-16adbcd69827.dc1 127.0.0.1 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.029Z [INFO] TestMembersCommand_verticalBar: Started DNS server: address=127.0.0.1:25328 network=tcp === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.030Z [INFO] TestMembersCommand_statusFilter.server.raft: entering follower state: follower="Node at 127.0.0.1:25339 [Follower]" leader= === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.038Z [INFO] TestMembersCommand_verticalBar: Started HTTP server: address=127.0.0.1:25329 network=tcp writer.go:29: 2021-01-29T19:34:37.038Z [INFO] TestMembersCommand_verticalBar: started state syncer === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:37.039Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.042Z [INFO] TestMembersCommand_statusFilter.server.serf.lan: serf: EventMemberJoin: Node-8968bb3b-1e85-f0aa-975d-16adbcd69827 127.0.0.1 writer.go:29: 2021-01-29T19:34:37.045Z [INFO] TestMembersCommand_statusFilter.server: Adding LAN server: server="Node-8968bb3b-1e85-f0aa-975d-16adbcd69827 (Addr: tcp/127.0.0.1:25339) (DC: dc1)" writer.go:29: 2021-01-29T19:34:37.046Z [INFO] TestMembersCommand_statusFilter.server: Handled event for server in area: event=member-join server=Node-8968bb3b-1e85-f0aa-975d-16adbcd69827.dc1 area=wan writer.go:29: 2021-01-29T19:34:37.049Z [INFO] TestMembersCommand_statusFilter: Started DNS server: address=127.0.0.1:25334 network=tcp writer.go:29: 2021-01-29T19:34:37.050Z [INFO] TestMembersCommand_statusFilter: Started DNS server: address=127.0.0.1:25334 network=udp writer.go:29: 2021-01-29T19:34:37.051Z [INFO] TestMembersCommand_statusFilter: Started HTTP server: address=127.0.0.1:25335 network=tcp writer.go:29: 2021-01-29T19:34:37.052Z [INFO] TestMembersCommand_statusFilter: started state syncer === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:37.056Z [WARN] TestMembersCommand_statusFilter_failed.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:37.057Z [INFO] TestMembersCommand_statusFilter_failed.server.raft: entering candidate state: node="Node at 127.0.0.1:25321 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:37.060Z [DEBUG] TestMembersCommand_statusFilter_failed.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:37.061Z [DEBUG] TestMembersCommand_statusFilter_failed.server.raft: vote granted: from=d985e87b-c997-00d1-3b15-7f3d2bba8a72 term=2 tally=1 writer.go:29: 2021-01-29T19:34:37.062Z [INFO] TestMembersCommand_statusFilter_failed.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:37.064Z [INFO] TestMembersCommand_statusFilter_failed.server.raft: entering leader state: leader="Node at 127.0.0.1:25321 [Leader]" writer.go:29: 2021-01-29T19:34:37.066Z [INFO] TestMembersCommand_statusFilter_failed.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:37.068Z [INFO] TestMembersCommand_statusFilter_failed.server: New leader elected: payload=Node-d985e87b-c997-00d1-3b15-7f3d2bba8a72 writer.go:29: 2021-01-29T19:34:37.070Z [DEBUG] TestMembersCommand_statusFilter_failed.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25321 writer.go:29: 2021-01-29T19:34:37.078Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.079Z [WARN] TestMembersCommand_verticalBar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:37.080Z [INFO] TestMembersCommand_verticalBar.server.raft: entering candidate state: node="Node at 127.0.0.1:25333 [Candidate]" term=2 === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.081Z [WARN] TestMembersCommand_statusFilter.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:37.082Z [INFO] TestMembersCommand_statusFilter.server.raft: entering candidate state: node="Node at 127.0.0.1:25339 [Candidate]" term=2 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.084Z [DEBUG] TestMembersCommand_verticalBar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:37.085Z [DEBUG] TestMembersCommand_verticalBar.server.raft: vote granted: from=47223be1-7d72-f7de-b191-cb4d500ecbb6 term=2 tally=1 === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.084Z [DEBUG] TestMembersCommand_statusFilter.server.raft: votes: needed=1 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.086Z [INFO] TestMembersCommand_verticalBar.server.raft: election won: tally=1 === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.086Z [DEBUG] TestMembersCommand_statusFilter.server.raft: vote granted: from=8968bb3b-1e85-f0aa-975d-16adbcd69827 term=2 tally=1 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.088Z [INFO] TestMembersCommand_verticalBar.server.raft: entering leader state: leader="Node at 127.0.0.1:25333 [Leader]" writer.go:29: 2021-01-29T19:34:37.090Z [INFO] TestMembersCommand_verticalBar.server: cluster leadership acquired === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.088Z [INFO] TestMembersCommand_statusFilter.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:37.103Z [INFO] TestMembersCommand_statusFilter.server.raft: entering leader state: leader="Node at 127.0.0.1:25339 [Leader]" === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.092Z [INFO] TestMembersCommand_verticalBar.server: New leader elected: payload=name|with|bars === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.106Z [INFO] TestMembersCommand_statusFilter.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:37.106Z [INFO] TestMembersCommand_statusFilter.server: New leader elected: payload=Node-8968bb3b-1e85-f0aa-975d-16adbcd69827 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.133Z [DEBUG] TestMembersCommand_verticalBar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25333 === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.141Z [DEBUG] TestMembersCommand_statusFilter.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25339 === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:37.157Z [INFO] TestMembersCommand_statusFilter_failed.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:37.159Z [INFO] TestMembersCommand_statusFilter_failed.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.161Z [DEBUG] TestMembersCommand_statusFilter_failed.server: Skipping self join check for node since the cluster is too small: node=Node-d985e87b-c997-00d1-3b15-7f3d2bba8a72 writer.go:29: 2021-01-29T19:34:37.163Z [INFO] TestMembersCommand_statusFilter_failed.server: member joined, marking health alive: member=Node-d985e87b-c997-00d1-3b15-7f3d2bba8a72 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.164Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.170Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:37.172Z [INFO] TestMembersCommand_WAN.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:37.182Z [INFO] TestMembersCommand_WAN.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.188Z [DEBUG] TestMembersCommand_WAN.server: Skipping self join check for node since the cluster is too small: node=Node-3b234eb8-b9c9-4944-f924-3abc4825b4f4 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.188Z [INFO] TestMembersCommand_verticalBar.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:37.198Z [INFO] TestMembersCommand_WAN.server: member joined, marking health alive: member=Node-3b234eb8-b9c9-4944-f924-3abc4825b4f4 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.199Z [INFO] TestMembersCommand_verticalBar.leader: started routine: routine="CA root pruning" === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.205Z [INFO] TestMembersCommand_statusFilter.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:37.206Z [INFO] TestMembersCommand_statusFilter.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.208Z [DEBUG] TestMembersCommand_statusFilter.server: Skipping self join check for node since the cluster is too small: node=Node-8968bb3b-1e85-f0aa-975d-16adbcd69827 writer.go:29: 2021-01-29T19:34:37.209Z [INFO] TestMembersCommand_statusFilter.server: member joined, marking health alive: member=Node-8968bb3b-1e85-f0aa-975d-16adbcd69827 === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.205Z [DEBUG] TestMembersCommand_verticalBar.server: Skipping self join check for node since the cluster is too small: node=name|with|bars writer.go:29: 2021-01-29T19:34:37.213Z [INFO] TestMembersCommand_verticalBar.server: member joined, marking health alive: member=name|with|bars writer.go:29: 2021-01-29T19:34:37.272Z [DEBUG] TestMembersCommand_verticalBar.http: Request finished: method=GET url=/v1/agent/members?segment=_all from=127.0.0.1:39892 latency=4.690393ms writer.go:29: 2021-01-29T19:34:37.279Z [INFO] TestMembersCommand_verticalBar: Requesting shutdown writer.go:29: 2021-01-29T19:34:37.282Z [INFO] TestMembersCommand_verticalBar.server: shutting down server writer.go:29: 2021-01-29T19:34:37.284Z [DEBUG] TestMembersCommand_verticalBar.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.287Z [WARN] TestMembersCommand_verticalBar.server.serf.lan: serf: Shutdown without a Leave === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.293Z [DEBUG] TestMembersCommand_statusFilter.http: Request finished: method=GET url=/v1/agent/members?segment=_all from=127.0.0.1:39288 latency=2.095688ms writer.go:29: 2021-01-29T19:34:37.295Z [INFO] TestMembersCommand_statusFilter: Requesting shutdown writer.go:29: 2021-01-29T19:34:37.297Z [INFO] TestMembersCommand_statusFilter.server: shutting down server writer.go:29: 2021-01-29T19:34:37.298Z [DEBUG] TestMembersCommand_statusFilter.leader: stopping routine: routine="CA root pruning" === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.284Z [ERROR] TestMembersCommand_verticalBar.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.300Z [WARN] TestMembersCommand_statusFilter.server.serf.lan: serf: Shutdown without a Leave === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.287Z [DEBUG] TestMembersCommand_verticalBar.leader: stopped routine: routine="CA root pruning" === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.302Z [ERROR] TestMembersCommand_statusFilter.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:37.302Z [DEBUG] TestMembersCommand_statusFilter.leader: stopped routine: routine="CA root pruning" === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.302Z [WARN] TestMembersCommand_verticalBar.server.serf.wan: serf: Shutdown without a Leave === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.312Z [WARN] TestMembersCommand_statusFilter.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:37.316Z [INFO] TestMembersCommand_statusFilter.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:37.318Z [INFO] TestMembersCommand_statusFilter: consul server down === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.316Z [INFO] TestMembersCommand_verticalBar.server.router.manager: shutting down === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.320Z [INFO] TestMembersCommand_statusFilter: shutdown complete === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.321Z [INFO] TestMembersCommand_verticalBar: consul server down === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:37.321Z [DEBUG] TestMembersCommand_statusFilter_failed: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.323Z [INFO] TestMembersCommand_verticalBar: shutdown complete writer.go:29: 2021-01-29T19:34:37.324Z [INFO] TestMembersCommand_verticalBar: Stopping server: protocol=DNS address=127.0.0.1:25328 network=tcp === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:37.325Z [INFO] TestMembersCommand_statusFilter_failed: Synced node info === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.326Z [INFO] TestMembersCommand_verticalBar: Stopping server: protocol=DNS address=127.0.0.1:25328 network=udp === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.321Z [INFO] TestMembersCommand_statusFilter: Stopping server: protocol=DNS address=127.0.0.1:25334 network=tcp === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.328Z [INFO] TestMembersCommand_verticalBar: Stopping server: protocol=HTTP address=127.0.0.1:25329 network=tcp === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.329Z [INFO] TestMembersCommand_statusFilter: Stopping server: protocol=DNS address=127.0.0.1:25334 network=udp writer.go:29: 2021-01-29T19:34:37.331Z [INFO] TestMembersCommand_statusFilter: Stopping server: protocol=HTTP address=127.0.0.1:25335 network=tcp === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:37.367Z [DEBUG] TestMembersCommand_statusFilter_failed.http: Request finished: method=GET url=/v1/agent/members?segment=_all from=127.0.0.1:55494 latency=256.344µs writer.go:29: 2021-01-29T19:34:37.369Z [INFO] TestMembersCommand_statusFilter_failed: Requesting shutdown writer.go:29: 2021-01-29T19:34:37.371Z [INFO] TestMembersCommand_statusFilter_failed.server: shutting down server writer.go:29: 2021-01-29T19:34:37.373Z [DEBUG] TestMembersCommand_statusFilter_failed.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.375Z [WARN] TestMembersCommand_statusFilter_failed.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:37.375Z [DEBUG] TestMembersCommand_statusFilter_failed.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.380Z [WARN] TestMembersCommand_statusFilter_failed.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:37.383Z [INFO] TestMembersCommand_statusFilter_failed.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:37.384Z [INFO] TestMembersCommand_statusFilter_failed: consul server down writer.go:29: 2021-01-29T19:34:37.387Z [INFO] TestMembersCommand_statusFilter_failed: shutdown complete writer.go:29: 2021-01-29T19:34:37.389Z [INFO] TestMembersCommand_statusFilter_failed: Stopping server: protocol=DNS address=127.0.0.1:25316 network=tcp writer.go:29: 2021-01-29T19:34:37.391Z [INFO] TestMembersCommand_statusFilter_failed: Stopping server: protocol=DNS address=127.0.0.1:25316 network=udp writer.go:29: 2021-01-29T19:34:37.392Z [INFO] TestMembersCommand_statusFilter_failed: Stopping server: protocol=HTTP address=127.0.0.1:25317 network=tcp === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:37.408Z [DEBUG] TestMembersCommand_WAN.http: Request finished: method=GET url=/v1/agent/members?segment=_all&wan=1 from=127.0.0.1:34718 latency=115.542µs writer.go:29: 2021-01-29T19:34:37.410Z [INFO] TestMembersCommand_WAN: Requesting shutdown writer.go:29: 2021-01-29T19:34:37.411Z [INFO] TestMembersCommand_WAN.server: shutting down server writer.go:29: 2021-01-29T19:34:37.417Z [DEBUG] TestMembersCommand_WAN.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.422Z [WARN] TestMembersCommand_WAN.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:37.417Z [ERROR] TestMembersCommand_WAN.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:37.422Z [DEBUG] TestMembersCommand_WAN.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.429Z [WARN] TestMembersCommand_WAN.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:37.439Z [INFO] TestMembersCommand_WAN.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:37.441Z [INFO] TestMembersCommand_WAN: consul server down writer.go:29: 2021-01-29T19:34:37.442Z [INFO] TestMembersCommand_WAN: shutdown complete writer.go:29: 2021-01-29T19:34:37.444Z [INFO] TestMembersCommand_WAN: Stopping server: protocol=DNS address=127.0.0.1:25322 network=tcp writer.go:29: 2021-01-29T19:34:37.445Z [INFO] TestMembersCommand_WAN: Stopping server: protocol=DNS address=127.0.0.1:25322 network=udp writer.go:29: 2021-01-29T19:34:37.447Z [INFO] TestMembersCommand_WAN: Stopping server: protocol=HTTP address=127.0.0.1:25323 network=tcp === CONT TestMembersCommand_verticalBar writer.go:29: 2021-01-29T19:34:37.830Z [INFO] TestMembersCommand_verticalBar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:37.833Z [INFO] TestMembersCommand_verticalBar: Endpoints down --- PASS: TestMembersCommand_verticalBar (0.99s) === CONT TestMembersCommand === CONT TestMembersCommand_statusFilter writer.go:29: 2021-01-29T19:34:37.841Z [INFO] TestMembersCommand_statusFilter: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:37.843Z [INFO] TestMembersCommand_statusFilter: Endpoints down --- PASS: TestMembersCommand_statusFilter (1.00s) === CONT TestMembersCommand writer.go:29: 2021-01-29T19:34:37.849Z [WARN] TestMembersCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:37.852Z [DEBUG] TestMembersCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:37.854Z [DEBUG] TestMembersCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:37.860Z [INFO] TestMembersCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:1b1c9cd0-3590-3e6d-48a4-0b1ae5a982db Address:127.0.0.1:25345}]" writer.go:29: 2021-01-29T19:34:37.863Z [INFO] TestMembersCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:25345 [Follower]" leader= writer.go:29: 2021-01-29T19:34:37.865Z [INFO] TestMembersCommand.server.serf.wan: serf: EventMemberJoin: Node-1b1c9cd0-3590-3e6d-48a4-0b1ae5a982db.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:37.868Z [INFO] TestMembersCommand.server.serf.lan: serf: EventMemberJoin: Node-1b1c9cd0-3590-3e6d-48a4-0b1ae5a982db 127.0.0.1 writer.go:29: 2021-01-29T19:34:37.871Z [INFO] TestMembersCommand.server: Adding LAN server: server="Node-1b1c9cd0-3590-3e6d-48a4-0b1ae5a982db (Addr: tcp/127.0.0.1:25345) (DC: dc1)" writer.go:29: 2021-01-29T19:34:37.872Z [INFO] TestMembersCommand.server: Handled event for server in area: event=member-join server=Node-1b1c9cd0-3590-3e6d-48a4-0b1ae5a982db.dc1 area=wan writer.go:29: 2021-01-29T19:34:37.874Z [INFO] TestMembersCommand: Started DNS server: address=127.0.0.1:25340 network=tcp writer.go:29: 2021-01-29T19:34:37.875Z [INFO] TestMembersCommand: Started DNS server: address=127.0.0.1:25340 network=udp writer.go:29: 2021-01-29T19:34:37.877Z [INFO] TestMembersCommand: Started HTTP server: address=127.0.0.1:25341 network=tcp writer.go:29: 2021-01-29T19:34:37.878Z [INFO] TestMembersCommand: started state syncer === CONT TestMembersCommand_statusFilter_failed writer.go:29: 2021-01-29T19:34:37.894Z [INFO] TestMembersCommand_statusFilter_failed: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:37.896Z [INFO] TestMembersCommand_statusFilter_failed: Endpoints down --- PASS: TestMembersCommand_statusFilter_failed (1.06s) === CONT TestMembersCommand writer.go:29: 2021-01-29T19:34:37.925Z [WARN] TestMembersCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:37.927Z [INFO] TestMembersCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:25345 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:37.929Z [DEBUG] TestMembersCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:37.931Z [DEBUG] TestMembersCommand.server.raft: vote granted: from=1b1c9cd0-3590-3e6d-48a4-0b1ae5a982db term=2 tally=1 writer.go:29: 2021-01-29T19:34:37.932Z [INFO] TestMembersCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:37.933Z [INFO] TestMembersCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:25345 [Leader]" writer.go:29: 2021-01-29T19:34:37.934Z [INFO] TestMembersCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:37.935Z [INFO] TestMembersCommand.server: New leader elected: payload=Node-1b1c9cd0-3590-3e6d-48a4-0b1ae5a982db writer.go:29: 2021-01-29T19:34:37.938Z [DEBUG] TestMembersCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:25345 writer.go:29: 2021-01-29T19:34:37.943Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestMembersCommand_WAN writer.go:29: 2021-01-29T19:34:37.949Z [INFO] TestMembersCommand_WAN: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:37.955Z [INFO] TestMembersCommand_WAN: Endpoints down --- PASS: TestMembersCommand_WAN (1.12s) === CONT TestMembersCommand writer.go:29: 2021-01-29T19:34:37.957Z [INFO] TestMembersCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:37.959Z [INFO] TestMembersCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:37.961Z [DEBUG] TestMembersCommand.server: Skipping self join check for node since the cluster is too small: node=Node-1b1c9cd0-3590-3e6d-48a4-0b1ae5a982db writer.go:29: 2021-01-29T19:34:37.963Z [INFO] TestMembersCommand.server: member joined, marking health alive: member=Node-1b1c9cd0-3590-3e6d-48a4-0b1ae5a982db writer.go:29: 2021-01-29T19:34:38.020Z [DEBUG] TestMembersCommand.http: Request finished: method=GET url=/v1/agent/members?segment=_all from=127.0.0.1:45150 latency=217.54µs writer.go:29: 2021-01-29T19:34:38.025Z [INFO] TestMembersCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:38.028Z [INFO] TestMembersCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:38.032Z [DEBUG] TestMembersCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:38.034Z [ERROR] TestMembersCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:38.037Z [DEBUG] TestMembersCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:38.038Z [WARN] TestMembersCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:38.052Z [WARN] TestMembersCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:38.057Z [INFO] TestMembersCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:38.057Z [INFO] TestMembersCommand: consul server down writer.go:29: 2021-01-29T19:34:38.062Z [INFO] TestMembersCommand: shutdown complete writer.go:29: 2021-01-29T19:34:38.064Z [INFO] TestMembersCommand: Stopping server: protocol=DNS address=127.0.0.1:25340 network=tcp writer.go:29: 2021-01-29T19:34:38.066Z [INFO] TestMembersCommand: Stopping server: protocol=DNS address=127.0.0.1:25340 network=udp writer.go:29: 2021-01-29T19:34:38.068Z [INFO] TestMembersCommand: Stopping server: protocol=HTTP address=127.0.0.1:25341 network=tcp writer.go:29: 2021-01-29T19:34:38.570Z [INFO] TestMembersCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:38.573Z [INFO] TestMembersCommand: Endpoints down --- PASS: TestMembersCommand (0.74s) PASS ok github.com/hashicorp/consul/command/members 1.890s === RUN TestMonitorCommand_exitsOnSignalBeforeLinesArrive monitor_test.go:13: DM-skipped --- SKIP: TestMonitorCommand_exitsOnSignalBeforeLinesArrive (0.00s) === RUN TestMonitorCommand_LogJSONValidFlag monitor_test.go:67: DM-skipped --- SKIP: TestMonitorCommand_LogJSONValidFlag (0.00s) PASS ok github.com/hashicorp/consul/command/monitor 0.247s === RUN TestOperatorCommand_noTabs === PAUSE TestOperatorCommand_noTabs === CONT TestOperatorCommand_noTabs --- PASS: TestOperatorCommand_noTabs (0.00s) PASS ok github.com/hashicorp/consul/command/operator 0.007s === RUN TestOperatorAutopilotCommand_noTabs === PAUSE TestOperatorAutopilotCommand_noTabs === CONT TestOperatorAutopilotCommand_noTabs --- PASS: TestOperatorAutopilotCommand_noTabs (0.00s) PASS ok github.com/hashicorp/consul/command/operator/autopilot 0.013s === RUN TestOperatorAutopilotGetConfigCommand_noTabs === PAUSE TestOperatorAutopilotGetConfigCommand_noTabs === RUN TestOperatorAutopilotGetConfigCommand === PAUSE TestOperatorAutopilotGetConfigCommand === CONT TestOperatorAutopilotGetConfigCommand_noTabs === CONT TestOperatorAutopilotGetConfigCommand --- PASS: TestOperatorAutopilotGetConfigCommand_noTabs (0.00s) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestOperatorAutopilotGetConfigCommand writer.go:29: 2021-01-29T19:34:43.341Z [WARN] TestOperatorAutopilotGetConfigCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:43.343Z [DEBUG] TestOperatorAutopilotGetConfigCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:43.347Z [DEBUG] TestOperatorAutopilotGetConfigCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:43.366Z [INFO] TestOperatorAutopilotGetConfigCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c517b556-f4c2-d820-9eae-6e198c1f3d1a Address:127.0.0.1:30426}]" writer.go:29: 2021-01-29T19:34:43.368Z [INFO] TestOperatorAutopilotGetConfigCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:30426 [Follower]" leader= writer.go:29: 2021-01-29T19:34:43.376Z [INFO] TestOperatorAutopilotGetConfigCommand.server.serf.wan: serf: EventMemberJoin: Node-c517b556-f4c2-d820-9eae-6e198c1f3d1a.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:43.382Z [INFO] TestOperatorAutopilotGetConfigCommand.server.serf.lan: serf: EventMemberJoin: Node-c517b556-f4c2-d820-9eae-6e198c1f3d1a 127.0.0.1 writer.go:29: 2021-01-29T19:34:43.387Z [INFO] TestOperatorAutopilotGetConfigCommand.server: Adding LAN server: server="Node-c517b556-f4c2-d820-9eae-6e198c1f3d1a (Addr: tcp/127.0.0.1:30426) (DC: dc1)" writer.go:29: 2021-01-29T19:34:43.389Z [INFO] TestOperatorAutopilotGetConfigCommand.server: Handled event for server in area: event=member-join server=Node-c517b556-f4c2-d820-9eae-6e198c1f3d1a.dc1 area=wan writer.go:29: 2021-01-29T19:34:43.392Z [INFO] TestOperatorAutopilotGetConfigCommand: Started DNS server: address=127.0.0.1:30421 network=tcp writer.go:29: 2021-01-29T19:34:43.396Z [INFO] TestOperatorAutopilotGetConfigCommand: Started DNS server: address=127.0.0.1:30421 network=udp writer.go:29: 2021-01-29T19:34:43.401Z [INFO] TestOperatorAutopilotGetConfigCommand: Started HTTP server: address=127.0.0.1:30422 network=tcp writer.go:29: 2021-01-29T19:34:43.402Z [INFO] TestOperatorAutopilotGetConfigCommand: started state syncer writer.go:29: 2021-01-29T19:34:43.429Z [WARN] TestOperatorAutopilotGetConfigCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:43.430Z [INFO] TestOperatorAutopilotGetConfigCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:30426 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:43.439Z [DEBUG] TestOperatorAutopilotGetConfigCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:43.441Z [DEBUG] TestOperatorAutopilotGetConfigCommand.server.raft: vote granted: from=c517b556-f4c2-d820-9eae-6e198c1f3d1a term=2 tally=1 writer.go:29: 2021-01-29T19:34:43.442Z [INFO] TestOperatorAutopilotGetConfigCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:43.444Z [INFO] TestOperatorAutopilotGetConfigCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:30426 [Leader]" writer.go:29: 2021-01-29T19:34:43.445Z [INFO] TestOperatorAutopilotGetConfigCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:43.447Z [INFO] TestOperatorAutopilotGetConfigCommand.server: New leader elected: payload=Node-c517b556-f4c2-d820-9eae-6e198c1f3d1a writer.go:29: 2021-01-29T19:34:43.449Z [DEBUG] TestOperatorAutopilotGetConfigCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30426 writer.go:29: 2021-01-29T19:34:43.457Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:43.539Z [INFO] TestOperatorAutopilotGetConfigCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:43.545Z [INFO] TestOperatorAutopilotGetConfigCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:43.551Z [DEBUG] TestOperatorAutopilotGetConfigCommand.server: Skipping self join check for node since the cluster is too small: node=Node-c517b556-f4c2-d820-9eae-6e198c1f3d1a writer.go:29: 2021-01-29T19:34:43.556Z [INFO] TestOperatorAutopilotGetConfigCommand.server: member joined, marking health alive: member=Node-c517b556-f4c2-d820-9eae-6e198c1f3d1a writer.go:29: 2021-01-29T19:34:43.666Z [DEBUG] TestOperatorAutopilotGetConfigCommand.http: Request finished: method=GET url=/v1/operator/autopilot/configuration from=127.0.0.1:57182 latency=3.005228ms writer.go:29: 2021-01-29T19:34:43.673Z [INFO] TestOperatorAutopilotGetConfigCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:43.675Z [INFO] TestOperatorAutopilotGetConfigCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:43.677Z [DEBUG] TestOperatorAutopilotGetConfigCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:43.693Z [WARN] TestOperatorAutopilotGetConfigCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:43.679Z [ERROR] TestOperatorAutopilotGetConfigCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:43.695Z [DEBUG] TestOperatorAutopilotGetConfigCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:43.701Z [WARN] TestOperatorAutopilotGetConfigCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:43.718Z [INFO] TestOperatorAutopilotGetConfigCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:43.718Z [INFO] TestOperatorAutopilotGetConfigCommand: consul server down writer.go:29: 2021-01-29T19:34:43.722Z [INFO] TestOperatorAutopilotGetConfigCommand: shutdown complete writer.go:29: 2021-01-29T19:34:43.726Z [INFO] TestOperatorAutopilotGetConfigCommand: Stopping server: protocol=DNS address=127.0.0.1:30421 network=tcp writer.go:29: 2021-01-29T19:34:43.729Z [INFO] TestOperatorAutopilotGetConfigCommand: Stopping server: protocol=DNS address=127.0.0.1:30421 network=udp writer.go:29: 2021-01-29T19:34:43.732Z [INFO] TestOperatorAutopilotGetConfigCommand: Stopping server: protocol=HTTP address=127.0.0.1:30422 network=tcp writer.go:29: 2021-01-29T19:34:44.236Z [INFO] TestOperatorAutopilotGetConfigCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:44.237Z [INFO] TestOperatorAutopilotGetConfigCommand: Endpoints down --- PASS: TestOperatorAutopilotGetConfigCommand (0.99s) PASS ok github.com/hashicorp/consul/command/operator/autopilot/get 1.264s === RUN TestOperatorAutopilotSetConfigCommand_noTabs === PAUSE TestOperatorAutopilotSetConfigCommand_noTabs === RUN TestOperatorAutopilotSetConfigCommand === PAUSE TestOperatorAutopilotSetConfigCommand === CONT TestOperatorAutopilotSetConfigCommand_noTabs === CONT TestOperatorAutopilotSetConfigCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestOperatorAutopilotSetConfigCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestOperatorAutopilotSetConfigCommand writer.go:29: 2021-01-29T19:34:48.571Z [WARN] TestOperatorAutopilotSetConfigCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:48.575Z [DEBUG] TestOperatorAutopilotSetConfigCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:48.580Z [DEBUG] TestOperatorAutopilotSetConfigCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:48.591Z [INFO] TestOperatorAutopilotSetConfigCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:3f7ef8f4-42c8-4c3a-856b-da4b396bede6 Address:127.0.0.1:21237}]" writer.go:29: 2021-01-29T19:34:48.595Z [INFO] TestOperatorAutopilotSetConfigCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:21237 [Follower]" leader= writer.go:29: 2021-01-29T19:34:48.599Z [INFO] TestOperatorAutopilotSetConfigCommand.server.serf.wan: serf: EventMemberJoin: Node-3f7ef8f4-42c8-4c3a-856b-da4b396bede6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:48.607Z [INFO] TestOperatorAutopilotSetConfigCommand.server.serf.lan: serf: EventMemberJoin: Node-3f7ef8f4-42c8-4c3a-856b-da4b396bede6 127.0.0.1 writer.go:29: 2021-01-29T19:34:48.612Z [INFO] TestOperatorAutopilotSetConfigCommand.server: Handled event for server in area: event=member-join server=Node-3f7ef8f4-42c8-4c3a-856b-da4b396bede6.dc1 area=wan writer.go:29: 2021-01-29T19:34:48.615Z [INFO] TestOperatorAutopilotSetConfigCommand.server: Adding LAN server: server="Node-3f7ef8f4-42c8-4c3a-856b-da4b396bede6 (Addr: tcp/127.0.0.1:21237) (DC: dc1)" writer.go:29: 2021-01-29T19:34:48.616Z [INFO] TestOperatorAutopilotSetConfigCommand: Started DNS server: address=127.0.0.1:21232 network=udp writer.go:29: 2021-01-29T19:34:48.633Z [INFO] TestOperatorAutopilotSetConfigCommand: Started DNS server: address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:34:48.635Z [INFO] TestOperatorAutopilotSetConfigCommand: Started HTTP server: address=127.0.0.1:21233 network=tcp writer.go:29: 2021-01-29T19:34:48.638Z [INFO] TestOperatorAutopilotSetConfigCommand: started state syncer writer.go:29: 2021-01-29T19:34:48.665Z [WARN] TestOperatorAutopilotSetConfigCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:48.667Z [INFO] TestOperatorAutopilotSetConfigCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:21237 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:48.671Z [DEBUG] TestOperatorAutopilotSetConfigCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:48.674Z [DEBUG] TestOperatorAutopilotSetConfigCommand.server.raft: vote granted: from=3f7ef8f4-42c8-4c3a-856b-da4b396bede6 term=2 tally=1 writer.go:29: 2021-01-29T19:34:48.679Z [INFO] TestOperatorAutopilotSetConfigCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:48.685Z [INFO] TestOperatorAutopilotSetConfigCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:21237 [Leader]" writer.go:29: 2021-01-29T19:34:48.688Z [INFO] TestOperatorAutopilotSetConfigCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:48.692Z [INFO] TestOperatorAutopilotSetConfigCommand.server: New leader elected: payload=Node-3f7ef8f4-42c8-4c3a-856b-da4b396bede6 writer.go:29: 2021-01-29T19:34:48.696Z [DEBUG] TestOperatorAutopilotSetConfigCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:21237 writer.go:29: 2021-01-29T19:34:48.705Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:48.813Z [INFO] TestOperatorAutopilotSetConfigCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:48.816Z [INFO] TestOperatorAutopilotSetConfigCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:48.818Z [DEBUG] TestOperatorAutopilotSetConfigCommand.server: Skipping self join check for node since the cluster is too small: node=Node-3f7ef8f4-42c8-4c3a-856b-da4b396bede6 writer.go:29: 2021-01-29T19:34:48.820Z [INFO] TestOperatorAutopilotSetConfigCommand.server: member joined, marking health alive: member=Node-3f7ef8f4-42c8-4c3a-856b-da4b396bede6 writer.go:29: 2021-01-29T19:34:48.866Z [DEBUG] TestOperatorAutopilotSetConfigCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:48.869Z [INFO] TestOperatorAutopilotSetConfigCommand: Synced node info writer.go:29: 2021-01-29T19:34:48.975Z [DEBUG] TestOperatorAutopilotSetConfigCommand.http: Request finished: method=GET url=/v1/operator/autopilot/configuration from=127.0.0.1:43500 latency=25.690042ms writer.go:29: 2021-01-29T19:34:49.023Z [DEBUG] TestOperatorAutopilotSetConfigCommand.http: Request finished: method=PUT url=/v1/operator/autopilot/configuration?cas=4 from=127.0.0.1:43500 latency=30.878535ms writer.go:29: 2021-01-29T19:34:49.030Z [INFO] TestOperatorAutopilotSetConfigCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:49.032Z [INFO] TestOperatorAutopilotSetConfigCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:49.036Z [DEBUG] TestOperatorAutopilotSetConfigCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:49.041Z [WARN] TestOperatorAutopilotSetConfigCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:49.041Z [DEBUG] TestOperatorAutopilotSetConfigCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:49.052Z [WARN] TestOperatorAutopilotSetConfigCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:49.058Z [INFO] TestOperatorAutopilotSetConfigCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:49.058Z [INFO] TestOperatorAutopilotSetConfigCommand: consul server down writer.go:29: 2021-01-29T19:34:49.062Z [INFO] TestOperatorAutopilotSetConfigCommand: shutdown complete writer.go:29: 2021-01-29T19:34:49.064Z [INFO] TestOperatorAutopilotSetConfigCommand: Stopping server: protocol=DNS address=127.0.0.1:21232 network=tcp writer.go:29: 2021-01-29T19:34:49.066Z [INFO] TestOperatorAutopilotSetConfigCommand: Stopping server: protocol=DNS address=127.0.0.1:21232 network=udp writer.go:29: 2021-01-29T19:34:49.068Z [INFO] TestOperatorAutopilotSetConfigCommand: Stopping server: protocol=HTTP address=127.0.0.1:21233 network=tcp writer.go:29: 2021-01-29T19:34:49.570Z [INFO] TestOperatorAutopilotSetConfigCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:49.572Z [INFO] TestOperatorAutopilotSetConfigCommand: Endpoints down --- PASS: TestOperatorAutopilotSetConfigCommand (1.07s) PASS ok github.com/hashicorp/consul/command/operator/autopilot/set 1.245s === RUN TestOperatorRaftCommand_noTabs === PAUSE TestOperatorRaftCommand_noTabs === CONT TestOperatorRaftCommand_noTabs --- PASS: TestOperatorRaftCommand_noTabs (0.00s) PASS ok github.com/hashicorp/consul/command/operator/raft 0.015s === RUN TestOperatorRaftListPeersCommand_noTabs === PAUSE TestOperatorRaftListPeersCommand_noTabs === RUN TestOperatorRaftListPeersCommand === PAUSE TestOperatorRaftListPeersCommand === RUN TestOperatorRaftListPeersCommand_verticalBar === PAUSE TestOperatorRaftListPeersCommand_verticalBar === CONT TestOperatorRaftListPeersCommand_noTabs === CONT TestOperatorRaftListPeersCommand_verticalBar === CONT TestOperatorRaftListPeersCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestOperatorRaftListPeersCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:49.864Z [WARN] TestOperatorRaftListPeersCommand_verticalBar: bootstrap = true: do not enable unless necessary === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:49.868Z [WARN] TestOperatorRaftListPeersCommand: bootstrap = true: do not enable unless necessary === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:49.876Z [WARN] TestOperatorRaftListPeersCommand_verticalBar: Node name will not be discoverable via DNS due to invalid characters. Valid characters include all alpha-numerics and dashes.: node_name=name|with|bars writer.go:29: 2021-01-29T19:34:49.892Z [DEBUG] TestOperatorRaftListPeersCommand_verticalBar.tlsutil: Update: version=1 === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:49.873Z [DEBUG] TestOperatorRaftListPeersCommand.tlsutil: Update: version=1 === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:49.896Z [DEBUG] TestOperatorRaftListPeersCommand_verticalBar.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:49.897Z [DEBUG] TestOperatorRaftListPeersCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:49.910Z [INFO] TestOperatorRaftListPeersCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e170d241-2437-f73e-f302-3130011f359c Address:127.0.0.1:15117}]" writer.go:29: 2021-01-29T19:34:49.914Z [INFO] TestOperatorRaftListPeersCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:15117 [Follower]" leader= === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:49.917Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:4d2d375f-c5d3-fe95-8eb7-9a1b22af41da Address:127.0.0.1:15111}]" === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:49.916Z [INFO] TestOperatorRaftListPeersCommand.server.serf.wan: serf: EventMemberJoin: Node-e170d241-2437-f73e-f302-3130011f359c.dc1 127.0.0.1 === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:49.920Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server.raft: entering follower state: follower="Node at 127.0.0.1:15111 [Follower]" leader= === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:49.930Z [INFO] TestOperatorRaftListPeersCommand.server.serf.lan: serf: EventMemberJoin: Node-e170d241-2437-f73e-f302-3130011f359c 127.0.0.1 writer.go:29: 2021-01-29T19:34:49.934Z [INFO] TestOperatorRaftListPeersCommand: Started DNS server: address=127.0.0.1:15112 network=udp writer.go:29: 2021-01-29T19:34:49.935Z [INFO] TestOperatorRaftListPeersCommand.server: Handled event for server in area: event=member-join server=Node-e170d241-2437-f73e-f302-3130011f359c.dc1 area=wan === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:49.927Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server.serf.wan: serf: EventMemberJoin: name|with|bars.dc1 127.0.0.1 === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:49.938Z [INFO] TestOperatorRaftListPeersCommand.server: Adding LAN server: server="Node-e170d241-2437-f73e-f302-3130011f359c (Addr: tcp/127.0.0.1:15117) (DC: dc1)" writer.go:29: 2021-01-29T19:34:49.955Z [INFO] TestOperatorRaftListPeersCommand: Started DNS server: address=127.0.0.1:15112 network=tcp === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:49.956Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server.serf.lan: serf: EventMemberJoin: name|with|bars 127.0.0.1 writer.go:29: 2021-01-29T19:34:49.964Z [WARN] TestOperatorRaftListPeersCommand_verticalBar.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:49.967Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server.raft: entering candidate state: node="Node at 127.0.0.1:15111 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:49.972Z [WARN] TestOperatorRaftListPeersCommand_verticalBar.server.raft: unable to get address for sever, using fallback address: id=4d2d375f-c5d3-fe95-8eb7-9a1b22af41da fallback=127.0.0.1:15111 error="Could not find address for server id 4d2d375f-c5d3-fe95-8eb7-9a1b22af41da" writer.go:29: 2021-01-29T19:34:49.974Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server: Adding LAN server: server="name|with|bars (Addr: tcp/127.0.0.1:15111) (DC: dc1)" writer.go:29: 2021-01-29T19:34:49.976Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: Started DNS server: address=127.0.0.1:15106 network=udp writer.go:29: 2021-01-29T19:34:49.978Z [DEBUG] TestOperatorRaftListPeersCommand_verticalBar.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:49.981Z [DEBUG] TestOperatorRaftListPeersCommand_verticalBar.server.raft: vote granted: from=4d2d375f-c5d3-fe95-8eb7-9a1b22af41da term=2 tally=1 writer.go:29: 2021-01-29T19:34:49.983Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server.raft: election won: tally=1 === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:49.984Z [INFO] TestOperatorRaftListPeersCommand: Started HTTP server: address=127.0.0.1:15113 network=tcp === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:49.982Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: Started DNS server: address=127.0.0.1:15106 network=tcp === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:49.987Z [WARN] TestOperatorRaftListPeersCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:49.991Z [INFO] TestOperatorRaftListPeersCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:15117 [Candidate]" term=2 === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:49.983Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server: Handled event for server in area: event=member-join server=name|with|bars.dc1 area=wan === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:49.989Z [INFO] TestOperatorRaftListPeersCommand: started state syncer === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:49.985Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server.raft: entering leader state: leader="Node at 127.0.0.1:15111 [Leader]" writer.go:29: 2021-01-29T19:34:49.994Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: Started HTTP server: address=127.0.0.1:15107 network=tcp writer.go:29: 2021-01-29T19:34:49.998Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: started state syncer writer.go:29: 2021-01-29T19:34:49.996Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:49.997Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server: New leader elected: payload=name|with|bars === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.005Z [DEBUG] TestOperatorRaftListPeersCommand.server.raft: votes: needed=1 === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:50.005Z [DEBUG] TestOperatorRaftListPeersCommand_verticalBar.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15111 === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.007Z [DEBUG] TestOperatorRaftListPeersCommand.server.raft: vote granted: from=e170d241-2437-f73e-f302-3130011f359c term=2 tally=1 writer.go:29: 2021-01-29T19:34:50.009Z [INFO] TestOperatorRaftListPeersCommand.server.raft: election won: tally=1 === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:50.017Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: Synced node info === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.011Z [INFO] TestOperatorRaftListPeersCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:15117 [Leader]" writer.go:29: 2021-01-29T19:34:50.020Z [INFO] TestOperatorRaftListPeersCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:50.021Z [INFO] TestOperatorRaftListPeersCommand.server: New leader elected: payload=Node-e170d241-2437-f73e-f302-3130011f359c === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:50.023Z [DEBUG] TestOperatorRaftListPeersCommand_verticalBar.http: Request finished: method=GET url=/v1/operator/raft/configuration from=127.0.0.1:37750 latency=13.171139ms writer.go:29: 2021-01-29T19:34:50.029Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:50.030Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: Requesting shutdown writer.go:29: 2021-01-29T19:34:50.033Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server: shutting down server writer.go:29: 2021-01-29T19:34:50.035Z [WARN] TestOperatorRaftListPeersCommand_verticalBar.server.serf.lan: serf: Shutdown without a Leave === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.025Z [DEBUG] TestOperatorRaftListPeersCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:15117 === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:50.040Z [WARN] TestOperatorRaftListPeersCommand_verticalBar.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:50.043Z [INFO] TestOperatorRaftListPeersCommand_verticalBar.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:50.044Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: consul server down writer.go:29: 2021-01-29T19:34:50.048Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: shutdown complete writer.go:29: 2021-01-29T19:34:50.050Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: Stopping server: protocol=DNS address=127.0.0.1:15106 network=tcp === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.051Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:50.056Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: Stopping server: protocol=DNS address=127.0.0.1:15106 network=udp === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.064Z [DEBUG] TestOperatorRaftListPeersCommand.http: Request finished: method=GET url=/v1/operator/raft/configuration from=127.0.0.1:39208 latency=373.44µs === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:50.064Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: Stopping server: protocol=HTTP address=127.0.0.1:15107 network=tcp === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.066Z [INFO] TestOperatorRaftListPeersCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:50.068Z [INFO] TestOperatorRaftListPeersCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:50.070Z [WARN] TestOperatorRaftListPeersCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:50.070Z [ERROR] TestOperatorRaftListPeersCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:50.077Z [WARN] TestOperatorRaftListPeersCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:50.094Z [INFO] TestOperatorRaftListPeersCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:50.095Z [INFO] TestOperatorRaftListPeersCommand: consul server down writer.go:29: 2021-01-29T19:34:50.097Z [INFO] TestOperatorRaftListPeersCommand: shutdown complete writer.go:29: 2021-01-29T19:34:50.099Z [INFO] TestOperatorRaftListPeersCommand: Stopping server: protocol=DNS address=127.0.0.1:15112 network=tcp writer.go:29: 2021-01-29T19:34:50.100Z [INFO] TestOperatorRaftListPeersCommand: Stopping server: protocol=DNS address=127.0.0.1:15112 network=udp writer.go:29: 2021-01-29T19:34:50.110Z [INFO] TestOperatorRaftListPeersCommand: Stopping server: protocol=HTTP address=127.0.0.1:15113 network=tcp === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:50.114Z [ERROR] TestOperatorRaftListPeersCommand_verticalBar.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: raft is already shutdown" === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.114Z [ERROR] TestOperatorRaftListPeersCommand.server: failed to establish leadership: error="error generating CA root certificate: error computing next serial number: raft is already shutdown" === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:50.116Z [ERROR] TestOperatorRaftListPeersCommand_verticalBar.server: failed to transfer leadership attempt, will retry: attempt=0 retry_limit=3 error="raft is already shutdown" === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.116Z [ERROR] TestOperatorRaftListPeersCommand.server: failed to transfer leadership attempt, will retry: attempt=0 retry_limit=3 error="raft is already shutdown" === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:50.118Z [ERROR] TestOperatorRaftListPeersCommand_verticalBar.server: failed to transfer leadership attempt, will retry: attempt=1 retry_limit=3 error="raft is already shutdown" === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.119Z [ERROR] TestOperatorRaftListPeersCommand.server: failed to transfer leadership attempt, will retry: attempt=1 retry_limit=3 error="raft is already shutdown" === CONT TestOperatorRaftListPeersCommand_verticalBar writer.go:29: 2021-01-29T19:34:50.119Z [ERROR] TestOperatorRaftListPeersCommand_verticalBar.server: failed to transfer leadership attempt, will retry: attempt=2 retry_limit=3 error="raft is already shutdown" writer.go:29: 2021-01-29T19:34:50.121Z [ERROR] TestOperatorRaftListPeersCommand_verticalBar.server: failed to transfer leadership: error="failed to transfer leadership in 3 attempts" writer.go:29: 2021-01-29T19:34:50.568Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:50.571Z [INFO] TestOperatorRaftListPeersCommand_verticalBar: Endpoints down --- PASS: TestOperatorRaftListPeersCommand_verticalBar (0.80s) === CONT TestOperatorRaftListPeersCommand writer.go:29: 2021-01-29T19:34:50.612Z [INFO] TestOperatorRaftListPeersCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:50.614Z [INFO] TestOperatorRaftListPeersCommand: Endpoints down --- PASS: TestOperatorRaftListPeersCommand (0.84s) PASS ok github.com/hashicorp/consul/command/operator/raft/listpeers 1.085s === RUN TestOperatorRaftRemovePeerCommand_noTabs === PAUSE TestOperatorRaftRemovePeerCommand_noTabs === RUN TestOperatorRaftRemovePeerCommand === PAUSE TestOperatorRaftRemovePeerCommand === CONT TestOperatorRaftRemovePeerCommand_noTabs === CONT TestOperatorRaftRemovePeerCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestOperatorRaftRemovePeerCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestOperatorRaftRemovePeerCommand writer.go:29: 2021-01-29T19:34:53.531Z [WARN] TestOperatorRaftRemovePeerCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:53.532Z [DEBUG] TestOperatorRaftRemovePeerCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:53.534Z [DEBUG] TestOperatorRaftRemovePeerCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:53.539Z [INFO] TestOperatorRaftRemovePeerCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:279ac6fa-1133-13da-6aed-3e6b3d2d1429 Address:127.0.0.1:13069}]" writer.go:29: 2021-01-29T19:34:53.551Z [INFO] TestOperatorRaftRemovePeerCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:13069 [Follower]" leader= writer.go:29: 2021-01-29T19:34:53.571Z [INFO] TestOperatorRaftRemovePeerCommand.server.serf.wan: serf: EventMemberJoin: Node-279ac6fa-1133-13da-6aed-3e6b3d2d1429.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:53.574Z [INFO] TestOperatorRaftRemovePeerCommand.server.serf.lan: serf: EventMemberJoin: Node-279ac6fa-1133-13da-6aed-3e6b3d2d1429 127.0.0.1 writer.go:29: 2021-01-29T19:34:53.577Z [INFO] TestOperatorRaftRemovePeerCommand.server: Adding LAN server: server="Node-279ac6fa-1133-13da-6aed-3e6b3d2d1429 (Addr: tcp/127.0.0.1:13069) (DC: dc1)" writer.go:29: 2021-01-29T19:34:53.578Z [INFO] TestOperatorRaftRemovePeerCommand.server: Handled event for server in area: event=member-join server=Node-279ac6fa-1133-13da-6aed-3e6b3d2d1429.dc1 area=wan writer.go:29: 2021-01-29T19:34:53.584Z [INFO] TestOperatorRaftRemovePeerCommand: Started DNS server: address=127.0.0.1:13064 network=tcp writer.go:29: 2021-01-29T19:34:53.585Z [INFO] TestOperatorRaftRemovePeerCommand: Started DNS server: address=127.0.0.1:13064 network=udp writer.go:29: 2021-01-29T19:34:53.588Z [INFO] TestOperatorRaftRemovePeerCommand: Started HTTP server: address=127.0.0.1:13065 network=tcp writer.go:29: 2021-01-29T19:34:53.589Z [INFO] TestOperatorRaftRemovePeerCommand: started state syncer writer.go:29: 2021-01-29T19:34:53.604Z [WARN] TestOperatorRaftRemovePeerCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:53.605Z [INFO] TestOperatorRaftRemovePeerCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:13069 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:53.606Z [DEBUG] TestOperatorRaftRemovePeerCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:53.607Z [DEBUG] TestOperatorRaftRemovePeerCommand.server.raft: vote granted: from=279ac6fa-1133-13da-6aed-3e6b3d2d1429 term=2 tally=1 writer.go:29: 2021-01-29T19:34:53.608Z [INFO] TestOperatorRaftRemovePeerCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:53.609Z [INFO] TestOperatorRaftRemovePeerCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:13069 [Leader]" writer.go:29: 2021-01-29T19:34:53.610Z [INFO] TestOperatorRaftRemovePeerCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:53.610Z [INFO] TestOperatorRaftRemovePeerCommand.server: New leader elected: payload=Node-279ac6fa-1133-13da-6aed-3e6b3d2d1429 writer.go:29: 2021-01-29T19:34:53.611Z [DEBUG] TestOperatorRaftRemovePeerCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:13069 writer.go:29: 2021-01-29T19:34:53.622Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:53.690Z [INFO] TestOperatorRaftRemovePeerCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:53.691Z [INFO] TestOperatorRaftRemovePeerCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.691Z [DEBUG] TestOperatorRaftRemovePeerCommand.server: Skipping self join check for node since the cluster is too small: node=Node-279ac6fa-1133-13da-6aed-3e6b3d2d1429 writer.go:29: 2021-01-29T19:34:53.692Z [INFO] TestOperatorRaftRemovePeerCommand.server: member joined, marking health alive: member=Node-279ac6fa-1133-13da-6aed-3e6b3d2d1429 writer.go:29: 2021-01-29T19:34:53.710Z [DEBUG] TestOperatorRaftRemovePeerCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:53.722Z [INFO] TestOperatorRaftRemovePeerCommand: Synced node info === RUN TestOperatorRaftRemovePeerCommand/Test_the_remove-peer_subcommand_directly === CONT TestOperatorRaftRemovePeerCommand writer.go:29: 2021-01-29T19:34:53.908Z [ERROR] TestOperatorRaftRemovePeerCommand.http: Request error: method=DELETE url=/v1/operator/raft/peer?address=nope from=127.0.0.1:60278 error="address "nope" was not found in the Raft configuration" writer.go:29: 2021-01-29T19:34:53.911Z [DEBUG] TestOperatorRaftRemovePeerCommand.http: Request finished: method=DELETE url=/v1/operator/raft/peer?address=nope from=127.0.0.1:60278 latency=3.247813ms === RUN TestOperatorRaftRemovePeerCommand/Test_the_remove-peer_subcommand_with_-id === CONT TestOperatorRaftRemovePeerCommand writer.go:29: 2021-01-29T19:34:53.928Z [ERROR] TestOperatorRaftRemovePeerCommand.http: Request error: method=DELETE url=/v1/operator/raft/peer?id=nope from=127.0.0.1:60280 error="id "nope" was not found in the Raft configuration" writer.go:29: 2021-01-29T19:34:53.930Z [DEBUG] TestOperatorRaftRemovePeerCommand.http: Request finished: method=DELETE url=/v1/operator/raft/peer?id=nope from=127.0.0.1:60280 latency=1.999564ms writer.go:29: 2021-01-29T19:34:53.932Z [INFO] TestOperatorRaftRemovePeerCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:53.936Z [INFO] TestOperatorRaftRemovePeerCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:53.937Z [DEBUG] TestOperatorRaftRemovePeerCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.940Z [WARN] TestOperatorRaftRemovePeerCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:53.940Z [DEBUG] TestOperatorRaftRemovePeerCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:53.944Z [WARN] TestOperatorRaftRemovePeerCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:53.948Z [INFO] TestOperatorRaftRemovePeerCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:53.950Z [INFO] TestOperatorRaftRemovePeerCommand: consul server down writer.go:29: 2021-01-29T19:34:53.952Z [INFO] TestOperatorRaftRemovePeerCommand: shutdown complete writer.go:29: 2021-01-29T19:34:53.954Z [INFO] TestOperatorRaftRemovePeerCommand: Stopping server: protocol=DNS address=127.0.0.1:13064 network=tcp writer.go:29: 2021-01-29T19:34:53.956Z [INFO] TestOperatorRaftRemovePeerCommand: Stopping server: protocol=DNS address=127.0.0.1:13064 network=udp writer.go:29: 2021-01-29T19:34:53.958Z [INFO] TestOperatorRaftRemovePeerCommand: Stopping server: protocol=HTTP address=127.0.0.1:13065 network=tcp writer.go:29: 2021-01-29T19:34:54.460Z [INFO] TestOperatorRaftRemovePeerCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:54.462Z [INFO] TestOperatorRaftRemovePeerCommand: Endpoints down --- PASS: TestOperatorRaftRemovePeerCommand (1.01s) --- PASS: TestOperatorRaftRemovePeerCommand/Test_the_remove-peer_subcommand_directly (0.02s) --- PASS: TestOperatorRaftRemovePeerCommand/Test_the_remove-peer_subcommand_with_-id (0.01s) PASS ok github.com/hashicorp/consul/command/operator/raft/removepeer 1.254s === RUN TestReloadCommand_noTabs === PAUSE TestReloadCommand_noTabs === RUN TestReloadCommand === PAUSE TestReloadCommand === CONT TestReloadCommand_noTabs === CONT TestReloadCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestReloadCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestReloadCommand writer.go:29: 2021-01-29T19:34:58.108Z [WARN] TestReloadCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:58.110Z [DEBUG] TestReloadCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:58.114Z [DEBUG] TestReloadCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:34:58.121Z [INFO] TestReloadCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:c9124fa3-5c4e-cb8b-28d9-44ac4692efff Address:127.0.0.1:23279}]" writer.go:29: 2021-01-29T19:34:58.124Z [INFO] TestReloadCommand.server.serf.wan: serf: EventMemberJoin: Node-c9124fa3-5c4e-cb8b-28d9-44ac4692efff.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:34:58.125Z [INFO] TestReloadCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:23279 [Follower]" leader= writer.go:29: 2021-01-29T19:34:58.128Z [INFO] TestReloadCommand.server.serf.lan: serf: EventMemberJoin: Node-c9124fa3-5c4e-cb8b-28d9-44ac4692efff 127.0.0.1 writer.go:29: 2021-01-29T19:34:58.135Z [INFO] TestReloadCommand.server: Handled event for server in area: event=member-join server=Node-c9124fa3-5c4e-cb8b-28d9-44ac4692efff.dc1 area=wan writer.go:29: 2021-01-29T19:34:58.136Z [INFO] TestReloadCommand.server: Adding LAN server: server="Node-c9124fa3-5c4e-cb8b-28d9-44ac4692efff (Addr: tcp/127.0.0.1:23279) (DC: dc1)" writer.go:29: 2021-01-29T19:34:58.140Z [INFO] TestReloadCommand: Started DNS server: address=127.0.0.1:23274 network=udp writer.go:29: 2021-01-29T19:34:58.142Z [INFO] TestReloadCommand: Started DNS server: address=127.0.0.1:23274 network=tcp writer.go:29: 2021-01-29T19:34:58.145Z [INFO] TestReloadCommand: Started HTTP server: address=127.0.0.1:23275 network=tcp writer.go:29: 2021-01-29T19:34:58.148Z [INFO] TestReloadCommand: started state syncer writer.go:29: 2021-01-29T19:34:58.172Z [WARN] TestReloadCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:58.174Z [INFO] TestReloadCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:23279 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:58.177Z [DEBUG] TestReloadCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:58.179Z [DEBUG] TestReloadCommand.server.raft: vote granted: from=c9124fa3-5c4e-cb8b-28d9-44ac4692efff term=2 tally=1 writer.go:29: 2021-01-29T19:34:58.181Z [INFO] TestReloadCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:58.183Z [INFO] TestReloadCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:23279 [Leader]" writer.go:29: 2021-01-29T19:34:58.185Z [INFO] TestReloadCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:58.186Z [INFO] TestReloadCommand.server: New leader elected: payload=Node-c9124fa3-5c4e-cb8b-28d9-44ac4692efff writer.go:29: 2021-01-29T19:34:58.186Z [DEBUG] TestReloadCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23279 writer.go:29: 2021-01-29T19:34:58.192Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:58.256Z [INFO] TestReloadCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:58.258Z [INFO] TestReloadCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:58.260Z [DEBUG] TestReloadCommand.server: Skipping self join check for node since the cluster is too small: node=Node-c9124fa3-5c4e-cb8b-28d9-44ac4692efff writer.go:29: 2021-01-29T19:34:58.262Z [INFO] TestReloadCommand.server: member joined, marking health alive: member=Node-c9124fa3-5c4e-cb8b-28d9-44ac4692efff writer.go:29: 2021-01-29T19:34:58.545Z [DEBUG] TestReloadCommand.http: Request finished: method=PUT url=/v1/agent/reload from=127.0.0.1:36020 latency=160.772µs writer.go:29: 2021-01-29T19:34:58.548Z [INFO] TestReloadCommand: Requesting shutdown writer.go:29: 2021-01-29T19:34:58.551Z [INFO] TestReloadCommand.server: shutting down server writer.go:29: 2021-01-29T19:34:58.553Z [DEBUG] TestReloadCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:58.556Z [WARN] TestReloadCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:58.559Z [ERROR] TestReloadCommand.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:34:58.561Z [DEBUG] TestReloadCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:58.563Z [WARN] TestReloadCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:58.574Z [INFO] TestReloadCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:58.574Z [INFO] TestReloadCommand: consul server down writer.go:29: 2021-01-29T19:34:58.582Z [INFO] TestReloadCommand: shutdown complete writer.go:29: 2021-01-29T19:34:58.585Z [INFO] TestReloadCommand: Stopping server: protocol=DNS address=127.0.0.1:23274 network=tcp writer.go:29: 2021-01-29T19:34:58.588Z [INFO] TestReloadCommand: Stopping server: protocol=DNS address=127.0.0.1:23274 network=udp writer.go:29: 2021-01-29T19:34:58.591Z [INFO] TestReloadCommand: Stopping server: protocol=HTTP address=127.0.0.1:23275 network=tcp writer.go:29: 2021-01-29T19:34:59.094Z [INFO] TestReloadCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:59.096Z [INFO] TestReloadCommand: Endpoints down --- PASS: TestReloadCommand (1.03s) PASS ok github.com/hashicorp/consul/command/reload 1.277s === RUN TestRTTCommand_noTabs === PAUSE TestRTTCommand_noTabs === RUN TestRTTCommand_BadArgs === PAUSE TestRTTCommand_BadArgs === RUN TestRTTCommand_LAN === PAUSE TestRTTCommand_LAN === RUN TestRTTCommand_WAN === PAUSE TestRTTCommand_WAN === CONT TestRTTCommand_noTabs === CONT TestRTTCommand_WAN === CONT TestRTTCommand_LAN === CONT TestRTTCommand_BadArgs === RUN TestRTTCommand_BadArgs/#00 [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === RUN TestRTTCommand_BadArgs/node1_node2_node3 --- PASS: TestRTTCommand_noTabs (0.01s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === RUN TestRTTCommand_BadArgs/-wan_node1_node2 === RUN TestRTTCommand_BadArgs/-wan_node1.dc1_node2 === RUN TestRTTCommand_BadArgs/-wan_node1_node2.dc1 --- PASS: TestRTTCommand_BadArgs (0.01s) --- PASS: TestRTTCommand_BadArgs/#00 (0.00s) --- PASS: TestRTTCommand_BadArgs/node1_node2_node3 (0.00s) --- PASS: TestRTTCommand_BadArgs/-wan_node1_node2 (0.00s) --- PASS: TestRTTCommand_BadArgs/-wan_node1.dc1_node2 (0.00s) --- PASS: TestRTTCommand_BadArgs/-wan_node1_node2.dc1 (0.00s) === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.843Z [WARN] TestRTTCommand_WAN: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:58.846Z [DEBUG] TestRTTCommand_WAN.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:58.851Z [DEBUG] TestRTTCommand_WAN.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:58.854Z [WARN] TestRTTCommand_LAN: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:34:58.856Z [DEBUG] TestRTTCommand_LAN.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:34:58.857Z [DEBUG] TestRTTCommand_LAN.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.866Z [INFO] TestRTTCommand_WAN.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ecb75b46-aa94-f814-d6cf-3c9f4ba12b39 Address:127.0.0.1:17159}]" === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:58.866Z [INFO] TestRTTCommand_LAN.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e6577bca-e497-0844-ed66-8782d91013f1 Address:127.0.0.1:17153}]" === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.869Z [INFO] TestRTTCommand_WAN.server.serf.wan: serf: EventMemberJoin: Node-ecb75b46-aa94-f814-d6cf-3c9f4ba12b39.dc1 127.0.0.1 === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:58.869Z [INFO] TestRTTCommand_LAN.server.serf.wan: serf: EventMemberJoin: Node-e6577bca-e497-0844-ed66-8782d91013f1.dc1 127.0.0.1 === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.871Z [INFO] TestRTTCommand_WAN.server.serf.lan: serf: EventMemberJoin: Node-ecb75b46-aa94-f814-d6cf-3c9f4ba12b39 127.0.0.1 === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:58.872Z [INFO] TestRTTCommand_LAN.server.serf.lan: serf: EventMemberJoin: Node-e6577bca-e497-0844-ed66-8782d91013f1 127.0.0.1 === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.874Z [INFO] TestRTTCommand_WAN: Started DNS server: address=127.0.0.1:17154 network=udp === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:58.876Z [INFO] TestRTTCommand_LAN: Started DNS server: address=127.0.0.1:17148 network=udp === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.876Z [INFO] TestRTTCommand_WAN.server.raft: entering follower state: follower="Node at 127.0.0.1:17159 [Follower]" leader= === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:58.877Z [INFO] TestRTTCommand_LAN.server.raft: entering follower state: follower="Node at 127.0.0.1:17153 [Follower]" leader= writer.go:29: 2021-01-29T19:34:58.882Z [INFO] TestRTTCommand_LAN.server: Adding LAN server: server="Node-e6577bca-e497-0844-ed66-8782d91013f1 (Addr: tcp/127.0.0.1:17153) (DC: dc1)" writer.go:29: 2021-01-29T19:34:58.883Z [INFO] TestRTTCommand_LAN.server: Handled event for server in area: event=member-join server=Node-e6577bca-e497-0844-ed66-8782d91013f1.dc1 area=wan === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.884Z [INFO] TestRTTCommand_WAN.server: Adding LAN server: server="Node-ecb75b46-aa94-f814-d6cf-3c9f4ba12b39 (Addr: tcp/127.0.0.1:17159) (DC: dc1)" writer.go:29: 2021-01-29T19:34:58.884Z [INFO] TestRTTCommand_WAN.server: Handled event for server in area: event=member-join server=Node-ecb75b46-aa94-f814-d6cf-3c9f4ba12b39.dc1 area=wan writer.go:29: 2021-01-29T19:34:58.885Z [INFO] TestRTTCommand_WAN: Started DNS server: address=127.0.0.1:17154 network=tcp === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:58.885Z [INFO] TestRTTCommand_LAN: Started DNS server: address=127.0.0.1:17148 network=tcp === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.889Z [INFO] TestRTTCommand_WAN: Started HTTP server: address=127.0.0.1:17155 network=tcp writer.go:29: 2021-01-29T19:34:58.890Z [INFO] TestRTTCommand_WAN: started state syncer === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:58.893Z [INFO] TestRTTCommand_LAN: Started HTTP server: address=127.0.0.1:17149 network=tcp writer.go:29: 2021-01-29T19:34:58.894Z [INFO] TestRTTCommand_LAN: started state syncer writer.go:29: 2021-01-29T19:34:58.932Z [WARN] TestRTTCommand_LAN.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:58.933Z [INFO] TestRTTCommand_LAN.server.raft: entering candidate state: node="Node at 127.0.0.1:17153 [Candidate]" term=2 writer.go:29: 2021-01-29T19:34:58.935Z [DEBUG] TestRTTCommand_LAN.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:58.937Z [DEBUG] TestRTTCommand_LAN.server.raft: vote granted: from=e6577bca-e497-0844-ed66-8782d91013f1 term=2 tally=1 writer.go:29: 2021-01-29T19:34:58.939Z [INFO] TestRTTCommand_LAN.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:58.941Z [INFO] TestRTTCommand_LAN.server.raft: entering leader state: leader="Node at 127.0.0.1:17153 [Leader]" writer.go:29: 2021-01-29T19:34:58.943Z [INFO] TestRTTCommand_LAN.server: cluster leadership acquired writer.go:29: 2021-01-29T19:34:58.944Z [INFO] TestRTTCommand_LAN.server: New leader elected: payload=Node-e6577bca-e497-0844-ed66-8782d91013f1 === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.947Z [WARN] TestRTTCommand_WAN.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:34:58.949Z [INFO] TestRTTCommand_WAN.server.raft: entering candidate state: node="Node at 127.0.0.1:17159 [Candidate]" term=2 === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:58.949Z [DEBUG] TestRTTCommand_LAN.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:17153 === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.951Z [DEBUG] TestRTTCommand_WAN.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:34:58.952Z [DEBUG] TestRTTCommand_WAN.server.raft: vote granted: from=ecb75b46-aa94-f814-d6cf-3c9f4ba12b39 term=2 tally=1 writer.go:29: 2021-01-29T19:34:58.954Z [INFO] TestRTTCommand_WAN.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:34:58.955Z [INFO] TestRTTCommand_WAN.server.raft: entering leader state: leader="Node at 127.0.0.1:17159 [Leader]" writer.go:29: 2021-01-29T19:34:58.956Z [INFO] TestRTTCommand_WAN.server: cluster leadership acquired === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:58.954Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:58.957Z [DEBUG] TestRTTCommand_WAN.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:17159 writer.go:29: 2021-01-29T19:34:58.957Z [INFO] TestRTTCommand_WAN.server: New leader elected: payload=Node-ecb75b46-aa94-f814-d6cf-3c9f4ba12b39 writer.go:29: 2021-01-29T19:34:58.972Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:34:59.014Z [INFO] TestRTTCommand_WAN.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:59.015Z [INFO] TestRTTCommand_WAN.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:59.017Z [DEBUG] TestRTTCommand_WAN.server: Skipping self join check for node since the cluster is too small: node=Node-ecb75b46-aa94-f814-d6cf-3c9f4ba12b39 writer.go:29: 2021-01-29T19:34:59.018Z [INFO] TestRTTCommand_WAN.server: member joined, marking health alive: member=Node-ecb75b46-aa94-f814-d6cf-3c9f4ba12b39 === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:59.022Z [INFO] TestRTTCommand_LAN.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:34:59.023Z [INFO] TestRTTCommand_LAN.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:59.022Z [INFO] TestRTTCommand_LAN: Synced node info writer.go:29: 2021-01-29T19:34:59.024Z [DEBUG] TestRTTCommand_LAN.server: Skipping self join check for node since the cluster is too small: node=Node-e6577bca-e497-0844-ed66-8782d91013f1 writer.go:29: 2021-01-29T19:34:59.026Z [INFO] TestRTTCommand_LAN.server: member joined, marking health alive: member=Node-e6577bca-e497-0844-ed66-8782d91013f1 === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:59.049Z [DEBUG] TestRTTCommand_WAN.http: Request finished: method=GET url=/v1/coordinate/datacenters from=127.0.0.1:55754 latency=5.043057ms writer.go:29: 2021-01-29T19:34:59.060Z [DEBUG] TestRTTCommand_WAN.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:55756 latency=3.074519ms writer.go:29: 2021-01-29T19:34:59.067Z [DEBUG] TestRTTCommand_WAN.http: Request finished: method=GET url=/v1/coordinate/datacenters from=127.0.0.1:55756 latency=2.621201ms writer.go:29: 2021-01-29T19:34:59.077Z [DEBUG] TestRTTCommand_WAN.http: Request finished: method=GET url=/v1/coordinate/datacenters from=127.0.0.1:55758 latency=288.077µs writer.go:29: 2021-01-29T19:34:59.080Z [INFO] TestRTTCommand_WAN: Requesting shutdown writer.go:29: 2021-01-29T19:34:59.081Z [INFO] TestRTTCommand_WAN.server: shutting down server writer.go:29: 2021-01-29T19:34:59.080Z [DEBUG] TestRTTCommand_WAN: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:34:59.086Z [DEBUG] TestRTTCommand_WAN.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:59.087Z [WARN] TestRTTCommand_WAN.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:59.093Z [WARN] TestRTTCommand_WAN.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:59.093Z [DEBUG] TestRTTCommand_WAN.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:59.098Z [INFO] TestRTTCommand_WAN: Synced node info writer.go:29: 2021-01-29T19:34:59.100Z [INFO] TestRTTCommand_WAN.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:59.101Z [INFO] TestRTTCommand_WAN: consul server down writer.go:29: 2021-01-29T19:34:59.110Z [INFO] TestRTTCommand_WAN: shutdown complete writer.go:29: 2021-01-29T19:34:59.116Z [INFO] TestRTTCommand_WAN: Stopping server: protocol=DNS address=127.0.0.1:17154 network=tcp writer.go:29: 2021-01-29T19:34:59.117Z [INFO] TestRTTCommand_WAN: Stopping server: protocol=DNS address=127.0.0.1:17154 network=udp writer.go:29: 2021-01-29T19:34:59.119Z [INFO] TestRTTCommand_WAN: Stopping server: protocol=HTTP address=127.0.0.1:17155 network=tcp === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:59.308Z [DEBUG] TestRTTCommand_LAN.http: Request finished: method=GET url=/v1/coordinate/nodes from=127.0.0.1:60440 latency=218.635µs writer.go:29: 2021-01-29T19:34:59.343Z [DEBUG] TestRTTCommand_LAN.http: Request finished: method=GET url=/v1/coordinate/nodes from=127.0.0.1:60442 latency=2.550016ms writer.go:29: 2021-01-29T19:34:59.360Z [DEBUG] TestRTTCommand_LAN.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:60444 latency=684.834µs writer.go:29: 2021-01-29T19:34:59.364Z [DEBUG] TestRTTCommand_LAN.http: Request finished: method=GET url=/v1/coordinate/nodes from=127.0.0.1:60444 latency=71.193µs writer.go:29: 2021-01-29T19:34:59.370Z [DEBUG] TestRTTCommand_LAN.http: Request finished: method=GET url=/v1/coordinate/nodes from=127.0.0.1:60446 latency=154.029µs writer.go:29: 2021-01-29T19:34:59.373Z [INFO] TestRTTCommand_LAN: Requesting shutdown writer.go:29: 2021-01-29T19:34:59.374Z [INFO] TestRTTCommand_LAN.server: shutting down server writer.go:29: 2021-01-29T19:34:59.375Z [DEBUG] TestRTTCommand_LAN.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:59.377Z [WARN] TestRTTCommand_LAN.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:59.378Z [DEBUG] TestRTTCommand_LAN.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:34:59.380Z [WARN] TestRTTCommand_LAN.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:34:59.383Z [INFO] TestRTTCommand_LAN.server.router.manager: shutting down writer.go:29: 2021-01-29T19:34:59.383Z [INFO] TestRTTCommand_LAN: consul server down writer.go:29: 2021-01-29T19:34:59.386Z [INFO] TestRTTCommand_LAN: shutdown complete writer.go:29: 2021-01-29T19:34:59.387Z [INFO] TestRTTCommand_LAN: Stopping server: protocol=DNS address=127.0.0.1:17148 network=tcp writer.go:29: 2021-01-29T19:34:59.388Z [INFO] TestRTTCommand_LAN: Stopping server: protocol=DNS address=127.0.0.1:17148 network=udp writer.go:29: 2021-01-29T19:34:59.389Z [INFO] TestRTTCommand_LAN: Stopping server: protocol=HTTP address=127.0.0.1:17149 network=tcp === CONT TestRTTCommand_WAN writer.go:29: 2021-01-29T19:34:59.621Z [INFO] TestRTTCommand_WAN: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:59.623Z [INFO] TestRTTCommand_WAN: Endpoints down --- PASS: TestRTTCommand_WAN (0.82s) === CONT TestRTTCommand_LAN writer.go:29: 2021-01-29T19:34:59.891Z [INFO] TestRTTCommand_LAN: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:34:59.893Z [INFO] TestRTTCommand_LAN: Endpoints down --- PASS: TestRTTCommand_LAN (1.09s) PASS ok github.com/hashicorp/consul/command/rtt 1.247s === RUN TestDevModeHasNoServices === PAUSE TestDevModeHasNoServices === RUN TestStructsToAgentService === PAUSE TestStructsToAgentService === RUN TestCommand_noTabs === PAUSE TestCommand_noTabs === CONT TestDevModeHasNoServices === CONT TestCommand_noTabs === CONT TestStructsToAgentService --- PASS: TestCommand_noTabs (0.00s) === RUN TestStructsToAgentService/Basic_service_with_port === PAUSE TestStructsToAgentService/Basic_service_with_port === RUN TestStructsToAgentService/Service_with_a_check === PAUSE TestStructsToAgentService/Service_with_a_check === RUN TestStructsToAgentService/Service_with_an_unnamed_check === PAUSE TestStructsToAgentService/Service_with_an_unnamed_check === RUN TestStructsToAgentService/Service_with_a_zero-value_check === PAUSE TestStructsToAgentService/Service_with_a_zero-value_check === RUN TestStructsToAgentService/Service_with_checks === PAUSE TestStructsToAgentService/Service_with_checks === RUN TestStructsToAgentService/Proxy_service === PAUSE TestStructsToAgentService/Proxy_service === CONT TestStructsToAgentService/Basic_service_with_port === CONT TestStructsToAgentService/Service_with_a_zero-value_check === CONT TestStructsToAgentService/Service_with_an_unnamed_check === CONT TestStructsToAgentService/Proxy_service === CONT TestStructsToAgentService/Service_with_a_check === CONT TestStructsToAgentService/Service_with_checks --- PASS: TestDevModeHasNoServices (0.04s) --- PASS: TestStructsToAgentService (0.01s) --- PASS: TestStructsToAgentService/Basic_service_with_port (0.00s) --- PASS: TestStructsToAgentService/Service_with_a_zero-value_check (0.00s) --- PASS: TestStructsToAgentService/Proxy_service (0.01s) --- PASS: TestStructsToAgentService/Service_with_checks (0.01s) --- PASS: TestStructsToAgentService/Service_with_an_unnamed_check (0.03s) --- PASS: TestStructsToAgentService/Service_with_a_check (0.03s) PASS ok github.com/hashicorp/consul/command/services 0.088s === RUN TestCommand_noTabs === PAUSE TestCommand_noTabs === RUN TestCommand_Validation === PAUSE TestCommand_Validation === RUN TestCommand_File_id === PAUSE TestCommand_File_id === RUN TestCommand_File_nameOnly === PAUSE TestCommand_File_nameOnly === RUN TestCommand_Flag === PAUSE TestCommand_Flag === CONT TestCommand_noTabs === CONT TestCommand_Flag --- PASS: TestCommand_noTabs (0.01s) === CONT TestCommand_File_id [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestCommand_File_nameOnly === CONT TestCommand_Validation === RUN TestCommand_Validation/no_args_or_id === RUN TestCommand_Validation/args_and_-id [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range --- PASS: TestCommand_Validation (0.02s) --- PASS: TestCommand_Validation/no_args_or_id (0.00s) --- PASS: TestCommand_Validation/args_and_-id (0.01s) === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.181Z [WARN] TestCommand_File_id: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:07.184Z [DEBUG] TestCommand_File_id.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:07.194Z [DEBUG] TestCommand_File_id.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:35:07.222Z [INFO] TestCommand_File_id.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:35b98fde-1b7e-5345-d2d8-b4fa66595767 Address:127.0.0.1:23279}]" === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.235Z [WARN] TestCommand_File_nameOnly: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:07.237Z [DEBUG] TestCommand_File_nameOnly.tlsutil: Update: version=1 === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.230Z [WARN] TestCommand_Flag: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:07.254Z [DEBUG] TestCommand_Flag.tlsutil: Update: version=1 === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.227Z [INFO] TestCommand_File_id.server.raft: entering follower state: follower="Node at 127.0.0.1:23279 [Follower]" leader= === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.256Z [DEBUG] TestCommand_File_nameOnly.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.276Z [DEBUG] TestCommand_Flag.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.274Z [INFO] TestCommand_File_id.server.serf.wan: serf: EventMemberJoin: Node-35b98fde-1b7e-5345-d2d8-b4fa66595767.dc1 127.0.0.1 === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.305Z [INFO] TestCommand_Flag.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:e552f78f-86ac-e8cd-4572-4675046b74aa Address:127.0.0.1:23291}]" === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.308Z [INFO] TestCommand_File_id.server.serf.lan: serf: EventMemberJoin: Node-35b98fde-1b7e-5345-d2d8-b4fa66595767 127.0.0.1 === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.308Z [INFO] TestCommand_File_nameOnly.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:091cc349-b774-7c17-2070-d22f6c2acc5f Address:127.0.0.1:23285}]" writer.go:29: 2021-01-29T19:35:07.323Z [INFO] TestCommand_File_nameOnly.server.raft: entering follower state: follower="Node at 127.0.0.1:23285 [Follower]" leader= === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.325Z [INFO] TestCommand_File_id.server: Adding LAN server: server="Node-35b98fde-1b7e-5345-d2d8-b4fa66595767 (Addr: tcp/127.0.0.1:23279) (DC: dc1)" writer.go:29: 2021-01-29T19:35:07.328Z [INFO] TestCommand_File_id.server: Handled event for server in area: event=member-join server=Node-35b98fde-1b7e-5345-d2d8-b4fa66595767.dc1 area=wan === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.308Z [INFO] TestCommand_Flag.server.raft: entering follower state: follower="Node at 127.0.0.1:23291 [Follower]" leader= writer.go:29: 2021-01-29T19:35:07.331Z [INFO] TestCommand_Flag.server.serf.wan: serf: EventMemberJoin: Node-e552f78f-86ac-e8cd-4572-4675046b74aa.dc1 127.0.0.1 === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.329Z [INFO] TestCommand_File_id: Started DNS server: address=127.0.0.1:23274 network=udp === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.340Z [INFO] TestCommand_File_nameOnly.server.serf.wan: serf: EventMemberJoin: Node-091cc349-b774-7c17-2070-d22f6c2acc5f.dc1 127.0.0.1 === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.339Z [INFO] TestCommand_File_id: Started DNS server: address=127.0.0.1:23274 network=tcp === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.342Z [INFO] TestCommand_Flag.server.serf.lan: serf: EventMemberJoin: Node-e552f78f-86ac-e8cd-4572-4675046b74aa 127.0.0.1 === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.336Z [WARN] TestCommand_File_id.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:07.345Z [INFO] TestCommand_File_id.server.raft: entering candidate state: node="Node at 127.0.0.1:23279 [Candidate]" term=2 === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.346Z [INFO] TestCommand_Flag: Started DNS server: address=127.0.0.1:23286 network=udp writer.go:29: 2021-01-29T19:35:07.347Z [INFO] TestCommand_Flag.server: Adding LAN server: server="Node-e552f78f-86ac-e8cd-4572-4675046b74aa (Addr: tcp/127.0.0.1:23291) (DC: dc1)" writer.go:29: 2021-01-29T19:35:07.349Z [INFO] TestCommand_Flag.server: Handled event for server in area: event=member-join server=Node-e552f78f-86ac-e8cd-4572-4675046b74aa.dc1 area=wan === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.348Z [INFO] TestCommand_File_id: Started HTTP server: address=127.0.0.1:23275 network=tcp === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.352Z [INFO] TestCommand_Flag: Started DNS server: address=127.0.0.1:23286 network=tcp writer.go:29: 2021-01-29T19:35:07.355Z [INFO] TestCommand_Flag: Started HTTP server: address=127.0.0.1:23287 network=tcp === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.353Z [INFO] TestCommand_File_nameOnly.server.serf.lan: serf: EventMemberJoin: Node-091cc349-b774-7c17-2070-d22f6c2acc5f 127.0.0.1 === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.357Z [INFO] TestCommand_Flag: started state syncer === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.351Z [INFO] TestCommand_File_id: started state syncer writer.go:29: 2021-01-29T19:35:07.364Z [DEBUG] TestCommand_File_id.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:07.366Z [DEBUG] TestCommand_File_id.server.raft: vote granted: from=35b98fde-1b7e-5345-d2d8-b4fa66595767 term=2 tally=1 writer.go:29: 2021-01-29T19:35:07.368Z [INFO] TestCommand_File_id.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:07.370Z [INFO] TestCommand_File_id.server.raft: entering leader state: leader="Node at 127.0.0.1:23279 [Leader]" writer.go:29: 2021-01-29T19:35:07.372Z [INFO] TestCommand_File_id.server: cluster leadership acquired === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.360Z [INFO] TestCommand_File_nameOnly.server: Handled event for server in area: event=member-join server=Node-091cc349-b774-7c17-2070-d22f6c2acc5f.dc1 area=wan === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.375Z [INFO] TestCommand_File_id.server: New leader elected: payload=Node-35b98fde-1b7e-5345-d2d8-b4fa66595767 === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.363Z [INFO] TestCommand_File_nameOnly: Started DNS server: address=127.0.0.1:23280 network=tcp writer.go:29: 2021-01-29T19:35:07.379Z [INFO] TestCommand_File_nameOnly: Started DNS server: address=127.0.0.1:23280 network=udp writer.go:29: 2021-01-29T19:35:07.365Z [INFO] TestCommand_File_nameOnly.server: Adding LAN server: server="Node-091cc349-b774-7c17-2070-d22f6c2acc5f (Addr: tcp/127.0.0.1:23285) (DC: dc1)" === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.389Z [WARN] TestCommand_Flag.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.393Z [INFO] TestCommand_File_nameOnly: Started HTTP server: address=127.0.0.1:23281 network=tcp === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.395Z [INFO] TestCommand_Flag.server.raft: entering candidate state: node="Node at 127.0.0.1:23291 [Candidate]" term=2 === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.397Z [INFO] TestCommand_File_nameOnly: started state syncer === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.405Z [DEBUG] TestCommand_Flag.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:07.407Z [DEBUG] TestCommand_Flag.server.raft: vote granted: from=e552f78f-86ac-e8cd-4572-4675046b74aa term=2 tally=1 === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.401Z [WARN] TestCommand_File_nameOnly.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.408Z [INFO] TestCommand_Flag.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:07.410Z [INFO] TestCommand_Flag.server.raft: entering leader state: leader="Node at 127.0.0.1:23291 [Leader]" === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.403Z [DEBUG] TestCommand_File_id.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23279 === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.412Z [INFO] TestCommand_Flag.server: cluster leadership acquired === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.409Z [INFO] TestCommand_File_nameOnly.server.raft: entering candidate state: node="Node at 127.0.0.1:23285 [Candidate]" term=2 === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.415Z [INFO] TestCommand_Flag.server: New leader elected: payload=Node-e552f78f-86ac-e8cd-4572-4675046b74aa === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.417Z [DEBUG] TestCommand_File_nameOnly.server.raft: votes: needed=1 === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.415Z [DEBUG] TestCommand_Flag.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23291 === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.422Z [DEBUG] TestCommand_File_nameOnly.server.raft: vote granted: from=091cc349-b774-7c17-2070-d22f6c2acc5f term=2 tally=1 === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.432Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.427Z [INFO] TestCommand_File_nameOnly.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:07.437Z [INFO] TestCommand_File_nameOnly.server.raft: entering leader state: leader="Node at 127.0.0.1:23285 [Leader]" writer.go:29: 2021-01-29T19:35:07.444Z [INFO] TestCommand_File_nameOnly.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:07.445Z [INFO] TestCommand_File_nameOnly.server: New leader elected: payload=Node-091cc349-b774-7c17-2070-d22f6c2acc5f === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.450Z [INFO] TestCommand_Flag: Synced node info writer.go:29: 2021-01-29T19:35:07.450Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.445Z [DEBUG] TestCommand_File_nameOnly.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23285 === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.452Z [INFO] TestCommand_Flag: Synced service: service=web === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.465Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.462Z [DEBUG] TestCommand_Flag.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:51954 latency=22.939086ms writer.go:29: 2021-01-29T19:35:07.475Z [DEBUG] TestCommand_Flag: Node info in sync writer.go:29: 2021-01-29T19:35:07.478Z [DEBUG] TestCommand_Flag: Service in sync: service=web writer.go:29: 2021-01-29T19:35:07.483Z [INFO] TestCommand_Flag: Synced service: service=db writer.go:29: 2021-01-29T19:35:07.488Z [DEBUG] TestCommand_Flag.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:51954 latency=13.181319ms writer.go:29: 2021-01-29T19:35:07.516Z [DEBUG] TestCommand_Flag: removed service: service=web writer.go:29: 2021-01-29T19:35:07.519Z [DEBUG] TestCommand_Flag: Node info in sync writer.go:29: 2021-01-29T19:35:07.522Z [DEBUG] TestCommand_Flag: Service in sync: service=db writer.go:29: 2021-01-29T19:35:07.528Z [INFO] TestCommand_Flag: Deregistered service: service=web writer.go:29: 2021-01-29T19:35:07.530Z [DEBUG] TestCommand_Flag.http: Request finished: method=PUT url=/v1/agent/service/deregister/web from=127.0.0.1:51958 latency=14.29982ms === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.536Z [INFO] TestCommand_File_nameOnly.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:07.538Z [INFO] TestCommand_File_nameOnly.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.540Z [DEBUG] TestCommand_File_nameOnly.server: Skipping self join check for node since the cluster is too small: node=Node-091cc349-b774-7c17-2070-d22f6c2acc5f writer.go:29: 2021-01-29T19:35:07.542Z [INFO] TestCommand_File_nameOnly.server: member joined, marking health alive: member=Node-091cc349-b774-7c17-2070-d22f6c2acc5f === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.554Z [INFO] TestCommand_File_id: Synced node info === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.562Z [INFO] TestCommand_Flag.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:07.566Z [INFO] TestCommand_Flag.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.567Z [DEBUG] TestCommand_Flag.server: Skipping self join check for node since the cluster is too small: node=Node-e552f78f-86ac-e8cd-4572-4675046b74aa writer.go:29: 2021-01-29T19:35:07.568Z [INFO] TestCommand_Flag.server: member joined, marking health alive: member=Node-e552f78f-86ac-e8cd-4572-4675046b74aa === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.584Z [INFO] TestCommand_File_id.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:07.586Z [INFO] TestCommand_File_id.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.588Z [DEBUG] TestCommand_File_id.server: Skipping self join check for node since the cluster is too small: node=Node-35b98fde-1b7e-5345-d2d8-b4fa66595767 === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.585Z [INFO] TestCommand_File_nameOnly: Synced node info === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.590Z [INFO] TestCommand_File_id.server: member joined, marking health alive: member=Node-35b98fde-1b7e-5345-d2d8-b4fa66595767 === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.602Z [INFO] TestCommand_File_nameOnly: Synced service: service=web === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.610Z [DEBUG] TestCommand_Flag.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:51954 latency=65.236392ms === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.610Z [DEBUG] TestCommand_File_nameOnly.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:57506 latency=58.588725ms writer.go:29: 2021-01-29T19:35:07.610Z [DEBUG] TestCommand_File_nameOnly: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:35:07.613Z [DEBUG] TestCommand_File_nameOnly: Node info in sync writer.go:29: 2021-01-29T19:35:07.615Z [DEBUG] TestCommand_File_nameOnly: Service in sync: service=web writer.go:29: 2021-01-29T19:35:07.625Z [DEBUG] TestCommand_File_nameOnly: Node info in sync writer.go:29: 2021-01-29T19:35:07.627Z [DEBUG] TestCommand_File_nameOnly: Service in sync: service=web === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.628Z [INFO] TestCommand_Flag: Requesting shutdown writer.go:29: 2021-01-29T19:35:07.632Z [DEBUG] TestCommand_Flag: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:35:07.635Z [DEBUG] TestCommand_Flag: Node info in sync === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.640Z [INFO] TestCommand_File_nameOnly: Synced service: service=db === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.638Z [DEBUG] TestCommand_Flag: Service in sync: service=db === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.641Z [DEBUG] TestCommand_File_nameOnly.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:57506 latency=16.54351ms === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.633Z [INFO] TestCommand_Flag.server: shutting down server writer.go:29: 2021-01-29T19:35:07.652Z [DEBUG] TestCommand_Flag.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.642Z [DEBUG] TestCommand_Flag: Node info in sync writer.go:29: 2021-01-29T19:35:07.654Z [DEBUG] TestCommand_Flag: Service in sync: service=db writer.go:29: 2021-01-29T19:35:07.653Z [WARN] TestCommand_Flag.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:07.653Z [DEBUG] TestCommand_Flag.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.657Z [WARN] TestCommand_Flag.server.serf.wan: serf: Shutdown without a Leave === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.660Z [DEBUG] TestCommand_File_nameOnly: removed service: service=web === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.660Z [INFO] TestCommand_Flag.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:07.661Z [INFO] TestCommand_Flag: consul server down === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.661Z [DEBUG] TestCommand_File_nameOnly: Node info in sync === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.663Z [INFO] TestCommand_Flag: shutdown complete === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.664Z [INFO] TestCommand_File_nameOnly: Deregistered service: service=web === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.665Z [INFO] TestCommand_Flag: Stopping server: protocol=DNS address=127.0.0.1:23286 network=tcp === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.666Z [DEBUG] TestCommand_File_nameOnly: Service in sync: service=db === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.666Z [INFO] TestCommand_Flag: Stopping server: protocol=DNS address=127.0.0.1:23286 network=udp === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.667Z [DEBUG] TestCommand_File_nameOnly.http: Request finished: method=PUT url=/v1/agent/service/deregister/web from=127.0.0.1:57510 latency=8.096383ms === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:07.668Z [INFO] TestCommand_Flag: Stopping server: protocol=HTTP address=127.0.0.1:23287 network=tcp === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:07.670Z [DEBUG] TestCommand_File_nameOnly.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:57506 latency=136.72µs writer.go:29: 2021-01-29T19:35:07.676Z [INFO] TestCommand_File_nameOnly: Requesting shutdown writer.go:29: 2021-01-29T19:35:07.678Z [INFO] TestCommand_File_nameOnly.server: shutting down server writer.go:29: 2021-01-29T19:35:07.680Z [DEBUG] TestCommand_File_nameOnly.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.692Z [WARN] TestCommand_File_nameOnly.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:07.692Z [DEBUG] TestCommand_File_nameOnly.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.695Z [WARN] TestCommand_File_nameOnly.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:07.697Z [INFO] TestCommand_File_nameOnly.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:07.697Z [INFO] TestCommand_File_nameOnly: consul server down writer.go:29: 2021-01-29T19:35:07.699Z [INFO] TestCommand_File_nameOnly: shutdown complete writer.go:29: 2021-01-29T19:35:07.700Z [INFO] TestCommand_File_nameOnly: Stopping server: protocol=DNS address=127.0.0.1:23280 network=tcp writer.go:29: 2021-01-29T19:35:07.701Z [INFO] TestCommand_File_nameOnly: Stopping server: protocol=DNS address=127.0.0.1:23280 network=udp writer.go:29: 2021-01-29T19:35:07.702Z [INFO] TestCommand_File_nameOnly: Stopping server: protocol=HTTP address=127.0.0.1:23281 network=tcp === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:07.801Z [DEBUG] TestCommand_File_id: Node info in sync writer.go:29: 2021-01-29T19:35:07.804Z [INFO] TestCommand_File_id: Synced service: service=web writer.go:29: 2021-01-29T19:35:07.805Z [DEBUG] TestCommand_File_id.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:36060 latency=8.530584ms writer.go:29: 2021-01-29T19:35:07.808Z [DEBUG] TestCommand_File_id: Node info in sync writer.go:29: 2021-01-29T19:35:07.810Z [DEBUG] TestCommand_File_id: Service in sync: service=web writer.go:29: 2021-01-29T19:35:07.814Z [INFO] TestCommand_File_id: Synced service: service=db writer.go:29: 2021-01-29T19:35:07.815Z [DEBUG] TestCommand_File_id.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:36060 latency=8.337345ms writer.go:29: 2021-01-29T19:35:07.858Z [DEBUG] TestCommand_File_id: removed service: service=web writer.go:29: 2021-01-29T19:35:07.860Z [DEBUG] TestCommand_File_id: Node info in sync writer.go:29: 2021-01-29T19:35:07.864Z [INFO] TestCommand_File_id: Deregistered service: service=web writer.go:29: 2021-01-29T19:35:07.866Z [DEBUG] TestCommand_File_id: Service in sync: service=db writer.go:29: 2021-01-29T19:35:07.868Z [DEBUG] TestCommand_File_id.http: Request finished: method=PUT url=/v1/agent/service/deregister/web from=127.0.0.1:36062 latency=9.682539ms writer.go:29: 2021-01-29T19:35:07.871Z [DEBUG] TestCommand_File_id.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:36060 latency=276.115µs writer.go:29: 2021-01-29T19:35:07.876Z [INFO] TestCommand_File_id: Requesting shutdown writer.go:29: 2021-01-29T19:35:07.877Z [INFO] TestCommand_File_id.server: shutting down server writer.go:29: 2021-01-29T19:35:07.879Z [DEBUG] TestCommand_File_id.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.880Z [WARN] TestCommand_File_id.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:07.882Z [DEBUG] TestCommand_File_id.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.883Z [WARN] TestCommand_File_id.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:07.887Z [INFO] TestCommand_File_id.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:07.889Z [INFO] TestCommand_File_id: consul server down writer.go:29: 2021-01-29T19:35:07.890Z [INFO] TestCommand_File_id: shutdown complete writer.go:29: 2021-01-29T19:35:07.891Z [INFO] TestCommand_File_id: Stopping server: protocol=DNS address=127.0.0.1:23274 network=tcp writer.go:29: 2021-01-29T19:35:07.893Z [INFO] TestCommand_File_id: Stopping server: protocol=DNS address=127.0.0.1:23274 network=udp writer.go:29: 2021-01-29T19:35:07.894Z [INFO] TestCommand_File_id: Stopping server: protocol=HTTP address=127.0.0.1:23275 network=tcp === CONT TestCommand_Flag writer.go:29: 2021-01-29T19:35:08.170Z [INFO] TestCommand_Flag: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:08.171Z [INFO] TestCommand_Flag: Endpoints down --- PASS: TestCommand_Flag (1.06s) === CONT TestCommand_File_nameOnly writer.go:29: 2021-01-29T19:35:08.203Z [INFO] TestCommand_File_nameOnly: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:08.204Z [INFO] TestCommand_File_nameOnly: Endpoints down --- PASS: TestCommand_File_nameOnly (1.08s) === CONT TestCommand_File_id writer.go:29: 2021-01-29T19:35:08.395Z [INFO] TestCommand_File_id: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:08.397Z [INFO] TestCommand_File_id: Endpoints down --- PASS: TestCommand_File_id (1.28s) PASS ok github.com/hashicorp/consul/command/services/deregister 1.485s === RUN TestCommand_noTabs === PAUSE TestCommand_noTabs === RUN TestCommand_Validation === PAUSE TestCommand_Validation === RUN TestCommand_File === PAUSE TestCommand_File === RUN TestCommand_Flags === PAUSE TestCommand_Flags === RUN TestCommand_Flags_TaggedAddresses === PAUSE TestCommand_Flags_TaggedAddresses === RUN TestCommand_FileWithUnnamedCheck === PAUSE TestCommand_FileWithUnnamedCheck === CONT TestCommand_noTabs === CONT TestCommand_Flags_TaggedAddresses === CONT TestCommand_File [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestCommand_FileWithUnnamedCheck --- PASS: TestCommand_noTabs (0.00s) === CONT TestCommand_Validation === RUN TestCommand_Validation/no_args_or_id === RUN TestCommand_Validation/args_and_-name --- PASS: TestCommand_Validation (0.01s) --- PASS: TestCommand_Validation/no_args_or_id (0.00s) --- PASS: TestCommand_Validation/args_and_-name (0.00s) === CONT TestCommand_Flags [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.028Z [WARN] TestCommand_Flags_TaggedAddresses: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:07.029Z [DEBUG] TestCommand_Flags_TaggedAddresses.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:07.042Z [DEBUG] TestCommand_Flags_TaggedAddresses.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.052Z [WARN] TestCommand_File: bootstrap = true: do not enable unless necessary === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.052Z [WARN] TestCommand_Flags: bootstrap = true: do not enable unless necessary === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.053Z [DEBUG] TestCommand_File.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:07.054Z [DEBUG] TestCommand_File.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.054Z [DEBUG] TestCommand_Flags.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:07.056Z [DEBUG] TestCommand_Flags.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.058Z [INFO] TestCommand_File.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:25a664de-1922-871b-a28d-95c7031396e7 Address:127.0.0.1:22258}]" === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.058Z [WARN] TestCommand_FileWithUnnamedCheck: bootstrap = true: do not enable unless necessary === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.060Z [INFO] TestCommand_File.server.serf.wan: serf: EventMemberJoin: Node-25a664de-1922-871b-a28d-95c7031396e7.dc1 127.0.0.1 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.060Z [DEBUG] TestCommand_FileWithUnnamedCheck.tlsutil: Update: version=1 === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.060Z [INFO] TestCommand_Flags.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:b29a106a-fe34-7cfb-694e-da1397be3cf6 Address:127.0.0.1:22276}]" === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.063Z [INFO] TestCommand_File.server.serf.lan: serf: EventMemberJoin: Node-25a664de-1922-871b-a28d-95c7031396e7 127.0.0.1 writer.go:29: 2021-01-29T19:35:07.069Z [INFO] TestCommand_File: Started DNS server: address=127.0.0.1:22253 network=udp writer.go:29: 2021-01-29T19:35:07.072Z [INFO] TestCommand_File.server.raft: entering follower state: follower="Node at 127.0.0.1:22258 [Follower]" leader= writer.go:29: 2021-01-29T19:35:07.077Z [INFO] TestCommand_File: Started DNS server: address=127.0.0.1:22253 network=tcp writer.go:29: 2021-01-29T19:35:07.082Z [INFO] TestCommand_File.server: Adding LAN server: server="Node-25a664de-1922-871b-a28d-95c7031396e7 (Addr: tcp/127.0.0.1:22258) (DC: dc1)" === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.078Z [DEBUG] TestCommand_FileWithUnnamedCheck.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.090Z [INFO] TestCommand_File: Started HTTP server: address=127.0.0.1:22254 network=tcp === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.073Z [INFO] TestCommand_Flags.server.raft: entering follower state: follower="Node at 127.0.0.1:22276 [Follower]" leader= === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.078Z [INFO] TestCommand_File.server: Handled event for server in area: event=member-join server=Node-25a664de-1922-871b-a28d-95c7031396e7.dc1 area=wan writer.go:29: 2021-01-29T19:35:07.097Z [INFO] TestCommand_File: started state syncer === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.098Z [INFO] TestCommand_FileWithUnnamedCheck.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:ef8b0aad-685f-4b57-c5e8-bcf4cc96b21d Address:127.0.0.1:22270}]" writer.go:29: 2021-01-29T19:35:07.100Z [INFO] TestCommand_FileWithUnnamedCheck.server.raft: entering follower state: follower="Node at 127.0.0.1:22270 [Follower]" leader= === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.095Z [INFO] TestCommand_Flags.server.serf.wan: serf: EventMemberJoin: Node-b29a106a-fe34-7cfb-694e-da1397be3cf6.dc1 127.0.0.1 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.107Z [INFO] TestCommand_FileWithUnnamedCheck.server.serf.wan: serf: EventMemberJoin: Node-ef8b0aad-685f-4b57-c5e8-bcf4cc96b21d.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:35:07.109Z [INFO] TestCommand_FileWithUnnamedCheck.server.serf.lan: serf: EventMemberJoin: Node-ef8b0aad-685f-4b57-c5e8-bcf4cc96b21d 127.0.0.1 === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.109Z [INFO] TestCommand_Flags.server.serf.lan: serf: EventMemberJoin: Node-b29a106a-fe34-7cfb-694e-da1397be3cf6 127.0.0.1 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.111Z [INFO] TestCommand_FileWithUnnamedCheck: Started DNS server: address=127.0.0.1:22265 network=udp === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.113Z [INFO] TestCommand_Flags: Started DNS server: address=127.0.0.1:22271 network=udp === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.114Z [INFO] TestCommand_FileWithUnnamedCheck.server: Adding LAN server: server="Node-ef8b0aad-685f-4b57-c5e8-bcf4cc96b21d (Addr: tcp/127.0.0.1:22270) (DC: dc1)" writer.go:29: 2021-01-29T19:35:07.116Z [INFO] TestCommand_FileWithUnnamedCheck.server: Handled event for server in area: event=member-join server=Node-ef8b0aad-685f-4b57-c5e8-bcf4cc96b21d.dc1 area=wan === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.117Z [INFO] TestCommand_Flags.server: Adding LAN server: server="Node-b29a106a-fe34-7cfb-694e-da1397be3cf6 (Addr: tcp/127.0.0.1:22276) (DC: dc1)" writer.go:29: 2021-01-29T19:35:07.118Z [INFO] TestCommand_Flags.server: Handled event for server in area: event=member-join server=Node-b29a106a-fe34-7cfb-694e-da1397be3cf6.dc1 area=wan === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.120Z [INFO] TestCommand_FileWithUnnamedCheck: Started DNS server: address=127.0.0.1:22265 network=tcp === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.122Z [INFO] TestCommand_Flags: Started DNS server: address=127.0.0.1:22271 network=tcp === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.124Z [INFO] TestCommand_FileWithUnnamedCheck: Started HTTP server: address=127.0.0.1:22266 network=tcp writer.go:29: 2021-01-29T19:35:07.126Z [INFO] TestCommand_FileWithUnnamedCheck: started state syncer === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.128Z [WARN] TestCommand_File.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:07.129Z [INFO] TestCommand_File.server.raft: entering candidate state: node="Node at 127.0.0.1:22258 [Candidate]" term=2 === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.132Z [INFO] TestCommand_Flags: Started HTTP server: address=127.0.0.1:22272 network=tcp writer.go:29: 2021-01-29T19:35:07.133Z [INFO] TestCommand_Flags: started state syncer writer.go:29: 2021-01-29T19:35:07.140Z [WARN] TestCommand_Flags.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:07.141Z [INFO] TestCommand_Flags.server.raft: entering candidate state: node="Node at 127.0.0.1:22276 [Candidate]" term=2 writer.go:29: 2021-01-29T19:35:07.143Z [DEBUG] TestCommand_Flags.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:07.144Z [DEBUG] TestCommand_Flags.server.raft: vote granted: from=b29a106a-fe34-7cfb-694e-da1397be3cf6 term=2 tally=1 === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.145Z [DEBUG] TestCommand_File.server.raft: votes: needed=1 === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.146Z [INFO] TestCommand_Flags.server.raft: election won: tally=1 === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.147Z [DEBUG] TestCommand_File.server.raft: vote granted: from=25a664de-1922-871b-a28d-95c7031396e7 term=2 tally=1 === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.147Z [INFO] TestCommand_Flags.server.raft: entering leader state: leader="Node at 127.0.0.1:22276 [Leader]" === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.148Z [INFO] TestCommand_File.server.raft: election won: tally=1 === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.148Z [INFO] TestCommand_Flags.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:07.150Z [INFO] TestCommand_Flags.server: New leader elected: payload=Node-b29a106a-fe34-7cfb-694e-da1397be3cf6 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.153Z [WARN] TestCommand_FileWithUnnamedCheck.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.150Z [DEBUG] TestCommand_Flags.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22276 === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.150Z [INFO] TestCommand_File.server.raft: entering leader state: leader="Node at 127.0.0.1:22258 [Leader]" === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.154Z [INFO] TestCommand_FileWithUnnamedCheck.server.raft: entering candidate state: node="Node at 127.0.0.1:22270 [Candidate]" term=2 === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.155Z [INFO] TestCommand_File.server: cluster leadership acquired === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.156Z [INFO] TestCommand_Flags_TaggedAddresses.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:d40a0ef7-a7e6-3111-6043-b88526baf8c5 Address:127.0.0.1:22264}]" === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.157Z [INFO] TestCommand_File.server: New leader elected: payload=Node-25a664de-1922-871b-a28d-95c7031396e7 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.158Z [DEBUG] TestCommand_FileWithUnnamedCheck.server.raft: votes: needed=1 === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.158Z [INFO] TestCommand_Flags_TaggedAddresses.server.raft: entering follower state: follower="Node at 127.0.0.1:22264 [Follower]" leader= === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.157Z [DEBUG] TestCommand_File.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22258 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.160Z [DEBUG] TestCommand_FileWithUnnamedCheck.server.raft: vote granted: from=ef8b0aad-685f-4b57-c5e8-bcf4cc96b21d term=2 tally=1 === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.163Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.163Z [INFO] TestCommand_FileWithUnnamedCheck.server.raft: election won: tally=1 === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.164Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.165Z [INFO] TestCommand_FileWithUnnamedCheck.server.raft: entering leader state: leader="Node at 127.0.0.1:22270 [Leader]" writer.go:29: 2021-01-29T19:35:07.168Z [INFO] TestCommand_FileWithUnnamedCheck.server: cluster leadership acquired === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.161Z [INFO] TestCommand_Flags_TaggedAddresses.server.serf.wan: serf: EventMemberJoin: Node-d40a0ef7-a7e6-3111-6043-b88526baf8c5.dc1 127.0.0.1 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.171Z [INFO] TestCommand_FileWithUnnamedCheck.server: New leader elected: payload=Node-ef8b0aad-685f-4b57-c5e8-bcf4cc96b21d writer.go:29: 2021-01-29T19:35:07.171Z [DEBUG] TestCommand_FileWithUnnamedCheck.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22270 === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.196Z [WARN] TestCommand_Flags_TaggedAddresses.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:07.197Z [INFO] TestCommand_Flags_TaggedAddresses.server.raft: entering candidate state: node="Node at 127.0.0.1:22264 [Candidate]" term=2 writer.go:29: 2021-01-29T19:35:07.199Z [WARN] TestCommand_Flags_TaggedAddresses.server.raft: unable to get address for sever, using fallback address: id=d40a0ef7-a7e6-3111-6043-b88526baf8c5 fallback=127.0.0.1:22264 error="Could not find address for server id d40a0ef7-a7e6-3111-6043-b88526baf8c5" writer.go:29: 2021-01-29T19:35:07.202Z [INFO] TestCommand_Flags_TaggedAddresses.server.serf.lan: serf: EventMemberJoin: Node-d40a0ef7-a7e6-3111-6043-b88526baf8c5 127.0.0.1 writer.go:29: 2021-01-29T19:35:07.203Z [DEBUG] TestCommand_Flags_TaggedAddresses.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:07.205Z [DEBUG] TestCommand_Flags_TaggedAddresses.server.raft: vote granted: from=d40a0ef7-a7e6-3111-6043-b88526baf8c5 term=2 tally=1 writer.go:29: 2021-01-29T19:35:07.206Z [INFO] TestCommand_Flags_TaggedAddresses.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:07.208Z [INFO] TestCommand_Flags_TaggedAddresses.server.raft: entering leader state: leader="Node at 127.0.0.1:22264 [Leader]" writer.go:29: 2021-01-29T19:35:07.205Z [INFO] TestCommand_Flags_TaggedAddresses.server: Handled event for server in area: event=member-join server=Node-d40a0ef7-a7e6-3111-6043-b88526baf8c5.dc1 area=wan === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.244Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.205Z [INFO] TestCommand_Flags_TaggedAddresses.server: Adding LAN server: server="Node-d40a0ef7-a7e6-3111-6043-b88526baf8c5 (Addr: tcp/127.0.0.1:22264) (DC: dc1)" writer.go:29: 2021-01-29T19:35:07.259Z [INFO] TestCommand_Flags_TaggedAddresses.server: New leader elected: payload=Node-d40a0ef7-a7e6-3111-6043-b88526baf8c5 writer.go:29: 2021-01-29T19:35:07.205Z [INFO] TestCommand_Flags_TaggedAddresses: Started DNS server: address=127.0.0.1:22259 network=udp writer.go:29: 2021-01-29T19:35:07.261Z [INFO] TestCommand_Flags_TaggedAddresses: Started DNS server: address=127.0.0.1:22259 network=tcp writer.go:29: 2021-01-29T19:35:07.240Z [INFO] TestCommand_Flags_TaggedAddresses.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:07.284Z [INFO] TestCommand_Flags_TaggedAddresses: Started HTTP server: address=127.0.0.1:22260 network=tcp writer.go:29: 2021-01-29T19:35:07.286Z [INFO] TestCommand_Flags_TaggedAddresses: started state syncer writer.go:29: 2021-01-29T19:35:07.311Z [DEBUG] TestCommand_Flags_TaggedAddresses.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:22264 writer.go:29: 2021-01-29T19:35:07.312Z [INFO] TestCommand_Flags_TaggedAddresses: Synced node info writer.go:29: 2021-01-29T19:35:07.335Z [DEBUG] TestCommand_Flags_TaggedAddresses: Node info in sync === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.318Z [INFO] TestCommand_File: Synced node info writer.go:29: 2021-01-29T19:35:07.337Z [INFO] TestCommand_File.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:07.343Z [INFO] TestCommand_File.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.347Z [DEBUG] TestCommand_File.server: Skipping self join check for node since the cluster is too small: node=Node-25a664de-1922-871b-a28d-95c7031396e7 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.352Z [INFO] TestCommand_FileWithUnnamedCheck: Synced node info writer.go:29: 2021-01-29T19:35:07.354Z [INFO] TestCommand_FileWithUnnamedCheck.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.350Z [INFO] TestCommand_File.server: member joined, marking health alive: member=Node-25a664de-1922-871b-a28d-95c7031396e7 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.355Z [INFO] TestCommand_FileWithUnnamedCheck.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.357Z [DEBUG] TestCommand_FileWithUnnamedCheck.server: Skipping self join check for node since the cluster is too small: node=Node-ef8b0aad-685f-4b57-c5e8-bcf4cc96b21d writer.go:29: 2021-01-29T19:35:07.358Z [INFO] TestCommand_FileWithUnnamedCheck.server: member joined, marking health alive: member=Node-ef8b0aad-685f-4b57-c5e8-bcf4cc96b21d === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.362Z [INFO] TestCommand_Flags.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.362Z [INFO] TestCommand_FileWithUnnamedCheck: Synced service: service=web === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.363Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.363Z [INFO] TestCommand_Flags.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.364Z [DEBUG] TestCommand_Flags.server: Skipping self join check for node since the cluster is too small: node=Node-b29a106a-fe34-7cfb-694e-da1397be3cf6 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.364Z [DEBUG] TestCommand_FileWithUnnamedCheck: Check in sync: check=service:web === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.365Z [INFO] TestCommand_Flags.server: member joined, marking health alive: member=Node-b29a106a-fe34-7cfb-694e-da1397be3cf6 === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.365Z [DEBUG] TestCommand_FileWithUnnamedCheck.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:37260 latency=45.162978ms writer.go:29: 2021-01-29T19:35:07.367Z [DEBUG] TestCommand_FileWithUnnamedCheck: Node info in sync === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.369Z [INFO] TestCommand_Flags: Synced node info === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.371Z [INFO] TestCommand_FileWithUnnamedCheck: Synced service: service=web writer.go:29: 2021-01-29T19:35:07.373Z [DEBUG] TestCommand_FileWithUnnamedCheck: Check in sync: check=service:web === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.373Z [INFO] TestCommand_Flags_TaggedAddresses: Synced service: service=web === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.375Z [DEBUG] TestCommand_FileWithUnnamedCheck.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:37264 latency=645.642µs === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.377Z [DEBUG] TestCommand_Flags_TaggedAddresses.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:53974 latency=54.079117ms === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.384Z [DEBUG] TestCommand_FileWithUnnamedCheck.http: Request finished: method=GET url=/v1/agent/checks from=127.0.0.1:37264 latency=1.497623ms === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.390Z [INFO] TestCommand_Flags: Synced service: service=web === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.392Z [DEBUG] TestCommand_Flags_TaggedAddresses.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:53978 latency=3.371231ms === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.394Z [DEBUG] TestCommand_Flags.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:36010 latency=63.187992ms === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.421Z [INFO] TestCommand_FileWithUnnamedCheck: Requesting shutdown writer.go:29: 2021-01-29T19:35:07.423Z [INFO] TestCommand_FileWithUnnamedCheck.server: shutting down server writer.go:29: 2021-01-29T19:35:07.424Z [DEBUG] TestCommand_FileWithUnnamedCheck.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.425Z [WARN] TestCommand_FileWithUnnamedCheck.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:07.428Z [DEBUG] TestCommand_FileWithUnnamedCheck.leader: stopped routine: routine="CA root pruning" === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.432Z [DEBUG] TestCommand_Flags.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:36020 latency=251.696µs === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.439Z [WARN] TestCommand_FileWithUnnamedCheck.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:07.444Z [INFO] TestCommand_FileWithUnnamedCheck.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:07.448Z [INFO] TestCommand_FileWithUnnamedCheck: consul server down writer.go:29: 2021-01-29T19:35:07.454Z [INFO] TestCommand_FileWithUnnamedCheck: shutdown complete writer.go:29: 2021-01-29T19:35:07.471Z [INFO] TestCommand_FileWithUnnamedCheck: Stopping server: protocol=DNS address=127.0.0.1:22265 network=tcp === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.444Z [INFO] TestCommand_Flags_TaggedAddresses.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.473Z [INFO] TestCommand_FileWithUnnamedCheck: Stopping server: protocol=DNS address=127.0.0.1:22265 network=udp === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.476Z [INFO] TestCommand_Flags: Requesting shutdown writer.go:29: 2021-01-29T19:35:07.478Z [INFO] TestCommand_Flags.server: shutting down server writer.go:29: 2021-01-29T19:35:07.479Z [DEBUG] TestCommand_Flags.leader: stopping routine: routine="CA root pruning" === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.474Z [INFO] TestCommand_Flags_TaggedAddresses.leader: started routine: routine="CA root pruning" === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.481Z [WARN] TestCommand_Flags.server.serf.lan: serf: Shutdown without a Leave === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.479Z [INFO] TestCommand_FileWithUnnamedCheck: Stopping server: protocol=HTTP address=127.0.0.1:22266 network=tcp === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.494Z [DEBUG] TestCommand_Flags_TaggedAddresses.server: Skipping self join check for node since the cluster is too small: node=Node-d40a0ef7-a7e6-3111-6043-b88526baf8c5 writer.go:29: 2021-01-29T19:35:07.495Z [INFO] TestCommand_Flags_TaggedAddresses.server: member joined, marking health alive: member=Node-d40a0ef7-a7e6-3111-6043-b88526baf8c5 === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.483Z [DEBUG] TestCommand_Flags.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.484Z [ERROR] TestCommand_Flags.anti_entropy: failed to sync remote state: error="No cluster leader" === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.493Z [INFO] TestCommand_Flags_TaggedAddresses: Requesting shutdown writer.go:29: 2021-01-29T19:35:07.500Z [INFO] TestCommand_Flags_TaggedAddresses.server: shutting down server writer.go:29: 2021-01-29T19:35:07.501Z [DEBUG] TestCommand_Flags_TaggedAddresses.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.503Z [WARN] TestCommand_Flags_TaggedAddresses.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:07.514Z [DEBUG] TestCommand_Flags_TaggedAddresses.leader: stopped routine: routine="CA root pruning" === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.520Z [WARN] TestCommand_Flags.server.serf.wan: serf: Shutdown without a Leave === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.523Z [WARN] TestCommand_Flags_TaggedAddresses.server.serf.wan: serf: Shutdown without a Leave === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.521Z [DEBUG] TestCommand_File: Node info in sync === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.528Z [DEBUG] TestCommand_Flags_TaggedAddresses.server: Skipping self join check for node since the cluster is too small: node=Node-d40a0ef7-a7e6-3111-6043-b88526baf8c5 writer.go:29: 2021-01-29T19:35:07.541Z [INFO] TestCommand_Flags_TaggedAddresses.server.router.manager: shutting down === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.541Z [INFO] TestCommand_Flags.server.router.manager: shutting down === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.542Z [INFO] TestCommand_Flags_TaggedAddresses: consul server down writer.go:29: 2021-01-29T19:35:07.543Z [INFO] TestCommand_Flags_TaggedAddresses: shutdown complete === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.543Z [INFO] TestCommand_Flags: consul server down writer.go:29: 2021-01-29T19:35:07.545Z [INFO] TestCommand_Flags: shutdown complete writer.go:29: 2021-01-29T19:35:07.552Z [INFO] TestCommand_Flags: Stopping server: protocol=DNS address=127.0.0.1:22271 network=tcp === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.548Z [INFO] TestCommand_File: Synced service: service=web === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.544Z [INFO] TestCommand_Flags_TaggedAddresses: Stopping server: protocol=DNS address=127.0.0.1:22259 network=tcp === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.553Z [DEBUG] TestCommand_File.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:48768 latency=66.969634ms === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.553Z [INFO] TestCommand_Flags: Stopping server: protocol=DNS address=127.0.0.1:22271 network=udp === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.554Z [INFO] TestCommand_Flags_TaggedAddresses: Stopping server: protocol=DNS address=127.0.0.1:22259 network=udp === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:07.555Z [INFO] TestCommand_Flags: Stopping server: protocol=HTTP address=127.0.0.1:22272 network=tcp === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:07.555Z [INFO] TestCommand_Flags_TaggedAddresses: Stopping server: protocol=HTTP address=127.0.0.1:22260 network=tcp === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:07.564Z [DEBUG] TestCommand_File.http: Request finished: method=GET url=/v1/agent/services from=127.0.0.1:48774 latency=133.418µs writer.go:29: 2021-01-29T19:35:07.567Z [INFO] TestCommand_File: Requesting shutdown writer.go:29: 2021-01-29T19:35:07.568Z [INFO] TestCommand_File.server: shutting down server writer.go:29: 2021-01-29T19:35:07.568Z [DEBUG] TestCommand_File.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.569Z [WARN] TestCommand_File.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:07.567Z [DEBUG] TestCommand_File: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:35:07.571Z [DEBUG] TestCommand_File: Node info in sync writer.go:29: 2021-01-29T19:35:07.573Z [DEBUG] TestCommand_File: Service in sync: service=web writer.go:29: 2021-01-29T19:35:07.572Z [DEBUG] TestCommand_File.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:07.573Z [WARN] TestCommand_File.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:07.576Z [DEBUG] TestCommand_File: Node info in sync writer.go:29: 2021-01-29T19:35:07.576Z [DEBUG] TestCommand_File: Service in sync: service=web writer.go:29: 2021-01-29T19:35:07.578Z [INFO] TestCommand_File.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:07.579Z [INFO] TestCommand_File: consul server down writer.go:29: 2021-01-29T19:35:07.580Z [INFO] TestCommand_File: shutdown complete writer.go:29: 2021-01-29T19:35:07.581Z [INFO] TestCommand_File: Stopping server: protocol=DNS address=127.0.0.1:22253 network=tcp writer.go:29: 2021-01-29T19:35:07.582Z [INFO] TestCommand_File: Stopping server: protocol=DNS address=127.0.0.1:22253 network=udp writer.go:29: 2021-01-29T19:35:07.583Z [INFO] TestCommand_File: Stopping server: protocol=HTTP address=127.0.0.1:22254 network=tcp === CONT TestCommand_FileWithUnnamedCheck writer.go:29: 2021-01-29T19:35:07.995Z [INFO] TestCommand_FileWithUnnamedCheck: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:07.996Z [INFO] TestCommand_FileWithUnnamedCheck: Endpoints down --- PASS: TestCommand_FileWithUnnamedCheck (1.03s) === CONT TestCommand_Flags writer.go:29: 2021-01-29T19:35:08.056Z [INFO] TestCommand_Flags: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:08.057Z [INFO] TestCommand_Flags: Endpoints down --- PASS: TestCommand_Flags (1.08s) === CONT TestCommand_Flags_TaggedAddresses writer.go:29: 2021-01-29T19:35:08.057Z [INFO] TestCommand_Flags_TaggedAddresses: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:08.058Z [INFO] TestCommand_Flags_TaggedAddresses: Endpoints down --- PASS: TestCommand_Flags_TaggedAddresses (1.09s) === CONT TestCommand_File writer.go:29: 2021-01-29T19:35:08.084Z [INFO] TestCommand_File: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:08.085Z [INFO] TestCommand_File: Endpoints down --- PASS: TestCommand_File (1.12s) PASS ok github.com/hashicorp/consul/command/services/register 1.368s === RUN TestSnapshotCommand_noTabs === PAUSE TestSnapshotCommand_noTabs === CONT TestSnapshotCommand_noTabs --- PASS: TestSnapshotCommand_noTabs (0.01s) PASS ok github.com/hashicorp/consul/command/snapshot 0.018s === RUN TestSnapshotInspectCommand_noTabs === PAUSE TestSnapshotInspectCommand_noTabs === RUN TestSnapshotInspectCommand_Validation === PAUSE TestSnapshotInspectCommand_Validation === RUN TestSnapshotInspectCommand === PAUSE TestSnapshotInspectCommand === CONT TestSnapshotInspectCommand_noTabs === CONT TestSnapshotInspectCommand [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... === CONT TestSnapshotInspectCommand_Validation --- PASS: TestSnapshotInspectCommand_Validation (0.00s) --- PASS: TestSnapshotInspectCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestSnapshotInspectCommand writer.go:29: 2021-01-29T19:35:08.311Z [WARN] TestSnapshotInspectCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:08.317Z [DEBUG] TestSnapshotInspectCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:08.319Z [DEBUG] TestSnapshotInspectCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:35:08.338Z [INFO] TestSnapshotInspectCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:188bee4a-472b-f4fc-2cad-3a6300ea1b2c Address:127.0.0.1:29405}]" writer.go:29: 2021-01-29T19:35:08.340Z [INFO] TestSnapshotInspectCommand.server.serf.wan: serf: EventMemberJoin: Node-188bee4a-472b-f4fc-2cad-3a6300ea1b2c.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:35:08.341Z [INFO] TestSnapshotInspectCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:29405 [Follower]" leader= writer.go:29: 2021-01-29T19:35:08.343Z [INFO] TestSnapshotInspectCommand.server.serf.lan: serf: EventMemberJoin: Node-188bee4a-472b-f4fc-2cad-3a6300ea1b2c 127.0.0.1 writer.go:29: 2021-01-29T19:35:08.354Z [INFO] TestSnapshotInspectCommand.server: Adding LAN server: server="Node-188bee4a-472b-f4fc-2cad-3a6300ea1b2c (Addr: tcp/127.0.0.1:29405) (DC: dc1)" writer.go:29: 2021-01-29T19:35:08.354Z [INFO] TestSnapshotInspectCommand.server: Handled event for server in area: event=member-join server=Node-188bee4a-472b-f4fc-2cad-3a6300ea1b2c.dc1 area=wan writer.go:29: 2021-01-29T19:35:08.356Z [INFO] TestSnapshotInspectCommand: Started DNS server: address=127.0.0.1:29400 network=udp writer.go:29: 2021-01-29T19:35:08.358Z [INFO] TestSnapshotInspectCommand: Started DNS server: address=127.0.0.1:29400 network=tcp writer.go:29: 2021-01-29T19:35:08.360Z [INFO] TestSnapshotInspectCommand: Started HTTP server: address=127.0.0.1:29401 network=tcp writer.go:29: 2021-01-29T19:35:08.362Z [INFO] TestSnapshotInspectCommand: started state syncer writer.go:29: 2021-01-29T19:35:08.395Z [WARN] TestSnapshotInspectCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:08.397Z [INFO] TestSnapshotInspectCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:29405 [Candidate]" term=2 writer.go:29: 2021-01-29T19:35:08.406Z [DEBUG] TestSnapshotInspectCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:08.407Z [DEBUG] TestSnapshotInspectCommand.server.raft: vote granted: from=188bee4a-472b-f4fc-2cad-3a6300ea1b2c term=2 tally=1 writer.go:29: 2021-01-29T19:35:08.409Z [INFO] TestSnapshotInspectCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:08.410Z [INFO] TestSnapshotInspectCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:29405 [Leader]" writer.go:29: 2021-01-29T19:35:08.412Z [INFO] TestSnapshotInspectCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:08.413Z [DEBUG] TestSnapshotInspectCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:29405 writer.go:29: 2021-01-29T19:35:08.413Z [INFO] TestSnapshotInspectCommand.server: New leader elected: payload=Node-188bee4a-472b-f4fc-2cad-3a6300ea1b2c writer.go:29: 2021-01-29T19:35:08.439Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:35:08.482Z [INFO] TestSnapshotInspectCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:08.483Z [INFO] TestSnapshotInspectCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:08.485Z [DEBUG] TestSnapshotInspectCommand.server: Skipping self join check for node since the cluster is too small: node=Node-188bee4a-472b-f4fc-2cad-3a6300ea1b2c writer.go:29: 2021-01-29T19:35:08.487Z [INFO] TestSnapshotInspectCommand.server: member joined, marking health alive: member=Node-188bee4a-472b-f4fc-2cad-3a6300ea1b2c writer.go:29: 2021-01-29T19:35:08.713Z [DEBUG] TestSnapshotInspectCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:35:08.717Z [INFO] TestSnapshotInspectCommand: Synced node info writer.go:29: 2021-01-29T19:35:08.718Z [DEBUG] TestSnapshotInspectCommand: Node info in sync writer.go:29: 2021-01-29T19:35:08.722Z [INFO] TestSnapshotInspectCommand.server.fsm: snapshot created: duration=48.404µs writer.go:29: 2021-01-29T19:35:08.724Z [INFO] TestSnapshotInspectCommand.server.raft: starting snapshot up to: index=11 writer.go:29: 2021-01-29T19:35:08.727Z [INFO] snapshot: creating new snapshot: path=/tmp/TestSnapshotInspectCommand-agent844086364/raft/snapshots/2-11-1611948908727.tmp writer.go:29: 2021-01-29T19:35:08.732Z [INFO] TestSnapshotInspectCommand.server.raft: snapshot complete up to: index=11 writer.go:29: 2021-01-29T19:35:08.763Z [DEBUG] TestSnapshotInspectCommand.http: Request finished: method=GET url=/v1/snapshot from=127.0.0.1:38208 latency=41.331343ms writer.go:29: 2021-01-29T19:35:08.770Z [INFO] TestSnapshotInspectCommand: Requesting shutdown writer.go:29: 2021-01-29T19:35:08.771Z [INFO] TestSnapshotInspectCommand.server: shutting down server writer.go:29: 2021-01-29T19:35:08.772Z [DEBUG] TestSnapshotInspectCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:08.773Z [WARN] TestSnapshotInspectCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:08.773Z [DEBUG] TestSnapshotInspectCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:08.775Z [WARN] TestSnapshotInspectCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:08.783Z [INFO] TestSnapshotInspectCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:08.784Z [INFO] TestSnapshotInspectCommand: consul server down writer.go:29: 2021-01-29T19:35:08.785Z [INFO] TestSnapshotInspectCommand: shutdown complete writer.go:29: 2021-01-29T19:35:08.786Z [INFO] TestSnapshotInspectCommand: Stopping server: protocol=DNS address=127.0.0.1:29400 network=tcp writer.go:29: 2021-01-29T19:35:08.787Z [INFO] TestSnapshotInspectCommand: Stopping server: protocol=DNS address=127.0.0.1:29400 network=udp writer.go:29: 2021-01-29T19:35:08.789Z [INFO] TestSnapshotInspectCommand: Stopping server: protocol=HTTP address=127.0.0.1:29401 network=tcp writer.go:29: 2021-01-29T19:35:09.290Z [INFO] TestSnapshotInspectCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:09.292Z [INFO] TestSnapshotInspectCommand: Endpoints down --- PASS: TestSnapshotInspectCommand (1.03s) PASS ok github.com/hashicorp/consul/command/snapshot/inspect 1.291s === RUN TestSnapshotRestoreCommand_noTabs === PAUSE TestSnapshotRestoreCommand_noTabs === RUN TestSnapshotRestoreCommand_Validation === PAUSE TestSnapshotRestoreCommand_Validation === RUN TestSnapshotRestoreCommand === PAUSE TestSnapshotRestoreCommand === RUN TestSnapshotRestoreCommand_TruncatedSnapshot === PAUSE TestSnapshotRestoreCommand_TruncatedSnapshot === CONT TestSnapshotRestoreCommand_noTabs === CONT TestSnapshotRestoreCommand_TruncatedSnapshot === CONT TestSnapshotRestoreCommand === CONT TestSnapshotRestoreCommand_Validation --- PASS: TestSnapshotRestoreCommand_Validation (0.00s) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... --- PASS: TestSnapshotRestoreCommand_noTabs (0.00s) [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.914Z [WARN] TestSnapshotRestoreCommand_TruncatedSnapshot: bootstrap = true: do not enable unless necessary === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.914Z [WARN] TestSnapshotRestoreCommand: bootstrap = true: do not enable unless necessary === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.916Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.tlsutil: Update: version=1 === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.916Z [DEBUG] TestSnapshotRestoreCommand.tlsutil: Update: version=1 === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.918Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.918Z [DEBUG] TestSnapshotRestoreCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:35:08.924Z [INFO] TestSnapshotRestoreCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:2c844f76-75f6-4386-43b3-3957b3a093e3 Address:127.0.0.1:30426}]" writer.go:29: 2021-01-29T19:35:08.926Z [INFO] TestSnapshotRestoreCommand.server.serf.wan: serf: EventMemberJoin: Node-2c844f76-75f6-4386-43b3-3957b3a093e3.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:35:08.926Z [INFO] TestSnapshotRestoreCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:30426 [Follower]" leader= writer.go:29: 2021-01-29T19:35:08.929Z [INFO] TestSnapshotRestoreCommand.server.serf.lan: serf: EventMemberJoin: Node-2c844f76-75f6-4386-43b3-3957b3a093e3 127.0.0.1 === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.931Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:713b8310-4be9-f942-2f77-53cb9204fa22 Address:127.0.0.1:30432}]" === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.935Z [INFO] TestSnapshotRestoreCommand: Started DNS server: address=127.0.0.1:30421 network=udp === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.935Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.serf.wan: serf: EventMemberJoin: Node-713b8310-4be9-f942-2f77-53cb9204fa22.dc1 127.0.0.1 === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.943Z [INFO] TestSnapshotRestoreCommand.server: Adding LAN server: server="Node-2c844f76-75f6-4386-43b3-3957b3a093e3 (Addr: tcp/127.0.0.1:30426) (DC: dc1)" === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.940Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.serf.lan: serf: EventMemberJoin: Node-713b8310-4be9-f942-2f77-53cb9204fa22 127.0.0.1 === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.944Z [INFO] TestSnapshotRestoreCommand.server: Handled event for server in area: event=member-join server=Node-2c844f76-75f6-4386-43b3-3957b3a093e3.dc1 area=wan writer.go:29: 2021-01-29T19:35:08.948Z [INFO] TestSnapshotRestoreCommand: Started DNS server: address=127.0.0.1:30421 network=tcp === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.943Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.raft: entering follower state: follower="Node at 127.0.0.1:30432 [Follower]" leader= === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.955Z [INFO] TestSnapshotRestoreCommand: Started HTTP server: address=127.0.0.1:30422 network=tcp === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.949Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: Started DNS server: address=127.0.0.1:30427 network=udp writer.go:29: 2021-01-29T19:35:08.966Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: Started DNS server: address=127.0.0.1:30427 network=tcp === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.958Z [INFO] TestSnapshotRestoreCommand: started state syncer writer.go:29: 2021-01-29T19:35:08.975Z [WARN] TestSnapshotRestoreCommand.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.950Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server: Handled event for server in area: event=member-join server=Node-713b8310-4be9-f942-2f77-53cb9204fa22.dc1 area=wan === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.977Z [INFO] TestSnapshotRestoreCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:30426 [Candidate]" term=2 === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.980Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: Started HTTP server: address=127.0.0.1:30428 network=tcp writer.go:29: 2021-01-29T19:35:08.982Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: started state syncer === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.981Z [DEBUG] TestSnapshotRestoreCommand.server.raft: votes: needed=1 === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:08.985Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server: Adding LAN server: server="Node-713b8310-4be9-f942-2f77-53cb9204fa22 (Addr: tcp/127.0.0.1:30432) (DC: dc1)" === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:08.985Z [DEBUG] TestSnapshotRestoreCommand.server.raft: vote granted: from=2c844f76-75f6-4386-43b3-3957b3a093e3 term=2 tally=1 writer.go:29: 2021-01-29T19:35:08.987Z [INFO] TestSnapshotRestoreCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:08.989Z [INFO] TestSnapshotRestoreCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:30426 [Leader]" writer.go:29: 2021-01-29T19:35:08.990Z [INFO] TestSnapshotRestoreCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:08.991Z [INFO] TestSnapshotRestoreCommand.server: New leader elected: payload=Node-2c844f76-75f6-4386-43b3-3957b3a093e3 writer.go:29: 2021-01-29T19:35:08.992Z [DEBUG] TestSnapshotRestoreCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30426 writer.go:29: 2021-01-29T19:35:08.997Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.006Z [WARN] TestSnapshotRestoreCommand_TruncatedSnapshot.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:09.008Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.raft: entering candidate state: node="Node at 127.0.0.1:30432 [Candidate]" term=2 writer.go:29: 2021-01-29T19:35:09.011Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:09.013Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.server.raft: vote granted: from=713b8310-4be9-f942-2f77-53cb9204fa22 term=2 tally=1 writer.go:29: 2021-01-29T19:35:09.015Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:09.016Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.raft: entering leader state: leader="Node at 127.0.0.1:30432 [Leader]" writer.go:29: 2021-01-29T19:35:09.019Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:09.019Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server: New leader elected: payload=Node-713b8310-4be9-f942-2f77-53cb9204fa22 writer.go:29: 2021-01-29T19:35:09.023Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30432 === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.026Z [INFO] TestSnapshotRestoreCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.029Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.029Z [INFO] TestSnapshotRestoreCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:09.032Z [DEBUG] TestSnapshotRestoreCommand.server: Skipping self join check for node since the cluster is too small: node=Node-2c844f76-75f6-4386-43b3-3957b3a093e3 === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.035Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.035Z [INFO] TestSnapshotRestoreCommand.server: member joined, marking health alive: member=Node-2c844f76-75f6-4386-43b3-3957b3a093e3 === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.036Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:09.038Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.server: Skipping self join check for node since the cluster is too small: node=Node-713b8310-4be9-f942-2f77-53cb9204fa22 writer.go:29: 2021-01-29T19:35:09.040Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server: member joined, marking health alive: member=Node-713b8310-4be9-f942-2f77-53cb9204fa22 === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.168Z [DEBUG] TestSnapshotRestoreCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:35:09.170Z [INFO] TestSnapshotRestoreCommand: Synced node info writer.go:29: 2021-01-29T19:35:09.365Z [INFO] TestSnapshotRestoreCommand.server.fsm: snapshot created: duration=61.348µs writer.go:29: 2021-01-29T19:35:09.366Z [INFO] TestSnapshotRestoreCommand.server.raft: starting snapshot up to: index=11 writer.go:29: 2021-01-29T19:35:09.371Z [INFO] snapshot: creating new snapshot: path=/tmp/TestSnapshotRestoreCommand-agent331314210/raft/snapshots/2-11-1611948909371.tmp writer.go:29: 2021-01-29T19:35:09.380Z [INFO] TestSnapshotRestoreCommand.server.raft: snapshot complete up to: index=11 writer.go:29: 2021-01-29T19:35:09.389Z [DEBUG] TestSnapshotRestoreCommand.http: Request finished: method=GET url=/v1/snapshot from=127.0.0.1:57240 latency=25.078181ms writer.go:29: 2021-01-29T19:35:09.400Z [INFO] snapshot: creating new snapshot: path=/tmp/TestSnapshotRestoreCommand-agent331314210/raft/snapshots/2-12-1611948909400.tmp === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.409Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:35:09.411Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: Synced node info === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.404Z [INFO] TestSnapshotRestoreCommand.server.raft: copied to local snapshot: bytes=4882 writer.go:29: 2021-01-29T19:35:09.439Z [INFO] TestSnapshotRestoreCommand.server.raft: restored user snapshot: index=1 writer.go:29: 2021-01-29T19:35:09.456Z [DEBUG] TestSnapshotRestoreCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:09.463Z [DEBUG] TestSnapshotRestoreCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:30426 === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.463Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request finished: method=PUT url=/v1/kv/blob from=127.0.0.1:48656 latency=36.403679ms === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.466Z [INFO] TestSnapshotRestoreCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:09.467Z [DEBUG] TestSnapshotRestoreCommand.http: Request finished: method=PUT url=/v1/snapshot from=127.0.0.1:57242 latency=70.182169ms === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.468Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.fsm: snapshot created: duration=55.601µs === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.468Z [DEBUG] TestSnapshotRestoreCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:09.469Z [INFO] TestSnapshotRestoreCommand: Requesting shutdown === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.470Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.raft: starting snapshot up to: index=12 === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.473Z [INFO] TestSnapshotRestoreCommand.server: shutting down server === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.474Z [INFO] snapshot: creating new snapshot: path=/tmp/TestSnapshotRestoreCommand_TruncatedSnapshot-agent617098009/raft/snapshots/2-12-1611948909474.tmp === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.475Z [DEBUG] TestSnapshotRestoreCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:09.478Z [WARN] TestSnapshotRestoreCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:09.478Z [DEBUG] TestSnapshotRestoreCommand.leader: stopped routine: routine="CA root pruning" === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.485Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.raft: snapshot complete up to: index=12 === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.484Z [WARN] TestSnapshotRestoreCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:09.494Z [INFO] TestSnapshotRestoreCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:09.494Z [INFO] TestSnapshotRestoreCommand: consul server down writer.go:29: 2021-01-29T19:35:09.500Z [INFO] TestSnapshotRestoreCommand: shutdown complete writer.go:29: 2021-01-29T19:35:09.502Z [INFO] TestSnapshotRestoreCommand: Stopping server: protocol=DNS address=127.0.0.1:30421 network=tcp === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.503Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request finished: method=GET url=/v1/snapshot from=127.0.0.1:48656 latency=35.13343ms === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:09.504Z [INFO] TestSnapshotRestoreCommand: Stopping server: protocol=DNS address=127.0.0.1:30421 network=udp writer.go:29: 2021-01-29T19:35:09.506Z [INFO] TestSnapshotRestoreCommand: Stopping server: protocol=HTTP address=127.0.0.1:30422 network=tcp === RUN TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_200_bytes_from_end === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.528Z [ERROR] TestSnapshotRestoreCommand_TruncatedSnapshot.server: Failed to close snapshot decompressor: error="unexpected EOF" writer.go:29: 2021-01-29T19:35:09.529Z [ERROR] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request error: method=PUT url=/v1/snapshot from=127.0.0.1:48658 error="failed to read snapshot file: failed reading snapshot: unexpected EOF" writer.go:29: 2021-01-29T19:35:09.531Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request finished: method=PUT url=/v1/snapshot from=127.0.0.1:48658 latency=4.73285ms === RUN TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_16_bytes_from_end === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.546Z [ERROR] TestSnapshotRestoreCommand_TruncatedSnapshot.server: Failed to close snapshot decompressor: error="unexpected EOF" writer.go:29: 2021-01-29T19:35:09.550Z [ERROR] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request error: method=PUT url=/v1/snapshot from=127.0.0.1:48660 error="failed to read snapshot file: failed reading snapshot: unexpected EOF" writer.go:29: 2021-01-29T19:35:09.559Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request finished: method=PUT url=/v1/snapshot from=127.0.0.1:48660 latency=18.701597ms === RUN TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_8_bytes_from_end === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.576Z [ERROR] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request error: method=PUT url=/v1/snapshot from=127.0.0.1:48662 error="unexpected EOF" writer.go:29: 2021-01-29T19:35:09.577Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request finished: method=PUT url=/v1/snapshot from=127.0.0.1:48662 latency=5.148414ms === RUN TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_4_bytes_from_end === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.677Z [ERROR] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request error: method=PUT url=/v1/snapshot from=127.0.0.1:48664 error="unexpected EOF" writer.go:29: 2021-01-29T19:35:09.687Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request finished: method=PUT url=/v1/snapshot from=127.0.0.1:48664 latency=63.280934ms === RUN TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_2_bytes_from_end === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.765Z [ERROR] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request error: method=PUT url=/v1/snapshot from=127.0.0.1:48666 error="unexpected EOF" writer.go:29: 2021-01-29T19:35:09.770Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request finished: method=PUT url=/v1/snapshot from=127.0.0.1:48666 latency=10.58749ms === RUN TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_1_bytes_from_end === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:09.806Z [ERROR] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request error: method=PUT url=/v1/snapshot from=127.0.0.1:48668 error="unexpected EOF" writer.go:29: 2021-01-29T19:35:09.808Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.http: Request finished: method=PUT url=/v1/snapshot from=127.0.0.1:48668 latency=4.806053ms writer.go:29: 2021-01-29T19:35:09.811Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: Requesting shutdown writer.go:29: 2021-01-29T19:35:09.812Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server: shutting down server writer.go:29: 2021-01-29T19:35:09.813Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:09.814Z [WARN] TestSnapshotRestoreCommand_TruncatedSnapshot.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:09.815Z [DEBUG] TestSnapshotRestoreCommand_TruncatedSnapshot.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:09.817Z [WARN] TestSnapshotRestoreCommand_TruncatedSnapshot.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:09.819Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:09.819Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: consul server down writer.go:29: 2021-01-29T19:35:09.823Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: shutdown complete writer.go:29: 2021-01-29T19:35:09.825Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: Stopping server: protocol=DNS address=127.0.0.1:30427 network=tcp writer.go:29: 2021-01-29T19:35:09.827Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: Stopping server: protocol=DNS address=127.0.0.1:30427 network=udp writer.go:29: 2021-01-29T19:35:09.829Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: Stopping server: protocol=HTTP address=127.0.0.1:30428 network=tcp === CONT TestSnapshotRestoreCommand writer.go:29: 2021-01-29T19:35:10.008Z [INFO] TestSnapshotRestoreCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:10.011Z [INFO] TestSnapshotRestoreCommand: Endpoints down --- PASS: TestSnapshotRestoreCommand (1.16s) === CONT TestSnapshotRestoreCommand_TruncatedSnapshot writer.go:29: 2021-01-29T19:35:10.331Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:10.333Z [INFO] TestSnapshotRestoreCommand_TruncatedSnapshot: Endpoints down --- PASS: TestSnapshotRestoreCommand_TruncatedSnapshot (1.48s) --- PASS: TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_200_bytes_from_end (0.02s) --- PASS: TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_16_bytes_from_end (0.03s) --- PASS: TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_8_bytes_from_end (0.02s) --- PASS: TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_4_bytes_from_end (0.13s) --- PASS: TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_2_bytes_from_end (0.07s) --- PASS: TestSnapshotRestoreCommand_TruncatedSnapshot/truncate_1_bytes_from_end (0.02s) PASS ok github.com/hashicorp/consul/command/snapshot/restore 1.769s === RUN TestSnapshotSaveCommand_noTabs === PAUSE TestSnapshotSaveCommand_noTabs === RUN TestSnapshotSaveCommand_Validation === PAUSE TestSnapshotSaveCommand_Validation === RUN TestSnapshotSaveCommand === PAUSE TestSnapshotSaveCommand === RUN TestSnapshotSaveCommand_TruncatedStream snapshot_save_test.go:106: DM-skipped --- SKIP: TestSnapshotSaveCommand_TruncatedStream (0.00s) === CONT TestSnapshotSaveCommand_noTabs === CONT TestSnapshotSaveCommand === CONT TestSnapshotSaveCommand_Validation --- PASS: TestSnapshotSaveCommand_noTabs (0.00s) --- PASS: TestSnapshotSaveCommand_Validation (0.00s) [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range === CONT TestSnapshotSaveCommand writer.go:29: 2021-01-29T19:35:17.569Z [WARN] TestSnapshotSaveCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:17.570Z [DEBUG] TestSnapshotSaveCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:17.572Z [DEBUG] TestSnapshotSaveCommand.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:35:17.589Z [INFO] TestSnapshotSaveCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:100c6f7b-fe01-aeb6-4e5b-6ff54aef1d43 Address:127.0.0.1:12048}]" writer.go:29: 2021-01-29T19:35:17.591Z [INFO] TestSnapshotSaveCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:12048 [Follower]" leader= writer.go:29: 2021-01-29T19:35:17.593Z [INFO] TestSnapshotSaveCommand.server.serf.wan: serf: EventMemberJoin: Node-100c6f7b-fe01-aeb6-4e5b-6ff54aef1d43.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:35:17.599Z [INFO] TestSnapshotSaveCommand.server.serf.lan: serf: EventMemberJoin: Node-100c6f7b-fe01-aeb6-4e5b-6ff54aef1d43 127.0.0.1 writer.go:29: 2021-01-29T19:35:17.608Z [INFO] TestSnapshotSaveCommand.server: Handled event for server in area: event=member-join server=Node-100c6f7b-fe01-aeb6-4e5b-6ff54aef1d43.dc1 area=wan writer.go:29: 2021-01-29T19:35:17.608Z [INFO] TestSnapshotSaveCommand.server: Adding LAN server: server="Node-100c6f7b-fe01-aeb6-4e5b-6ff54aef1d43 (Addr: tcp/127.0.0.1:12048) (DC: dc1)" writer.go:29: 2021-01-29T19:35:17.613Z [INFO] TestSnapshotSaveCommand: Started DNS server: address=127.0.0.1:12043 network=tcp writer.go:29: 2021-01-29T19:35:17.618Z [INFO] TestSnapshotSaveCommand: Started DNS server: address=127.0.0.1:12043 network=udp writer.go:29: 2021-01-29T19:35:17.624Z [INFO] TestSnapshotSaveCommand: Started HTTP server: address=127.0.0.1:12044 network=tcp writer.go:29: 2021-01-29T19:35:17.631Z [INFO] TestSnapshotSaveCommand: started state syncer writer.go:29: 2021-01-29T19:35:17.657Z [WARN] TestSnapshotSaveCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:17.658Z [INFO] TestSnapshotSaveCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:12048 [Candidate]" term=2 writer.go:29: 2021-01-29T19:35:17.673Z [DEBUG] TestSnapshotSaveCommand.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:17.674Z [DEBUG] TestSnapshotSaveCommand.server.raft: vote granted: from=100c6f7b-fe01-aeb6-4e5b-6ff54aef1d43 term=2 tally=1 writer.go:29: 2021-01-29T19:35:17.675Z [INFO] TestSnapshotSaveCommand.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:17.676Z [INFO] TestSnapshotSaveCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:12048 [Leader]" writer.go:29: 2021-01-29T19:35:17.677Z [INFO] TestSnapshotSaveCommand.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:17.678Z [INFO] TestSnapshotSaveCommand.server: New leader elected: payload=Node-100c6f7b-fe01-aeb6-4e5b-6ff54aef1d43 writer.go:29: 2021-01-29T19:35:17.685Z [DEBUG] TestSnapshotSaveCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:12048 writer.go:29: 2021-01-29T19:35:17.706Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:35:17.750Z [INFO] TestSnapshotSaveCommand.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:17.752Z [INFO] TestSnapshotSaveCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:17.754Z [DEBUG] TestSnapshotSaveCommand.server: Skipping self join check for node since the cluster is too small: node=Node-100c6f7b-fe01-aeb6-4e5b-6ff54aef1d43 writer.go:29: 2021-01-29T19:35:17.756Z [INFO] TestSnapshotSaveCommand.server: member joined, marking health alive: member=Node-100c6f7b-fe01-aeb6-4e5b-6ff54aef1d43 writer.go:29: 2021-01-29T19:35:17.791Z [INFO] TestSnapshotSaveCommand.server.fsm: snapshot created: duration=59.547µs writer.go:29: 2021-01-29T19:35:17.793Z [INFO] TestSnapshotSaveCommand.server.raft: starting snapshot up to: index=10 writer.go:29: 2021-01-29T19:35:17.797Z [INFO] snapshot: creating new snapshot: path=/tmp/TestSnapshotSaveCommand-agent895533355/raft/snapshots/2-10-1611948917797.tmp writer.go:29: 2021-01-29T19:35:17.813Z [INFO] TestSnapshotSaveCommand.server.raft: snapshot complete up to: index=10 writer.go:29: 2021-01-29T19:35:17.820Z [DEBUG] TestSnapshotSaveCommand.http: Request finished: method=GET url=/v1/snapshot from=127.0.0.1:39620 latency=29.06203ms writer.go:29: 2021-01-29T19:35:17.833Z [INFO] snapshot: creating new snapshot: path=/tmp/TestSnapshotSaveCommand-agent895533355/raft/snapshots/2-11-1611948917833.tmp writer.go:29: 2021-01-29T19:35:17.840Z [INFO] TestSnapshotSaveCommand.server.raft: copied to local snapshot: bytes=4612 writer.go:29: 2021-01-29T19:35:17.843Z [INFO] TestSnapshotSaveCommand.server.raft: restored user snapshot: index=1 writer.go:29: 2021-01-29T19:35:17.851Z [DEBUG] TestSnapshotSaveCommand: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:35:17.861Z [DEBUG] TestSnapshotSaveCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:17.863Z [DEBUG] TestSnapshotSaveCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:17.863Z [DEBUG] TestSnapshotSaveCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:12048 writer.go:29: 2021-01-29T19:35:17.861Z [INFO] TestSnapshotSaveCommand: Synced node info writer.go:29: 2021-01-29T19:35:17.866Z [INFO] TestSnapshotSaveCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:17.869Z [DEBUG] TestSnapshotSaveCommand.http: Request finished: method=PUT url=/v1/snapshot from=127.0.0.1:39622 latency=39.223568ms writer.go:29: 2021-01-29T19:35:17.871Z [INFO] TestSnapshotSaveCommand: Requesting shutdown writer.go:29: 2021-01-29T19:35:17.872Z [INFO] TestSnapshotSaveCommand.server: shutting down server writer.go:29: 2021-01-29T19:35:17.873Z [DEBUG] TestSnapshotSaveCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:17.873Z [WARN] TestSnapshotSaveCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:17.875Z [DEBUG] TestSnapshotSaveCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:17.877Z [WARN] TestSnapshotSaveCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:17.879Z [INFO] TestSnapshotSaveCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:17.880Z [INFO] TestSnapshotSaveCommand: consul server down writer.go:29: 2021-01-29T19:35:17.882Z [INFO] TestSnapshotSaveCommand: shutdown complete writer.go:29: 2021-01-29T19:35:17.883Z [INFO] TestSnapshotSaveCommand: Stopping server: protocol=DNS address=127.0.0.1:12043 network=tcp writer.go:29: 2021-01-29T19:35:17.884Z [INFO] TestSnapshotSaveCommand: Stopping server: protocol=DNS address=127.0.0.1:12043 network=udp writer.go:29: 2021-01-29T19:35:17.885Z [INFO] TestSnapshotSaveCommand: Stopping server: protocol=HTTP address=127.0.0.1:12044 network=tcp writer.go:29: 2021-01-29T19:35:18.387Z [INFO] TestSnapshotSaveCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:18.389Z [INFO] TestSnapshotSaveCommand: Endpoints down --- PASS: TestSnapshotSaveCommand (0.91s) PASS ok github.com/hashicorp/consul/command/snapshot/save 1.205s === RUN TestValidateCommand_noTabs === PAUSE TestValidateCommand_noTabs === RUN TestValidateCommand_FailOnEmptyFile === PAUSE TestValidateCommand_FailOnEmptyFile === RUN TestValidateCommand_SucceedOnMinimalConfigFile === PAUSE TestValidateCommand_SucceedOnMinimalConfigFile === RUN TestValidateCommand_SucceedWithMinimalJSONConfigFormat === PAUSE TestValidateCommand_SucceedWithMinimalJSONConfigFormat === RUN TestValidateCommand_SucceedWithMinimalHCLConfigFormat === PAUSE TestValidateCommand_SucceedWithMinimalHCLConfigFormat === RUN TestValidateCommand_SucceedWithJSONAsHCL === PAUSE TestValidateCommand_SucceedWithJSONAsHCL === RUN TestValidateCommand_SucceedOnMinimalConfigDir === PAUSE TestValidateCommand_SucceedOnMinimalConfigDir === RUN TestValidateCommand_FailForInvalidJSONConfigFormat === PAUSE TestValidateCommand_FailForInvalidJSONConfigFormat === RUN TestValidateCommand_Quiet === PAUSE TestValidateCommand_Quiet === CONT TestValidateCommand_noTabs === CONT TestValidateCommand_Quiet === CONT TestValidateCommand_SucceedWithMinimalJSONConfigFormat === CONT TestValidateCommand_SucceedWithJSONAsHCL --- PASS: TestValidateCommand_noTabs (0.02s) === CONT TestValidateCommand_FailForInvalidJSONConfigFormat --- PASS: TestValidateCommand_FailForInvalidJSONConfigFormat (0.01s) === CONT TestValidateCommand_SucceedOnMinimalConfigDir --- PASS: TestValidateCommand_Quiet (0.05s) === CONT TestValidateCommand_SucceedWithMinimalHCLConfigFormat --- PASS: TestValidateCommand_SucceedWithJSONAsHCL (0.05s) === CONT TestValidateCommand_SucceedOnMinimalConfigFile --- PASS: TestValidateCommand_SucceedWithMinimalJSONConfigFormat (0.07s) === CONT TestValidateCommand_FailOnEmptyFile --- PASS: TestValidateCommand_SucceedOnMinimalConfigFile (0.02s) --- PASS: TestValidateCommand_SucceedWithMinimalHCLConfigFormat (0.04s) --- PASS: TestValidateCommand_SucceedOnMinimalConfigDir (0.09s) --- PASS: TestValidateCommand_FailOnEmptyFile (0.05s) PASS ok github.com/hashicorp/consul/command/validate 0.201s === RUN TestVersionCommand_noTabs === PAUSE TestVersionCommand_noTabs === CONT TestVersionCommand_noTabs --- PASS: TestVersionCommand_noTabs (0.00s) PASS ok github.com/hashicorp/consul/command/version 0.038s === RUN TestWatchCommand_noTabs === PAUSE TestWatchCommand_noTabs === RUN TestWatchCommand === PAUSE TestWatchCommand === RUN TestWatchCommand_loadToken [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:35:19.627Z [WARN] TestWatchCommand_loadToken: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:19.629Z [DEBUG] TestWatchCommand_loadToken.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:19.631Z [DEBUG] TestWatchCommand_loadToken.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:35:19.664Z [INFO] TestWatchCommand_loadToken.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:877ebfc1-c6d3-4ff3-a12f-496fe10e956f Address:127.0.0.1:17153}]" writer.go:29: 2021-01-29T19:35:19.667Z [INFO] TestWatchCommand_loadToken.server.serf.wan: serf: EventMemberJoin: Node-877ebfc1-c6d3-4ff3-a12f-496fe10e956f.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:35:19.669Z [INFO] TestWatchCommand_loadToken.server.raft: entering follower state: follower="Node at 127.0.0.1:17153 [Follower]" leader= writer.go:29: 2021-01-29T19:35:19.670Z [INFO] TestWatchCommand_loadToken.server.serf.lan: serf: EventMemberJoin: Node-877ebfc1-c6d3-4ff3-a12f-496fe10e956f 127.0.0.1 writer.go:29: 2021-01-29T19:35:19.679Z [INFO] TestWatchCommand_loadToken.server: Adding LAN server: server="Node-877ebfc1-c6d3-4ff3-a12f-496fe10e956f (Addr: tcp/127.0.0.1:17153) (DC: dc1)" writer.go:29: 2021-01-29T19:35:19.681Z [INFO] TestWatchCommand_loadToken.server: Handled event for server in area: event=member-join server=Node-877ebfc1-c6d3-4ff3-a12f-496fe10e956f.dc1 area=wan writer.go:29: 2021-01-29T19:35:19.683Z [INFO] TestWatchCommand_loadToken: Started DNS server: address=127.0.0.1:17148 network=tcp writer.go:29: 2021-01-29T19:35:19.684Z [INFO] TestWatchCommand_loadToken: Started DNS server: address=127.0.0.1:17148 network=udp writer.go:29: 2021-01-29T19:35:19.686Z [INFO] TestWatchCommand_loadToken: Started HTTP server: address=127.0.0.1:17149 network=tcp writer.go:29: 2021-01-29T19:35:19.688Z [INFO] TestWatchCommand_loadToken: started state syncer writer.go:29: 2021-01-29T19:35:19.731Z [WARN] TestWatchCommand_loadToken.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:19.734Z [INFO] TestWatchCommand_loadToken.server.raft: entering candidate state: node="Node at 127.0.0.1:17153 [Candidate]" term=2 writer.go:29: 2021-01-29T19:35:19.737Z [DEBUG] TestWatchCommand_loadToken.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:19.739Z [DEBUG] TestWatchCommand_loadToken.server.raft: vote granted: from=877ebfc1-c6d3-4ff3-a12f-496fe10e956f term=2 tally=1 writer.go:29: 2021-01-29T19:35:19.741Z [INFO] TestWatchCommand_loadToken.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:19.743Z [INFO] TestWatchCommand_loadToken.server.raft: entering leader state: leader="Node at 127.0.0.1:17153 [Leader]" writer.go:29: 2021-01-29T19:35:19.745Z [INFO] TestWatchCommand_loadToken.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:19.746Z [INFO] TestWatchCommand_loadToken.server: New leader elected: payload=Node-877ebfc1-c6d3-4ff3-a12f-496fe10e956f writer.go:29: 2021-01-29T19:35:19.746Z [DEBUG] TestWatchCommand_loadToken.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:17153 writer.go:29: 2021-01-29T19:35:19.757Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:35:19.812Z [INFO] TestWatchCommand_loadToken: Synced node info writer.go:29: 2021-01-29T19:35:19.815Z [INFO] TestWatchCommand_loadToken.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:19.817Z [INFO] TestWatchCommand_loadToken.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:19.819Z [DEBUG] TestWatchCommand_loadToken.server: Skipping self join check for node since the cluster is too small: node=Node-877ebfc1-c6d3-4ff3-a12f-496fe10e956f writer.go:29: 2021-01-29T19:35:19.820Z [INFO] TestWatchCommand_loadToken.server: member joined, marking health alive: member=Node-877ebfc1-c6d3-4ff3-a12f-496fe10e956f === RUN TestWatchCommand_loadToken/token_arg === RUN TestWatchCommand_loadToken/token_env === RUN TestWatchCommand_loadToken/token_file_arg === RUN TestWatchCommand_loadToken/token_file_env === CONT TestWatchCommand_loadToken writer.go:29: 2021-01-29T19:35:20.039Z [INFO] TestWatchCommand_loadToken: Requesting shutdown writer.go:29: 2021-01-29T19:35:20.040Z [INFO] TestWatchCommand_loadToken.server: shutting down server writer.go:29: 2021-01-29T19:35:20.041Z [DEBUG] TestWatchCommand_loadToken.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:20.042Z [WARN] TestWatchCommand_loadToken.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:20.042Z [DEBUG] TestWatchCommand_loadToken.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:20.060Z [WARN] TestWatchCommand_loadToken.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:20.063Z [INFO] TestWatchCommand_loadToken.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:20.063Z [INFO] TestWatchCommand_loadToken: consul server down writer.go:29: 2021-01-29T19:35:20.066Z [INFO] TestWatchCommand_loadToken: shutdown complete writer.go:29: 2021-01-29T19:35:20.068Z [INFO] TestWatchCommand_loadToken: Stopping server: protocol=DNS address=127.0.0.1:17148 network=tcp writer.go:29: 2021-01-29T19:35:20.069Z [INFO] TestWatchCommand_loadToken: Stopping server: protocol=DNS address=127.0.0.1:17148 network=udp writer.go:29: 2021-01-29T19:35:20.070Z [INFO] TestWatchCommand_loadToken: Stopping server: protocol=HTTP address=127.0.0.1:17149 network=tcp writer.go:29: 2021-01-29T19:35:20.572Z [INFO] TestWatchCommand_loadToken: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:20.574Z [INFO] TestWatchCommand_loadToken: Endpoints down --- PASS: TestWatchCommand_loadToken (1.01s) --- PASS: TestWatchCommand_loadToken/token_arg (0.03s) --- PASS: TestWatchCommand_loadToken/token_env (0.01s) --- PASS: TestWatchCommand_loadToken/token_file_arg (0.00s) --- PASS: TestWatchCommand_loadToken/token_file_env (0.00s) === RUN TestWatchCommandNoConnect === PAUSE TestWatchCommandNoConnect === RUN TestWatchCommandNoAgentService === PAUSE TestWatchCommandNoAgentService === CONT TestWatchCommand_noTabs --- PASS: TestWatchCommand_noTabs (0.00s) === CONT TestWatchCommandNoAgentService === CONT TestWatchCommandNoConnect === CONT TestWatchCommand === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.615Z [WARN] TestWatchCommandNoConnect: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:20.617Z [DEBUG] TestWatchCommandNoConnect.tlsutil: Update: version=1 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.626Z [WARN] TestWatchCommand: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:20.627Z [DEBUG] TestWatchCommand.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:20.630Z [DEBUG] TestWatchCommand.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.636Z [DEBUG] TestWatchCommandNoConnect.tlsutil: OutgoingRPCWrapper: version=1 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.636Z [INFO] TestWatchCommand.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:817e0cde-4d70-7506-71ab-79ee3f3104cd Address:127.0.0.1:17171}]" === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.637Z [WARN] TestWatchCommandNoAgentService: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:20.638Z [DEBUG] TestWatchCommandNoAgentService.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:20.640Z [DEBUG] TestWatchCommandNoAgentService.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:35:20.644Z [INFO] TestWatchCommandNoAgentService.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:622139cc-e751-d334-a34c-30cf76cc8fe2 Address:127.0.0.1:17159}]" === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.639Z [INFO] TestWatchCommand.server.raft: entering follower state: follower="Node at 127.0.0.1:17171 [Follower]" leader= === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.647Z [INFO] TestWatchCommandNoAgentService.server.serf.wan: serf: EventMemberJoin: Node-622139cc-e751-d334-a34c-30cf76cc8fe2.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:35:20.650Z [INFO] TestWatchCommandNoAgentService.server.serf.lan: serf: EventMemberJoin: Node-622139cc-e751-d334-a34c-30cf76cc8fe2 127.0.0.1 writer.go:29: 2021-01-29T19:35:20.647Z [INFO] TestWatchCommandNoAgentService.server.raft: entering follower state: follower="Node at 127.0.0.1:17159 [Follower]" leader= === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.640Z [INFO] TestWatchCommand.server.serf.wan: serf: EventMemberJoin: Node-817e0cde-4d70-7506-71ab-79ee3f3104cd.dc1 127.0.0.1 === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.655Z [INFO] TestWatchCommandNoAgentService.server: Adding LAN server: server="Node-622139cc-e751-d334-a34c-30cf76cc8fe2 (Addr: tcp/127.0.0.1:17159) (DC: dc1)" writer.go:29: 2021-01-29T19:35:20.656Z [INFO] TestWatchCommandNoAgentService.server: Handled event for server in area: event=member-join server=Node-622139cc-e751-d334-a34c-30cf76cc8fe2.dc1 area=wan === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.656Z [INFO] TestWatchCommandNoConnect.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:5a48e56e-ecb3-0112-5291-3e3fadfb4395 Address:127.0.0.1:17165}]" writer.go:29: 2021-01-29T19:35:20.661Z [INFO] TestWatchCommandNoConnect.server.raft: entering follower state: follower="Node at 127.0.0.1:17165 [Follower]" leader= === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.659Z [INFO] TestWatchCommand.server.serf.lan: serf: EventMemberJoin: Node-817e0cde-4d70-7506-71ab-79ee3f3104cd 127.0.0.1 writer.go:29: 2021-01-29T19:35:20.666Z [INFO] TestWatchCommand: Started DNS server: address=127.0.0.1:17166 network=udp writer.go:29: 2021-01-29T19:35:20.669Z [INFO] TestWatchCommand.server: Adding LAN server: server="Node-817e0cde-4d70-7506-71ab-79ee3f3104cd (Addr: tcp/127.0.0.1:17171) (DC: dc1)" writer.go:29: 2021-01-29T19:35:20.669Z [INFO] TestWatchCommand.server: Handled event for server in area: event=member-join server=Node-817e0cde-4d70-7506-71ab-79ee3f3104cd.dc1 area=wan writer.go:29: 2021-01-29T19:35:20.672Z [INFO] TestWatchCommand: Started DNS server: address=127.0.0.1:17166 network=tcp === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.672Z [INFO] TestWatchCommandNoAgentService: Started DNS server: address=127.0.0.1:17154 network=tcp === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.662Z [INFO] TestWatchCommandNoConnect.server.serf.wan: serf: EventMemberJoin: Node-5a48e56e-ecb3-0112-5291-3e3fadfb4395.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:35:20.675Z [INFO] TestWatchCommandNoConnect.server.serf.lan: serf: EventMemberJoin: Node-5a48e56e-ecb3-0112-5291-3e3fadfb4395 127.0.0.1 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.677Z [INFO] TestWatchCommand: Started HTTP server: address=127.0.0.1:17167 network=tcp === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.674Z [INFO] TestWatchCommandNoAgentService: Started DNS server: address=127.0.0.1:17154 network=udp === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.681Z [INFO] TestWatchCommand: started state syncer === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.678Z [INFO] TestWatchCommandNoConnect: Started DNS server: address=127.0.0.1:17160 network=udp writer.go:29: 2021-01-29T19:35:20.682Z [INFO] TestWatchCommandNoConnect: Started DNS server: address=127.0.0.1:17160 network=tcp writer.go:29: 2021-01-29T19:35:20.679Z [INFO] TestWatchCommandNoConnect.server: Adding LAN server: server="Node-5a48e56e-ecb3-0112-5291-3e3fadfb4395 (Addr: tcp/127.0.0.1:17165) (DC: dc1)" === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.687Z [INFO] TestWatchCommandNoAgentService: Started HTTP server: address=127.0.0.1:17155 network=tcp === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.680Z [INFO] TestWatchCommandNoConnect.server: Handled event for server in area: event=member-join server=Node-5a48e56e-ecb3-0112-5291-3e3fadfb4395.dc1 area=wan writer.go:29: 2021-01-29T19:35:20.686Z [INFO] TestWatchCommandNoConnect: Started HTTP server: address=127.0.0.1:17161 network=tcp writer.go:29: 2021-01-29T19:35:20.690Z [INFO] TestWatchCommandNoConnect: started state syncer === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.712Z [INFO] TestWatchCommandNoAgentService: started state syncer === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.714Z [WARN] TestWatchCommand.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:20.715Z [INFO] TestWatchCommand.server.raft: entering candidate state: node="Node at 127.0.0.1:17171 [Candidate]" term=2 === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.727Z [WARN] TestWatchCommandNoAgentService.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.735Z [WARN] TestWatchCommandNoConnect.server.raft: heartbeat timeout reached, starting election: last-leader= === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.728Z [INFO] TestWatchCommandNoAgentService.server.raft: entering candidate state: node="Node at 127.0.0.1:17159 [Candidate]" term=2 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.737Z [DEBUG] TestWatchCommand.server.raft: votes: needed=1 === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.736Z [INFO] TestWatchCommandNoConnect.server.raft: entering candidate state: node="Node at 127.0.0.1:17165 [Candidate]" term=2 === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.739Z [DEBUG] TestWatchCommandNoAgentService.server.raft: votes: needed=1 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.738Z [DEBUG] TestWatchCommand.server.raft: vote granted: from=817e0cde-4d70-7506-71ab-79ee3f3104cd term=2 tally=1 === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.741Z [DEBUG] TestWatchCommandNoAgentService.server.raft: vote granted: from=622139cc-e751-d334-a34c-30cf76cc8fe2 term=2 tally=1 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.741Z [INFO] TestWatchCommand.server.raft: election won: tally=1 === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.742Z [INFO] TestWatchCommandNoAgentService.server.raft: election won: tally=1 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.742Z [INFO] TestWatchCommand.server.raft: entering leader state: leader="Node at 127.0.0.1:17171 [Leader]" writer.go:29: 2021-01-29T19:35:20.744Z [INFO] TestWatchCommand.server: cluster leadership acquired === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.744Z [DEBUG] TestWatchCommandNoConnect.server.raft: votes: needed=1 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.746Z [INFO] TestWatchCommand.server: New leader elected: payload=Node-817e0cde-4d70-7506-71ab-79ee3f3104cd === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.746Z [DEBUG] TestWatchCommandNoConnect.server.raft: vote granted: from=5a48e56e-ecb3-0112-5291-3e3fadfb4395 term=2 tally=1 writer.go:29: 2021-01-29T19:35:20.748Z [INFO] TestWatchCommandNoConnect.server.raft: election won: tally=1 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.748Z [DEBUG] TestWatchCommand.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:17171 === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.749Z [INFO] TestWatchCommandNoConnect.server.raft: entering leader state: leader="Node at 127.0.0.1:17165 [Leader]" === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.743Z [INFO] TestWatchCommandNoAgentService.server.raft: entering leader state: leader="Node at 127.0.0.1:17159 [Leader]" === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.751Z [INFO] TestWatchCommandNoConnect.server: cluster leadership acquired === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.751Z [INFO] TestWatchCommandNoAgentService.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:20.752Z [INFO] TestWatchCommandNoAgentService.server: New leader elected: payload=Node-622139cc-e751-d334-a34c-30cf76cc8fe2 === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.751Z [INFO] TestWatchCommandNoConnect.server: New leader elected: payload=Node-5a48e56e-ecb3-0112-5291-3e3fadfb4395 writer.go:29: 2021-01-29T19:35:20.754Z [DEBUG] TestWatchCommandNoConnect.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:17165 === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.754Z [DEBUG] TestWatchCommandNoAgentService.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:17159 === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.760Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.760Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.764Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.766Z [INFO] TestWatchCommandNoConnect.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.767Z [INFO] TestWatchCommand.server.connect: initialized primary datacenter CA with provider: provider=consul === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.768Z [INFO] TestWatchCommandNoConnect.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:20.769Z [DEBUG] TestWatchCommandNoConnect.server: Skipping self join check for node since the cluster is too small: node=Node-5a48e56e-ecb3-0112-5291-3e3fadfb4395 writer.go:29: 2021-01-29T19:35:20.770Z [INFO] TestWatchCommandNoConnect.server: member joined, marking health alive: member=Node-5a48e56e-ecb3-0112-5291-3e3fadfb4395 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.770Z [INFO] TestWatchCommand.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:20.771Z [DEBUG] TestWatchCommand.server: Skipping self join check for node since the cluster is too small: node=Node-817e0cde-4d70-7506-71ab-79ee3f3104cd writer.go:29: 2021-01-29T19:35:20.773Z [INFO] TestWatchCommand.server: member joined, marking health alive: member=Node-817e0cde-4d70-7506-71ab-79ee3f3104cd === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:20.776Z [INFO] TestWatchCommandNoAgentService.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:20.777Z [INFO] TestWatchCommandNoAgentService.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:20.778Z [DEBUG] TestWatchCommandNoAgentService.server: Skipping self join check for node since the cluster is too small: node=Node-622139cc-e751-d334-a34c-30cf76cc8fe2 writer.go:29: 2021-01-29T19:35:20.779Z [INFO] TestWatchCommandNoAgentService.server: member joined, marking health alive: member=Node-622139cc-e751-d334-a34c-30cf76cc8fe2 === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.862Z [DEBUG] TestWatchCommand: Skipping remote check since it is managed automatically: check=serfHealth === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.860Z [INFO] TestWatchCommandNoConnect: Requesting shutdown === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.865Z [INFO] TestWatchCommand: Synced node info === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:20.868Z [INFO] TestWatchCommandNoConnect.server: shutting down server writer.go:29: 2021-01-29T19:35:20.875Z [DEBUG] TestWatchCommandNoConnect.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:20.877Z [WARN] TestWatchCommandNoConnect.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:20.875Z [ERROR] TestWatchCommandNoConnect.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:35:20.877Z [DEBUG] TestWatchCommandNoConnect.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:20.890Z [WARN] TestWatchCommandNoConnect.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:20.895Z [INFO] TestWatchCommandNoConnect.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:20.896Z [INFO] TestWatchCommandNoConnect: consul server down writer.go:29: 2021-01-29T19:35:20.897Z [INFO] TestWatchCommandNoConnect: shutdown complete writer.go:29: 2021-01-29T19:35:20.898Z [INFO] TestWatchCommandNoConnect: Stopping server: protocol=DNS address=127.0.0.1:17160 network=tcp writer.go:29: 2021-01-29T19:35:20.899Z [INFO] TestWatchCommandNoConnect: Stopping server: protocol=DNS address=127.0.0.1:17160 network=udp writer.go:29: 2021-01-29T19:35:20.900Z [INFO] TestWatchCommandNoConnect: Stopping server: protocol=HTTP address=127.0.0.1:17161 network=tcp === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:20.948Z [DEBUG] TestWatchCommand.http: Request finished: method=GET url=/v1/agent/self from=127.0.0.1:56304 latency=4.436944ms writer.go:29: 2021-01-29T19:35:20.955Z [DEBUG] TestWatchCommand.http: Request finished: method=GET url=/v1/catalog/nodes from=127.0.0.1:56306 latency=867.94µs writer.go:29: 2021-01-29T19:35:20.957Z [INFO] TestWatchCommand: Requesting shutdown writer.go:29: 2021-01-29T19:35:20.959Z [INFO] TestWatchCommand.server: shutting down server writer.go:29: 2021-01-29T19:35:20.960Z [DEBUG] TestWatchCommand.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:20.961Z [WARN] TestWatchCommand.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:20.963Z [DEBUG] TestWatchCommand.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:20.964Z [WARN] TestWatchCommand.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:20.969Z [INFO] TestWatchCommand.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:20.969Z [INFO] TestWatchCommand: consul server down writer.go:29: 2021-01-29T19:35:20.972Z [INFO] TestWatchCommand: shutdown complete writer.go:29: 2021-01-29T19:35:20.974Z [INFO] TestWatchCommand: Stopping server: protocol=DNS address=127.0.0.1:17166 network=tcp writer.go:29: 2021-01-29T19:35:20.975Z [INFO] TestWatchCommand: Stopping server: protocol=DNS address=127.0.0.1:17166 network=udp writer.go:29: 2021-01-29T19:35:20.976Z [INFO] TestWatchCommand: Stopping server: protocol=HTTP address=127.0.0.1:17167 network=tcp === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:21.058Z [INFO] TestWatchCommandNoAgentService: Requesting shutdown writer.go:29: 2021-01-29T19:35:21.060Z [INFO] TestWatchCommandNoAgentService.server: shutting down server writer.go:29: 2021-01-29T19:35:21.061Z [DEBUG] TestWatchCommandNoAgentService.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:21.062Z [WARN] TestWatchCommandNoAgentService.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:21.062Z [ERROR] TestWatchCommandNoAgentService.anti_entropy: failed to sync remote state: error="No cluster leader" writer.go:29: 2021-01-29T19:35:21.063Z [DEBUG] TestWatchCommandNoAgentService.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:21.070Z [WARN] TestWatchCommandNoAgentService.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:21.072Z [INFO] TestWatchCommandNoAgentService.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:21.074Z [INFO] TestWatchCommandNoAgentService: consul server down writer.go:29: 2021-01-29T19:35:21.074Z [INFO] TestWatchCommandNoAgentService: shutdown complete writer.go:29: 2021-01-29T19:35:21.075Z [INFO] TestWatchCommandNoAgentService: Stopping server: protocol=DNS address=127.0.0.1:17154 network=tcp writer.go:29: 2021-01-29T19:35:21.076Z [INFO] TestWatchCommandNoAgentService: Stopping server: protocol=DNS address=127.0.0.1:17154 network=udp writer.go:29: 2021-01-29T19:35:21.076Z [INFO] TestWatchCommandNoAgentService: Stopping server: protocol=HTTP address=127.0.0.1:17155 network=tcp === CONT TestWatchCommandNoConnect writer.go:29: 2021-01-29T19:35:21.403Z [INFO] TestWatchCommandNoConnect: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:21.406Z [INFO] TestWatchCommandNoConnect: Endpoints down --- PASS: TestWatchCommandNoConnect (0.83s) === CONT TestWatchCommand writer.go:29: 2021-01-29T19:35:21.478Z [INFO] TestWatchCommand: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:21.480Z [INFO] TestWatchCommand: Endpoints down --- PASS: TestWatchCommand (0.90s) === CONT TestWatchCommandNoAgentService writer.go:29: 2021-01-29T19:35:21.577Z [INFO] TestWatchCommandNoAgentService: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:21.579Z [INFO] TestWatchCommandNoAgentService: Endpoints down --- PASS: TestWatchCommandNoAgentService (1.00s) PASS ok github.com/hashicorp/consul/command/watch 2.221s === RUN TestStaticResolver_Resolve === RUN TestStaticResolver_Resolve/simples --- PASS: TestStaticResolver_Resolve (0.01s) --- PASS: TestStaticResolver_Resolve/simples (0.00s) === RUN TestConsulResolver_Resolve [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range writer.go:29: 2021-01-29T19:35:21.276Z [WARN] test-consul: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:21.278Z [DEBUG] test-consul.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:21.280Z [DEBUG] test-consul.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:35:21.291Z [INFO] test-consul.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:366399ad-433d-20e5-451b-c2703be48c08 Address:127.0.0.1:23279}]" writer.go:29: 2021-01-29T19:35:21.295Z [INFO] test-consul.server.serf.wan: serf: EventMemberJoin: Node-366399ad-433d-20e5-451b-c2703be48c08.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:35:21.298Z [INFO] test-consul.server.serf.lan: serf: EventMemberJoin: Node-366399ad-433d-20e5-451b-c2703be48c08 127.0.0.1 writer.go:29: 2021-01-29T19:35:21.300Z [INFO] test-consul.server.raft: entering follower state: follower="Node at 127.0.0.1:23279 [Follower]" leader= writer.go:29: 2021-01-29T19:35:21.303Z [INFO] test-consul: Started DNS server: address=127.0.0.1:23274 network=udp writer.go:29: 2021-01-29T19:35:21.324Z [INFO] test-consul: Started DNS server: address=127.0.0.1:23274 network=tcp writer.go:29: 2021-01-29T19:35:21.310Z [INFO] test-consul.server: Handled event for server in area: event=member-join server=Node-366399ad-433d-20e5-451b-c2703be48c08.dc1 area=wan writer.go:29: 2021-01-29T19:35:21.312Z [INFO] test-consul.server: Adding LAN server: server="Node-366399ad-433d-20e5-451b-c2703be48c08 (Addr: tcp/127.0.0.1:23279) (DC: dc1)" writer.go:29: 2021-01-29T19:35:21.349Z [INFO] test-consul: Started HTTP server: address=127.0.0.1:23275 network=tcp writer.go:29: 2021-01-29T19:35:21.362Z [WARN] test-consul.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:21.369Z [INFO] test-consul.server.raft: entering candidate state: node="Node at 127.0.0.1:23279 [Candidate]" term=2 writer.go:29: 2021-01-29T19:35:21.368Z [INFO] test-consul: started state syncer writer.go:29: 2021-01-29T19:35:21.372Z [DEBUG] test-consul.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:21.373Z [DEBUG] test-consul.server.raft: vote granted: from=366399ad-433d-20e5-451b-c2703be48c08 term=2 tally=1 writer.go:29: 2021-01-29T19:35:21.373Z [INFO] test-consul.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:21.374Z [INFO] test-consul.server.raft: entering leader state: leader="Node at 127.0.0.1:23279 [Leader]" writer.go:29: 2021-01-29T19:35:21.375Z [INFO] test-consul.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:21.376Z [INFO] test-consul.server: New leader elected: payload=Node-366399ad-433d-20e5-451b-c2703be48c08 writer.go:29: 2021-01-29T19:35:21.380Z [DEBUG] test-consul.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23279 writer.go:29: 2021-01-29T19:35:21.390Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:35:21.430Z [INFO] test-consul.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:21.431Z [INFO] test-consul.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:21.432Z [DEBUG] test-consul.server: Skipping self join check for node since the cluster is too small: node=Node-366399ad-433d-20e5-451b-c2703be48c08 writer.go:29: 2021-01-29T19:35:21.433Z [INFO] test-consul.server: member joined, marking health alive: member=Node-366399ad-433d-20e5-451b-c2703be48c08 writer.go:29: 2021-01-29T19:35:21.697Z [DEBUG] test-consul: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:35:21.717Z [INFO] test-consul: Synced node info writer.go:29: 2021-01-29T19:35:21.765Z [DEBUG] test-consul: Node info in sync writer.go:29: 2021-01-29T19:35:21.768Z [INFO] test-consul: Synced service: service=web writer.go:29: 2021-01-29T19:35:21.783Z [DEBUG] test-consul.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:36092 latency=24.813704ms writer.go:29: 2021-01-29T19:35:21.787Z [DEBUG] test-consul: Node info in sync writer.go:29: 2021-01-29T19:35:21.789Z [DEBUG] test-consul: Service in sync: service=web writer.go:29: 2021-01-29T19:35:21.793Z [INFO] test-consul: Synced service: service=web-proxy writer.go:29: 2021-01-29T19:35:21.796Z [DEBUG] test-consul.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:36092 latency=10.121026ms writer.go:29: 2021-01-29T19:35:21.796Z [ERROR] test-consul.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:35:21.807Z [DEBUG] test-consul: Node info in sync writer.go:29: 2021-01-29T19:35:21.808Z [DEBUG] test-consul: Service in sync: service=web writer.go:29: 2021-01-29T19:35:21.809Z [DEBUG] test-consul: Service in sync: service=web-proxy writer.go:29: 2021-01-29T19:35:21.808Z [ERROR] test-consul.proxycfg: watch error: id=service-http-checks: error="invalid type for service checks response: cache.FetchResult, want: []structs.CheckType" writer.go:29: 2021-01-29T19:35:21.812Z [INFO] test-consul: Synced service: service=web-proxy-2 writer.go:29: 2021-01-29T19:35:21.813Z [DEBUG] test-consul.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:36092 latency=10.796758ms writer.go:29: 2021-01-29T19:35:21.817Z [DEBUG] test-consul: Node info in sync writer.go:29: 2021-01-29T19:35:21.822Z [DEBUG] test-consul: Service in sync: service=web writer.go:29: 2021-01-29T19:35:21.827Z [DEBUG] test-consul: Service in sync: service=web-proxy writer.go:29: 2021-01-29T19:35:21.832Z [DEBUG] test-consul: Service in sync: service=web-proxy-2 writer.go:29: 2021-01-29T19:35:21.847Z [INFO] test-consul: Synced service: service=db writer.go:29: 2021-01-29T19:35:21.848Z [DEBUG] test-consul.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:36092 latency=32.109663ms writer.go:29: 2021-01-29T19:35:21.853Z [DEBUG] test-consul.http: Request finished: method=POST url=/v1/query from=127.0.0.1:36092 latency=3.426528ms === RUN TestConsulResolver_Resolve/basic_service_discovery === CONT TestConsulResolver_Resolve writer.go:29: 2021-01-29T19:35:21.876Z [DEBUG] test-consul.http: Request finished: method=GET url=/v1/health/connect/web?connect=true&passing=1&stale= from=127.0.0.1:36092 latency=8.785633ms === RUN TestConsulResolver_Resolve/basic_service_with_native_service === CONT TestConsulResolver_Resolve writer.go:29: 2021-01-29T19:35:21.895Z [DEBUG] test-consul.http: Request finished: method=GET url=/v1/health/connect/db?connect=true&passing=1&stale= from=127.0.0.1:36092 latency=1.26527ms === RUN TestConsulResolver_Resolve/Bad_Type_errors === RUN TestConsulResolver_Resolve/Non-existent_service_errors === CONT TestConsulResolver_Resolve writer.go:29: 2021-01-29T19:35:21.901Z [DEBUG] test-consul.http: Request finished: method=GET url=/v1/health/connect/foo?connect=true&passing=1&stale= from=127.0.0.1:36092 latency=227.789µs === RUN TestConsulResolver_Resolve/timeout_errors === RUN TestConsulResolver_Resolve/prepared_query_by_id === CONT TestConsulResolver_Resolve writer.go:29: 2021-01-29T19:35:21.907Z [DEBUG] test-consul.http: Request finished: method=GET url=/v1/query/ad42ad3c-82c4-7995-bc80-cf6e801c0000/execute?connect=true&stale= from=127.0.0.1:36092 latency=509.293µs === RUN TestConsulResolver_Resolve/prepared_query_by_name === CONT TestConsulResolver_Resolve writer.go:29: 2021-01-29T19:35:21.914Z [DEBUG] test-consul.http: Request finished: method=GET url=/v1/query/test-query/execute?connect=true&stale= from=127.0.0.1:36092 latency=1.439106ms writer.go:29: 2021-01-29T19:35:21.919Z [INFO] test-consul: Requesting shutdown writer.go:29: 2021-01-29T19:35:21.920Z [INFO] test-consul.server: shutting down server writer.go:29: 2021-01-29T19:35:21.921Z [DEBUG] test-consul.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:21.922Z [WARN] test-consul.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:21.922Z [DEBUG] test-consul.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:21.925Z [WARN] test-consul.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:21.929Z [INFO] test-consul.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:21.932Z [INFO] test-consul: consul server down writer.go:29: 2021-01-29T19:35:21.934Z [INFO] test-consul: shutdown complete writer.go:29: 2021-01-29T19:35:21.935Z [INFO] test-consul: Stopping server: protocol=DNS address=127.0.0.1:23274 network=tcp writer.go:29: 2021-01-29T19:35:21.936Z [INFO] test-consul: Stopping server: protocol=DNS address=127.0.0.1:23274 network=udp writer.go:29: 2021-01-29T19:35:21.938Z [INFO] test-consul: Stopping server: protocol=HTTP address=127.0.0.1:23275 network=tcp writer.go:29: 2021-01-29T19:35:22.451Z [INFO] test-consul: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:22.453Z [INFO] test-consul: Endpoints down --- PASS: TestConsulResolver_Resolve (1.21s) --- PASS: TestConsulResolver_Resolve/basic_service_discovery (0.03s) --- PASS: TestConsulResolver_Resolve/basic_service_with_native_service (0.01s) --- PASS: TestConsulResolver_Resolve/Bad_Type_errors (0.00s) --- PASS: TestConsulResolver_Resolve/Non-existent_service_errors (0.00s) --- PASS: TestConsulResolver_Resolve/timeout_errors (0.00s) --- PASS: TestConsulResolver_Resolve/prepared_query_by_id (0.01s) --- PASS: TestConsulResolver_Resolve/prepared_query_by_name (0.01s) === RUN TestConsulResolverFromAddrFunc === RUN TestConsulResolverFromAddrFunc/service === RUN TestConsulResolverFromAddrFunc/query === RUN TestConsulResolverFromAddrFunc/service_with_dc === RUN TestConsulResolverFromAddrFunc/query_with_dc === RUN TestConsulResolverFromAddrFunc/invalid_host:port === RUN TestConsulResolverFromAddrFunc/custom_domain === RUN TestConsulResolverFromAddrFunc/unsupported_query_type === RUN TestConsulResolverFromAddrFunc/unsupported_query_type_and_datacenter === RUN TestConsulResolverFromAddrFunc/unsupported_query_type_and_datacenter#01 === RUN TestConsulResolverFromAddrFunc/unsupported_tag_filter === RUN TestConsulResolverFromAddrFunc/unsupported_tag_filter_with_DC --- PASS: TestConsulResolverFromAddrFunc (0.06s) --- PASS: TestConsulResolverFromAddrFunc/service (0.02s) --- PASS: TestConsulResolverFromAddrFunc/query (0.00s) --- PASS: TestConsulResolverFromAddrFunc/service_with_dc (0.00s) --- PASS: TestConsulResolverFromAddrFunc/query_with_dc (0.00s) --- PASS: TestConsulResolverFromAddrFunc/invalid_host:port (0.00s) --- PASS: TestConsulResolverFromAddrFunc/custom_domain (0.00s) --- PASS: TestConsulResolverFromAddrFunc/unsupported_query_type (0.00s) --- PASS: TestConsulResolverFromAddrFunc/unsupported_query_type_and_datacenter (0.00s) --- PASS: TestConsulResolverFromAddrFunc/unsupported_query_type_and_datacenter#01 (0.00s) --- PASS: TestConsulResolverFromAddrFunc/unsupported_tag_filter (0.00s) --- PASS: TestConsulResolverFromAddrFunc/unsupported_tag_filter_with_DC (0.00s) === RUN TestService_Name --- PASS: TestService_Name (0.01s) === RUN TestService_Dial service_test.go:36: DM-skipped --- SKIP: TestService_Dial (0.00s) === RUN TestService_ServerTLSConfig service_test.go:129: DM-skipped --- SKIP: TestService_ServerTLSConfig (0.00s) === RUN TestService_HTTPClient 2021/01/29 19:35:22 starting test connect HTTPS server on 127.0.0.1:23280 2021/01/29 19:35:22 test connect service listening on 127.0.0.1:23280 --- PASS: TestService_HTTPClient (0.02s) === RUN TestService_HasDefaultHTTPResolverFromAddr --- PASS: TestService_HasDefaultHTTPResolverFromAddr (0.00s) === RUN Test_verifyServerCertMatchesURI 2021-01-29T19:35:22.563Z [ERROR] connect.watch: Watch errored: service=foo type=connect_roots error="Get "http://127.0.0.1:8500/v1/agent/connect/ca/roots": dial tcp 127.0.0.1:8500: connect: connection refused" retry=5s 2021-01-29T19:35:22.564Z [ERROR] connect.watch: Watch errored: service=foo type=connect_leaf error="Get "http://127.0.0.1:8500/v1/agent/connect/ca/leaf/foo": dial tcp 127.0.0.1:8500: connect: connection refused" retry=5s === RUN Test_verifyServerCertMatchesURI/simple_match === RUN Test_verifyServerCertMatchesURI/different_trust-domain_allowed === RUN Test_verifyServerCertMatchesURI/mismatch === RUN Test_verifyServerCertMatchesURI/no_certs === RUN Test_verifyServerCertMatchesURI/nil_certs --- PASS: Test_verifyServerCertMatchesURI (0.02s) --- PASS: Test_verifyServerCertMatchesURI/simple_match (0.00s) --- PASS: Test_verifyServerCertMatchesURI/different_trust-domain_allowed (0.00s) --- PASS: Test_verifyServerCertMatchesURI/mismatch (0.00s) --- PASS: Test_verifyServerCertMatchesURI/no_certs (0.00s) --- PASS: Test_verifyServerCertMatchesURI/nil_certs (0.00s) === RUN TestClientSideVerifier === RUN TestClientSideVerifier/ok_service_ca1 === RUN TestClientSideVerifier/untrusted_CA === RUN TestClientSideVerifier/cross_signed_intermediate === RUN TestClientSideVerifier/cross_signed_without_intermediate --- PASS: TestClientSideVerifier (0.06s) --- PASS: TestClientSideVerifier/ok_service_ca1 (0.01s) --- PASS: TestClientSideVerifier/untrusted_CA (0.01s) --- PASS: TestClientSideVerifier/cross_signed_intermediate (0.00s) --- PASS: TestClientSideVerifier/cross_signed_without_intermediate (0.00s) === RUN TestServerSideVerifier writer.go:29: 2021-01-29T19:35:22.703Z [WARN] test-consul: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:22.706Z [DEBUG] test-consul.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:22.708Z [DEBUG] test-consul.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:35:22.734Z [INFO] test-consul.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:a0f4ce57-c9e0-312c-5095-53ae7fa50d68 Address:127.0.0.1:23286}]" writer.go:29: 2021-01-29T19:35:22.736Z [INFO] test-consul.server.serf.wan: serf: EventMemberJoin: Node-a0f4ce57-c9e0-312c-5095-53ae7fa50d68.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:35:22.739Z [INFO] test-consul.server.serf.lan: serf: EventMemberJoin: Node-a0f4ce57-c9e0-312c-5095-53ae7fa50d68 127.0.0.1 writer.go:29: 2021-01-29T19:35:22.743Z [INFO] test-consul.server.raft: entering follower state: follower="Node at 127.0.0.1:23286 [Follower]" leader= writer.go:29: 2021-01-29T19:35:22.749Z [INFO] test-consul.server: Adding LAN server: server="Node-a0f4ce57-c9e0-312c-5095-53ae7fa50d68 (Addr: tcp/127.0.0.1:23286) (DC: dc1)" writer.go:29: 2021-01-29T19:35:22.765Z [INFO] test-consul: Started DNS server: address=127.0.0.1:23281 network=udp writer.go:29: 2021-01-29T19:35:22.751Z [INFO] test-consul.server: Handled event for server in area: event=member-join server=Node-a0f4ce57-c9e0-312c-5095-53ae7fa50d68.dc1 area=wan writer.go:29: 2021-01-29T19:35:22.769Z [INFO] test-consul: Started DNS server: address=127.0.0.1:23281 network=tcp writer.go:29: 2021-01-29T19:35:22.773Z [INFO] test-consul: Started HTTP server: address=127.0.0.1:23282 network=tcp writer.go:29: 2021-01-29T19:35:22.776Z [INFO] test-consul: started state syncer writer.go:29: 2021-01-29T19:35:22.811Z [WARN] test-consul.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:22.812Z [INFO] test-consul.server.raft: entering candidate state: node="Node at 127.0.0.1:23286 [Candidate]" term=2 writer.go:29: 2021-01-29T19:35:22.817Z [DEBUG] test-consul.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:22.819Z [DEBUG] test-consul.server.raft: vote granted: from=a0f4ce57-c9e0-312c-5095-53ae7fa50d68 term=2 tally=1 writer.go:29: 2021-01-29T19:35:22.821Z [INFO] test-consul.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:22.824Z [INFO] test-consul.server.raft: entering leader state: leader="Node at 127.0.0.1:23286 [Leader]" writer.go:29: 2021-01-29T19:35:22.826Z [INFO] test-consul.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:22.828Z [INFO] test-consul.server: New leader elected: payload=Node-a0f4ce57-c9e0-312c-5095-53ae7fa50d68 writer.go:29: 2021-01-29T19:35:22.830Z [DEBUG] test-consul.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:23286 writer.go:29: 2021-01-29T19:35:22.837Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:35:22.843Z [INFO] test-consul.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:22.853Z [INFO] test-consul.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:22.858Z [DEBUG] test-consul.server: Skipping self join check for node since the cluster is too small: node=Node-a0f4ce57-c9e0-312c-5095-53ae7fa50d68 writer.go:29: 2021-01-29T19:35:22.861Z [INFO] test-consul.server: member joined, marking health alive: member=Node-a0f4ce57-c9e0-312c-5095-53ae7fa50d68 writer.go:29: 2021-01-29T19:35:22.996Z [DEBUG] test-consul: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:35:23.000Z [INFO] test-consul: Synced node info writer.go:29: 2021-01-29T19:35:23.002Z [DEBUG] test-consul: Node info in sync writer.go:29: 2021-01-29T19:35:23.075Z [DEBUG] test-consul.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:39272 latency=3.94775ms writer.go:29: 2021-01-29T19:35:23.079Z [DEBUG] test-consul.http: Request finished: method=POST url=/v1/connect/intentions from=127.0.0.1:39272 latency=1.215887ms === RUN TestServerSideVerifier/ok_service_ca1,_allow === CONT TestServerSideVerifier writer.go:29: 2021-01-29T19:35:23.114Z [DEBUG] test-consul.http: Request finished: method=POST url=/v1/agent/connect/authorize from=127.0.0.1:39272 latency=2.212891ms === RUN TestServerSideVerifier/untrusted_CA writer.go:29: 2021-01-29T19:35:23.120Z [ERROR] TestServerSideVerifier/untrusted_CA: failed TLS verification: error="x509: certificate signed by unknown authority" === RUN TestServerSideVerifier/cross_signed_intermediate,_allow === CONT TestServerSideVerifier writer.go:29: 2021-01-29T19:35:23.128Z [DEBUG] test-consul.http: Request finished: method=POST url=/v1/agent/connect/authorize from=127.0.0.1:39272 latency=278.967µs === RUN TestServerSideVerifier/cross_signed_without_intermediate writer.go:29: 2021-01-29T19:35:23.134Z [ERROR] TestServerSideVerifier/cross_signed_without_intermediate: failed TLS verification: error="x509: certificate signed by unknown authority" === RUN TestServerSideVerifier/ok_service_ca1,_deny === CONT TestServerSideVerifier writer.go:29: 2021-01-29T19:35:23.141Z [DEBUG] test-consul.http: Request finished: method=POST url=/v1/agent/connect/authorize from=127.0.0.1:39272 latency=314.49µs === CONT TestServerSideVerifier/ok_service_ca1,_deny writer.go:29: 2021-01-29T19:35:23.143Z [ERROR] TestServerSideVerifier/ok_service_ca1,_deny: authz call denied: reason="Matched intention: DENY default/* => default/db (ID: 8a5e58c8-ac77-507d-800e-57e092199cf3, Precedence: 8)" === RUN TestServerSideVerifier/cross_signed_intermediate,_deny === CONT TestServerSideVerifier writer.go:29: 2021-01-29T19:35:23.149Z [DEBUG] test-consul.http: Request finished: method=POST url=/v1/agent/connect/authorize from=127.0.0.1:39272 latency=337.827µs === CONT TestServerSideVerifier/cross_signed_intermediate,_deny writer.go:29: 2021-01-29T19:35:23.152Z [ERROR] TestServerSideVerifier/cross_signed_intermediate,_deny: authz call denied: reason="Matched intention: DENY default/* => default/db (ID: 8a5e58c8-ac77-507d-800e-57e092199cf3, Precedence: 8)" === CONT TestServerSideVerifier writer.go:29: 2021-01-29T19:35:23.155Z [INFO] test-consul: Requesting shutdown writer.go:29: 2021-01-29T19:35:23.156Z [INFO] test-consul.server: shutting down server writer.go:29: 2021-01-29T19:35:23.158Z [DEBUG] test-consul.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:23.160Z [WARN] test-consul.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:23.160Z [DEBUG] test-consul.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:23.164Z [WARN] test-consul.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:23.167Z [INFO] test-consul.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:23.167Z [INFO] test-consul: consul server down writer.go:29: 2021-01-29T19:35:23.169Z [INFO] test-consul: shutdown complete writer.go:29: 2021-01-29T19:35:23.170Z [INFO] test-consul: Stopping server: protocol=DNS address=127.0.0.1:23281 network=tcp writer.go:29: 2021-01-29T19:35:23.171Z [INFO] test-consul: Stopping server: protocol=DNS address=127.0.0.1:23281 network=udp writer.go:29: 2021-01-29T19:35:23.172Z [INFO] test-consul: Stopping server: protocol=HTTP address=127.0.0.1:23282 network=tcp writer.go:29: 2021-01-29T19:35:23.673Z [INFO] test-consul: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:23.675Z [INFO] test-consul: Endpoints down --- PASS: TestServerSideVerifier (1.02s) --- PASS: TestServerSideVerifier/ok_service_ca1,_allow (0.01s) --- PASS: TestServerSideVerifier/untrusted_CA (0.01s) --- PASS: TestServerSideVerifier/cross_signed_intermediate,_allow (0.01s) --- PASS: TestServerSideVerifier/cross_signed_without_intermediate (0.01s) --- PASS: TestServerSideVerifier/ok_service_ca1,_deny (0.01s) --- PASS: TestServerSideVerifier/cross_signed_intermediate,_deny (0.01s) === RUN TestDynamicTLSConfig --- PASS: TestDynamicTLSConfig (0.10s) === RUN TestDynamicTLSConfig_Ready --- PASS: TestDynamicTLSConfig_Ready (0.01s) PASS ok github.com/hashicorp/consul/connect 2.734s ? github.com/hashicorp/consul/connect/certgen [no test files] === RUN TestUpstreamResolverFuncFromClient === PAUSE TestUpstreamResolverFuncFromClient === RUN TestAgentConfigWatcherSidecarProxy === PAUSE TestAgentConfigWatcherSidecarProxy === RUN TestConn conn_test.go:68: DM-skipped --- SKIP: TestConn (0.00s) === RUN TestConnSrcClosing === PAUSE TestConnSrcClosing === RUN TestConnDstClosing === PAUSE TestConnDstClosing === RUN TestPublicListener [INFO] freeport: blockSize 1500 too big for system limit 1024. Adjusting... [INFO] freeport: detected ephemeral port range of [32768, 60999] [INFO] freeport: reducing max blocks from 30 to 22 to avoid the ephemeral port range 2021/01/29 19:35:22 test tcp server listening on localhost:10002 2021-01-29T19:35:22.712Z [INFO] nil client provided 2021/01/29 19:35:22 test tcp echo server 127.0.0.1:10002 stopped --- PASS: TestPublicListener (0.09s) === RUN TestUpstreamListener listener_test.go:163: DM-skipped --- SKIP: TestUpstreamListener (0.00s) === RUN TestProxy_public proxy_test.go:21: DM-skipped --- SKIP: TestProxy_public (0.00s) === CONT TestUpstreamResolverFuncFromClient === RUN TestUpstreamResolverFuncFromClient/service === CONT TestConnDstClosing === CONT TestConnSrcClosing === RUN TestUpstreamResolverFuncFromClient/prepared_query === CONT TestAgentConfigWatcherSidecarProxy === RUN TestUpstreamResolverFuncFromClient/unknown_behaves_like_service --- PASS: TestUpstreamResolverFuncFromClient (0.02s) --- PASS: TestUpstreamResolverFuncFromClient/service (0.01s) --- PASS: TestUpstreamResolverFuncFromClient/prepared_query (0.00s) --- PASS: TestUpstreamResolverFuncFromClient/unknown_behaves_like_service (0.01s) --- PASS: TestConnDstClosing (0.03s) --- PASS: TestConnSrcClosing (0.03s) === CONT TestAgentConfigWatcherSidecarProxy writer.go:29: 2021-01-29T19:35:22.768Z [WARN] agent_smith: bootstrap = true: do not enable unless necessary writer.go:29: 2021-01-29T19:35:22.768Z [DEBUG] agent_smith.tlsutil: Update: version=1 writer.go:29: 2021-01-29T19:35:22.770Z [DEBUG] agent_smith.tlsutil: OutgoingRPCWrapper: version=1 writer.go:29: 2021-01-29T19:35:22.777Z [INFO] agent_smith.server.raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:8f808cf9-e84c-c7a0-7867-13520a6ec7d6 Address:127.0.0.1:10008}]" writer.go:29: 2021-01-29T19:35:22.779Z [INFO] agent_smith.server.raft: entering follower state: follower="Node at 127.0.0.1:10008 [Follower]" leader= writer.go:29: 2021-01-29T19:35:22.781Z [INFO] agent_smith.server.serf.wan: serf: EventMemberJoin: Node-8f808cf9-e84c-c7a0-7867-13520a6ec7d6.dc1 127.0.0.1 writer.go:29: 2021-01-29T19:35:22.785Z [INFO] agent_smith.server.serf.lan: serf: EventMemberJoin: Node-8f808cf9-e84c-c7a0-7867-13520a6ec7d6 127.0.0.1 writer.go:29: 2021-01-29T19:35:22.787Z [INFO] agent_smith: Started DNS server: address=127.0.0.1:10003 network=udp writer.go:29: 2021-01-29T19:35:22.788Z [INFO] agent_smith.server: Adding LAN server: server="Node-8f808cf9-e84c-c7a0-7867-13520a6ec7d6 (Addr: tcp/127.0.0.1:10008) (DC: dc1)" writer.go:29: 2021-01-29T19:35:22.789Z [INFO] agent_smith.server: Handled event for server in area: event=member-join server=Node-8f808cf9-e84c-c7a0-7867-13520a6ec7d6.dc1 area=wan writer.go:29: 2021-01-29T19:35:22.791Z [INFO] agent_smith: Started DNS server: address=127.0.0.1:10003 network=tcp writer.go:29: 2021-01-29T19:35:22.794Z [INFO] agent_smith: Started HTTP server: address=127.0.0.1:10004 network=tcp writer.go:29: 2021-01-29T19:35:22.795Z [INFO] agent_smith: started state syncer writer.go:29: 2021-01-29T19:35:22.844Z [WARN] agent_smith.server.raft: heartbeat timeout reached, starting election: last-leader= writer.go:29: 2021-01-29T19:35:22.847Z [INFO] agent_smith.server.raft: entering candidate state: node="Node at 127.0.0.1:10008 [Candidate]" term=2 writer.go:29: 2021-01-29T19:35:22.853Z [DEBUG] agent_smith.server.raft: votes: needed=1 writer.go:29: 2021-01-29T19:35:22.854Z [DEBUG] agent_smith.server.raft: vote granted: from=8f808cf9-e84c-c7a0-7867-13520a6ec7d6 term=2 tally=1 writer.go:29: 2021-01-29T19:35:22.856Z [INFO] agent_smith.server.raft: election won: tally=1 writer.go:29: 2021-01-29T19:35:22.858Z [INFO] agent_smith.server.raft: entering leader state: leader="Node at 127.0.0.1:10008 [Leader]" writer.go:29: 2021-01-29T19:35:22.860Z [INFO] agent_smith.server: cluster leadership acquired writer.go:29: 2021-01-29T19:35:22.861Z [INFO] agent_smith.server: New leader elected: payload=Node-8f808cf9-e84c-c7a0-7867-13520a6ec7d6 writer.go:29: 2021-01-29T19:35:22.863Z [DEBUG] agent_smith.server: Cannot upgrade to new ACLs: leaderMode=0 mode=0 found=true leader=127.0.0.1:10008 writer.go:29: 2021-01-29T19:35:22.880Z [DEBUG] connect.ca.consul: consul CA provider configured: id=07:80:c8:de:f6:41:86:29:8f:9c:b8:17:d6:48:c2:d5:c5:5c:7f:0c:03:f7:cf:97:5a:a7:c1:68:aa:23:ae:81 is_primary=true writer.go:29: 2021-01-29T19:35:22.889Z [INFO] agent_smith.server.connect: initialized primary datacenter CA with provider: provider=consul writer.go:29: 2021-01-29T19:35:22.891Z [INFO] agent_smith.leader: started routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:22.893Z [DEBUG] agent_smith.server: Skipping self join check for node since the cluster is too small: node=Node-8f808cf9-e84c-c7a0-7867-13520a6ec7d6 writer.go:29: 2021-01-29T19:35:22.895Z [INFO] agent_smith.server: member joined, marking health alive: member=Node-8f808cf9-e84c-c7a0-7867-13520a6ec7d6 writer.go:29: 2021-01-29T19:35:22.899Z [INFO] agent_smith: Synced node info writer.go:29: 2021-01-29T19:35:22.901Z [DEBUG] agent_smith: Node info in sync [WARN] freeport: 1 out of 2 pending ports are still in use; something probably didn't wait around for the port to be closed! writer.go:29: 2021-01-29T19:35:22.965Z [DEBUG] agent_smith: Skipping remote check since it is managed automatically: check=serfHealth writer.go:29: 2021-01-29T19:35:22.967Z [INFO] agent_smith: Synced node info [WARN] freeport: 1 out of 1 pending ports are still in use; something probably didn't wait around for the port to be closed! writer.go:29: 2021-01-29T19:35:23.195Z [DEBUG] agent_smith: Node info in sync writer.go:29: 2021-01-29T19:35:23.220Z [INFO] agent_smith: Synced service: service=web writer.go:29: 2021-01-29T19:35:23.226Z [INFO] agent_smith: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:35:23.228Z [DEBUG] agent_smith: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:35:23.229Z [DEBUG] agent_smith: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:35:23.230Z [DEBUG] agent_smith.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:58022 latency=43.546345ms writer.go:29: 2021-01-29T19:35:23.230Z [DEBUG] agent_smith: Node info in sync writer.go:29: 2021-01-29T19:35:23.232Z [DEBUG] agent_smith: Service in sync: service=web writer.go:29: 2021-01-29T19:35:23.233Z [DEBUG] agent_smith: Service in sync: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:35:23.236Z [DEBUG] agent_smith: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:35:23.237Z [DEBUG] agent_smith: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:35:23.240Z [DEBUG] agent_smith.http: Request finished: method=GET url=/v1/agent/service/web-sidecar-proxy from=127.0.0.1:58022 latency=2.83772ms writer.go:29: 2021-01-29T19:35:23.306Z [DEBUG] agent_smith: Node info in sync writer.go:29: 2021-01-29T19:35:23.310Z [INFO] agent_smith: Synced service: service=web writer.go:29: 2021-01-29T19:35:23.315Z [INFO] agent_smith: Synced service: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:35:23.316Z [DEBUG] agent_smith: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:35:23.318Z [DEBUG] agent_smith: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:35:23.319Z [DEBUG] agent_smith.http: Request finished: method=PUT url=/v1/agent/service/register from=127.0.0.1:58024 latency=22.900597ms writer.go:29: 2021-01-29T19:35:23.320Z [DEBUG] agent_smith.http: Request finished: method=GET url=/v1/agent/service/web-sidecar-proxy?hash=e79fad2be70b72e8 from=127.0.0.1:58022 latency=41.795066ms writer.go:29: 2021-01-29T19:35:23.320Z [DEBUG] agent_smith: Node info in sync writer.go:29: 2021-01-29T19:35:23.324Z [DEBUG] agent_smith: Service in sync: service=web writer.go:29: 2021-01-29T19:35:23.324Z [DEBUG] agent_smith: Service in sync: service=web-sidecar-proxy writer.go:29: 2021-01-29T19:35:23.325Z [DEBUG] agent_smith: Check in sync: check=service:web-sidecar-proxy:1 writer.go:29: 2021-01-29T19:35:23.326Z [DEBUG] agent_smith: Check in sync: check=service:web-sidecar-proxy:2 writer.go:29: 2021-01-29T19:35:23.323Z [INFO] agent_smith: Requesting shutdown writer.go:29: 2021-01-29T19:35:23.328Z [INFO] agent_smith.server: shutting down server writer.go:29: 2021-01-29T19:35:23.328Z [DEBUG] agent_smith.leader: stopping routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:23.329Z [WARN] agent_smith.server.serf.lan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:23.330Z [DEBUG] agent_smith.leader: stopped routine: routine="CA root pruning" writer.go:29: 2021-01-29T19:35:23.332Z [WARN] agent_smith.server.serf.wan: serf: Shutdown without a Leave writer.go:29: 2021-01-29T19:35:23.334Z [INFO] agent_smith.server.router.manager: shutting down writer.go:29: 2021-01-29T19:35:23.334Z [INFO] agent_smith: consul server down writer.go:29: 2021-01-29T19:35:23.336Z [INFO] agent_smith: shutdown complete writer.go:29: 2021-01-29T19:35:23.337Z [INFO] agent_smith: Stopping server: protocol=DNS address=127.0.0.1:10003 network=tcp writer.go:29: 2021-01-29T19:35:23.338Z [INFO] agent_smith: Stopping server: protocol=DNS address=127.0.0.1:10003 network=udp writer.go:29: 2021-01-29T19:35:23.339Z [INFO] agent_smith: Stopping server: protocol=HTTP address=127.0.0.1:10004 network=tcp [WARN] freeport: 1 out of 1 pending ports are still in use; something probably didn't wait around for the port to be closed! [WARN] freeport: 1 out of 1 pending ports are still in use; something probably didn't wait around for the port to be closed! [WARN] freeport: 1 out of 1 pending ports are still in use; something probably didn't wait around for the port to be closed! [WARN] freeport: 1 out of 1 pending ports are still in use; something probably didn't wait around for the port to be closed! writer.go:29: 2021-01-29T19:35:24.343Z [WARN] agent_smith: Timeout stopping server: protocol=HTTP address=127.0.0.1:10004 network=tcp writer.go:29: 2021-01-29T19:35:24.345Z [INFO] agent_smith: Waiting for endpoints to shut down writer.go:29: 2021-01-29T19:35:24.346Z [INFO] agent_smith: Endpoints down --- PASS: TestAgentConfigWatcherSidecarProxy (1.61s) PASS ok github.com/hashicorp/consul/connect/proxy 1.950s === RUN TestIsPrivateIP === RUN TestIsPrivateIP/10.0.0.1 === RUN TestIsPrivateIP/100.64.0.1 === RUN TestIsPrivateIP/172.16.0.1 === RUN TestIsPrivateIP/192.168.0.1 === RUN TestIsPrivateIP/192.0.0.1 === RUN TestIsPrivateIP/192.0.2.1 === RUN TestIsPrivateIP/127.0.0.1 === RUN TestIsPrivateIP/169.254.0.1 === RUN TestIsPrivateIP/1.2.3.4 === RUN TestIsPrivateIP/::1 === RUN TestIsPrivateIP/fe80::1 === RUN TestIsPrivateIP/fc00::1 === RUN TestIsPrivateIP/fec0::1 === RUN TestIsPrivateIP/2001:db8::1 === RUN TestIsPrivateIP/2004:db6::1 --- PASS: TestIsPrivateIP (0.00s) --- PASS: TestIsPrivateIP/10.0.0.1 (0.00s) --- PASS: TestIsPrivateIP/100.64.0.1 (0.00s) --- PASS: TestIsPrivateIP/172.16.0.1 (0.00s) --- PASS: TestIsPrivateIP/192.168.0.1 (0.00s) --- PASS: TestIsPrivateIP/192.0.0.1 (0.00s) --- PASS: TestIsPrivateIP/192.0.2.1 (0.00s) --- PASS: TestIsPrivateIP/127.0.0.1 (0.00s) --- PASS: TestIsPrivateIP/169.254.0.1 (0.00s) --- PASS: TestIsPrivateIP/1.2.3.4 (0.00s) --- PASS: TestIsPrivateIP/::1 (0.00s) --- PASS: TestIsPrivateIP/fe80::1 (0.00s) --- PASS: TestIsPrivateIP/fc00::1 (0.00s) --- PASS: TestIsPrivateIP/fec0::1 (0.00s) --- PASS: TestIsPrivateIP/2001:db8::1 (0.00s) --- PASS: TestIsPrivateIP/2004:db6::1 (0.00s) === RUN TestIsIPv6 === RUN TestIsIPv6/10.0.0.1 === RUN TestIsIPv6/100.64.0.1 === RUN TestIsIPv6/172.16.0.1 === RUN TestIsIPv6/192.168.0.1 === RUN TestIsIPv6/192.0.0.1 === RUN TestIsIPv6/192.0.2.1 === RUN TestIsIPv6/127.0.0.1 === RUN TestIsIPv6/169.254.0.1 === RUN TestIsIPv6/1.2.3.4 === RUN TestIsIPv6/::1 === RUN TestIsIPv6/fe80::1 === RUN TestIsIPv6/fc00::1 === RUN TestIsIPv6/fec0::1 === RUN TestIsIPv6/2001:db8::1 === RUN TestIsIPv6/2004:db6::1 === RUN TestIsIPv6/example.com === RUN TestIsIPv6/localhost === RUN TestIsIPv6/1.257.0.1 --- PASS: TestIsIPv6 (0.00s) --- PASS: TestIsIPv6/10.0.0.1 (0.00s) --- PASS: TestIsIPv6/100.64.0.1 (0.00s) --- PASS: TestIsIPv6/172.16.0.1 (0.00s) --- PASS: TestIsIPv6/192.168.0.1 (0.00s) --- PASS: TestIsIPv6/192.0.0.1 (0.00s) --- PASS: TestIsIPv6/192.0.2.1 (0.00s) --- PASS: TestIsIPv6/127.0.0.1 (0.00s) --- PASS: TestIsIPv6/169.254.0.1 (0.00s) --- PASS: TestIsIPv6/1.2.3.4 (0.00s) --- PASS: TestIsIPv6/::1 (0.00s) --- PASS: TestIsIPv6/fe80::1 (0.00s) --- PASS: TestIsIPv6/fc00::1 (0.00s) --- PASS: TestIsIPv6/fec0::1 (0.00s) --- PASS: TestIsIPv6/2001:db8::1 (0.00s) --- PASS: TestIsIPv6/2004:db6::1 (0.00s) --- PASS: TestIsIPv6/example.com (0.00s) --- PASS: TestIsIPv6/localhost (0.00s) --- PASS: TestIsIPv6/1.257.0.1 (0.00s) PASS ok github.com/hashicorp/consul/ipaddr 0.028s === RUN TestDurationMinusBuffer --- PASS: TestDurationMinusBuffer (0.00s) === RUN TestDurationMinusBufferDomain --- PASS: TestDurationMinusBufferDomain (0.00s) === RUN TestRandomStagger --- PASS: TestRandomStagger (0.00s) === RUN TestRateScaledInterval --- PASS: TestRateScaledInterval (0.00s) === RUN TestMapWalk map_walker_test.go:10: DM-skipped --- SKIP: TestMapWalk (0.00s) === RUN TestPatchSliceOfMaps === RUN TestPatchSliceOfMaps/00:_{"a":{"b":"c"}}_->_{"a":{"b":"c"}}_skip:_[] === RUN TestPatchSliceOfMaps/01:_{"a":[{"b":"c"}]}_->_{"a":{"b":"c"}}_skip:_[] === RUN TestPatchSliceOfMaps/02:_{"a":[{"b":[{"c":"d"}]}]}_->_{"a":{"b":{"c":"d"}}}_skip:_[] === RUN TestPatchSliceOfMaps/03:_{"a":[{"b":"c"}]}_->_{"a":[{"b":"c"}]}_skip:_[a] === RUN TestPatchSliceOfMaps/04:_{_____"Services":_[______{_______"checks":_[________{_________"header":_[__________{"a":"b"}_________]________}_______]______}_____]____}_->_{_____"Services":_[______{_______"checks":_[________{_________"header":_{"a":"b"}________}_______]______}_____]____}_skip:_[services_services.checks] === RUN TestPatchSliceOfMaps/05:_____{_____"a":_[______{_______"b":_[________{_________"c":_"val1",_________"d":_{__________"foo":_"bar"_________},_________"e":_[__________{___________"super":_"duper"__________}_________]________}_______]______}_____]____}_____->_____{_____"a":_{______"b":_[_______{________"c":_"val1",________"d":_{_________"foo":_"bar"________},________"e":_[_________{__________"super":_"duper"_________}________]_______}______]_____}____}_____skip:_[] --- PASS: TestPatchSliceOfMaps (0.00s) --- PASS: TestPatchSliceOfMaps/00:_{"a":{"b":"c"}}_->_{"a":{"b":"c"}}_skip:_[] (0.00s) --- PASS: TestPatchSliceOfMaps/01:_{"a":[{"b":"c"}]}_->_{"a":{"b":"c"}}_skip:_[] (0.00s) --- PASS: TestPatchSliceOfMaps/02:_{"a":[{"b":[{"c":"d"}]}]}_->_{"a":{"b":{"c":"d"}}}_skip:_[] (0.00s) --- PASS: TestPatchSliceOfMaps/03:_{"a":[{"b":"c"}]}_->_{"a":[{"b":"c"}]}_skip:_[a] (0.00s) --- PASS: TestPatchSliceOfMaps/04:_{_____"Services":_[______{_______"checks":_[________{_________"header":_[__________{"a":"b"}_________]________}_______]______}_____]____}_->_{_____"Services":_[______{_______"checks":_[________{_________"header":_{"a":"b"}________}_______]______}_____]____}_skip:_[services_services.checks] (0.00s) --- PASS: TestPatchSliceOfMaps/05:_____{_____"a":_[______{_______"b":_[________{_________"c":_"val1",_________"d":_{__________"foo":_"bar"_________},_________"e":_[__________{___________"super":_"duper"__________}_________]________}_______]______}_____]____}_____->_____{_____"a":_{______"b":_[_______{________"c":_"val1",________"d":_{_________"foo":_"bar"________},________"e":_[_________{__________"super":_"duper"_________}________]_______}______]_____}____}_____skip:_[] (0.00s) === RUN TestJitterRandomStagger === PAUSE TestJitterRandomStagger === RUN TestRetryWaiter_calculateWait === PAUSE TestRetryWaiter_calculateWait === RUN TestRetryWaiter_WaitChans === PAUSE TestRetryWaiter_WaitChans === RUN TestRTT_ComputeDistance === RUN TestRTT_ComputeDistance/10_ms === RUN TestRTT_ComputeDistance/0_ms === RUN TestRTT_ComputeDistance/2_ms === RUN TestRTT_ComputeDistance/2_ms_reversed === RUN TestRTT_ComputeDistance/a_nil === RUN TestRTT_ComputeDistance/b_nil === RUN TestRTT_ComputeDistance/both_nil --- PASS: TestRTT_ComputeDistance (0.00s) --- PASS: TestRTT_ComputeDistance/10_ms (0.00s) --- PASS: TestRTT_ComputeDistance/0_ms (0.00s) --- PASS: TestRTT_ComputeDistance/2_ms (0.00s) --- PASS: TestRTT_ComputeDistance/2_ms_reversed (0.00s) --- PASS: TestRTT_ComputeDistance/a_nil (0.00s) --- PASS: TestRTT_ComputeDistance/b_nil (0.00s) --- PASS: TestRTT_ComputeDistance/both_nil (0.00s) === RUN TestRTT_Intersect === RUN TestRTT_Intersect/nil_maps === RUN TestRTT_Intersect/two_servers === RUN TestRTT_Intersect/two_clients === RUN TestRTT_Intersect/server1_and_client_alpha === RUN TestRTT_Intersect/server1_and_client_beta_1 === RUN TestRTT_Intersect/server1_and_client_alpha_reversed === RUN TestRTT_Intersect/server1_and_client_beta_1_reversed === RUN TestRTT_Intersect/nothing_in_common === RUN TestRTT_Intersect/nothing_in_common_reversed --- PASS: TestRTT_Intersect (0.00s) --- PASS: TestRTT_Intersect/nil_maps (0.00s) --- PASS: TestRTT_Intersect/two_servers (0.00s) --- PASS: TestRTT_Intersect/two_clients (0.00s) --- PASS: TestRTT_Intersect/server1_and_client_alpha (0.00s) --- PASS: TestRTT_Intersect/server1_and_client_beta_1 (0.00s) --- PASS: TestRTT_Intersect/server1_and_client_alpha_reversed (0.00s) --- PASS: TestRTT_Intersect/server1_and_client_beta_1_reversed (0.00s) --- PASS: TestRTT_Intersect/nothing_in_common (0.00s) --- PASS: TestRTT_Intersect/nothing_in_common_reversed (0.00s) === RUN TestStrContains --- PASS: TestStrContains (0.00s) === RUN TestTelemetryConfig_MergeDefaults === RUN TestTelemetryConfig_MergeDefaults/basic_merge === RUN TestTelemetryConfig_MergeDefaults/exhaustive --- PASS: TestTelemetryConfig_MergeDefaults (0.00s) --- PASS: TestTelemetryConfig_MergeDefaults/basic_merge (0.00s) --- PASS: TestTelemetryConfig_MergeDefaults/exhaustive (0.00s) === RUN TestTranslateKeys === RUN TestTranslateKeys/x->y === RUN TestTranslateKeys/discard_x === RUN TestTranslateKeys/b.x->b.y === RUN TestTranslateKeys/json:_x->y === RUN TestTranslateKeys/json:_X->y === RUN TestTranslateKeys/json:_discard_x === RUN TestTranslateKeys/json:_b.x->b.y === RUN TestTranslateKeys/json:_b[0].x->b[0].y --- PASS: TestTranslateKeys (0.01s) --- PASS: TestTranslateKeys/x->y (0.00s) --- PASS: TestTranslateKeys/discard_x (0.00s) --- PASS: TestTranslateKeys/b.x->b.y (0.00s) --- PASS: TestTranslateKeys/json:_x->y (0.00s) --- PASS: TestTranslateKeys/json:_X->y (0.00s) --- PASS: TestTranslateKeys/json:_discard_x (0.00s) --- PASS: TestTranslateKeys/json:_b.x->b.y (0.00s) --- PASS: TestTranslateKeys/json:_b[0].x->b[0].y (0.00s) === RUN TestUserAgent --- PASS: TestUserAgent (0.00s) === RUN TestMathAbsInt --- PASS: TestMathAbsInt (0.00s) === RUN TestMathMaxInt --- PASS: TestMathMaxInt (0.00s) === RUN TestMathMinInt --- PASS: TestMathMinInt (0.00s) === CONT TestJitterRandomStagger === RUN TestJitterRandomStagger/0_percent === PAUSE TestJitterRandomStagger/0_percent === RUN TestJitterRandomStagger/10_percent === PAUSE TestJitterRandomStagger/10_percent === RUN TestJitterRandomStagger/100_percent === PAUSE TestJitterRandomStagger/100_percent === CONT TestJitterRandomStagger/0_percent === CONT TestRetryWaiter_WaitChans === RUN TestRetryWaiter_WaitChans/Minimum_Wait_-_Success === PAUSE TestRetryWaiter_WaitChans/Minimum_Wait_-_Success === RUN TestRetryWaiter_WaitChans/Minimum_Wait_-_WaitIf === PAUSE TestRetryWaiter_WaitChans/Minimum_Wait_-_WaitIf === RUN TestRetryWaiter_WaitChans/Minimum_Wait_-_WaitIfErr === PAUSE TestRetryWaiter_WaitChans/Minimum_Wait_-_WaitIfErr === RUN TestRetryWaiter_WaitChans/Maximum_Wait_-_Failed === PAUSE TestRetryWaiter_WaitChans/Maximum_Wait_-_Failed === RUN TestRetryWaiter_WaitChans/Maximum_Wait_-_WaitIf === PAUSE TestRetryWaiter_WaitChans/Maximum_Wait_-_WaitIf === RUN TestRetryWaiter_WaitChans/Maximum_Wait_-_WaitIfErr === PAUSE TestRetryWaiter_WaitChans/Maximum_Wait_-_WaitIfErr === CONT TestRetryWaiter_WaitChans/Minimum_Wait_-_Success === CONT TestRetryWaiter_calculateWait === RUN TestRetryWaiter_calculateWait/Defaults === PAUSE TestRetryWaiter_calculateWait/Defaults === RUN TestRetryWaiter_calculateWait/Minimum_Wait === PAUSE TestRetryWaiter_calculateWait/Minimum_Wait === RUN TestRetryWaiter_calculateWait/Minimum_Failures === PAUSE TestRetryWaiter_calculateWait/Minimum_Failures === RUN TestRetryWaiter_calculateWait/Maximum_Wait === PAUSE TestRetryWaiter_calculateWait/Maximum_Wait === CONT TestRetryWaiter_calculateWait/Defaults === CONT TestRetryWaiter_WaitChans/Maximum_Wait_-_Failed === CONT TestJitterRandomStagger/100_percent === CONT TestJitterRandomStagger/10_percent === CONT TestRetryWaiter_WaitChans/Maximum_Wait_-_WaitIfErr --- PASS: TestJitterRandomStagger (0.00s) --- PASS: TestJitterRandomStagger/0_percent (0.00s) --- PASS: TestJitterRandomStagger/10_percent (0.00s) --- PASS: TestJitterRandomStagger/100_percent (0.01s) === CONT TestRetryWaiter_WaitChans/Maximum_Wait_-_WaitIf === CONT TestRetryWaiter_calculateWait/Maximum_Wait === CONT TestRetryWaiter_WaitChans/Minimum_Wait_-_WaitIfErr === CONT TestRetryWaiter_WaitChans/Minimum_Wait_-_WaitIf === CONT TestRetryWaiter_calculateWait/Minimum_Failures === CONT TestRetryWaiter_calculateWait/Minimum_Wait --- PASS: TestRetryWaiter_calculateWait (0.00s) --- PASS: TestRetryWaiter_calculateWait/Defaults (0.00s) --- PASS: TestRetryWaiter_calculateWait/Maximum_Wait (0.00s) --- PASS: TestRetryWaiter_calculateWait/Minimum_Failures (0.00s) --- PASS: TestRetryWaiter_calculateWait/Minimum_Wait (0.00s) --- PASS: TestRetryWaiter_WaitChans (0.00s) --- PASS: TestRetryWaiter_WaitChans/Minimum_Wait_-_Success (0.20s) --- PASS: TestRetryWaiter_WaitChans/Maximum_Wait_-_Failed (0.25s) --- PASS: TestRetryWaiter_WaitChans/Maximum_Wait_-_WaitIfErr (0.25s) --- PASS: TestRetryWaiter_WaitChans/Maximum_Wait_-_WaitIf (0.25s) --- PASS: TestRetryWaiter_WaitChans/Minimum_Wait_-_WaitIfErr (0.20s) --- PASS: TestRetryWaiter_WaitChans/Minimum_Wait_-_WaitIf (0.20s) PASS ok github.com/hashicorp/consul/lib 0.483s === RUN TestWriteAtomic --- PASS: TestWriteAtomic (0.00s) PASS ok github.com/hashicorp/consul/lib/file 0.010s === RUN TestDynamic === PAUSE TestDynamic === RUN TestDynamicPanic === PAUSE TestDynamicPanic === RUN TestDynamicAcquire === PAUSE TestDynamicAcquire === CONT TestDynamic === CONT TestDynamicPanic --- PASS: TestDynamicPanic (0.00s) === CONT TestDynamicAcquire --- PASS: TestDynamicAcquire (0.05s) --- PASS: TestDynamic (1.61s) PASS ok github.com/hashicorp/consul/lib/semaphore 1.633s # github.com/hashicorp/consul/sdk/testutil/retry.test loadinternal: cannot find runtime/cgo === RUN TestGatedWriter_impl --- PASS: TestGatedWriter_impl (0.00s) === RUN TestGatedWriter --- PASS: TestGatedWriter (0.00s) === RUN TestGRPCLogger --- PASS: TestGRPCLogger (0.00s) === RUN TestGRPCLogger_V === RUN TestGRPCLogger_V/ERR,-1 === RUN TestGRPCLogger_V/ERR,0 === RUN TestGRPCLogger_V/ERR,1 === RUN TestGRPCLogger_V/ERR,2 === RUN TestGRPCLogger_V/ERR,3 === RUN TestGRPCLogger_V/WARN,-1 === RUN TestGRPCLogger_V/WARN,0 === RUN TestGRPCLogger_V/WARN,1 === RUN TestGRPCLogger_V/WARN,2 === RUN TestGRPCLogger_V/WARN,3 === RUN TestGRPCLogger_V/INFO,-1 === RUN TestGRPCLogger_V/INFO,0 === RUN TestGRPCLogger_V/INFO,1 === RUN TestGRPCLogger_V/INFO,2 === RUN TestGRPCLogger_V/INFO,3 === RUN TestGRPCLogger_V/DEBUG,-1 === RUN TestGRPCLogger_V/DEBUG,0 === RUN TestGRPCLogger_V/DEBUG,1 === RUN TestGRPCLogger_V/DEBUG,2 === RUN TestGRPCLogger_V/DEBUG,3 === RUN TestGRPCLogger_V/TRACE,-1 === RUN TestGRPCLogger_V/TRACE,0 === RUN TestGRPCLogger_V/TRACE,1 === RUN TestGRPCLogger_V/TRACE,2 === RUN TestGRPCLogger_V/TRACE,3 --- PASS: TestGRPCLogger_V (0.00s) --- PASS: TestGRPCLogger_V/ERR,-1 (0.00s) --- PASS: TestGRPCLogger_V/ERR,0 (0.00s) --- PASS: TestGRPCLogger_V/ERR,1 (0.00s) --- PASS: TestGRPCLogger_V/ERR,2 (0.00s) --- PASS: TestGRPCLogger_V/ERR,3 (0.00s) --- PASS: TestGRPCLogger_V/WARN,-1 (0.00s) --- PASS: TestGRPCLogger_V/WARN,0 (0.00s) --- PASS: TestGRPCLogger_V/WARN,1 (0.00s) --- PASS: TestGRPCLogger_V/WARN,2 (0.00s) --- PASS: TestGRPCLogger_V/WARN,3 (0.00s) --- PASS: TestGRPCLogger_V/INFO,-1 (0.00s) --- PASS: TestGRPCLogger_V/INFO,0 (0.00s) --- PASS: TestGRPCLogger_V/INFO,1 (0.00s) --- PASS: TestGRPCLogger_V/INFO,2 (0.00s) --- PASS: TestGRPCLogger_V/INFO,3 (0.00s) --- PASS: TestGRPCLogger_V/DEBUG,-1 (0.00s) --- PASS: TestGRPCLogger_V/DEBUG,0 (0.00s) --- PASS: TestGRPCLogger_V/DEBUG,1 (0.00s) --- PASS: TestGRPCLogger_V/DEBUG,2 (0.00s) --- PASS: TestGRPCLogger_V/DEBUG,3 (0.00s) --- PASS: TestGRPCLogger_V/TRACE,-1 (0.00s) --- PASS: TestGRPCLogger_V/TRACE,0 (0.00s) --- PASS: TestGRPCLogger_V/TRACE,1 (0.00s) --- PASS: TestGRPCLogger_V/TRACE,2 (0.00s) --- PASS: TestGRPCLogger_V/TRACE,3 (0.00s) === RUN TestLogFile_timeRotation === PAUSE TestLogFile_timeRotation === RUN TestLogFile_openNew === PAUSE TestLogFile_openNew === RUN TestLogFile_byteRotation === PAUSE TestLogFile_byteRotation === RUN TestLogFile_deleteArchives === PAUSE TestLogFile_deleteArchives === RUN TestLogFile_deleteArchivesDisabled === PAUSE TestLogFile_deleteArchivesDisabled === RUN TestLogFile_rotationDisabled === PAUSE TestLogFile_rotationDisabled === RUN TestLogger_SetupBasic === PAUSE TestLogger_SetupBasic === RUN TestLogger_SetupInvalidLogLevel === PAUSE TestLogger_SetupInvalidLogLevel === RUN TestLogger_SetupLoggerErrorLevel === PAUSE TestLogger_SetupLoggerErrorLevel === RUN TestLogger_SetupLoggerDebugLevel === PAUSE TestLogger_SetupLoggerDebugLevel === RUN TestLogger_SetupLoggerWithName === PAUSE TestLogger_SetupLoggerWithName === RUN TestLogger_SetupLoggerWithJSON === PAUSE TestLogger_SetupLoggerWithJSON === CONT TestLogFile_timeRotation === CONT TestLogger_SetupInvalidLogLevel --- PASS: TestLogger_SetupInvalidLogLevel (0.00s) === CONT TestLogFile_deleteArchivesDisabled === CONT TestLogger_SetupBasic --- PASS: TestLogger_SetupBasic (0.00s) === CONT TestLogFile_deleteArchives --- PASS: TestLogFile_deleteArchives (0.00s) === CONT TestLogFile_byteRotation --- PASS: TestLogFile_byteRotation (0.00s) === CONT TestLogFile_openNew --- PASS: TestLogFile_openNew (0.00s) === CONT TestLogger_SetupLoggerDebugLevel --- PASS: TestLogger_SetupLoggerDebugLevel (0.00s) === CONT TestLogger_SetupLoggerErrorLevel === RUN TestLogger_SetupLoggerErrorLevel/ERR_log_level === RUN TestLogger_SetupLoggerErrorLevel/ERROR_log_level --- PASS: TestLogger_SetupLoggerErrorLevel (0.00s) --- PASS: TestLogger_SetupLoggerErrorLevel/ERR_log_level (0.00s) --- PASS: TestLogger_SetupLoggerErrorLevel/ERROR_log_level (0.00s) === CONT TestLogFile_rotationDisabled --- PASS: TestLogFile_rotationDisabled (0.00s) --- PASS: TestLogFile_deleteArchivesDisabled (0.00s) === CONT TestLogger_SetupLoggerWithJSON === CONT TestLogger_SetupLoggerWithName --- PASS: TestLogger_SetupLoggerWithJSON (0.00s) --- PASS: TestLogger_SetupLoggerWithName (0.00s) --- PASS: TestLogFile_timeRotation (2.00s) PASS ok github.com/hashicorp/consul/logging 2.007s === RUN TestMonitor_Start === PAUSE TestMonitor_Start === RUN TestMonitor_Stop === PAUSE TestMonitor_Stop === RUN TestMonitor_DroppedMessages === PAUSE TestMonitor_DroppedMessages === RUN TestMonitor_ZeroBufSizeDefault === PAUSE TestMonitor_ZeroBufSizeDefault === RUN TestMonitor_WriteStopped === PAUSE TestMonitor_WriteStopped === CONT TestMonitor_Start --- PASS: TestMonitor_Start (0.00s) === CONT TestMonitor_WriteStopped --- PASS: TestMonitor_WriteStopped (0.00s) === CONT TestMonitor_ZeroBufSizeDefault === CONT TestMonitor_DroppedMessages === CONT TestMonitor_Stop --- PASS: TestMonitor_Stop (0.10s) --- PASS: TestMonitor_ZeroBufSizeDefault (0.10s) --- PASS: TestMonitor_DroppedMessages (0.10s) PASS ok github.com/hashicorp/consul/logging/monitor 0.132s === RUN TestGetEphemeralPortRange ephemeral_linux_test.go:17: min=32768, max=60999 --- PASS: TestGetEphemeralPortRange (0.01s) === RUN TestTakeReturn freeport_test.go:13: DM-skipped --- SKIP: TestTakeReturn (0.00s) === RUN TestIntervalOverlap === RUN TestIntervalOverlap/0:0_vs_0:0 === RUN TestIntervalOverlap/1:1_vs_1:1 === RUN TestIntervalOverlap/1:3_vs_1:3 === RUN TestIntervalOverlap/1:3_vs_4:6 === RUN TestIntervalOverlap/1:4_vs_3:6 === RUN TestIntervalOverlap/1:6_vs_3:4 --- PASS: TestIntervalOverlap (0.00s) --- PASS: TestIntervalOverlap/0:0_vs_0:0 (0.00s) --- PASS: TestIntervalOverlap/1:1_vs_1:1 (0.00s) --- PASS: TestIntervalOverlap/1:3_vs_1:3 (0.00s) --- PASS: TestIntervalOverlap/1:3_vs_4:6 (0.00s) --- PASS: TestIntervalOverlap/1:4_vs_3:6 (0.00s) --- PASS: TestIntervalOverlap/1:6_vs_3:4 (0.00s) PASS ok github.com/hashicorp/consul/sdk/freeport 0.013s ? github.com/hashicorp/consul/sdk/testutil [no test files] === RUN TestRetryer retry_test.go:12: DM-skipped --- SKIP: TestRetryer (0.00s) PASS ok github.com/hashicorp/consul/sdk/testutil/retry 0.003s ? github.com/hashicorp/consul/sentinel [no test files] ? github.com/hashicorp/consul/service_os [no test files] === RUN TestArchive --- PASS: TestArchive (0.00s) === RUN TestArchive_GoodData --- PASS: TestArchive_GoodData (0.00s) === RUN TestArchive_BadData --- PASS: TestArchive_BadData (0.00s) === RUN TestArchive_hashList --- PASS: TestArchive_hashList (0.00s) === RUN TestSnapshot 2021-01-29T19:35:26.171Z [INFO] raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:server-910a5011-6550-1126-f870-04cf1dd8975f Address:910a5011-6550-1126-f870-04cf1dd8975f}]" 2021-01-29T19:35:26.171Z [INFO] raft: entering follower state: follower="Node at 910a5011-6550-1126-f870-04cf1dd8975f [Follower]" leader= 2021-01-29T19:35:27.430Z [WARN] raft: heartbeat timeout reached, starting election: last-leader= 2021-01-29T19:35:27.430Z [INFO] raft: entering candidate state: node="Node at 910a5011-6550-1126-f870-04cf1dd8975f [Candidate]" term=2 2021-01-29T19:35:27.431Z [DEBUG] raft: votes: needed=1 2021-01-29T19:35:27.431Z [DEBUG] raft: vote granted: from=server-910a5011-6550-1126-f870-04cf1dd8975f term=2 tally=1 2021-01-29T19:35:27.431Z [INFO] raft: election won: tally=1 2021-01-29T19:35:27.431Z [INFO] raft: entering leader state: leader="Node at 910a5011-6550-1126-f870-04cf1dd8975f [Leader]" 2021-01-29T19:35:28.739Z [INFO] raft: starting snapshot up to: index=65538 2021-01-29T19:35:28.739Z [INFO] snapshot: creating new snapshot: path=/tmp/consul-test/TestSnapshot-snapshot569633851/before/snapshots/2-65538-1611948928739.tmp 2021-01-29T19:35:28.861Z [INFO] raft: compacting logs: from=1 to=55298 2021-01-29T19:35:28.866Z [INFO] raft: snapshot complete up to: index=65538 2021-01-29T19:35:29.546Z [INFO] raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:server-1f983d89-106e-9ae5-1c10-ad0485e43111 Address:1f983d89-106e-9ae5-1c10-ad0485e43111}]" 2021-01-29T19:35:29.547Z [INFO] raft: entering follower state: follower="Node at 1f983d89-106e-9ae5-1c10-ad0485e43111 [Follower]" leader= 2021-01-29T19:35:31.384Z [WARN] raft: heartbeat timeout reached, starting election: last-leader= 2021-01-29T19:35:31.384Z [INFO] raft: entering candidate state: node="Node at 1f983d89-106e-9ae5-1c10-ad0485e43111 [Candidate]" term=2 2021-01-29T19:35:31.384Z [DEBUG] raft: votes: needed=1 2021-01-29T19:35:31.384Z [DEBUG] raft: vote granted: from=server-1f983d89-106e-9ae5-1c10-ad0485e43111 term=2 tally=1 2021-01-29T19:35:31.384Z [INFO] raft: election won: tally=1 2021-01-29T19:35:31.384Z [INFO] raft: entering leader state: leader="Node at 1f983d89-106e-9ae5-1c10-ad0485e43111 [Leader]" 2021-01-29T19:35:31.439Z [INFO] snapshot: creating new snapshot: path=/tmp/consul-test/TestSnapshot-snapshot569633851/after/snapshots/2-65539-1611948931439.tmp 2021-01-29T19:35:31.490Z [INFO] raft: copied to local snapshot: bytes=16973829 2021-01-29T19:35:31.530Z [INFO] raft: restored user snapshot: index=1 --- PASS: TestSnapshot (5.39s) === RUN TestSnapshot_Nil --- PASS: TestSnapshot_Nil (0.00s) === RUN TestSnapshot_BadVerify --- PASS: TestSnapshot_BadVerify (0.00s) === RUN TestSnapshot_TruncatedVerify 2021-01-29T19:35:31.558Z [INFO] raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:server-e28d71e3-3b72-9882-ebb4-ba0735b641fb Address:e28d71e3-3b72-9882-ebb4-ba0735b641fb}]" 2021-01-29T19:35:31.558Z [INFO] raft: entering follower state: follower="Node at e28d71e3-3b72-9882-ebb4-ba0735b641fb [Follower]" leader= 2021-01-29T19:35:33.508Z [WARN] raft: heartbeat timeout reached, starting election: last-leader= 2021-01-29T19:35:33.509Z [INFO] raft: entering candidate state: node="Node at e28d71e3-3b72-9882-ebb4-ba0735b641fb [Candidate]" term=2 2021-01-29T19:35:33.509Z [DEBUG] raft: votes: needed=1 2021-01-29T19:35:33.509Z [DEBUG] raft: vote granted: from=server-e28d71e3-3b72-9882-ebb4-ba0735b641fb term=2 tally=1 2021-01-29T19:35:33.509Z [INFO] raft: election won: tally=1 2021-01-29T19:35:33.509Z [INFO] raft: entering leader state: leader="Node at e28d71e3-3b72-9882-ebb4-ba0735b641fb [Leader]" 2021-01-29T19:35:44.085Z [INFO] raft: starting snapshot up to: index=65538 2021-01-29T19:35:44.085Z [INFO] snapshot: creating new snapshot: path=/tmp/consul-test/TestSnapshot_TruncatedVerify-snapshot076829504/before/snapshots/2-65538-1611948944085.tmp 2021-01-29T19:35:44.166Z [INFO] raft: compacting logs: from=1 to=55298 2021-01-29T19:35:44.170Z [INFO] raft: snapshot complete up to: index=65538 === RUN TestSnapshot_TruncatedVerify/truncate_200_bytes_from_end === RUN TestSnapshot_TruncatedVerify/truncate_16_bytes_from_end === RUN TestSnapshot_TruncatedVerify/truncate_8_bytes_from_end === RUN TestSnapshot_TruncatedVerify/truncate_4_bytes_from_end === RUN TestSnapshot_TruncatedVerify/truncate_2_bytes_from_end === RUN TestSnapshot_TruncatedVerify/truncate_1_bytes_from_end --- PASS: TestSnapshot_TruncatedVerify (13.31s) --- PASS: TestSnapshot_TruncatedVerify/truncate_200_bytes_from_end (0.01s) --- PASS: TestSnapshot_TruncatedVerify/truncate_16_bytes_from_end (0.01s) --- PASS: TestSnapshot_TruncatedVerify/truncate_8_bytes_from_end (0.01s) --- PASS: TestSnapshot_TruncatedVerify/truncate_4_bytes_from_end (0.01s) --- PASS: TestSnapshot_TruncatedVerify/truncate_2_bytes_from_end (0.01s) --- PASS: TestSnapshot_TruncatedVerify/truncate_1_bytes_from_end (0.01s) === RUN TestSnapshot_BadRestore 2021-01-29T19:35:44.868Z [INFO] raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:server-201ff669-eb46-78ce-c663-c42535c6c156 Address:201ff669-eb46-78ce-c663-c42535c6c156}]" 2021-01-29T19:35:44.868Z [INFO] raft: entering follower state: follower="Node at 201ff669-eb46-78ce-c663-c42535c6c156 [Follower]" leader= 2021-01-29T19:35:46.235Z [WARN] raft: heartbeat timeout reached, starting election: last-leader= 2021-01-29T19:35:46.235Z [INFO] raft: entering candidate state: node="Node at 201ff669-eb46-78ce-c663-c42535c6c156 [Candidate]" term=2 2021-01-29T19:35:46.235Z [DEBUG] raft: votes: needed=1 2021-01-29T19:35:46.235Z [DEBUG] raft: vote granted: from=server-201ff669-eb46-78ce-c663-c42535c6c156 term=2 tally=1 2021-01-29T19:35:46.235Z [INFO] raft: election won: tally=1 2021-01-29T19:35:46.235Z [INFO] raft: entering leader state: leader="Node at 201ff669-eb46-78ce-c663-c42535c6c156 [Leader]" 2021-01-29T19:35:46.552Z [INFO] raft: starting snapshot up to: index=16386 2021-01-29T19:35:46.552Z [INFO] snapshot: creating new snapshot: path=/tmp/consul-test/TestSnapshot_BadRestore-snapshot837055602/before/snapshots/2-16386-1611948946552.tmp 2021-01-29T19:35:46.582Z [INFO] raft: compacting logs: from=1 to=6146 2021-01-29T19:35:46.582Z [INFO] raft: snapshot complete up to: index=16386 2021-01-29T19:35:46.760Z [INFO] raft: initial configuration: index=1 servers="[{Suffrage:Voter ID:server-f6d098ad-68ae-1eac-3156-d6a49e439785 Address:f6d098ad-68ae-1eac-3156-d6a49e439785}]" 2021-01-29T19:35:46.760Z [INFO] raft: entering follower state: follower="Node at f6d098ad-68ae-1eac-3156-d6a49e439785 [Follower]" leader= 2021-01-29T19:35:48.431Z [WARN] raft: heartbeat timeout reached, starting election: last-leader= 2021-01-29T19:35:48.431Z [INFO] raft: entering candidate state: node="Node at f6d098ad-68ae-1eac-3156-d6a49e439785 [Candidate]" term=2 2021-01-29T19:35:48.431Z [DEBUG] raft: votes: needed=1 2021-01-29T19:35:48.431Z [DEBUG] raft: vote granted: from=server-f6d098ad-68ae-1eac-3156-d6a49e439785 term=2 tally=1 2021-01-29T19:35:48.431Z [INFO] raft: election won: tally=1 2021-01-29T19:35:48.431Z [INFO] raft: entering leader state: leader="Node at f6d098ad-68ae-1eac-3156-d6a49e439785 [Leader]" writer.go:29: 2021-01-29T19:35:48.432Z [ERROR] TestSnapshot_BadRestore: Failed to close snapshot decompressor: error="unexpected EOF" --- PASS: TestSnapshot_BadRestore (3.57s) PASS ok github.com/hashicorp/consul/snapshot 22.299s ? github.com/hashicorp/consul/testrpc [no test files] === RUN TestConfigurator_outgoingWrapper_OK 2021-01-29T19:35:26.291Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.291Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 2021-01-29T19:35:26.292Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.292Z [DEBUG] tlsutil: OutgoingRPCWrapper: version=1 2021-01-29T19:35:26.292Z [DEBUG] tlsutil: OutgoingRPCConfig: version=1 2021-01-29T19:35:26.293Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 --- PASS: TestConfigurator_outgoingWrapper_OK (0.01s) === RUN TestConfigurator_outgoingWrapper_noverify_OK 2021-01-29T19:35:26.305Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.305Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 2021-01-29T19:35:26.306Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.306Z [DEBUG] tlsutil: OutgoingRPCWrapper: version=1 2021-01-29T19:35:26.306Z [DEBUG] tlsutil: OutgoingRPCConfig: version=1 2021-01-29T19:35:26.306Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 --- PASS: TestConfigurator_outgoingWrapper_noverify_OK (0.01s) === RUN TestConfigurator_outgoingWrapper_BadDC 2021-01-29T19:35:26.313Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.313Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 2021-01-29T19:35:26.313Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.314Z [DEBUG] tlsutil: OutgoingRPCWrapper: version=1 2021-01-29T19:35:26.314Z [DEBUG] tlsutil: OutgoingRPCConfig: version=1 2021-01-29T19:35:26.314Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 --- PASS: TestConfigurator_outgoingWrapper_BadDC (0.01s) === RUN TestConfigurator_outgoingWrapper_BadCert 2021-01-29T19:35:26.320Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.320Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 2021-01-29T19:35:26.321Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.321Z [DEBUG] tlsutil: OutgoingRPCWrapper: version=1 2021-01-29T19:35:26.321Z [DEBUG] tlsutil: OutgoingRPCConfig: version=1 2021-01-29T19:35:26.322Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 --- PASS: TestConfigurator_outgoingWrapper_BadCert (0.01s) === RUN TestConfigurator_wrapTLS_OK 2021-01-29T19:35:26.329Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.329Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 2021-01-29T19:35:26.330Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.330Z [DEBUG] tlsutil: OutgoingRPCConfig: version=1 2021-01-29T19:35:26.330Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 --- PASS: TestConfigurator_wrapTLS_OK (0.01s) === RUN TestConfigurator_wrapTLS_BadCert 2021-01-29T19:35:26.341Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.341Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 2021-01-29T19:35:26.342Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.342Z [DEBUG] tlsutil: OutgoingRPCConfig: version=1 2021-01-29T19:35:26.342Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 --- PASS: TestConfigurator_wrapTLS_BadCert (0.01s) === RUN TestConfig_ParseCiphers --- PASS: TestConfig_ParseCiphers (0.00s) === RUN TestConfigurator_loadKeyPair --- PASS: TestConfigurator_loadKeyPair (0.00s) === RUN TestConfig_SpecifyDC --- PASS: TestConfig_SpecifyDC (0.00s) === RUN TestConfigurator_NewConfigurator writer.go:29: 2021-01-29T19:35:26.353Z [DEBUG] TestConfigurator_NewConfigurator.tlsutil: Update: version=1 2021-01-29T19:35:26.354Z [DEBUG] tlsutil: Update: version=0 --- PASS: TestConfigurator_NewConfigurator (0.00s) === RUN TestConfigurator_ErrorPropagation 2021-01-29T19:35:26.354Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.355Z [DEBUG] tlsutil: Update: version=0 2021-01-29T19:35:26.355Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.356Z [DEBUG] tlsutil: Update: version=0 2021-01-29T19:35:26.357Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.357Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.359Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.360Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.360Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.361Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.364Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.365Z [DEBUG] tlsutil: Update: version=0 2021-01-29T19:35:26.365Z [DEBUG] tlsutil: Update: version=0 2021-01-29T19:35:26.366Z [DEBUG] tlsutil: Update: version=0 2021-01-29T19:35:26.366Z [DEBUG] tlsutil: Update: version=0 2021-01-29T19:35:26.367Z [DEBUG] tlsutil: Update: version=0 2021-01-29T19:35:26.368Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.370Z [DEBUG] tlsutil: Update: version=0 2021-01-29T19:35:26.370Z [DEBUG] tlsutil: Update: version=0 2021-01-29T19:35:26.370Z [DEBUG] tlsutil: Update: version=0 2021-01-29T19:35:26.370Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.370Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.371Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.371Z [DEBUG] tlsutil: Update: version=1 --- PASS: TestConfigurator_ErrorPropagation (0.02s) === RUN TestConfigurator_CommonTLSConfigServerNameNodeName 2021-01-29T19:35:26.372Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.372Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.372Z [DEBUG] tlsutil: Update: version=1 --- PASS: TestConfigurator_CommonTLSConfigServerNameNodeName (0.00s) === RUN TestConfigurator_loadCAs --- PASS: TestConfigurator_loadCAs (0.01s) === RUN TestConfigurator_CommonTLSConfigInsecureSkipVerify 2021-01-29T19:35:26.379Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.379Z [DEBUG] tlsutil: Update: version=2 2021-01-29T19:35:26.379Z [DEBUG] tlsutil: Update: version=3 --- PASS: TestConfigurator_CommonTLSConfigInsecureSkipVerify (0.00s) === RUN TestConfigurator_CommonTLSConfigPreferServerCipherSuites 2021-01-29T19:35:26.380Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.380Z [DEBUG] tlsutil: Update: version=2 2021-01-29T19:35:26.380Z [DEBUG] tlsutil: Update: version=3 --- PASS: TestConfigurator_CommonTLSConfigPreferServerCipherSuites (0.00s) === RUN TestConfigurator_CommonTLSConfigCipherSuites 2021-01-29T19:35:26.380Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.381Z [DEBUG] tlsutil: Update: version=2 --- PASS: TestConfigurator_CommonTLSConfigCipherSuites (0.00s) === RUN TestConfigurator_CommonTLSConfigGetClientCertificate 2021-01-29T19:35:26.381Z [DEBUG] tlsutil: Update: version=1 --- PASS: TestConfigurator_CommonTLSConfigGetClientCertificate (0.02s) === RUN TestConfigurator_CommonTLSConfigGetCertificate 2021-01-29T19:35:26.405Z [DEBUG] tlsutil: Update: version=1 --- PASS: TestConfigurator_CommonTLSConfigGetCertificate (0.00s) === RUN TestConfigurator_CommonTLSConfigCAs 2021-01-29T19:35:26.407Z [DEBUG] tlsutil: Update: version=1 --- PASS: TestConfigurator_CommonTLSConfigCAs (0.00s) === RUN TestConfigurator_CommonTLSConfigTLSMinVersion 2021-01-29T19:35:26.407Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.407Z [DEBUG] tlsutil: Update: version=2 2021-01-29T19:35:26.407Z [DEBUG] tlsutil: Update: version=3 2021-01-29T19:35:26.407Z [DEBUG] tlsutil: Update: version=4 2021-01-29T19:35:26.407Z [DEBUG] tlsutil: Update: version=5 2021-01-29T19:35:26.408Z [DEBUG] tlsutil: Update: version=5 --- PASS: TestConfigurator_CommonTLSConfigTLSMinVersion (0.00s) === RUN TestConfigurator_CommonTLSConfigVerifyIncoming --- PASS: TestConfigurator_CommonTLSConfigVerifyIncoming (0.00s) === RUN TestConfigurator_OutgoingRPCTLSDisabled --- PASS: TestConfigurator_OutgoingRPCTLSDisabled (0.00s) === RUN TestConfigurator_VerifyIncomingRPC --- PASS: TestConfigurator_VerifyIncomingRPC (0.00s) === RUN TestConfigurator_VerifyIncomingHTTPS --- PASS: TestConfigurator_VerifyIncomingHTTPS (0.00s) === RUN TestConfigurator_EnableAgentTLSForChecks --- PASS: TestConfigurator_EnableAgentTLSForChecks (0.00s) === RUN TestConfigurator_IncomingRPCConfig 2021-01-29T19:35:26.410Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.410Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 2021-01-29T19:35:26.410Z [DEBUG] tlsutil: IncomingRPCConfig: version=1 --- PASS: TestConfigurator_IncomingRPCConfig (0.00s) === RUN TestConfigurator_IncomingHTTPSConfig --- PASS: TestConfigurator_IncomingHTTPSConfig (0.00s) === RUN TestConfigurator_OutgoingTLSConfigForChecks --- PASS: TestConfigurator_OutgoingTLSConfigForChecks (0.00s) === RUN TestConfigurator_OutgoingRPCConfig --- PASS: TestConfigurator_OutgoingRPCConfig (0.00s) === RUN TestConfigurator_OutgoingRPCWrapper 2021-01-29T19:35:26.411Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.412Z [DEBUG] tlsutil: OutgoingRPCWrapper: version=1 2021-01-29T19:35:26.412Z [DEBUG] tlsutil: OutgoingRPCConfig: version=1 --- PASS: TestConfigurator_OutgoingRPCWrapper (0.00s) === RUN TestConfigurator_UpdateChecks 2021-01-29T19:35:26.412Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.412Z [DEBUG] tlsutil: Update: version=2 2021-01-29T19:35:26.412Z [DEBUG] tlsutil: Update: version=2 2021-01-29T19:35:26.413Z [DEBUG] tlsutil: Update: version=2 --- PASS: TestConfigurator_UpdateChecks (0.00s) === RUN TestConfigurator_UpdateSetsStuff 2021-01-29T19:35:26.413Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.413Z [DEBUG] tlsutil: Update: version=1 2021-01-29T19:35:26.414Z [DEBUG] tlsutil: Update: version=2 --- PASS: TestConfigurator_UpdateSetsStuff (0.00s) === RUN TestConfigurator_ServerNameOrNodeName --- PASS: TestConfigurator_ServerNameOrNodeName (0.00s) === RUN TestConfigurator_VerifyOutgoing --- PASS: TestConfigurator_VerifyOutgoing (0.00s) === RUN TestConfigurator_Domain --- PASS: TestConfigurator_Domain (0.00s) === RUN TestConfigurator_VerifyServerHostname --- PASS: TestConfigurator_VerifyServerHostname (0.00s) === RUN TestConfigurator_AutoEncrytCertExpired --- PASS: TestConfigurator_AutoEncrytCertExpired (0.00s) === RUN TestConfig_tlsVersions --- PASS: TestConfig_tlsVersions (0.00s) === RUN TestSerialNumber --- PASS: TestSerialNumber (0.00s) === RUN TestGeneratePrivateKey === PAUSE TestGeneratePrivateKey === RUN TestGenerateCA === PAUSE TestGenerateCA === RUN TestGenerateCert generate_test.go:101: DM-skipped --- SKIP: TestGenerateCert (0.00s) === CONT TestGeneratePrivateKey === CONT TestGenerateCA --- PASS: TestGeneratePrivateKey (0.00s) --- PASS: TestGenerateCA (0.00s) PASS ok github.com/hashicorp/consul/tlsutil 0.146s ? github.com/hashicorp/consul/types [no test files] ? github.com/hashicorp/consul/version [no test files] FAIL dh_auto_test: error: cd _build && go test -vet=off -v -p 4 -short -failfast -timeout 8m github.com/hashicorp/consul github.com/hashicorp/consul/acl github.com/hashicorp/consul/agent github.com/hashicorp/consul/agent/ae github.com/hashicorp/consul/agent/agentpb github.com/hashicorp/consul/agent/config github.com/hashicorp/consul/agent/debug github.com/hashicorp/consul/agent/exec github.com/hashicorp/consul/agent/local github.com/hashicorp/consul/agent/metadata github.com/hashicorp/consul/agent/mock github.com/hashicorp/consul/agent/pool github.com/hashicorp/consul/agent/router github.com/hashicorp/consul/agent/structs github.com/hashicorp/consul/agent/systemd github.com/hashicorp/consul/agent/token github.com/hashicorp/consul/agent/xds github.com/hashicorp/consul/command github.com/hashicorp/consul/command/acl github.com/hashicorp/consul/command/acl/agenttokens github.com/hashicorp/consul/command/acl/authmethod github.com/hashicorp/consul/command/acl/authmethod/create github.com/hashicorp/consul/command/acl/authmethod/delete github.com/hashicorp/consul/command/acl/authmethod/list github.com/hashicorp/consul/command/acl/authmethod/read github.com/hashicorp/consul/command/acl/authmethod/update github.com/hashicorp/consul/command/acl/bindingrule github.com/hashicorp/consul/command/acl/bindingrule/create github.com/hashicorp/consul/command/acl/bindingrule/delete github.com/hashicorp/consul/command/acl/bindingrule/list github.com/hashicorp/consul/command/acl/bindingrule/read github.com/hashicorp/consul/command/acl/bindingrule/update github.com/hashicorp/consul/command/acl/bootstrap github.com/hashicorp/consul/command/acl/policy github.com/hashicorp/consul/command/acl/policy/create github.com/hashicorp/consul/command/acl/policy/delete github.com/hashicorp/consul/command/acl/policy/list github.com/hashicorp/consul/command/acl/policy/read github.com/hashicorp/consul/command/acl/policy/update github.com/hashicorp/consul/command/acl/role github.com/hashicorp/consul/command/acl/role/create github.com/hashicorp/consul/command/acl/role/delete github.com/hashicorp/consul/command/acl/role/list github.com/hashicorp/consul/command/acl/role/read github.com/hashicorp/consul/command/acl/role/update github.com/hashicorp/consul/command/acl/rules github.com/hashicorp/consul/command/acl/token github.com/hashicorp/consul/command/acl/token/clone github.com/hashicorp/consul/command/acl/token/create github.com/hashicorp/consul/command/acl/token/delete github.com/hashicorp/consul/command/acl/token/list github.com/hashicorp/consul/command/acl/token/read github.com/hashicorp/consul/command/acl/token/update github.com/hashicorp/consul/command/agent github.com/hashicorp/consul/command/catalog github.com/hashicorp/consul/command/catalog/list/dc github.com/hashicorp/consul/command/catalog/list/nodes github.com/hashicorp/consul/command/catalog/list/services github.com/hashicorp/consul/command/config github.com/hashicorp/consul/command/config/delete github.com/hashicorp/consul/command/config/list github.com/hashicorp/consul/command/config/read github.com/hashicorp/consul/command/config/write github.com/hashicorp/consul/command/connect github.com/hashicorp/consul/command/connect/ca github.com/hashicorp/consul/command/connect/ca/get github.com/hashicorp/consul/command/connect/ca/set github.com/hashicorp/consul/command/connect/envoy github.com/hashicorp/consul/command/connect/envoy/pipe-bootstrap github.com/hashicorp/consul/command/connect/proxy github.com/hashicorp/consul/command/event github.com/hashicorp/consul/command/exec github.com/hashicorp/consul/command/flags github.com/hashicorp/consul/command/forceleave github.com/hashicorp/consul/command/helpers github.com/hashicorp/consul/command/info github.com/hashicorp/consul/command/intention github.com/hashicorp/consul/command/intention/check github.com/hashicorp/consul/command/intention/create github.com/hashicorp/consul/command/intention/delete github.com/hashicorp/consul/command/intention/finder github.com/hashicorp/consul/command/intention/get github.com/hashicorp/consul/command/intention/match github.com/hashicorp/consul/command/join github.com/hashicorp/consul/command/keygen github.com/hashicorp/consul/command/keyring github.com/hashicorp/consul/command/kv github.com/hashicorp/consul/command/kv/del github.com/hashicorp/consul/command/kv/exp github.com/hashicorp/consul/command/kv/get github.com/hashicorp/consul/command/kv/imp github.com/hashicorp/consul/command/kv/impexp github.com/hashicorp/consul/command/kv/put github.com/hashicorp/consul/command/leave github.com/hashicorp/consul/command/lock github.com/hashicorp/consul/command/login github.com/hashicorp/consul/command/logout github.com/hashicorp/consul/command/maint github.com/hashicorp/consul/command/members github.com/hashicorp/consul/command/monitor github.com/hashicorp/consul/command/operator github.com/hashicorp/consul/command/operator/autopilot github.com/hashicorp/consul/command/operator/autopilot/get github.com/hashicorp/consul/command/operator/autopilot/set github.com/hashicorp/consul/command/operator/raft github.com/hashicorp/consul/command/operator/raft/listpeers github.com/hashicorp/consul/command/operator/raft/removepeer github.com/hashicorp/consul/command/reload github.com/hashicorp/consul/command/rtt github.com/hashicorp/consul/command/services github.com/hashicorp/consul/command/services/deregister github.com/hashicorp/consul/command/services/register github.com/hashicorp/consul/command/snapshot github.com/hashicorp/consul/command/snapshot/inspect github.com/hashicorp/consul/command/snapshot/restore github.com/hashicorp/consul/command/snapshot/save github.com/hashicorp/consul/command/validate github.com/hashicorp/consul/command/version github.com/hashicorp/consul/command/watch github.com/hashicorp/consul/connect github.com/hashicorp/consul/connect/certgen github.com/hashicorp/consul/connect/proxy github.com/hashicorp/consul/ipaddr github.com/hashicorp/consul/lib github.com/hashicorp/consul/lib/file github.com/hashicorp/consul/lib/semaphore github.com/hashicorp/consul/logging github.com/hashicorp/consul/logging/monitor github.com/hashicorp/consul/sdk/freeport github.com/hashicorp/consul/sdk/testutil github.com/hashicorp/consul/sdk/testutil/retry github.com/hashicorp/consul/sentinel github.com/hashicorp/consul/service_os github.com/hashicorp/consul/snapshot github.com/hashicorp/consul/testrpc github.com/hashicorp/consul/tlsutil github.com/hashicorp/consul/types github.com/hashicorp/consul/version returned exit code 1 make[1]: *** [debian/rules:63: override_dh_auto_test] Error 25 make[1]: Leaving directory '/<>/consul-1.7.4+dfsg1' make: *** [debian/rules:25: build-arch] Error 2 dpkg-buildpackage: error: debian/rules build-arch subprocess returned exit status 2 -------------------------------------------------------------------------------- Build finished at 2021-01-29T19:35:48Z Finished -------- +------------------------------------------------------------------------------+ | Cleanup | +------------------------------------------------------------------------------+ Purging /<> Not removing build depends: as requested E: Build failure (dpkg-buildpackage died) +------------------------------------------------------------------------------+ | Summary | +------------------------------------------------------------------------------+ Build Architecture: s390x Build Type: any Build-Space: n/a Build-Time: 408 Distribution: hirsute Fail-Stage: build Host Architecture: s390x Install-Time: 31 Job: consul_1.7.4+dfsg1-1.dsc Machine Architecture: s390x Package: consul Package-Time: 441 Source-Version: 1.7.4+dfsg1-1 Space: n/a Status: attempted Version: 1.7.4+dfsg1-1 -------------------------------------------------------------------------------- Finished at 2021-01-29T19:35:48Z Build needed 00:07:21, no disk space E: Build failure (dpkg-buildpackage died) RUN: /usr/share/launchpad-buildd/bin/in-target scan-for-processes --backend=chroot --series=hirsute --arch=s390x PACKAGEBUILD-20426680 Scanning for processes to kill in build PACKAGEBUILD-20426680